HDP on Linux – Installation Forum

Users Hosted in Winbind

  • #18441
    Anthony Mattas

    I’m trying to setup HDP using servers that host the usernames in Winbind. However, unfortunately when trying to install the services its unable to find those users in /etc/passwd so it won’t add them to the groups. They do show up in getent passwd however.

    Any thoughts?

to create new topics or reply. | New User Registration

  • Author
  • #18463
    Sasha J

    Winbind if not supported in the current release.
    why not to pre-create users ahead of the installation and try installation again?
    Or, just use local users…

    Thank you!

    Anthony Mattas

    So related question –

    How does inter-node communication happen? Is it all over SSH? Originally I did Winbind to ensure I had Kerberos available.


    Hi Anthony,
    In regards to your question, if you are referring to Ambari, ssh is originally used to get into the nodes to install the agents. If you don’t want to enable ssh, you should be able to manually install the ambari-agents on all the nodes. Now when agents are installed, the communication between ambari-server and ambari-agent is via https. As for Hadoop components (jobtracker, namenode, datanodes, task trackers), they communicate via remote procedure calls (RPC) protocol.


    Anthony Mattas

    How do nodes authenticate to each other over RPC? Is there any type of trust formed?


    Hi Anthony,
    In a non-secure mode cluster (default) there is no trust being formed, since the assumption the machines are behind a firewall and are on its own subnet. If you require additional security, hadoop can run also in secure mode, where kerberos is used for authentication.


You must be to reply to this topic. | Create Account

Support from the Experts

A HDP Support Subscription connects you experts with deep experience running Apache Hadoop in production, at-scale on the most demanding workloads.

Enterprise Support »

Become HDP Certified

Real world training designed by the core architects of Hadoop. Scenario-based training courses are available in-classroom or online from anywhere in the world

Training »

Hortonworks Data Platform
The Hortonworks Data Platform is a 100% open source distribution of Apache Hadoop that is truly enterprise grade having been built, tested and hardened with enterprise rigor.
Get started with Sandbox
Hortonworks Sandbox is a self-contained virtual machine with Apache Hadoop pre-configured alongside a set of hands-on, step-by-step Hadoop tutorials.
Modern Data Architecture
Tackle the challenges of big data. Hadoop integrates with existing EDW, RDBMS and MPP systems to deliver lower cost, higher capacity infrastructure.