The Hortonworks Community Connection is now live. A completely rebuilt Q&A forum, Knowledge Base, Code Hub and more, backed by the experts in the industry.

You will be redirected here in 10 seconds. If your are not redirected, click here to visit the new site.

The legacy Hortonworks Forum is now closed. You can view a read-only version of the former site by clicking here. The site will be taken offline on January 31,2016

HDP on Linux – Installation Forum

Ambari 1.4 Host Failure on Installation

  • #39838

    Slave nodes are failing on “Confirm Hosts”

    here is the main error: Failed to connect to https://master.hdp.hadoop:8440/cert/ca due to [Errno 111] Connection refused

    here is the log….thanks!

    STDOUT
    Error: No matching Packages to list
    Error: No matching Packages to list
    Restarting ambari-agent
    Verifying Python version compatibility…
    Using python /usr/bin/python2.6
    ambari-agent is not running. No PID found at /var/run/ambari-agent/ambari-agent.pid
    Verifying Python version compatibility…
    Using python /usr/bin/python2.6
    Checking for previously running Ambari Agent…
    Starting ambari-agent
    Verifying ambari-agent process status…
    Ambari Agent successfully started
    Agent PID at: /var/run/ambari-agent/ambari-agent.pid
    Agent out at: /var/log/ambari-agent/ambari-agent.out
    Agent log at: /var/log/ambari-agent/ambari-agent.log
    (‘INFO 2013-10-10 20:04:45,068 NetUtil.py:41 – Connecting to the following url https://master.hdp.hadoop:8440/cert/ca
    INFO 2013-10-10 20:04:45,070 NetUtil.py:55 – Failed to connect to https://master.hdp.hadoop:8440/cert/ca due to [Errno 111] Connection refused
    INFO 2013-10-10 20:04:45,070 NetUtil.py:74 – Server at https://master.hdp.hadoop:8440 is not reachable, sleeping for 10 seconds…
    INFO 2013-10-10 20:04:55,080 NetUtil.py:41 – Connecting to the following url https://master.hdp.hadoop:8440/cert/ca
    INFO 2013-10-10 20:04:55,082 NetUtil.py:55 – Failed to connect to https://master.hdp.hadoop:8440/cert/ca due to [Errno 111] Connection refused
    INFO 2013-10-10 20:04:55,082 NetUtil.py:74 – Server at https://master.hdp.hadoop:8440 is not reachable, sleeping for 10 seconds…
    INFO 2013-10-10 20:05:04,815 main.py:55 – signal received, exiting.
    INFO 2013-10-10 20:05:04,815 ProcessHelper.py:38 – Removing pid file
    INFO 2013-10-10 20:05:04,816 ProcessHelper.py:45 – Removing temp files
    INFO 2013-10-10 20:13:37,594 main.py:83 – loglevel=logging.INFO
    INFO 2013-10-10 20:13:37,606 DataCleaner.py:36 – Data cleanup thread started
    INFO 2013-10-10 20:13:37,608 shell.py:52 – Killing stale processes
    INFO 2013-10-10 20

  • Author
    Replies
  • #39843
    Dave
    Moderator

    Hi Dennis.

    Can you run a wget http://master.hdp.hadoop:8440
    or a telnet master.hdp.hadoop 8440

    If you get connection refused then you should investigate the firewall & connectivity issues between these 2 machines.

    Thanks

    Dave

    #39870

    not sure if this helps, but notice that when i run netstat -lan | egrep LISTEN for my ambari master i get this:

    [root@masterambari ~]# netstat -lan | egrep LISTEN

    tcp 0 0 0.0.0.0:22 0.0.0.0:* LIST EN
    tcp 0 0 0.0.0.0:5432 0.0.0.0:* LIST EN
    tcp 0 0 127.0.0.1:25 0.0.0.0:* LIST EN
    tcp 0 0 :::8080 :::* LIST EN
    tcp 0 0 :::22 :::* LIST EN
    tcp 0 0 :::8440 :::* LIST EN
    tcp 0 0 :::5432 :::* LIST EN
    tcp 0 0 :::8441 :::* LIST

    however when i run for master.hdp.hadoop i get

    tcp 0 0 0.0.0.0:22 0.0.0.0:* LISTEN
    tcp 0 0 127.0.0.1:631 0.0.0.0:* LISTEN
    tcp 0 0 127.0.0.1:25 0.0.0.0:* LISTEN
    tcp 0 0 127.0.0.1:199 0.0.0.0:* LISTEN
    tcp 0 0 :::22 :::* LISTEN
    tcp 0 0 ::1:631 :::* LISTEN

    I essentially have 5 nodes:
    masterambari.hdp.hadoop (but not part of the cluster)
    master.hdp.hadoop
    slave1.hdp.hadoop
    slave2.hdp.hadoop
    slave3.hdp.hadoop

    if i do a wget from a slave into masterambari.hdp.hadoop:8440 it seems to be working

    [root@slave3 ~]# wget http://masterambari.hdp.hadoop:8440
    –2013-10-10 22:32:01– http://masterambari.hdp.hadoop:8440/
    Resolving masterambari.hdp.hadoop… 10.29.37.230
    Connecting to masterambari.hdp.hadoop|10.29.37.230|:8440… connected.
    HTTP request sent, awaiting response… No data received.
    Retrying.

    however…wget from a slave into master.hdp.hadoop:8440 does not work

    [root@slave3 ~]# wget http://master.hdp.hadoop:8440
    –2013-10-10 22:34:34– http://master.hdp.hadoop:8440/
    Resolving master.hdp.hadoop… 10.154.185.54
    Connecting to master.hdp.hadoop|10.154.185.54|:8440… failed: Connection refused.
    [root@slave3 ~]#

    I think im confused on masterambari vs master, any help will be very appreciated. I have been following this: http://hortonworks.com/kb/ambari-on-ec2/

    Thanks!

    #39872
    Dave
    Moderator

    Hi Dennis,

    It looks like your ambari-agents are trying to talk to master and not masterambari.
    You can check this by looking at a slave and going to /etc/ambari-agent/cont/ambari-agent.ini

    Here the server should be masterambari.hdp.hadoop

    Let me know how you get on,

    Dave

    #39873

    worked! Thanks again Dave.

The topic ‘Ambari 1.4 Host Failure on Installation’ is closed to new replies.

Support from the Experts

A HDP Support Subscription connects you experts with deep experience running Apache Hadoop in production, at-scale on the most demanding workloads.

Enterprise Support »

Become HDP Certified

Real world training designed by the core architects of Hadoop. Scenario-based training courses are available in-classroom or online from anywhere in the world

Training »

Hortonworks Data Platform
The Hortonworks Data Platform is a 100% open source distribution of Apache Hadoop that is truly enterprise grade having been built, tested and hardened with enterprise rigor.
Get started with Sandbox
Hortonworks Sandbox is a self-contained virtual machine with Apache Hadoop pre-configured alongside a set of hands-on, step-by-step Hadoop tutorials.
Modern Data Architecture
Tackle the challenges of big data. Hadoop integrates with existing EDW, RDBMS and MPP systems to deliver lower cost, higher capacity infrastructure.