Home Forums HDP on Linux – Installation Adding additional services after the cluster has been setup

This topic contains 22 replies, has 6 voices, and was last updated by  Sasha J 2 years, 1 month ago.

  • Creator
    Topic
  • #6639

    I am setting up a single node cluster in CentOS 5.8. I initially tried installing all the services but it was failing. ( [2012:07:05 11:57:06][INFO][PuppetInvoker][PuppetInvoker.php:270][genKickWait]: Kick attempt (1/3)
    [2012:07:05 11:57:06][INFO][PuppetInvoker][PuppetInvoker.php:310][waitForResults]: Waiting for results )
    The above thing goes on forever and then it fails.

    I then tried installing only the basic services and it succeeded. I am able to see cluster management and monitoring in the web ui. Everything looks fine.
    NameNode
    Secondary NameNode
    JobTracker
    Ganglia Collector
    Nagios Server
    HMC Server
    HDFS Client
    MapReduce Client

    Can I install the rest of the services without going through uninstall of the whole cluster and start over again?
    Are there any options? I can’t find any.

    PS:Thrilled that the minimal hadoop services got installed I did uninstall and reinstall with all the services but I again got the puppet kick error.

Viewing 22 replies - 1 through 22 (of 22 total)

You must be logged in to reply to this topic.

  • Author
    Replies
  • #11162

    Sasha J
    Moderator

    Sahab,
    In the current implementation it is not possible…
    You can re-deploy the whole cluster or, you can add oozie manually, but you will have to stop using HMC to start/stop service after new service (oozie) deployed.
    This limitations suppose to be fixed in next release.
    Please watch the news, or just wait for a while for those new features become available.

    Collapse
    #11161

    sahabdeen k
    Member

    Hi All,

    Is there any possibility to add a service after installing HMC..Is present release of HMC will support for this?

    I want to add oozie service in single node cluster, I have skipped oozie due to frequent authentication failure
    of oozie test.

    Regards
    Sahab

    Collapse
    #8168

    Ben Cuthbert
    Member

    Thanks Sasha

    So I manage to get the installer working and the main control site up as you suggested. When I try and start HDFS it gives me some puppet errors

    [2012:08:11 07:09:27][INFO][HMC][HMC.php:476][internalTrigger]: HMC triggering action=startServices, Starting services, list=HDFS
    [2012:08:11 07:09:27][INFO][HMCTxnUtils][HMCTxnUtils.php:103][execBackgroundProcess]: Trying to background a new process, cluster=devcluster, txnId=10, command=/usr/bin/php /usr/share/hmc//php/orchestrator/ClusterMain.php, args= -c devcluster -d /var/db/hmc/data/data.db -a start -s HDFS -x 10, logFile=/var/log/hmc/hmc.txn.10.log, execCommand=/usr/bin/php /usr/share/hmc//php/util/BackgroundExecutor.php -t “10” -c “/usr/bin/php /usr/share/hmc//php/orchestrator/ClusterMain.php” -a ” -c devcluster -d /var/db/hmc/data/data.db -a start -s HDFS -x 10″ -l “/var/log/hmc/hmc.txn.10.log”
    [2012:08:11 07:09:27][INFO][HMCTxnUtils][HMCTxnUtils.php:127][execBackgroundProcess]: Output from process, command=/usr/bin/php /usr/share/hmc//php/orchestrator/ClusterMain.php, txnId=10, output=Executing /usr/bin/php /usr/share/hmc//php/orchestrator/ClusterMain.php -c devcluster -d /var/db/hmc/data/data.db -a start -s HDFS -x 10 > /var/log/hmc/hmc.txn.10.log 2>&1

    [2012:08:11 07:09:27][INFO][HMCTxnUtils][HMCTxnUtils.php:127][execBackgroundProcess]: Output from process, command=/usr/bin/php /usr/share/hmc//php/orchestrator/ClusterMain.php, txnId=10, output=Executing /usr/bin/php /usr/share/hmc//php/orchestrator/ClusterMain.php -c devcluster -d /var/db/hmc/data/data.db -a start -s HDFS -x 10 > /var/log/hmc/hmc.txn.10.log 2>&1
    Background Child Process PID:2449

    [2012:08:11 07:09:27][INFO][HMCTxnUtils][HMCTxnUtils.php:137][execBackgroundProcess]: Found child pid, command=/usr/bin/php /usr/share/hmc//php/orchestrator/ClusterMain.php, txnId=10, output=Executing /usr/bin/php /usr/share/hmc//php/orchestrator/ClusterMain.php -c devcluster -d /var/db/hmc/data/data.db -a start -s HDFS -x 10 > /var/log/hmc/hmc.txn.10.log 2>&1
    Background Child Process PID:2449
    , pid=2449
    [2012:08:11 07:09:27][INFO][ClusterState][clusterState.php:40][updateClusterState]: Update Cluster State with {“state”:”SERVICE_MANAGEMENT_IN_PROGRESS”,”displayName”:”Service management in progress”,”timeStamp”:1344668967,”context”:{“txnId”:10,”stashedDeployState”:{“state”:”DEPLOYED”,”displayName”:”Deployed successfully”,”timeStamp”:1344668672,”context”:{“status”:true,”txnId”:”3″,”isInPostProcess”:false,”postProcessSuccessful”:true}}}}
    [2012:08:11 07:09:27][INFO][ClusterMain:TxnId=10][ClusterMain.php:322][]: Taking action=start on cluster=devcluster, txn=10-0-0
    [2012:08:11 07:09:27][INFO][Cluster:devcluster][Cluster.php:726][_startAllServices]: Starting HDP with 1 services…
    [2012:08:11 07:09:27][INFO][Cluster:devcluster][Cluster.php:807][startService]: Starting service HDFS …
    [2012:08:11 07:09:27][INFO][OrchestratorDB][OrchestratorDB.php:610][persistTransaction]: persist: 10-2-0:PENDING:HDFS start:PENDING
    [2012:08:11 07:09:27][INFO][Service: HDFS (devcluster)][Service.php:130][setState]: HDFS – STARTING dryRun=1
    [2012:08:11 07:09:27][INFO][OrchestratorDB][OrchestratorDB.php:610][persistTransaction]: persist: 10-4-3:PENDING:NameNode start:PENDING
    [2012:08:11 07:09:27][INFO][ServiceComponent:NAMENODE][ServiceComponent.php:118][setState]: NAMENODE – STARTING dryRun=1
    [2012:08:11 07:09:27][INFO][OrchestratorDB][OrchestratorDB.php:610][persistTransaction]: persist: 10-4-3:PENDING:NameNode start:PENDING
    [2012:08:11 07:09:27][INFO][ServiceComponent:NAMENODE][ServiceComponent.php:118][setState]: NAMENODE – STARTED dryRun=1
    [2012:08:11 07:09:27][INFO][OrchestratorDB][OrchestratorDB.php:610][persistTransaction]: persist: 10-3-2:PENDING:Datanode start:PENDING
    [2012:08:11 07:09:27][INFO][ServiceComponent:DATANODE][ServiceComponent.php:118][setState]: DATANODE – STARTING dryRun=1
    [2012:08:11 07:09:27][INFO][OrchestratorDB][OrchestratorDB.php:610][persistTransaction]: persist: 10-3-2:PENDING:Datanode start:PENDING
    [2012:08:11 07:09:27][INFO][ServiceComponent:DATANODE][ServiceComponent.php:118][setState]: DATANODE – STARTED dryRun=1
    [2012:08:11 07:09:27][INFO][ServiceComponent:NAMENODE][ServiceComponent.php:199][start]: ServiceComponent NAMENODE is already STARTED!
    [2012:08:11 07:09:27][INFO][ServiceComponent:NAMENODE][ServiceComponent.php:199][start]: ServiceComponent NAMENODE is already STARTED!
    [2012:08:11 07:09:27][INFO][OrchestratorDB][OrchestratorDB.php:610][persistTransaction]: persist: 10-7-2:PENDING:Secondary NameNode start:PENDING
    [2012:08:11 07:09:27][INFO][ServiceComponent:SNAMENODE][ServiceComponent.php:118][setState]: SNAMENODE – STARTING dryRun=1
    [2012:08:11 07:09:27][INFO][OrchestratorDB][OrchestratorDB.php:610][persistTransaction]: persist: 10-7-2:PENDING:Secondary NameNode start:PENDING
    [2012:08:11 07:09:27][INFO][ServiceComponent:SNAMENODE][ServiceComponent.php:118][setState]: SNAMENODE – STARTED dryRun=1
    [2012:08:11 07:09:27][INFO][OrchestratorDB][OrchestratorDB.php:610][persistTransaction]: persist: 10-2-0:PENDING:HDFS start:PENDING
    [2012:08:11 07:09:27][INFO][Service: HDFS (devcluster)][Service.php:130][setState]: HDFS – STARTED dryRun=1
    [2012:08:11 07:09:27][INFO][OrchestratorDB][OrchestratorDB.php:610][persistTransaction]: persist: 10-9-0:PENDING:HDFS test:PENDING
    [2012:08:11 07:09:27][INFO][OrchestratorDB][OrchestratorDB.php:610][persistTransaction]: persist: 10-9-0:PENDING:HDFS test:PENDING
    [2012:08:11 07:09:27][INFO][Cluster:devcluster][Cluster.php:810][startService]: Starting service HDFS complete. Result=0
    [2012:08:11 07:09:27][INFO][Cluster:devcluster][Cluster.php:735][_startAllServices]: Starting HDP with 1 services complete.
    [2012:08:11 07:09:27][INFO][Cluster:devcluster][Cluster.php:726][_startAllServices]: Starting HDP with 1 services…
    [2012:08:11 07:09:27][INFO][Cluster:devcluster][Cluster.php:807][startService]: Starting service HDFS …
    [2012:08:11 07:09:27][INFO][OrchestratorDB][OrchestratorDB.php:610][persistTransaction]: persist: 10-2-0:STARTING:HDFS start:IN_PROGRESS
    [2012:08:11 07:09:27][INFO][OrchestratorDB][OrchestratorDB.php:556][setServiceState]: HDFS – STARTING
    [2012:08:11 07:09:27][INFO][Service: HDFS (devcluster)][Service.php:130][setState]: HDFS – STARTING dryRun=
    [2012:08:11 07:09:27][INFO][OrchestratorDB][OrchestratorDB.php:610][persistTransaction]: persist: 10-4-3:STARTING:NameNode start:IN_PROGRESS
    [2012:08:11 07:09:27][INFO][OrchestratorDB][OrchestratorDB.php:577][setServiceComponentState]: Update ServiceComponentState HDFS – NAMENODE – STARTING
    [2012:08:11 07:09:27][INFO][ServiceComponent:NAMENODE][ServiceComponent.php:118][setState]: NAMENODE – STARTING dryRun=
    [2012:08:11 07:09:27][INFO][ServiceComponent:NAMENODE][ServiceComponent.php:245][start]: Kicking puppet for starting component on cluster=devcluster, servicecomponent=NAMENODE, txn=10-4-3
    [2012:08:11 07:09:27][INFO][PuppetInvoker][PuppetInvoker.php:277][genKickWait]: rm -f /etc/puppet/master/modules/catalog/files/modules.tgz
    [2012:08:11 07:09:27][INFO][PuppetInvoker][PuppetInvoker.php:280][genKickWait]: tar zcf /etc/puppet/master/manifestloader/modules.tgz /etc/puppet/master/modules
    [2012:08:11 07:09:28][INFO][PuppetInvoker][PuppetInvoker.php:283][genKickWait]: mv /etc/puppet/master/manifestloader/modules.tgz /etc/puppet/master/modules/catalog/files
    [2012:08:11 07:09:28][INFO][PuppetInvoker][PuppetInvoker.php:292][genKickWait]: Kick attempt (1/3)
    [2012:08:11 07:09:28][INFO][PuppetInvoker][PuppetInvoker.php:100][sendKick]: devhortonworks: Kick failed with Host devhortonworks failed: Connection refused – connect(2)
    Triggering devhortonworks
    devhortonworks finished with exit code 2
    Failed: devhortonworks

    [2012:08:11 07:09:28][WARN][PuppetInvoker][PuppetInvoker.php:114][sendKickWithRetry]: Retrying kick after 10 seconds on Array
    (
    [0] => devhortonworks
    )

    [2012:08:11 07:09:38][INFO][PuppetInvoker][PuppetInvoker.php:100][sendKick]: devhortonworks: Kick failed with Host devhortonworks failed: Connection refused – connect(2)
    Triggering devhortonworks
    devhortonworks finished with exit code 2
    Failed: devhortonworks

    [2012:08:11 07:09:38][WARN][PuppetInvoker][PuppetInvoker.php:114][sendKickWithRetry]: Retrying kick after 10 seconds on Array
    (
    [0] => devhortonworks
    )

    [2012:08:11 07:09:49][INFO][PuppetInvoker][PuppetInvoker.php:100][sendKick]: devhortonworks: Kick failed with Host devhortonworks failed: Connection refused – connect(2)
    Triggering devhortonworks
    devhortonworks finished with exit code 2
    Failed: devhortonworks

    [2012:08:11 07:09:49][WARN][PuppetInvoker][PuppetInvoker.php:114][sendKickWithRetry]: Retrying kick after 10 seconds on Array
    (
    [0] => devhortonworks
    )

    [2012:08:11 07:09:59][INFO][PuppetInvoker][PuppetInvoker.php:100][sendKick]: devhortonworks: Kick failed with Host devhortonworks failed: Connection refused – connect(2)
    Triggering devhortonworks
    devhortonworks finished with exit code 2
    Failed: devhortonworks

    [2012:08:11 07:09:59][INFO][PuppetInvoker][PuppetInvoker.php:332][waitForResults]: Waiting for results from
    [2012:08:11 07:09:59][INFO][PuppetInvoker][PuppetInvoker.php:336][waitForResults]: 0 out of 0 nodes have reported for txn 10-4-3
    [2012:08:11 07:10:00][INFO][PuppetInvoker][PuppetInvoker.php:237][createGenKickWaitResponse]: Response of genKickWait:
    Array
    (
    [result] => -1
    [error] => All kicks failed
    [nokick] => Array
    (
    [0] => devhortonworks
    )

    [failed] => Array
    (
    )

    [success] => Array
    (
    )

    [timedoutnodes] => Array
    (
    )

    )

    [2012:08:11 07:10:00][INFO][ServiceComponent:NAMENODE][ServiceComponent.php:254][start]: Puppet kick response for starting component on cluster=devcluster, servicecomponent=NAMENODE, txn=10-4-3, response=Array
    (
    [result] => -1
    [error] => All kicks failed
    [nokick] => Array
    (
    [0] => devhortonworks
    )

    [failed] => Array
    (
    )

    [success] => Array
    (
    )

    [timedoutnodes] => Array
    (
    )

    )

    [2012:08:11 07:10:00][INFO][ServiceComponent:NAMENODE][ServiceComponent.php:270][start]: Persisting puppet report for starting NAMENODE
    [2012:08:11 07:10:00][ERROR][ServiceComponent:NAMENODE][ServiceComponent.php:276][start]: Puppet kick failed, result=-1
    [2012:08:11 07:10:00][INFO][OrchestratorDB][OrchestratorDB.php:610][persistTransaction]: persist: 10-4-3:FAILED:NameNode start:FAILED
    [2012:08:11 07:10:00][INFO][OrchestratorDB][OrchestratorDB.php:577][setServiceComponentState]: Update ServiceComponentState HDFS – NAMENODE – FAILED
    [2012:08:11 07:10:00][INFO][ServiceComponent:NAMENODE][ServiceComponent.php:118][setState]: NAMENODE – FAILED dryRun=
    [2012:08:11 07:10:00][INFO][OrchestratorDB][OrchestratorDB.php:610][persistTransaction]: persist: 10-2-0:FAILED:HDFS start:FAILED
    [2012:08:11 07:10:00][INFO][OrchestratorDB][OrchestratorDB.php:556][setServiceState]: HDFS – FAILED
    [2012:08:11 07:10:00][INFO][Service: HDFS (devcluster)][Service.php:130][setState]: HDFS – FAILED dryRun=
    [2012:08:11 07:10:00][INFO][Cluster:devcluster][Cluster.php:810][startService]: Starting service HDFS complete. Result=-1
    [2012:08:11 07:10:00][INFO][ClusterMain:TxnId=10][ClusterMain.php:353][]: Completed action=start on cluster=devcluster, txn=10-0-0, result=-1, error=Failed to start DATANODE with -1 (\’Failed to start NAMENODE with -1 (\’All kicks failed\’)\’)
    [2012:08:11 07:10:01][INFO][ClusterState][clusterState.php:40][updateClusterState]: Update Cluster State with {“state”:”SERVICE_MANAGEMENT_IN_PROGRESS”,”displayName”:”Service management in progress”,”timeStamp”:1344669000,”context”:{“txnId”:10,”stashedDeployState”:{“state”:”DEPLOYED”,”displayName”:”Deployed successfully”,”timeStamp”:1344668672,”context”:{“status”:true,”txnId”:”3″,”isInPostProcess”:false,”postProcessSuccessful”:true}},”isInPostProcess”:true}}
    [2012:08:11 07:10:01][INFO][ClusterState][clusterState.php:40][updateClusterState]: Update Cluster State with {“state”:”DEPLOYED”,”displayName”:”Deployed successfully”,”timeStamp”:1344669001,”context”:{“status”:true,”txnId”:”3″,”isInPostProcess”:false,”postProcessSuccessful”:true}}
    [2012:08:11 07:10:01][INFO][ClusterState][clusterState.php:40][updateClusterState]: Update Cluster State with {“state”:”DEPLOYED”,”displayName”:”Deployed successfully”,”timeStamp”:1344669001,”context”:{“status”:true,”txnId”:”3″,”isInPostProcess”:false,”postProcessSuccessful”:true}}

    Collapse
    #8142

    Sasha J
    Moderator

    Ben,
    please do this

    yum -y erase hmc puppet mysql
    reboot
    yum -y install hmc
    service hmc start

    then start installer again.

    As you can see, your user in mysql is not created.
    check form iptables and selinux as well.

    Run check script mentioned in sticky post and provide us with the output.

    Thank you!
    Sasha

    Collapse
    #8133

    Ben Cuthbert
    Member

    Thanks for the post. So I have just the base mysql installed as per the requirements. And I ran the suggestions as you stated above. No luck really. I did notice this in the log

    Fri Aug 10 18:49:40 +0100 2012 Puppet (debug): Executing ‘test -e /usr/jdk32/jdk1.6.0_31/bin/java’
    Fri Aug 10 18:49:40 +0100 2012 Exec[mysqld start](provider=posix) (debug): Executing check ‘/etc/init.d/mysqld status’
    Fri Aug 10 18:49:40 +0100 2012 Puppet (debug): Executing ‘/etc/init.d/mysqld status’
    Fri Aug 10 18:49:40 +0100 2012 Exec[mysqld start](provider=posix) (debug): Executing ‘/etc/init.d/mysqld start’
    Fri Aug 10 18:49:40 +0100 2012 Puppet (debug): Executing ‘/etc/init.d/mysqld start’
    Fri Aug 10 18:49:42 +0100 2012 /Stage[13]/Hdp-mysql::Server/Hdp::Exec[mysqld start]/Exec[mysqld start]/returns (notice): executed successfully
    Fri Aug 10 18:49:42 +0100 2012 /Stage[13]/Hdp-mysql::Server/Hdp::Exec[mysqld start]/Exec[mysqld start] (debug): The container Hdp::Exec[mysqld start] will propagate my refresh event
    Fri Aug 10 18:49:42 +0100 2012 Hdp::Exec[mysqld start] (debug): The container Class[Hdp-mysql::Server] will propagate my refresh event
    Fri Aug 10 18:49:42 +0100 2012 /Stage[13]/Hdp-mysql::Server/Exec[/tmp/startMysql.sh]/returns (debug): Exec try 1/3
    Fri Aug 10 18:49:42 +0100 2012 Exec[/tmp/startMysql.sh](provider=posix) (debug): Executing ‘sh /tmp/startMysql.sh hive hive localhost’
    Fri Aug 10 18:49:42 +0100 2012 Puppet (debug): Executing ‘sh /tmp/startMysql.sh hive hive localhost’
    Fri Aug 10 18:49:42 +0100 2012 /Stage[13]/Hdp-mysql::Server/Exec[/tmp/startMysql.sh]/returns (notice): Adding user hive@localhost
    Fri Aug 10 18:49:42 +0100 2012 /Stage[13]/Hdp-mysql::Server/Exec[/tmp/startMysql.sh]/returns (notice): ERROR 1396 (HY000) at line 1: Operation CREATE USER failed for ‘hive’@’localhost’
    Fri Aug 10 18:49:42 +0100 2012 /Stage[13]/Hdp-mysql::Server/Exec[/tmp/startMysql.sh]/returns (notice): executed successfully
    Fri Aug 10 18:49:42 +0100 2012 /Stage[13]/Hdp-mysql::Server/Exec[/tmp/startMysql.sh] (debug): The container Class[Hdp-mysql::Server] will propagate my refresh event
    Fri Aug 10 18:49:42 +0100 2012 Class[Hdp-mysql::Server] (debug): The container Stage[13] will propagate my refresh event

    Then the apply.log shows the same as above and then it just fails.

    Collapse
    #8130

    Ben Cuthbert
    Member

    I have removed all the services and run the installer again. It still hangs the puppet_apple.log I just see this at the end with no errors or failures.

    Fri Aug 10 18:32:36 +0100 2012 Puppet (debug): Finishing transaction 23456252572840
    Fri Aug 10 18:32:36 +0100 2012 Puppet (debug): Storing state
    Fri Aug 10 18:32:36 +0100 2012 Puppet (debug): Stored state in 0.07 seconds
    Fri Aug 10 18:32:36 +0100 2012 Puppet (notice): Finished catalog run in 26.42 seconds
    Fri Aug 10 18:32:36 +0100 2012 Puppet (debug): Finishing transaction 23456250482540
    Fri Aug 10 18:32:36 +0100 2012 Puppet (debug): Received report to process from devhortonworks.localdomain
    Fri Aug 10 18:32:36 +0100 2012 Puppet (debug): Processing report from devhortonworks.localdomain with processor Puppet::Reports::Store

    Collapse
    #8117

    Sasha J
    Moderator

    When both install and uninstall failed, tou have to do the following:

    yum -y erase hmc puppet
    yum -y install hmc
    service hmc start

    this will give you ability to start over.
    Also, when you install single node , just unilstall MySQL if it is installed, and do not point to any mysql server during the HMC installation, but provide connectivity information.
    HMC will install mysql and make all needed configurations for you.

    Thank you!
    Sasha

    Collapse
    #8116

    Ben Cuthbert
    Member

    OKay think I found the issue. It has to do with the local MYSQL not being able to called /tmp/startMysql.sh
    access denied.

    Just a note. When this fails and then it says installation failed. You try and run the uninstall and that fails as well. I can did out the logs if required.

    Collapse
    #8115

    Ben Cuthbert
    Member

    If you are talking about pre-installing all the packages before running the HMC install, yes I did that. But I still receive the timeouts.

    Collapse
    #8113

    Ben Cuthbert
    Member

    HI Sasha I can’t see any reference to the script? Would you mind sending me the link or script name?

    Collapse
    #8112

    Sasha J
    Moderator

    Take a closer look on the log.
    as you have this:
    Skipping because of failed dependencies

    there should be something failed before.

    did you run that script mentioned in sticky note in the beginning or this thread?
    Please, do and upload results to us.

    Thank you!
    Sasha

    Collapse
    #8105

    Ben Cuthbert
    Member

    The last entry from the puppet_apply.log is

    ri Aug 10 12:20:12 +0100 2012 /Stage[25]/Hdp-ganglia::Monitor::Config-gen/Anchor[hdp-ganglia::monitor::config-gen::end] (notice): Dependency Exec[/tmp/startMysql.sh] has failures: true
    Fri Aug 10 12:20:12 +0100 2012 /Stage[25]/Hdp-ganglia::Monitor::Config-gen/Anchor[hdp-ganglia::monitor::config-gen::end] (warning): Skipping because of failed dependencies
    Fri Aug 10 12:20:12 +0100 2012 Puppet (debug): Finishing transaction 23456252487400
    Fri Aug 10 12:20:12 +0100 2012 Puppet (debug): Storing state
    Fri Aug 10 12:20:12 +0100 2012 Puppet (debug): Stored state in 0.04 seconds
    Fri Aug 10 12:20:12 +0100 2012 Puppet (notice): Finished catalog run in 74.28 seconds
    Fri Aug 10 12:20:12 +0100 2012 Puppet (debug): Finishing transaction 23456254414440
    Fri Aug 10 12:20:12 +0100 2012 Puppet (debug): Received report to process from devhortonworks.localdomain
    Fri Aug 10 12:20:12 +0100 2012 Puppet (debug): Processing report from devhortonworks.localdomain with processor Puppet::Reports::Store

    There are no FAILED messages in the log.

    Collapse
    #8104

    Ben Cuthbert
    Member

    Hi I tried what you suggested by installing all the packages. But when setting up a single node cluster I am still getting

    2012:08:10 10:39:56][INFO][PuppetInvoker][PuppetInvoker.php:280][genKickWait]: tar zcf /etc/puppet/master/manifestloader/modules.tgz /etc/puppet/master/modules
    [2012:08:10 10:39:56][INFO][PuppetInvoker][PuppetInvoker.php:283][genKickWait]: mv /etc/puppet/master/manifestloader/modules.tgz /etc/puppet/master/modules/catalog/files
    [2012:08:10 10:39:56][INFO][PuppetInvoker][PuppetInvoker.php:292][genKickWait]: Kick attempt (1/3)
    [2012:08:10 10:39:57][INFO][PuppetInvoker][PuppetInvoker.php:332][waitForResults]: Waiting for results from devhortonworks
    [2012:08:10 10:39:57][INFO][PuppetInvoker][PuppetInvoker.php:336][waitForResults]: 0 out of 1 nodes have reported for txn 3-2-0
    [2012:08:10 10:40:02][INFO][PuppetInvoker][PuppetInvoker.php:336][waitForResults]: 0 out of 1 nodes have reported for txn 3-2-0
    [2012:08:10 10:40:07][INFO][PuppetInvoker][PuppetInvoker.php:336][waitForResults]: 0 out of 1 nodes have reported for txn 3-2-0
    [2012:08:10 10:40:12][INFO][PuppetInvoker][PuppetInvoker.php:336][waitForResults]: 0 out of 1 nodes have reported for txn 3-2-0
    [2012:08:10 10:40:17][INFO][PuppetInvoker][PuppetInvoker.php:336][waitForResults]: 0 out of 1 nodes have reported for txn 3-2-0
    [2012:08:10 10:40:22][INFO][PuppetInvoker][PuppetInvoker.php:336][waitForResults]: 0 out of 1 nodes have reported for txn 3-2-0
    [2012:08:10 10:40:27][INFO][PuppetInvoker][PuppetInvoker.php:336][waitForResults]: 0 out of 1 nodes have reported for txn 3-2-0

    Collapse
    #7929

    Also, are there plans to detect / work with existing hadoop installations. Or simple work with a different version of hadoop.

    Thanks

    Collapse
    #7849

    Are there plans to allow the atomic migration of existing processes in the next release?

    example use case:

    Migrate namenode or jobtracker to a different / new node in order to perform maintenance on the current node.

    Collapse
    #6847

    Thanks Rohit! That would be great.

    Collapse
    #6756

    Rohit Bakhshi
    Moderator

    Hi Austin – I’m in the PM group here at Hortonworks. Thanks for downloading and working with HDP!

    We have CentOS/RHEL 6.x support for HMC on our radar, and I will follow up and let you know as soon as it is available for download and use.

    Thanks!

    Collapse
    #6664

    Sasha J
    Moderator

    I need to check with product management on this…WIll get back to you.

    Sasha

    Collapse
    #6659

    Thanks Sasha! that clears a lot of my doubts. The web interface looks awesome by the way.
    Can I also expect support for CentOS 6.2 in this next release? August – September?

    Collapse
    #6644

    Sasha J
    Moderator

    Austin,
    The problem happens because of timeout exceeding during the “yum install” execution. Command takes longer than defined timeout. This may be related to network speed, as all packages are downloaded from the HDP repository.
    You may avoid this by setting up your local repository server. http://hortonworks.com/download/thankyou_hdp1a/installing-the-software-via-local-repository/

    When you setup HDP on multiple nodes, each of them have less packages to install, and time to run “yum install” command is also less, so you should not hit timeout. However, it still may happen…

    So, you can setup local repository, of pre-install all packages on each node (you may have all packages included into the image you build systems from). Timeout is not configurable at this time.

    Next version planned to be released in August – September timeframe…

    Please, let me know if the any other things you may need help with.

    Thank you!
    Sasha

    Collapse
    #6642

    Hi Sasha,
    I have installed those packages and I have started the HMC, Thanks a lot, I didn’t get kicked by the puppet!
    1. But currently this is a single node. So I gather that if I am running this with multiple nodes I will have to do install of these packages on all the nodes before I start HMC?
    2. Why is this puppet kick error happening when I am selecting all the services?

    And when can I expect the next release of HMC?
    Thanks!
    Austin

    Collapse
    #6640

    Sasha J
    Moderator

    Austin,
    you can not add services later, this feature will be implemented in next HMC release.
    As of not, you should make installation from the beginning.
    In case if you hit puppet kick error again, there is a workaround:
    run the following command from the terminal before starting installation by HMC:

    yum install -y hadoop hadoop-libhdfs.x86_64 hadoop-native.x86_64 hadoop-pipes.x86_64 hadoop-sbin.x86_64 hadoop-lzo hadoop hadoop-libhdfs.i386 hadoop-native.i386 hadoop-pipes.i386 hadoop-sbin.i386 hadoop-lzo hive hcatalog oozie-client.noarch hdp_mon_dashboard hdp_mon_nagios_addons nagios-3.2.3 nagios-plugins-1.4.9 fping net-snmp-utils ganglia-gmetad-3.2.0 ganglia-gmond-3.2.0 gweb hdp_mon_ganglia_addons ganglia-gmond-3.2.0 hdp_mon_ganglia_addons snappy snappy-devel lzo lzo.i386 lzo-devel lzo-devel.i386 hadoop-secondarynamenode.x86_64

    This will install packages on your system, and then HMC will detect this and go directly to configuring and starting part.

    Please, try this way and let me know if there are still problems or questions.

    Thank you!
    Sasha

    Collapse
Viewing 22 replies - 1 through 22 (of 22 total)