MapReduce Forum

MapReduce2 Service Check has failures

  • #48440
    Pavel Hladik

    Hi, I would like to ask you why MapReduce2 Service Check failed even I have working (green) HDP2 services, MapReduce2 as well. There is a log:

    err: /Stage[2]/Hdp-yarn::Mapred2::Service_check/Hdp-hadoop::Exec-hadoop[mapred::service_check::run_wordcount]/Hdp::Exec[hadoop –config /etc/hadoop/conf jar /usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples-2.*.jar wordcount /user/ambari-qa/mapredsmokeinput /user/ambari-qa/mapredsmokeoutput]/Exec[hadoop –config /etc/hadoop/conf jar /usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples-2.*.jar wordcount /user/ambari-qa/mapredsmokeinput /user/ambari-qa/mapredsmokeoutput]/returns: change from notrun to 0 failed: hadoop –config /etc/hadoop/conf jar /usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples-2.*.jar wordcount /user/ambari-qa/mapredsmokeinput /user/ambari-qa/mapredsmokeoutput returned 255 instead of one of [0] at /var/lib/ambari-agent/puppet/modules/hdp/manifests/init.pp:487

to create new topics or reply. | New User Registration

  • Author
  • #48530
    Pavel Hladik

    I use HDP 2.0.6 with latest Ambari on CentOS 6.5. I have no that issue on test environment, when I was using dedicated host for client install, in production environment I installed client on master host and later to slave host.

    Pavel Hladik

    I found this issue is related to setting property dfs.replication.max to 4 (50 from default Ambari 1.4.4 installation and 512 from Hortonworks docs).

    We have 6 nodes, so value 4 will be enough.

You must be to reply to this topic. | Create Account

Support from the Experts

A HDP Support Subscription connects you experts with deep experience running Apache Hadoop in production, at-scale on the most demanding workloads.

Enterprise Support »

Become HDP Certified

Real world training designed by the core architects of Hadoop. Scenario-based training courses are available in-classroom or online from anywhere in the world

Training »

Hortonworks Data Platform
The Hortonworks Data Platform is a 100% open source distribution of Apache Hadoop that is truly enterprise grade having been built, tested and hardened with enterprise rigor.
Get started with Sandbox
Hortonworks Sandbox is a self-contained virtual machine with Apache Hadoop pre-configured alongside a set of hands-on, step-by-step Hadoop tutorials.
Modern Data Architecture
Tackle the challenges of big data. Hadoop integrates with existing EDW, RDBMS and MPP systems to deliver lower cost, higher capacity infrastructure.