HDP on Linux – Installation Forum

Hive check execute failed HDP1.3

  • #41876
    Kobi Attia


    I ran the installation of cluster in ambari, everything was installed correctly until the point “Hive check execute”, I get this Error:
    notice: /Stage[2]/Hdp-hcat::Hcat::Service_check/Exec[hcatSmoke.sh prepare]/returns: SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
    notice: /Stage[2]/Hdp-hcat::Hcat::Service_check/Exec[hcatSmoke.sh prepare]/returns: SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
    notice: /Stage[2]/Hdp-hcat::Hcat::Service_check/Exec[hcatSmoke.sh prepare]/returns: FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. MetaException(message:javax.jdo.JDODataStoreException: An exception was thrown while adding/validating class(es) : Specified key was too long; max key length is 767 bytes
    notice: /Stage[2]/Hdp-hcat::Hcat::Service_check/Exec[hcatSmoke.sh prepare]/returns: com.mysql.jdbc.exceptions.MySQLSyntaxErrorException: Specified key was too long; max key length is 767 bytes
    notice: /Stage[2]/Hdp-hcat::Hcat::Service_check/Exec[hcatSmoke.sh prepare]/returns: at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:936)
    notice: /Stage[2]/Hdp-hcat::Hcat::Service_check/Exec[hcatSmoke.sh prepare]/returns: at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:2985)
    notice: /Stage[2]/Hdp-hcat::Hcat::Service_check/Exec[hcatSmoke.sh prepare]/returns: at com.mysql.jdbc.MysqlIO.sendCommand(MysqlIO.java:1631)
    notice: /Stage[2]/Hdp-hcat::Hcat::Service_check/Exec[hcatSmoke.sh prepare]/returns: at com.mysql.jdbc.MysqlIO.sqlQueryDirect(MysqlIO.java:1723)

    Thank for any help!


to create new topics or reply. | New User Registration

  • Author
  • #42228
    Kobi Attia


    Problem: Hive throws above exception when old version of MySQL is used as hive metastore.

    Solution: Set Latin1 as the charset for metastore
    mysql> alter database metastore character set latin1.


The topic ‘Hive check execute failed HDP1.3’ is closed to new replies.

Support from the Experts

A HDP Support Subscription connects you experts with deep experience running Apache Hadoop in production, at-scale on the most demanding workloads.

Enterprise Support »

Become HDP Certified

Real world training designed by the core architects of Hadoop. Scenario-based training courses are available in-classroom or online from anywhere in the world

Training »

Hortonworks Data Platform
The Hortonworks Data Platform is a 100% open source distribution of Apache Hadoop that is truly enterprise grade having been built, tested and hardened with enterprise rigor.
Get started with Sandbox
Hortonworks Sandbox is a self-contained virtual machine with Apache Hadoop pre-configured alongside a set of hands-on, step-by-step Hadoop tutorials.
Modern Data Architecture
Tackle the challenges of big data. Hadoop integrates with existing EDW, RDBMS and MPP systems to deliver lower cost, higher capacity infrastructure.