Hive / HCatalog Forum

MapReduce Error when using HCatInputFormat process data

  • #28822
    Andy Miao
    Participant

    HI guys,
    I tried to create map/reduce program to practice Hcatalog, I followed the guidence on http://incubator.apache.org/hcatalog/docs/r0.5.0/inputoutput.html Site.
    The Map/Reduce program just read data from HCatalog table and wirte to another one. But the simple program always raise error on following code snipt
    HCatInputFormat.setInput(job, InputJobInfo.create(DbName,InputTbName,null));

    Here is the dtailed Error Message:
    ClassNotFoundException:org.apache.hadoop.hive.ql.metadataq.hiveStorageHandler.

    I have already package hive-exec.jar into my map/Reduce program.

    MRCommand:
    hadoop jar d:\hcat\HCatInterface.jar hcatalog.HCatInterfacePOC student studentg
    BTW, What i am using is HDP for windows.

to create new topics or reply. | New User Registration

  • Author
    Replies
  • #28824
    abdelrahman
    Moderator

    Hi Andy,

    It is most likely because it can’t find some Hive jars. You may want to try to add the Hive jars into the CLASSPATH.

    Thanks
    -Abdelrahman

You must be to reply to this topic. | Create Account

Support from the Experts

A HDP Support Subscription connects you experts with deep experience running Apache Hadoop in production, at-scale on the most demanding workloads.

Enterprise Support »

Become HDP Certified

Real world training designed by the core architects of Hadoop. Scenario-based training courses are available in-classroom or online from anywhere in the world

Training »

Hortonworks Data Platform
The Hortonworks Data Platform is a 100% open source distribution of Apache Hadoop that is truly enterprise grade having been built, tested and hardened with enterprise rigor.
Get started with Sandbox
Hortonworks Sandbox is a self-contained virtual machine with Apache Hadoop pre-configured alongside a set of hands-on, step-by-step Hadoop tutorials.
Modern Data Architecture
Tackle the challenges of big data. Hadoop integrates with existing EDW, RDBMS and MPP systems to deliver lower cost, higher capacity infrastructure.