HiveColumnarLoader not loading int columns

to create new topics or reply. | New User Registration

This topic contains 1 reply, has 2 voices, and was last updated by  Thejas Nair 1 year, 2 months ago.

  • Creator
  • #52648

    Greg Smorag

    I have a problem with weird behaviour of HiveColumnadLoader and I wonder, if anybody can help me with that?

    The issue is, when I am attempting to load data from my hive table via pig, which looks more of less like the code below, pig never loads int columns. Everything is completed successfully, but the output of dumping Z, which has null values in place of ints. When I change ‘dump Z’ into something else, e.g. Mongo DB store, if just saves nulls in place of int values.

    register /opt/hadoop/pig/contrib/piggybank/java/piggybank.jar;
    register /opt/hadoop/hive/lib/hive-common-0.12.0.jar;
    register /opt/hadoop/hive/lib/hive-exec-0.12.0.jar;
    A = load '<VALID_HIVE_PATH>/<HIVE_TABLE>' USING'column_a int,column_b int,column_c string,column_d string,column_e int');
    L = foreach A generate *;
    Z = filter L by fk_client == '<PARTITION VALUE>';
    dump Z;

    HIVE_TABLE is a hive partitioned tableof type RCfile. The partition denominator column is fk_client column.

    I am using hadoop 2.3, hive 0.12.0 and pig 0.12.1.

    Can you please help with me and point out, what might be the reason or what I am possibily doing wrong, as I am running out of options?

    Thank you very much.

Viewing 1 replies (of 1 total)

You must be to reply to this topic. | Create Account

  • Author
  • #52742

    Thejas Nair

    I doubt if HiveColumnadLoader is widely used. You might want to try hcatalog’s HCatLoader with pig instead.

Viewing 1 replies (of 1 total)
Hortonworks Data Platform
The Hortonworks Data Platform is a 100% open source distribution of Apache Hadoop that is truly enterprise grade having been built, tested and hardened with enterprise rigor.
Get started with Sandbox
Hortonworks Sandbox is a self-contained virtual machine with Apache Hadoop pre-configured alongside a set of hands-on, step-by-step Hadoop tutorials.
Modern Data Architecture
Tackle the challenges of big data. Hadoop integrates with existing EDW, RDBMS and MPP systems to deliver lower cost, higher capacity infrastructure.