Sqoop Forum

Trouble with Teradata Connector

  • #34119
    Chris Schrader
    Participant

    I’m trying to test out the teradata connector in my sandbox environment by doing a simple extract of dbc.tables to a file. The problem I’m running into seems to be related to Sqoop or the TD adaptor trying to an “rm” instead of an “rm -r” on the staging directory it creates on hdfs.

    I’ve gotten it to connect to Teradata and display tables, run queries, etc. Just when trying to export data to file seems to be an issue.

    Command I am using:


    sqoop import \
    -libjars $LIB_JARS \
    -Dteradata.db.input.job.type=hdfs \
    -Dteradata.db.input.source.table=dbc.tables \
    -Dteradata.db.input.target.paths=/users/chris \
    --connect jdbc:***MyConnectString*** \
    --connection-manager org.apache.sqoop.teradata.TeradataConnManager \
    --table=dbc.tables\
    --username ***Myuser*** \
    --password ***mypass***\
    --target-dir /user/chris \
    --split-by databasename

    Output I am getting:


    13/09/03 10:17:39 WARN tool.BaseSqoopTool: Setting your password on the command-line is insecure. Consider using -P instead.
    13/09/03 10:17:39 INFO manager.SqlManager: Using default fetchSize of 1000
    13/09/03 10:17:39 INFO tool.CodeGenTool: Beginning code generation
    13/09/03 10:17:43 INFO manager.SqlManager: Executing SQL statement: SELECT t.* FROM dbc.tables AS t WHERE 1=0
    13/09/03 10:17:44 INFO orm.CompilationManager: HADOOP_MAPRED_HOME is /usr/lib/hadoop
    13/09/03 10:17:44 INFO orm.CompilationManager: Found hadoop core jar at: /usr/lib/hadoop/hadoop-core.jar
    Note: /tmp/sqoop-hdfs/compile/2133daa3bc737b3d7257cfea97a9fc47/dbc_tables.java uses or overrides a deprecated API.
    Note: Recompile with -Xlint:deprecation for details.
    ....

    13/09/03 10:18:02 INFO mapred.JobClient: Cleaning up the staging area hdfs://sandbox:8020/user/hdfs/.staging/job_201308280904_0005
    13/09/03 10:18:02 ERROR security.UserGroupInformation: PriviledgedActionException as:hdfs cause:java.io.IOException: Target hdfs://sandbox:8020/user/hdfs/.staging/job_201308280904_0005/libjars/conf/conf is a directory
    13/09/03 10:18:02 INFO mapreduce.TeradataInputProcessor: job cleanup starts at 1378228682948
    13/09/03 10:18:04 INFO mapreduce.TeradataInputProcessor: job cleanup ends at 1378228684426
    13/09/03 10:18:04 INFO mapreduce.TeradataInputProcessor: job cleanup time is 1s
    13/09/03 10:18:04 ERROR teradata.TeradataSqoopImportJob: Exception running Teradata import job
    com.teradata.hadoop.exception.TeradataHadoopException: java.io.IOException: Target hdfs://sandbox:8020/user/hdfs/.staging/job_201308280904_0005/libjars/conf/conf is a directory
    3/09/03 10:18:02 INFO mapreduce.TeradataInputProcessor: job cleanup starts at 1378228682948
    13/09/03 10:18:04 INFO mapreduce.TeradataInputProcessor: job cleanup ends at 1378228684426
    13/09/03 10:18:04 INFO mapreduce.TeradataInputProcessor: job cleanup time is 1s
    13/09/03 10:18:04 ERROR teradata.TeradataSqoopImportJob: Exception running Teradata import job
    com.teradata.hadoop.exception.TeradataHadoopException: java.io.IOException: Target hdfs

to create new topics or reply. | New User Registration

  • Author
    Replies
  • #37764
    Jack Liu
    Member

    Hi Chris Schrader,
    I wonder what is your hdp version?
    I failed to find the class org.apache.sqoop.teradata.TeradataConnManager, could you please share where does the jar come from?

    and I am using hdp 2.1.x.

    Appreciate so much for your help.

    #40808
    Robert Molina
    Moderator

    Hi Chris,
    What user are you running the sqoop job as? Have you tried running it as hdfs user?

    Regards,
    Robert

The topic ‘Trouble with Teradata Connector’ is closed to new replies.

Support from the Experts

A HDP Support Subscription connects you experts with deep experience running Apache Hadoop in production, at-scale on the most demanding workloads.

Enterprise Support »

Become HDP Certified

Real world training designed by the core architects of Hadoop. Scenario-based training courses are available in-classroom or online from anywhere in the world

Training »

Hortonworks Data Platform
The Hortonworks Data Platform is a 100% open source distribution of Apache Hadoop that is truly enterprise grade having been built, tested and hardened with enterprise rigor.
Get started with Sandbox
Hortonworks Sandbox is a self-contained virtual machine with Apache Hadoop pre-configured alongside a set of hands-on, step-by-step Hadoop tutorials.
Modern Data Architecture
Tackle the challenges of big data. Hadoop integrates with existing EDW, RDBMS and MPP systems to deliver lower cost, higher capacity infrastructure.