Sqoop import failed after initial import

to create new topics or reply. | New User Registration

This topic contains 3 replies, has 2 voices, and was last updated by  Seth Lyubich 1 year, 11 months ago.

  • Creator
  • #31577

    I’ve imported a table from SQL Server 2008 R2 to Hive using SQOOP. It worked but when I dropped the table in Hive and reimport I’m getting following error
    13/08/13 09:56:17 INFO tool.BaseSqoopTool: Using Hive-specific delimiters for o
    tput. You can override
    13/08/13 09:56:17 INFO tool.BaseSqoopTool: delimiters with –fields-terminated-
    y, etc.
    13/08/13 09:56:17 INFO manager.SqlManager: Using default fetchSize of 1000
    13/08/13 09:56:17 INFO tool.CodeGenTool: Beginning code generation
    13/08/13 09:56:17 INFO manager.SqlManager: Executing SQL statement: SELECT t.*
    ROM [MPTest] AS t WHERE 1=0
    13/08/13 09:56:18 INFO orm.CompilationManager: HADOOP_HOME is C:\hadoop\hadoop-
    13/08/13 09:56:18 INFO orm.CompilationManager: Found hadoop core jar at: C:\had
    Note: \tmp\sqoop-SBalakumar\compile\94b485ecc82edaa79223f07782127527\MPTest.java uses or overrides a deprecated API.
    Note: Recompile with -Xlint:deprecation for details.
    13/08/13 09:56:21 INFO orm.CompilationManager: Writing jar file: \tmp\sqoop-SBa
    13/08/13 09:56:21 INFO mapreduce.ImportJobBase: Beginning import of MPTest
    13/08/13 09:56:24 INFO mapred.JobClient: Cleaning up the staging area hdfs://AD
    13/08/13 09:56:24 ERROR security.UserGroupInformation: PriviledgedActionExcepti
    n as:SBalakumar cause:org.apache.hadoop.mapred.FileAlreadyExistsException: Outp
    t directory MPTest already exists

    13/08/13 09:56:24 ERROR tool.ImportTool: Encountered IOException running import
    job: org.apache.hadoop.mapred.FileAlreadyExistsException: Output directory MPTest already exists
    at org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.checkOutputS
    at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:887)
    at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:850)
    at java.security.AccessController.doPrivileged(Native Method)
    at javax.security.auth.Subject.doAs(Subject.java:396)
    at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInform
    at org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:
    at org.apache.hadoop.mapreduce.Job.submit(Job.java:500)
    at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:530)
    at org.apache.sqoop.mapreduce.ImportJobBase.doSubmitJob(ImportJobBase.j
    at org.apache.sqoop.mapreduce.ImportJobBase.runJob(ImportJobBase.java:1
    at org.apache.sqoop.mapreduce.ImportJobBase.runImport(ImportJobBase.jav
    at org.apache.sqoop.manager.SqlManager.importTable(SqlManager.java:465)
    at org.apache.sqoop.manager.SQLServerManager.importTable(SQLServerManag
    at org.apache.sqoop.tool.ImportTool.importTable(ImportTool.java:403)
    at org.apache.sqoop.tool.ImportTool.run

Viewing 3 replies - 1 through 3 (of 3 total)

You must be to reply to this topic. | Create Account

  • Author
  • #32665

    Seth Lyubich

    Hi Sundar,

    Good to know that your issue is resolved. Can you please open new thread for Groovy issue?



    Thanks. I deleted the directory and it is working fine. I’m trying to add a xml mapper in Groovy to Transform XML in hive Select. I’m getting an error
    Caused by: java.io.IOException: Cannot run program “C:\BigData\sampleMapper.groovy”: CreateProcess error=193, %1 is not a valid Win32 application
    at java.lang.ProcessBuilder.start(ProcessBuilder.java:460)
    at org.apache.hadoop.hive.ql.exec.ScriptOperator.processOp(ScriptOperator.java:287)
    … 18 more
    Caused by: java.io.IOException: CreateProcess error=193, %1 is not a valid Win32 application
    at java.lang.ProcessImpl.create(Native Method)
    at java.lang.ProcessImpl.(ProcessImpl.java:81)
    at java.lang.ProcessImpl.start(ProcessImpl.java:30)
    at java.lang.ProcessBuilder.start(ProcessBuilder.java:453)
    … 19 more

    Can you please let me know how to use Groovy script in Hive?


    Seth Lyubich


    Looks like the error is :
    Output directory MPTest already exists

    Can you please try specifying new directory?

    Hope this helps,


Viewing 3 replies - 1 through 3 (of 3 total)
Hortonworks Data Platform
The Hortonworks Data Platform is a 100% open source distribution of Apache Hadoop that is truly enterprise grade having been built, tested and hardened with enterprise rigor.
Get started with Sandbox
Hortonworks Sandbox is a self-contained virtual machine with Apache Hadoop pre-configured alongside a set of hands-on, step-by-step Hadoop tutorials.
Modern Data Architecture
Tackle the challenges of big data. Hadoop integrates with existing EDW, RDBMS and MPP systems to deliver lower cost, higher capacity infrastructure.