The Hortonworks Community Connection is now live. A completely rebuilt Q&A forum, Knowledge Base, Code Hub and more, backed by the experts in the industry.

You will be redirected here in 10 seconds. If your are not redirected, click here to visit the new site.

The legacy Hortonworks Forum is now closed. You can view a read-only version of the former site by clicking here. The site will be taken offline on January 31,2016

Hortonworks Sandbox Forum

importing data with file browser times out

  • #27603
    alex Gordon

    I am trying to upload a 400 mb google ngram zip file using the file browser into the sandbox.

    I’m getting this error:

    The following error(s) occurred:
    timed out

    Is there a limit on the size we can import? What am I doing wrong?

  • Author
  • #27604
    alex Gordon

    I apologize, my description of the problem is inaccurate.

    I actually was able to upload 100% of the zip file.

    However, right after uploading it, it gets stuck on the unzip:
    Uploading to: /user/hue
    The file will then be extracted in the path specified above.

    at this point it just errors out.

    are we not allowed to import large zip files?


    Hi Alex,

    As far as I know there isn’t a limit to the files you upload, only the space on the disk. if the file is 400MB when zipped how much space is it going to take when unzipped? On a new Sandbox there is just under 40GB of space in HDFS.


The forum ‘Hortonworks Sandbox’ is closed to new topics and replies.

Support from the Experts

A HDP Support Subscription connects you experts with deep experience running Apache Hadoop in production, at-scale on the most demanding workloads.

Enterprise Support »

Become HDP Certified

Real world training designed by the core architects of Hadoop. Scenario-based training courses are available in-classroom or online from anywhere in the world

Training »

Hortonworks Data Platform
The Hortonworks Data Platform is a 100% open source distribution of Apache Hadoop that is truly enterprise grade having been built, tested and hardened with enterprise rigor.
Get started with Sandbox
Hortonworks Sandbox is a self-contained virtual machine with Apache Hadoop pre-configured alongside a set of hands-on, step-by-step Hadoop tutorials.
Modern Data Architecture
Tackle the challenges of big data. Hadoop integrates with existing EDW, RDBMS and MPP systems to deliver lower cost, higher capacity infrastructure.