Bulk read from Hbase table

to create new topics or reply. | New User Registration

This topic contains 2 replies, has 2 voices, and was last updated by  SambaSivaRao Y 1 year, 2 months ago.

  • Creator
  • #51885

    SambaSivaRao Y


    We need to read Hbase table rows of volume about 100Million.
    For mapper, It is taking 2 minutes to read 1Lakh rows on a single machine.
    We are using “TableMapReduceUtil” for reading from the Hbase table.

    Could you please help us to tune this job.


Viewing 2 replies - 1 through 2 (of 2 total)

You must be to reply to this topic. | Create Account

  • Author
  • #54886

    SambaSivaRao Y

    Thanks for your reply CaoYu, I’ll check and let you know.


    Cao Yu

    Are you rows is seriation ?
    if you rows is seriation,you can using HbaseAPI.
    Java class for “scan” and function for setStartKey,setEndKey

    that can be query a row region

    English is not good,don’t care my syntax error,just you can understand is nice

Viewing 2 replies - 1 through 2 (of 2 total)
Hortonworks Data Platform
The Hortonworks Data Platform is a 100% open source distribution of Apache Hadoop that is truly enterprise grade having been built, tested and hardened with enterprise rigor.
Get started with Sandbox
Hortonworks Sandbox is a self-contained virtual machine with Apache Hadoop pre-configured alongside a set of hands-on, step-by-step Hadoop tutorials.
Modern Data Architecture
Tackle the challenges of big data. Hadoop integrates with existing EDW, RDBMS and MPP systems to deliver lower cost, higher capacity infrastructure.