Bulk read from Hbase table

This topic contains 2 replies, has 2 voices, and was last updated by  SambaSivaRao Y 9 months ago.

  • Creator
    Topic
  • #51885

    SambaSivaRao Y
    Participant

    Hi,

    We need to read Hbase table rows of volume about 100Million.
    For mapper, It is taking 2 minutes to read 1Lakh rows on a single machine.
    We are using “TableMapReduceUtil” for reading from the Hbase table.

    Could you please help us to tune this job.

    Thanks,
    SambaShiva

Viewing 2 replies - 1 through 2 (of 2 total)

You must be to reply to this topic. | Create Account

  • Author
    Replies
  • #54886

    SambaSivaRao Y
    Participant

    Thanks for your reply CaoYu, I’ll check and let you know.

    Collapse
    #54432

    Cao Yu
    Participant

    Hi,
    Are you rows is seriation ?
    if you rows is seriation,you can using HbaseAPI.
    Java class for “scan” and function for setStartKey,setEndKey

    that can be query a row region

    CaoYu
    ——————-
    English is not good,don’t care my syntax error,just you can understand is nice

    Collapse
Viewing 2 replies - 1 through 2 (of 2 total)
Hortonworks Data Platform
The Hortonworks Data Platform is a 100% open source distribution of Apache Hadoop that is truly enterprise grade having been built, tested and hardened with enterprise rigor.
Get started with Sandbox
Hortonworks Sandbox is a self-contained virtual machine with Apache Hadoop pre-configured alongside a set of hands-on, step-by-step Hadoop tutorials.
Modern Data Architecture
Tackle the challenges of big data. Hadoop integrates with existing EDW, RDBMS and MPP systems to deliver lower cost, higher capacity infrastructure.