hbase-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Nanheng Wu <nanhen...@gmail.com>
Subject How to implement efficient bulk query
Date Fri, 22 Jul 2011 13:46:41 GMT

  I have an use case for my data stored in HBase where I need to make
a query for 20K-30K keys at once. I know that the HBase client API
supports get operation with a list of "gets", so a naive
implementation would probably just make one or more batch get calls.
First of all I am wondering if I choose this implementation how should
I choose the batch size? Can I put all the keys in a single batch?
Secondly, is there a better implementation that's more efficient? For
instance I can sort the keys first and split them into groups of a
certain size, for each group do a scan using the first and last key of
the group and filter out retuned rows that are not in the group (kinda
like a merge join). Would the second implementation be faster that the
first? Are there better ways to go about it? I am using HBase 0.20.6.


View raw message