hbase-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Jonathan Gray <jg...@facebook.com>
Subject RE: Optimal block size for large columns
Date Tue, 18 May 2010 22:12:07 GMT
It would depend on your read patterns.

Is everything going to be single row gets, or will you also scan?

Single row lookups will be faster with smaller block sizes, at the expense of a larger index
size (and potentially slower scans as you have to deal with more block fetches).

> -----Original Message-----
> From: Jason Strutz [mailto:jason@cumuluscode.com]
> Sent: Tuesday, May 18, 2010 9:33 AM
> To: hbase-user@hadoop.apache.org
> Subject: Optimal block size for large columns
> I am working with a small cluster, trying to nail down appropriate
> settings for block size.  We will have a single table with a single
> column of data averaging 300k in size, sometimes upwards of 2mb, never
> more than 10mb.
> Is there any rule-of-thumb or other sage advice for block sizes for
> large columns?
> Thanks!

View raw message