hbase-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Gaojinchao <gaojinc...@huawei.com>
Subject Re: Hmaster is OutOfMemory
Date Thu, 12 May 2011 07:24:12 GMT
Thanks, Stack.
The heap dump is so big. I try to dig it and share the result. 
Then ,you can give me some suggestion.

Do you give me more suggestion about my cluster?

My application:
write operation with api put<list> is about 75k Puts/s( one put about 400 Bytes)
read opreation is rarely, but the latency time demands lower than 5s

cpu:    8 core 2.GHz
memory: 24G, Hbase use 8G
Disk:   2T*8 = 16T

node number: 13 nodes

dfs configure:

dfs.block.size 256M 
dfs.datanode.handler.count 10 
dfs.namenode.handler.count 30 
dfs.datanode.max.xcievers 2047 
dfs.support.append True 
hbase configure:

hbase.regionserver.handler.count 50 
hbase.regionserver.global.memstore.upperLimit 0.4 
hbase.regionserver.global.memstore.lowerLimit 0.35 
hbase.hregion.memstore.flush.size 128M 
hbase.hregion.max.filesize 512M 
hbase.client.scanner.caching 1 
hfile.block.cache.size 0.2 
hbase.hregion.memstore.block.multiplier 3 
hbase.hstore.blockingStoreFiles 10 
hbase.hstore.compaction.min.size 64M 

compress: gz

I am afraid of some problem:
1, One region server has about 12k reigons or more, if up the hregion.max.filesize , parallel
scalability will be low and affect scan latency .
2, If a region server crashed, It could bring on the other region server crashed?

Can you give some suggestion about the hbase parameter and my cluster ?

发件人: saint.ack@gmail.com [mailto:saint.ack@gmail.com] 代表 Stack
发送时间: 2011年5月10日 23:21
收件人: user@hbase.apache.org
主题: Re: Hmaster is OutOfMemory

2011/5/9 Gaojinchao <gaojinchao@huawei.com>:
> My first cluster needs save 147 TB data. If one region has 512M or 1 GB, It will be 300
K regions or 147K regions.
> In the future, If store several PB, It will more regions.

You might want to up the size of your regions to 4G (FB run w/ big
regions IIUC).

Do you want to put up a heap dump for me to take a look at.  That'd be
easier than me figuring time to try and replicate your scenario.

Thanks Gao,
View raw message