hbase-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From 苏铖 <such...@lietou.com>
Subject 答复: How to adjust hbase settings when too many store files?
Date Mon, 29 Oct 2012 07:52:31 GMT
I checked the region server log again, and I found something below:

2012-10-28 06:24:24,811 ERROR
org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest:
Compaction failed regionName=stati
stic_visit_detail,20120922|13984|451728,1351376659451.9b2bfae5d77109693a153e
b16fcb7793., storeName=cf1, fileCount=7, fileSize=1.1g (
681.7m, 168.6m, 139.9m, 36.2m, 53.0m, 26.5m, 5.9m), priority=0,
time=469259302083252
java.io.IOException: java.io.IOException: File
/hbase/statistic_visit_detail/9b2bfae5d77109693a153eb16fcb7793/.tmp/3a3e6ee8
8a524659b
9f9716e5ca21a74 could only be replicated to 0 nodes, instead of 1
        at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNam
esystem.java:1531)
        at
org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:685)
        at sun.reflect.GeneratedMethodAccessor7.invoke(Unknown Source)
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
        at java.lang.reflect.Method.invoke(Unknown Source)
        at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:563)
        at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
        at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)

It seems that, the fileSize already exceed the max size, which is 1G as
default.

So I add more region servers and enlarge the max file size, to see if it
works.

Thanks a lot.

-----邮件原件-----
发件人: Ramkrishna.S.Vasudevan [mailto:ramkrishna.vasudevan@huawei.com] 
发送时间: 2012年10月29日 14:06
收件人: user@hbase.apache.org
主题: RE: How to adjust hbase settings when too many store files?

Also check what is your heap size of RS?
When you say hTable.put(), how many such threads are there?

What is your region size?  Is your regions splitting continuously do to
heavy load?

Regards
Ram


> -----Original Message-----
> From: yuzhihong@gmail.com [mailto:yuzhihong@gmail.com]
> Sent: Monday, October 29, 2012 10:01 AM
> To: user@hbase.apache.org
> Cc: <user@hbase.apache.org>
> Subject: Re: How to adjust hbase settings when too many store files?
> 
> What version of hbase were you using ?
> Did you pre split the table before loading ?
> 
> Thanks
> 
> 
> 
> On Oct 28, 2012, at 8:33 PM, 苏铖 <sucheng@lietou.com> wrote:
> 
> > Hello. I encounter a region server error when I try to put bulk data
> from a
> > java client.
> >
> > The java client extracts data from a relational database and puts
> those data
> > into hbase.
> >
> > When I try to extract data from a large table(say, 1 billion
> records), the
> > error happens.
> >
> >
> >
> > The region server's log says:
> >
> >
> >
> >> 2012-10-28 00:00:02,169 WARN
> > org.apache.hadoop.hbase.regionserver.MemStoreFlusher: Region
> > statistic_visit_detail,20120804|72495|8549
> >
> > 56,1351353594195.ad2592ee7a3610c60c47cf8be77496c8. has too many store
> files;
> > delaying flush up to 90000ms
> >
> >> 2012-10-28 00:00:02,791 DEBUG
> > org.apache.hadoop.hbase.regionserver.MemStoreFlusher: Flush thread
> woke up
> > because memory above low wa
> >
> > ter=347.1m
> >
> >> 2012-10-28 00:00:02,791 DEBUG
> > org.apache.hadoop.hbase.regionserver.MemStoreFlusher: Under global
> heap
> > pressure: Region statistic_vis
> >
> >
> it_detail,20120804|72495|854956,1351353594195.ad2592ee7a3610c60c47cf8be
> 77496
> > c8. has too many store files, but is 141.5m vs best flus
> >
> > hable region's 46.8m. Choosing the bigger.
> >
> >> 2012-10-28 00:00:02,791 INFO
> > org.apache.hadoop.hbase.regionserver.MemStoreFlusher: Flush of region
> > statistic_visit_detail,20120804|7
> >
> > 2495|854956,1351353594195.ad2592ee7a3610c60c47cf8be77496c8. due to
> global
> > heap pressure
> >
> > ...
> >
> >
> >
> > And finally,
> >
> >
> >
> >> 2012-10-28 00:00:43,511 INFO
> org.apache.hadoop.hbase.regionserver.HRegion:
> > compaction interrupted by user
> >
> >> java.io.InterruptedIOException: Aborting compaction of store cf1 in
> region
> > statistic_visit_detail,20120804|72495|854956,135135359419
> >
> > 5.ad2592ee7a3610c60c47cf8be77496c8. because user requested stop.
> >
> >        at
> >
> org.apache.hadoop.hbase.regionserver.Store.compactStore(Store.java:1275
> )
> >
> >        at
> > org.apache.hadoop.hbase.regionserver.Store.compact(Store.java:765)
> >
> >        at
> >
> org.apache.hadoop.hbase.regionserver.HRegion.compact(HRegion.java:1023)
> >
> >        at
> >
> org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest.run(
> Compa
> > ctionRequest.java:177)
> >
> >        at
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecut
> or.ja
> > va:886)
> >
> >        at
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.j
> ava:9
> > 08)
> >
> >        at java.lang.Thread.run(Thread.java:662)
> >
> >
> >
> > Then the region server shuts down.
> >
> >
> >
> > It seems that too many store files(due to too many records from
> > relational-db) consumed too many memories, if I'm right.
> >
> > I'm new to hbase, what settings should I adjust? Or even increase
> region
> > servers?
> >
> > I'm going to do some research by myself, and any advise will be
> appreciated.
> >
> > Best regards,
> >
> >
> >
> > Su
> >
> >
> >



Mime
View raw message