hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From stack <st...@duboce.net>
Subject Re: hbase error - Too many open files
Date Tue, 18 Dec 2007 20:52:14 GMT
See if last item in FAQ fixes your issue Billy: 
http://wiki.apache.org/lucene-hadoop/Hbase/FAQ
St.Ack

Billy wrote:
> I have tried to load hbase several times and always keep filing
>
> 2007-12-18 14:21:45,062 FATAL org.apache.hadoop.hbase.HRegionServer: Replay 
> of hlog required. Forcing server restart
> org.apache.hadoop.hbase.DroppedSnapshotException: java.io.IOException: Too 
> many open files
>
> Thats the error I keep getting before all the problems start then when I 
> shutdown and restart everything the region server complane about not being 
> abile to find a block
>
> 2007-12-18 14:30:10,610 INFO org.apache.hadoop.fs.DFSClient: Could not 
> obtain block blk_-8122648511302257739 from any node:  java.io.IOException: 
> No live nodes contain current block
>
> 2007-12-18 14:30:22,221 ERROR org.apache.hadoop.hbase.HRegionServer: Unable 
> to close log in abort
> java.io.IOException: java.io.IOException: Too many open files
>         at java.lang.UNIXProcess.<init>(UNIXProcess.java:148)
>         at java.lang.ProcessImpl.start(ProcessImpl.java:65)
>         at java.lang.ProcessBuilder.start(ProcessBuilder.java:451)
>         at java.lang.Runtime.exec(Runtime.java:591)
>         at java.lang.Runtime.exec(Runtime.java:464)
>         at 
> org.apache.hadoop.fs.ShellCommand.runCommand(ShellCommand.java:63)
>         at org.apache.hadoop.fs.ShellCommand.run(ShellCommand.java:57)
>         at org.apache.hadoop.fs.DF.getAvailable(DF.java:72)
>         at 
> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWrite(LocalDirAllocator.java:264)
>         at 
> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.createTmpFileForWrite(LocalDirAllocator.java:294)
>         at 
> org.apache.hadoop.fs.LocalDirAllocator.createTmpFileForWrite(LocalDirAllocator.java:155)
>         at 
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.newBackupFile(DFSClient.java:1462)
>         at 
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.openBackupStream(DFSClient.java:1429)
>         at 
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.writeChunk(DFSClient.java:1571)
>         at 
> org.apache.hadoop.fs.FSOutputSummer.writeChecksumChunk(FSOutputSummer.java:141)
>         at 
> org.apache.hadoop.fs.FSOutputSummer.flushBuffer(FSOutputSummer.java:124)
>         at 
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.close(DFSClient.java:1724)
>         at 
> org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStream.java:49)
>         at 
> org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:64)
>         at 
> org.apache.hadoop.io.SequenceFile$Writer.close(SequenceFile.java:918)
>         at org.apache.hadoop.hbase.HLog.close(HLog.java:390)
>         at org.apache.hadoop.hbase.HRegionServer.run(HRegionServer.java:860)
>         at java.lang.Thread.run(Thread.java:595)
>
>
> I have my setup set to split the table at 16MB spits so I can do some 
> testing the problem starts to show up with the table get about 6-7 splits. I 
> have this on a 3 node setup 1 master/namenoe and 2 datanodes/region server I 
> attached a tail of the last 5000 line of the regional servers I could not 
> get much from the master log as it keep trying to connect and there is tons 
> of data I scanned for WARN and FATAL but got nothing that looks like an 
> error message.
>
> Billy 
>
>
>   


Mime
View raw message