See if last item in FAQ fixes your issue Billy: http://wiki.apache.org/lucene-hadoop/Hbase/FAQ
St.Ack

Billy wrote:
I have tried to load hbase several times and always keep filing

2007-12-18 14:21:45,062 FATAL org.apache.hadoop.hbase.HRegionServer: Replay of hlog required. Forcing server restart org.apache.hadoop.hbase.DroppedSnapshotException: java.io.IOException: Too many open files

Thats the error I keep getting before all the problems start then when I shutdown and restart everything the region server complane about not being abile to find a block

2007-12-18 14:30:10,610 INFO org.apache.hadoop.fs.DFSClient: Could not obtain block blk_-8122648511302257739 from any node: java.io.IOException: No live nodes contain current block

2007-12-18 14:30:22,221 ERROR org.apache.hadoop.hbase.HRegionServer: Unable to close log in abort
java.io.IOException: java.io.IOException: Too many open files
        at java.lang.UNIXProcess.<init>(UNIXProcess.java:148)
        at java.lang.ProcessImpl.start(ProcessImpl.java:65)
        at java.lang.ProcessBuilder.start(ProcessBuilder.java:451)
        at java.lang.Runtime.exec(Runtime.java:591)
        at java.lang.Runtime.exec(Runtime.java:464)
at org.apache.hadoop.fs.ShellCommand.runCommand(ShellCommand.java:63)
        at org.apache.hadoop.fs.ShellCommand.run(ShellCommand.java:57)
        at org.apache.hadoop.fs.DF.getAvailable(DF.java:72)
at org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWrite(LocalDirAllocator.java:264) at org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.createTmpFileForWrite(LocalDirAllocator.java:294) at org.apache.hadoop.fs.LocalDirAllocator.createTmpFileForWrite(LocalDirAllocator.java:155) at org.apache.hadoop.dfs.DFSClient$DFSOutputStream.newBackupFile(DFSClient.java:1462) at org.apache.hadoop.dfs.DFSClient$DFSOutputStream.openBackupStream(DFSClient.java:1429) at org.apache.hadoop.dfs.DFSClient$DFSOutputStream.writeChunk(DFSClient.java:1571) at org.apache.hadoop.fs.FSOutputSummer.writeChecksumChunk(FSOutputSummer.java:141) at org.apache.hadoop.fs.FSOutputSummer.flushBuffer(FSOutputSummer.java:124) at org.apache.hadoop.dfs.DFSClient$DFSOutputStream.close(DFSClient.java:1724) at org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStream.java:49) at org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:64) at org.apache.hadoop.io.SequenceFile$Writer.close(SequenceFile.java:918)
        at org.apache.hadoop.hbase.HLog.close(HLog.java:390)
        at org.apache.hadoop.hbase.HRegionServer.run(HRegionServer.java:860)
        at java.lang.Thread.run(Thread.java:595)


I have my setup set to split the table at 16MB spits so I can do some testing the problem starts to show up with the table get about 6-7 splits. I have this on a 3 node setup 1 master/namenoe and 2 datanodes/region server I attached a tail of the last 5000 line of the regional servers I could not get much from the master log as it keep trying to connect and there is tons of data I scanned for WARN and FATAL but got nothing that looks like an error message.

Billy


Reply via email to