[ https://issues.apache.org/jira/browse/HADOOP-2040?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12541356 ]
stack commented on HADOOP-2040: ------------------------------- Hudson is hung. Here is tail of log. {code} [junit] 2007-11-09 08:20:49,385 DEBUG [main] org.apache.hadoop.hbase.TestLogRolling.countLogFiles(TestLogRolling.java:174): number of log files: 1 [junit] 2007-11-09 08:20:49,386 INFO [main] org.apache.hadoop.hbase.TestLogRolling.testLogRolling(TestLogRolling.java:191): Finished writing. There are 1 log files. Sleeping to let cache flusher and log roller run [junit] 2007-11-09 08:20:49,386 DEBUG [main] org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:202): Shutting down HBase Cluster [junit] 2007-11-09 08:20:49,488 DEBUG [RegionServer:0] org.apache.hadoop.hbase.HRegionServer.run(HRegionServer.java:502): Got regionserver stop message [junit] 2007-11-09 08:20:49,488 INFO [RegionServer:0] org.apache.hadoop.hbase.Leases.close(Leases.java:109): RegionServer:0 closing leases [junit] 2007-11-09 08:20:49,489 INFO [RegionServer:0.leaseChecker] org.apache.hadoop.hbase.Chore.run(Chore.java:62): RegionServer:0.leaseChecker exiting [junit] 2007-11-09 08:20:49,489 INFO [RegionServer:0] org.apache.hadoop.hbase.Leases.close(Leases.java:123): RegionServer:0 closed leases [junit] 2007-11-09 08:20:49,490 INFO [RegionServer:0.logRoller] org.apache.hadoop.hbase.Chore.run(Chore.java:62): RegionServer:0.logRoller exiting [junit] 2007-11-09 08:20:49,607 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,608 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,608 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,608 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,608 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,608 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,609 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,609 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,609 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,609 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,609 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,610 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,610 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,610 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,610 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,610 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,611 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,611 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,611 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,611 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,612 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,612 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,612 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,612 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,612 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,613 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,613 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,613 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,613 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,613 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,614 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,614 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,614 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,614 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,615 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,615 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,615 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,615 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,814 WARN [IPC Server handler 5 on 58346] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,830 DEBUG [RegionServer:0.cacheFlusher] org.apache.hadoop.hbase.HStore.flushCacheHelper(HStore.java:504): Added -1547818355/info/8261001142386214874 with sequence id 2208 and size 16.8k [junit] 2007-11-09 08:20:49,830 DEBUG [RegionServer:0.cacheFlusher] org.apache.hadoop.hbase.HRegion.internalFlushcache(HRegion.java:919): Finished memcache flush for region testLogRolling,row1025,1194596368242 in 523ms [junit] 2007-11-09 08:20:49,831 DEBUG [RegionServer:0] org.apache.hadoop.hbase.HRegionServer.closeAllRegions(HRegionServer.java:971): closing region -ROOT-,,0 [junit] 2007-11-09 08:20:49,831 INFO [RegionServer:0.splitOrCompactChecker] org.apache.hadoop.hbase.Chore.run(Chore.java:62): RegionServer:0.splitOrCompactChecker exiting [junit] 2007-11-09 08:20:49,832 DEBUG [RegionServer:0] org.apache.hadoop.hbase.HRegion.internalFlushcache(HRegion.java:847): Started memcache flush for region -ROOT-,,0. Size 0.0 [junit] 2007-11-09 08:20:49,832 DEBUG [RegionServer:0] org.apache.hadoop.hbase.HRegion.internalFlushcache(HRegion.java:865): Finished memcache flush; empty snapshot [junit] 2007-11-09 08:20:49,833 DEBUG [RegionServer:0] org.apache.hadoop.hbase.HStore.close(HStore.java:419): closed -70236052/info [junit] 2007-11-09 08:20:49,833 INFO [RegionServer:0] org.apache.hadoop.hbase.HRegion.close(HRegion.java:402): closed -ROOT-,,0 [junit] 2007-11-09 08:20:49,833 DEBUG [RegionServer:0] org.apache.hadoop.hbase.HRegionServer.closeAllRegions(HRegionServer.java:971): closing region .META.,,1 [junit] 2007-11-09 08:20:49,833 DEBUG [RegionServer:0] org.apache.hadoop.hbase.HRegion.internalFlushcache(HRegion.java:847): Started memcache flush for region .META.,,1. Size 0.0 [junit] 2007-11-09 08:20:49,833 DEBUG [RegionServer:0] org.apache.hadoop.hbase.HRegion.internalFlushcache(HRegion.java:865): Finished memcache flush; empty snapshot [junit] 2007-11-09 08:20:49,833 DEBUG [RegionServer:0] org.apache.hadoop.hbase.HStore.close(HStore.java:419): closed 1028785192/info [junit] 2007-11-09 08:20:49,834 INFO [RegionServer:0] org.apache.hadoop.hbase.HRegion.close(HRegion.java:402): closed .META.,,1 [junit] 2007-11-09 08:20:49,834 DEBUG [RegionServer:0] org.apache.hadoop.hbase.HRegionServer.closeAllRegions(HRegionServer.java:971): closing region testLogRolling,,1194596277787 [junit] 2007-11-09 08:20:49,834 DEBUG [RegionServer:0] org.apache.hadoop.hbase.HRegion.internalFlushcache(HRegion.java:847): Started memcache flush for region testLogRolling,,1194596277787. Size 0.0 [junit] 2007-11-09 08:20:49,834 DEBUG [RegionServer:0] org.apache.hadoop.hbase.HRegion.internalFlushcache(HRegion.java:865): Finished memcache flush; empty snapshot [junit] 2007-11-09 08:20:49,835 DEBUG [RegionServer:0] org.apache.hadoop.hbase.HStore.close(HStore.java:419): closed 216611736/info [junit] 2007-11-09 08:20:49,835 INFO [RegionServer:0] org.apache.hadoop.hbase.HRegion.close(HRegion.java:402): closed testLogRolling,,1194596277787 [junit] 2007-11-09 08:20:49,835 DEBUG [RegionServer:0] org.apache.hadoop.hbase.HRegionServer.closeAllRegions(HRegionServer.java:971): closing region testLogRolling,row0513,1194596368241 [junit] 2007-11-09 08:20:49,835 DEBUG [RegionServer:0] org.apache.hadoop.hbase.HRegion.internalFlushcache(HRegion.java:847): Started memcache flush for region testLogRolling,row0513,1194596368241. Size 0.0 [junit] 2007-11-09 08:20:49,835 DEBUG [RegionServer:0] org.apache.hadoop.hbase.HRegion.internalFlushcache(HRegion.java:865): Finished memcache flush; empty snapshot [junit] 2007-11-09 08:20:49,836 DEBUG [RegionServer:0] org.apache.hadoop.hbase.HStore.close(HStore.java:419): closed 1463872906/info [junit] 2007-11-09 08:20:49,836 INFO [RegionServer:0] org.apache.hadoop.hbase.HRegion.close(HRegion.java:402): closed testLogRolling,row0513,1194596368241 [junit] 2007-11-09 08:20:49,836 DEBUG [RegionServer:0] org.apache.hadoop.hbase.HRegionServer.closeAllRegions(HRegionServer.java:971): closing region testLogRolling,row1025,1194596368242 [junit] 2007-11-09 08:20:49,836 DEBUG [RegionServer:0] org.apache.hadoop.hbase.HRegion.internalFlushcache(HRegion.java:847): Started memcache flush for region testLogRolling,row1025,1194596368242. Size 0.0 [junit] 2007-11-09 08:20:49,836 DEBUG [RegionServer:0] org.apache.hadoop.hbase.HRegion.internalFlushcache(HRegion.java:865): Finished memcache flush; empty snapshot [junit] 2007-11-09 08:20:49,837 DEBUG [RegionServer:0] org.apache.hadoop.hbase.HStore.close(HStore.java:419): closed -1547818355/info [junit] 2007-11-09 08:20:49,837 INFO [RegionServer:0] org.apache.hadoop.hbase.HRegion.close(HRegion.java:402): closed testLogRolling,row1025,1194596368242 [junit] 2007-11-09 08:20:49,837 DEBUG [RegionServer:0] org.apache.hadoop.hbase.HLog.close(HLog.java:382): closing log writer in /hbase/log_140.211.11.75_-2039724685788569167_58358 [junit] 2007-11-09 08:20:49,838 WARN [IPC Server handler 3 on 58346] org.apache.hadoop.dfs.ReplicationTargetChooser.chooseTarget(ReplicationTargetChooser.java:177): Not able to place enough replicas, still in need of 1 [junit] 2007-11-09 08:20:49,848 INFO [RegionServer:0] org.apache.hadoop.hbase.HRegionServer.run(HRegionServer.java:603): telling master that region server is shutting down at: 140.211.11.75:58358 [junit] 2007-11-09 08:20:49,849 DEBUG [IPC Server handler 4 on 60000] org.apache.hadoop.hbase.HMaster.regionServerReport(HMaster.java:1316): Region server 140.211.11.75:58358: MSG_REPORT_EXITING -- cancelling lease [junit] 2007-11-09 08:20:49,849 INFO [IPC Server handler 4 on 60000] org.apache.hadoop.hbase.HMaster.cancelLease(HMaster.java:1438): Cancelling lease for 140.211.11.75:58358 [junit] 2007-11-09 08:20:49,849 INFO [IPC Server handler 4 on 60000] org.apache.hadoop.hbase.HMaster.regionServerReport(HMaster.java:1323): Region server 140.211.11.75:58358: MSG_REPORT_EXITING -- lease cancelled [junit] 2007-11-09 08:20:49,850 INFO [RegionServer:0] org.apache.hadoop.hbase.HRegionServer.run(HRegionServer.java:610): stopping server at: 140.211.11.75:58358 [junit] 2007-11-09 08:20:49,977 INFO [RegionServer:0.worker] org.apache.hadoop.hbase.HRegionServer$Worker.run(HRegionServer.java:920): worker thread exiting [junit] 2007-11-09 08:20:49,977 INFO [RegionServer:0] org.apache.hadoop.hbase.HRegionServer.run(HRegionServer.java:615): RegionServer:0 exiting [junit] 2007-11-09 08:20:50,947 INFO [HMaster.metaScanner] org.apache.hadoop.hbase.Chore.run(Chore.java:62): HMaster.metaScanner exiting [junit] 2007-11-09 08:20:50,948 INFO [HMaster] org.apache.hadoop.hbase.Leases.close(Leases.java:109): HMaster closing leases [junit] 2007-11-09 08:20:50,947 INFO [HMaster.rootScanner] org.apache.hadoop.hbase.Chore.run(Chore.java:62): HMaster.rootScanner exiting [junit] 2007-11-09 08:20:50,949 INFO [HMaster.leaseChecker] org.apache.hadoop.hbase.Chore.run(Chore.java:62): HMaster.leaseChecker exiting [junit] 2007-11-09 08:20:50,949 INFO [HMaster] org.apache.hadoop.hbase.Leases.close(Leases.java:123): HMaster closed leases [junit] 2007-11-09 08:20:50,949 INFO [HMaster] org.apache.hadoop.hbase.HMaster.run(HMaster.java:1163): HMaster main thread exiting [junit] 2007-11-09 08:20:50,949 INFO [main] org.apache.hadoop.hbase.LocalHBaseCluster.shutdown(LocalHBaseCluster.java:226): Shutdown HMaster 1 region server(s) [junit] Shutting down the Mini HDFS Cluster [junit] Shutting down DataNode 1 [junit] 2007-11-09 08:20:51,709 WARN [DataNode: [/export/home/hudson/hudson/jobs/Hadoop-Patch/workspace/trunk/build/contrib/hbase/test/data/dfs/data/data3,/export/home/hudson/hudson/jobs/Hadoop-Patch/workspace/trunk/build/contrib/hbase/test/data/dfs/data/data4]] org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:617): java.io.IOException: java.lang.InterruptedException [junit] at org.apache.hadoop.fs.ShellCommand.runCommand(ShellCommand.java:59) [junit] at org.apache.hadoop.fs.ShellCommand.run(ShellCommand.java:42) [junit] at org.apache.hadoop.fs.DU.getUsed(DU.java:52) [junit] at org.apache.hadoop.dfs.FSDataset$FSVolume.getDfsUsed(FSDataset.java:299) [junit] at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getDfsUsed(FSDataset.java:396) [junit] at org.apache.hadoop.dfs.FSDataset.getDfsUsed(FSDataset.java:495) [junit] at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:532) [junit] at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1695) [junit] at java.lang.Thread.run(Thread.java:595) [junit] Shutting down DataNode 0 [junit] 2007-11-09 08:20:52,252 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.PendingReplicationBlocks$PendingReplicationMonitor.run(PendingReplicationBlocks.java:186): PendingReplicationMonitor thread received exception. java.lang.InterruptedException: sleep interrupted [junit] 2007-11-09 08:20:52,570 ERROR [EMAIL PROTECTED] org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:839): DataXceiver: java.io.IOException: df: (/export/home/hudson/hudson/jobs/Hadoop-Patch/workspace/trunk/build/contrib/hbase/test/data/dfs/data/data3) not a block device, directory or mounted resource [junit] at org.apache.hadoop.fs.ShellCommand.runCommand(ShellCommand.java:52) [junit] at org.apache.hadoop.fs.ShellCommand.run(ShellCommand.java:42) [junit] at org.apache.hadoop.fs.DF.getAvailable(DF.java:72) [junit] at org.apache.hadoop.dfs.FSDataset$FSVolume.getAvailable(FSDataset.java:308) [junit] at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getNextVolume(FSDataset.java:386) [junit] at org.apache.hadoop.dfs.FSDataset.writeToBlock(FSDataset.java:580) [junit] at org.apache.hadoop.dfs.DataNode$BlockReceiver.<init>(DataNode.java:1458) [junit] at org.apache.hadoop.dfs.DataNode$DataXceiver.writeBlock(DataNode.java:929) [junit] at org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:824) [junit] at java.lang.Thread.run(Thread.java:595) [junit] Exception! java.io.IOException: No such file or directory [junit] 2007-11-09 08:20:52,864 ERROR [EMAIL PROTECTED] org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:839): DataXceiver: java.io.IOException: df: (/export/home/hudson/hudson/jobs/Hadoop-Patch/workspace/trunk/build/contrib/hbase/test/data/dfs/data/data4) not a block device, directory or mounted resource [junit] at org.apache.hadoop.fs.ShellCommand.runCommand(ShellCommand.java:52) [junit] at org.apache.hadoop.fs.ShellCommand.run(ShellCommand.java:42) [junit] at org.apache.hadoop.fs.DF.getCapacity(DF.java:62) [junit] at org.apache.hadoop.dfs.FSDataset$FSVolume.getCapacity(FSDataset.java:303) [junit] at org.apache.hadoop.dfs.FSDataset$FSVolume.getAvailable(FSDataset.java:307) [junit] at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getNextVolume(FSDataset.java:386) [junit] at org.apache.hadoop.dfs.FSDataset.writeToBlock(FSDataset.java:580) [junit] at org.apache.hadoop.dfs.DataNode$BlockReceiver.<init>(DataNode.java:1458) [junit] at org.apache.hadoop.dfs.DataNode$DataXceiver.writeBlock(DataNode.java:929) [junit] at org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:824) [junit] at java.lang.Thread.run(Thread.java:595) [junit] 2007-11-09 08:20:52,864 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.DataNode$DataTransfer.run(DataNode.java:1668): Failed to transfer blk_-379738272651084333 to 127.0.0.1:50011 got java.io.IOException: operation failed at /127.0.0.1 [junit] at org.apache.hadoop.dfs.DataNode.receiveResponse(DataNode.java:725) [junit] at org.apache.hadoop.dfs.DataNode.access$200(DataNode.java:80) [junit] at org.apache.hadoop.dfs.DataNode$DataTransfer.run(DataNode.java:1664) [junit] at java.lang.Thread.run(Thread.java:595) [junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 344.356 sec [junit] 2007-11-09 08:20:52,864 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.DataNode$DataXceiver.writeBlock(DataNode.java:995): Error writing reply back to /127.0.0.1 for writing block blk_-3764842785131980349 [junit] java.net.SocketException: Broken pipe [junit] at java.net.SocketOutputStream.socketWrite0(Native Method) [junit] at java.net.SocketOutputStream.socketWrite(SocketOutputStream.java:92) [junit] at java.net.SocketOutputStream.write(SocketOutputStream.java:115) [junit] at java.io.DataOutputStream.writeShort(DataOutputStream.java:151) [junit] at org.apache.hadoop.dfs.DataNode.sendResponse(DataNode.java:737) [junit] at org.apache.hadoop.dfs.DataNode.access$300(DataNode.java:80) [junit] at org.apache.hadoop.dfs.DataNode$DataXceiver.writeBlock(DataNode.java:993) [junit] at org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:824) [junit] at java.lang.Thread.run(Thread.java:595) [junit] Exception! java.io.IOException: No such file or directory [junit] 2007-11-09 08:20:53,748 ERROR [EMAIL PROTECTED] org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:839): DataXceiver: java.io.IOException: No such file or directory [junit] at java.io.UnixFileSystem.createFileExclusively(Native Method) [junit] at java.io.File.createNewFile(File.java:850) [junit] at org.apache.hadoop.dfs.FSDataset$FSVolume.createTmpFile(FSDataset.java:329) [junit] at org.apache.hadoop.dfs.FSDataset.createTmpFile(FSDataset.java:606) [junit] at org.apache.hadoop.dfs.FSDataset.writeToBlock(FSDataset.java:582) [junit] at org.apache.hadoop.dfs.DataNode$BlockReceiver.<init>(DataNode.java:1458) [junit] at org.apache.hadoop.dfs.DataNode$DataXceiver.writeBlock(DataNode.java:929) [junit] at org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:824) [junit] at java.lang.Thread.run(Thread.java:595) [junit] 2007-11-09 08:20:53,748 ERROR [EMAIL PROTECTED] org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:839): DataXceiver: java.io.IOException: du: /export/home/hudson/hudson/jobs/Hadoop-Patch/workspace/trunk/build/contrib/hbase/test/data/dfs/data: No such file or directory [junit] at org.apache.hadoop.fs.ShellCommand.runCommand(ShellCommand.java:52) [junit] at org.apache.hadoop.fs.ShellCommand.run(ShellCommand.java:42) [junit] at org.apache.hadoop.fs.DU.getUsed(DU.java:52) [junit] at org.apache.hadoop.dfs.FSDataset$FSVolume.getDfsUsed(FSDataset.java:299) [junit] at org.apache.hadoop.dfs.FSDataset$FSVolume.getAvailable(FSDataset.java:307) [junit] at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getNextVolume(FSDataset.java:386) [junit] at org.apache.hadoop.dfs.FSDataset.writeToBlock(FSDataset.java:580) [junit] at org.apache.hadoop.dfs.DataNode$BlockReceiver.<init>(DataNode.java:1458) [junit] at org.apache.hadoop.dfs.DataNode$DataXceiver.writeBlock(DataNode.java:929) [junit] at org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:824) [junit] at java.lang.Thread.run(Thread.java:595) [junit] 2007-11-09 08:20:53,749 ERROR [EMAIL PROTECTED] org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:839): DataXceiver: java.io.IOException: No such file or directory [junit] at java.io.UnixFileSystem.createFileExclusively(Native Method) [junit] at java.io.File.createNewFile(File.java:850) [junit] at org.apache.hadoop.dfs.FSDataset$FSVolume.createTmpFile(FSDataset.java:329) [junit] at org.apache.hadoop.dfs.FSDataset.createTmpFile(FSDataset.java:606) [junit] at org.apache.hadoop.dfs.FSDataset.writeToBlock(FSDataset.java:582) [junit] at org.apache.hadoop.dfs.DataNode$BlockReceiver.<init>(DataNode.java:1458) [junit] at org.apache.hadoop.dfs.DataNode$DataXceiver.writeBlock(DataNode.java:929) [junit] at org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:824) [junit] at java.lang.Thread.run(Thread.java:595) [junit] Exception! java.io.IOException: No such file or directory [junit] 2007-11-09 08:20:53,800 ERROR [EMAIL PROTECTED] org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:839): DataXceiver: java.io.IOException: No such file or directory [junit] at java.io.UnixFileSystem.createFileExclusively(Native Method) [junit] at java.io.File.createNewFile(File.java:850) [junit] at org.apache.hadoop.dfs.FSDataset$FSVolume.createTmpFile(FSDataset.java:329) [junit] at org.apache.hadoop.dfs.FSDataset.createTmpFile(FSDataset.java:606) [junit] at org.apache.hadoop.dfs.FSDataset.writeToBlock(FSDataset.java:582) [junit] at org.apache.hadoop.dfs.DataNode$BlockReceiver.<init>(DataNode.java:1458) [junit] at org.apache.hadoop.dfs.DataNode$DataXceiver.writeBlock(DataNode.java:929) [junit] at org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:824) [junit] at java.lang.Thread.run(Thread.java:595) [junit] 2007-11-09 08:20:53,801 WARN [EMAIL PROTECTED] org.apache.hadoop.dfs.DataNode$DataTransfer.run(DataNode.java:1668): Failed to transfer blk_1352514030345870875 to 127.0.0.1:50011 got java.io.IOException: operation failed at /127.0.0.1 [junit] at org.apache.hadoop.dfs.DataNode.receiveResponse(DataNode.java:725) [junit] at org.apache.hadoop.dfs.DataNode.access$200(DataNode.java:80) [junit] at org.apache.hadoop.dfs.DataNode$DataTransfer.run(DataNode.java:1664) [junit] at java.lang.Thread.run(Thread.java:595) {code} Test had not reported itself done. Are these du'ings and unix process invocations of interest? (Check). > [hbase] TestHStoreFile/TestBloomFilter hang occasionally on hudson AFTER test > has finished > ------------------------------------------------------------------------------------------ > > Key: HADOOP-2040 > URL: https://issues.apache.org/jira/browse/HADOOP-2040 > Project: Hadoop > Issue Type: Bug > Components: contrib/hbase > Reporter: stack > Priority: Minor > Attachments: endoftesttd.patch > > > Weird. Last night TestBloomFilter was hung after junit had printed test had > completed without error. Just now, I noticed a hung TestHStore -- again > after junit had printed out test had succeeded (Nigel Daley has reported he's > seen at least two hangs in TestHStoreFile, perhaps in same location). > Last night and just now I was unable to get a thread dump. > Here is log from around this evenings hang: > {code} > ... > [junit] 2007-10-12 04:19:28,477 INFO [main] > org.apache.hadoop.hbase.TestHStoreFile.testOutOfRangeMidkeyHalfMapFile(TestHStoreFile.java:366): > Last bottom when key > top: zz/zz/1192162768317 > [junit] 2007-10-12 04:19:28,493 WARN [IPC Server handler 0 on 36620] > org.apache.hadoop.dfs.FSDirectory.unprotectedDelete(FSDirectory.java:400): > DIR* FSDirectory.unprotectedDelete: failed to remove > /testOutOfRangeMidkeyHalfMapFile because it does not exist > [junit] Shutting down the Mini HDFS Cluster > [junit] Shutting down DataNode 1 > [junit] Shutting down DataNode 0 > [junit] 2007-10-12 04:19:29,316 WARN [EMAIL PROTECTED] > org.apache.hadoop.dfs.PendingReplicationBlocks$PendingReplicationMonitor.run(PendingReplicationBlocks.java:186): > PendingReplicationMonitor thread received exception. > java.lang.InterruptedException: sleep interrupted > [junit] Tests run: 4, Failures: 0, Errors: 0, Time elapsed: 16.274 sec > [junit] Running org.apache.hadoop.hbase.TestHTable > [junit] Starting DataNode 0 with dfs.data.dir: > /export/home/hudson/hudson/jobs/Hadoop-Patch/workspace/trunk/build/contrib/hbase/test/data/dfs/data/data1,/export/home/hudson/hudson/jobs/Hadoop-Patch/workspace/trunk/build/contrib/hbase/test/data/dfs/data/data2 > [junit] Starting DataNode 1 with dfs.data.dir: > /export/home/hudson/hudson/jobs/Hadoop-Patch/workspace/trunk/build/contrib/hbase/test/data/dfs/data/data3,/export/home/hudson/hudson/jobs/Hadoop-Patch/workspace/trunk/build/contrib/hbase/test/data/dfs/data/data4 > [junit] 2007-10-12 05:21:48,332 INFO [main] > org.apache.hadoop.hbase.HMaster.<init>(HMaster.java:862): Root region dir: > /hbase/hregion_-ROOT-,,0 > ... > {code} > Notice the hour of elapsed (hung) time in above. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.