[
https://issues.apache.org/jira/browse/HADOOP-2040?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12534225
]
stack commented on HADOOP-2040:
-------------------------------
Looks like it hung again in same build -- #931 -- but this time in a test that
hasn't been prone to hanging, TestListTables. Again I can't get a thread dump
but log is interesting on the way out:
{code}
[junit] Shutting down the Mini HDFS Cluster
[junit] Shutting down DataNode 1
[junit] 2007-10-12 05:23:16,082 WARN [DataNode:
[/export/home/hudson/hudson/jobs/Hadoop-Patch/workspace/trunk/build/contrib/hbase/test/data/dfs/data/data3,/export/home/hudson/hudson/jobs/Hadoop-Patch/workspace/trunk/build/contrib/hbase/test/data/dfs/data/data4]]
org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:596):
java.io.IOException: java.lang.InterruptedException
[junit] Shutting down DataNode 0
[junit] at
org.apache.hadoop.fs.ShellCommand.runCommand(ShellCommand.java:59)
[junit] at org.apache.hadoop.fs.ShellCommand.run(ShellCommand.java:42)
[junit] at org.apache.hadoop.fs.DU.getUsed(DU.java:52)
[junit] at
org.apache.hadoop.dfs.FSDataset$FSVolume.getDfsUsed(FSDataset.java:299)
[junit] at
org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getDfsUsed(FSDataset.java:396)
[junit] at
org.apache.hadoop.dfs.FSDataset.getDfsUsed(FSDataset.java:495)
[junit] at
org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:520)
[junit] at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1494)
[junit] at java.lang.Thread.run(Thread.java:595)
[junit] 2007-10-12 05:23:16,349 WARN [DataNode:
[/export/home/hudson/hudson/jobs/Hadoop-Patch/workspace/trunk/build/contrib/hbase/test/data/dfs/data/data1,/export/home/hudson/hudson/jobs/Hadoop-Patch/workspace/trunk/build/contrib/hbase/test/data/dfs/data/data2]]
org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:596):
java.io.InterruptedIOException
[junit] at java.net.SocketOutputStream.socketWrite0(Native Method)
[junit] at
java.net.SocketOutputStream.socketWrite(SocketOutputStream.java:92)
[junit] at
java.net.SocketOutputStream.write(SocketOutputStream.java:136)
[junit] at
org.apache.hadoop.ipc.Client$Connection$2.write(Client.java:192)
[junit] at
java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:65)
[junit] at
java.io.BufferedOutputStream.flush(BufferedOutputStream.java:123)
[junit] at java.io.DataOutputStream.flush(DataOutputStream.java:106)
[junit] at
org.apache.hadoop.ipc.Client$Connection.sendParam(Client.java:327)
[junit] at org.apache.hadoop.ipc.Client.call(Client.java:474)
[junit] at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:184)
[junit] at org.apache.hadoop.dfs.$Proxy1.sendHeartbeat(Unknown Source)
[junit] at
org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:520)
[junit] at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1494)
[junit] at java.lang.Thread.run(Thread.java:595)
[junit] 2007-10-12 05:23:16,351 WARN [EMAIL PROTECTED]
org.apache.hadoop.dfs.PendingReplicationBlocks$PendingReplicationMonitor.run(PendingReplicationBlocks.java:186):
PendingReplicationMonitor thread received exception.
java.lang.InterruptedException: sleep interrupted
[junit] 2007-10-12 05:23:16,610 INFO [main]
org.apache.hadoop.hbase.MiniHBaseCluster.shutdown(MiniHBaseCluster.java:424):
Shutting down FileSystem
[junit] Tests run: 1, Failures: 0, Errors: 0, Time elapsed: 36.108 sec
{code}
It reports tests succeeded but just before hand its reporting and interrupted
flush. I wonder if interrupt broke the flush. It would be interesting to know
(for HADOOP-1924).
> [hbase] TestHStoreFile/TestBloomFilter hang occasionally on hudson AFTER test
> has finished
> ------------------------------------------------------------------------------------------
>
> Key: HADOOP-2040
> URL: https://issues.apache.org/jira/browse/HADOOP-2040
> Project: Hadoop
> Issue Type: Bug
> Components: contrib/hbase
> Reporter: stack
> Priority: Minor
>
> Weird. Last night TestBloomFilter was hung after junit had printed test had
> completed without error. Just now, I noticed a hung TestHStore -- again
> after junit had printed out test had succeeded (Nigel Daley has reported he's
> seen at least two hangs in TestHStoreFile, perhaps in same location).
> Last night and just now I was unable to get a thread dump.
> Here is log from around this evenings hang:
> {code}
> ...
> [junit] 2007-10-12 04:19:28,477 INFO [main]
> org.apache.hadoop.hbase.TestHStoreFile.testOutOfRangeMidkeyHalfMapFile(TestHStoreFile.java:366):
> Last bottom when key > top: zz/zz/1192162768317
> [junit] 2007-10-12 04:19:28,493 WARN [IPC Server handler 0 on 36620]
> org.apache.hadoop.dfs.FSDirectory.unprotectedDelete(FSDirectory.java:400):
> DIR* FSDirectory.unprotectedDelete: failed to remove
> /testOutOfRangeMidkeyHalfMapFile because it does not exist
> [junit] Shutting down the Mini HDFS Cluster
> [junit] Shutting down DataNode 1
> [junit] Shutting down DataNode 0
> [junit] 2007-10-12 04:19:29,316 WARN [EMAIL PROTECTED]
> org.apache.hadoop.dfs.PendingReplicationBlocks$PendingReplicationMonitor.run(PendingReplicationBlocks.java:186):
> PendingReplicationMonitor thread received exception.
> java.lang.InterruptedException: sleep interrupted
> [junit] Tests run: 4, Failures: 0, Errors: 0, Time elapsed: 16.274 sec
> [junit] Running org.apache.hadoop.hbase.TestHTable
> [junit] Starting DataNode 0 with dfs.data.dir:
> /export/home/hudson/hudson/jobs/Hadoop-Patch/workspace/trunk/build/contrib/hbase/test/data/dfs/data/data1,/export/home/hudson/hudson/jobs/Hadoop-Patch/workspace/trunk/build/contrib/hbase/test/data/dfs/data/data2
> [junit] Starting DataNode 1 with dfs.data.dir:
> /export/home/hudson/hudson/jobs/Hadoop-Patch/workspace/trunk/build/contrib/hbase/test/data/dfs/data/data3,/export/home/hudson/hudson/jobs/Hadoop-Patch/workspace/trunk/build/contrib/hbase/test/data/dfs/data/data4
> [junit] 2007-10-12 05:21:48,332 INFO [main]
> org.apache.hadoop.hbase.HMaster.<init>(HMaster.java:862): Root region dir:
> /hbase/hregion_-ROOT-,,0
> ...
> {code}
> Notice the hour of elapsed (hung) time in above.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.