[ https://issues.apache.org/jira/browse/HADOOP-3504?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12604724#action_12604724 ]
Tsz Wo (Nicholas), SZE commented on HADOOP-3504: ------------------------------------------------ Lohit and I are finally able to reproduce the problem. We ran randomwriter on a 500-node cluster and got a lot of map tasks failure. The problem is due to out of disk space in some datanodes. When a dfs client writes a block to datanode pipeline, if there is a datanode in the pipeline ran out of disk space. It throws {noformat} 2008-06-13 01:01:58,909 ERROR org.apache.hadoop.dfs.DataNode: DatanodeRegistration(xx.xx.xx.xx:52961, storageID=DS-yyyyyyy, infoPort=58576, ipcPort=50020):DataXceiver: org.apache.hadoop.util.DiskChecker$DiskOutOfSpaceException: No space left on device at org.apache.hadoop.dfs.DataNode.checkDiskError(DataNode.java:590) at org.apache.hadoop.dfs.DataNode.access$1400(DataNode.java:82) at org.apache.hadoop.dfs.DataNode$BlockReceiver.receivePacket(DataNode.java:2611) at org.apache.hadoop.dfs.DataNode$BlockReceiver.receiveBlock(DataNode.java:2664) at org.apache.hadoop.dfs.DataNode$DataXceiver.writeBlock(DataNode.java:1267) at org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:1027) at java.lang.Thread.run(Thread.java:619) {noformat} However, the datanode does not report the exception back to client and the client will time-out. That's why we see SocketTimeoutException from the client. > Reduce task hangs after java.net.SocketTimeoutException > ------------------------------------------------------- > > Key: HADOOP-3504 > URL: https://issues.apache.org/jira/browse/HADOOP-3504 > Project: Hadoop Core > Issue Type: Bug > Components: dfs > Affects Versions: 0.18.0 > Environment: Linux > Reporter: Mukund Madhugiri > Assignee: Tsz Wo (Nicholas), SZE > Priority: Blocker > Fix For: 0.18.0 > > > When running gridmix, I saw 11 reduce tasks hanging. I manually failed the > tasks and they re-ran and then finished. > Here is the task tracker logs: > syslog logs > al on-disk merge with 14 files > 2008-06-05 19:02:49,804 INFO org.apache.hadoop.mapred.Merger: Merging 14 > sorted segments > 2008-06-05 19:03:03,663 INFO org.apache.hadoop.mapred.Merger: Down to the > last merge-pass, with 14 segments left of total size: 1476315198 bytes > 2008-06-05 19:03:03,731 WARN org.apache.hadoop.fs.FileSystem: > "hostname:56007" is a deprecated filesystem name. Use > "hdfs://hostname:56007/" instead. > 2008-06-05 19:03:27,301 INFO org.apache.hadoop.streaming.PipeMapRed: > R/W/S=1/0/0 in:0=1/2626 [rec/s] out:0=0/2626 [rec/s] > 2008-06-05 19:03:27,347 INFO org.apache.hadoop.streaming.PipeMapRed: > R/W/S=10/0/0 in:0=10/2626 [rec/s] out:0=0/2626 [rec/s] > 2008-06-05 19:03:27,578 INFO org.apache.hadoop.streaming.PipeMapRed: > R/W/S=100/0/0 in:0=100/2627 [rec/s] out:0=0/2627 [rec/s] > 2008-06-05 19:03:28,380 INFO org.apache.hadoop.streaming.PipeMapRed: Records > R/W=226/1 > 2008-06-05 19:03:35,276 INFO org.apache.hadoop.streaming.PipeMapRed: > R/W/S=1000/842/0 in:0=1000/2634 [rec/s] out:0=842/2634 [rec/s] > 2008-06-05 19:03:38,667 INFO org.apache.hadoop.streaming.PipeMapRed: Records > R/W=2434/2274 > 2008-06-05 19:03:45,301 INFO org.apache.hadoop.streaming.PipeMapRed: > R/W/S=10000/9892/0 in:3=10000/2644 [rec/s] out:3=9892/2644 [rec/s] > 2008-06-05 19:03:48,716 INFO org.apache.hadoop.streaming.PipeMapRed: Records > R/W=15057/14957 > 2008-06-05 19:03:59,056 INFO org.apache.hadoop.streaming.PipeMapRed: Records > R/W=24946/24887 > 2008-06-05 19:04:11,742 INFO org.apache.hadoop.streaming.PipeMapRed: Records > R/W=34653/34433 > 2008-06-05 19:04:22,548 INFO org.apache.hadoop.streaming.PipeMapRed: Records > R/W=42930/42803 > 2008-06-05 19:04:32,635 INFO org.apache.hadoop.streaming.PipeMapRed: Records > R/W=57737/57686 > 2008-06-05 19:04:42,662 INFO org.apache.hadoop.streaming.PipeMapRed: Records > R/W=76224/76063 > 2008-06-05 19:04:52,666 INFO org.apache.hadoop.streaming.PipeMapRed: Records > R/W=99423/99307 > 2008-06-05 19:04:52,802 INFO org.apache.hadoop.streaming.PipeMapRed: > R/W/S=100000/99795/0 in:36=100000/2712 [rec/s] out:36=99795/2712 [rec/s] > 2008-06-05 19:05:02,754 INFO org.apache.hadoop.streaming.PipeMapRed: Records > R/W=127265/127145 > 2008-06-05 19:05:12,758 INFO org.apache.hadoop.streaming.PipeMapRed: Records > R/W=185310/185202 > 2008-06-05 19:05:15,858 INFO org.apache.hadoop.streaming.PipeMapRed: > R/W/S=200000/199974/0 in:73=200000/2735 [rec/s] out:73=199974/2735 [rec/s] > 2008-06-05 19:05:22,772 INFO org.apache.hadoop.streaming.PipeMapRed: Records > R/W=218164/218082 > 2008-06-05 19:05:55,316 INFO org.apache.hadoop.streaming.PipeMapRed: Records > R/W=242591/242411 > 2008-06-05 19:06:13,678 INFO org.apache.hadoop.streaming.PipeMapRed: Records > R/W=242591/242412 > 2008-06-05 19:07:23,173 WARN org.apache.hadoop.dfs.DFSClient: DFSOutputStream > ResponseProcessor exception for block > blk_-3463507617208131068_33273java.net.SocketTimeoutException: 69000 millis > timeout while waiting for channel to be ready for read. ch : > java.nio.channels.SocketChannel[connected local=/NNN.NNN.NNN.125:59802 > remote=/NNN.NNN.NNN.125:57834] > at > org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:162) > at > org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:150) > at > org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:123) > at java.io.DataInputStream.readFully(DataInputStream.java:178) > at java.io.DataInputStream.readLong(DataInputStream.java:399) > at > org.apache.hadoop.dfs.DFSClient$DFSOutputStream$ResponseProcessor.run(DFSClient.java:2045) > 2008-06-05 19:11:22,535 INFO org.apache.hadoop.streaming.PipeMapRed: Records > R/W=243534/243374 > 2008-06-05 19:11:22,536 WARN org.apache.hadoop.dfs.DFSClient: Error Recovery > for block blk_-3463507617208131068_33273 bad datanode[0] NNN.NNN.NNN.125:57834 > 2008-06-05 19:11:24,388 WARN org.apache.hadoop.dfs.DFSClient: Error Recovery > for block blk_-3463507617208131068_33273 in pipeline NNN.NNN.NNN.125:57834, > NNN.NNN.NNN.107:58706, NNN.NNN.NNN.122:52897: bad datanode > NNN.NNN.NNN.125:57834 -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.