[ https://issues.apache.org/jira/browse/HDFS-183?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Suresh Srinivas resolved HDFS-183. ---------------------------------- Resolution: Won't Fix This bug has been open for a while, with no reply to the last comment posted. Closing for now. Please re-open the bug or open a new bug, if the problem still persists. > MapReduce Streaming job hang when all replications of the input file has > corrupted! > ----------------------------------------------------------------------------------- > > Key: HDFS-183 > URL: https://issues.apache.org/jira/browse/HDFS-183 > Project: Hadoop HDFS > Issue Type: Bug > Reporter: ZhuGuanyin > Priority: Critical > > On some special cases, all replications of a given file has truncated to zero > but the namenode still hold the original size (we don't know why), the > mapreduce streaming job will hang if we don't specified mapred.task.timeout > when the input files contain this corrupted file, even the dfs shell "cat" > will hang when fetch data from this corrupted file. > We found that job hang at DFSInputStream.blockSeekTo() when chosing a > datanode. The following test will show: > 1) Copy a small file to hdfs. > 2) Get the file blocks and login to these datanodes, and truncate these > blocks to zero. > 3) Cat this file through dfs shell "cat" > 4) Cat command will enter dead loop. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira