[ 
https://issues.apache.org/jira/browse/HADOOP-5334?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12676936#action_12676936
 ] 

Suman Sehgal commented on HADOOP-5334:
--------------------------------------

Sort also gets failed when all the datanodes of one of the rack are killed  
displaying following error message:

java.io.IOException: Could not obtain block: blk_-5588658908635997065_2179 
file=/user/hadoopqa/input1/part-00848
        at 
org.apache.hadoop.dfs.DFSClient$DFSInputStream.chooseDataNode(DFSClient.java:1471)
        at 
org.apache.hadoop.dfs.DFSClient$DFSInputStream.blockSeekTo(DFSClient.java:1321)
        at 
org.apache.hadoop.dfs.DFSClient$DFSInputStream.read(DFSClient.java:1426)
        at java.io.DataInputStream.readFully(DataInputStream.java:178)
        at java.io.DataInputStream.readFully(DataInputStream.java:152)
        at org.apache.hadoop.io.SequenceFile$Reader.init(SequenceFile.java:1453)
        at org.apache.hadoop.io.SequenceFile$Reader.(SequenceFile.java:1431)
        at org.apache.hadoop.io.SequenceFile$Reader.(SequenceFile.java:1420)
        at org.apache.hadoop.io.SequenceFile$Reader.(SequenceFile.java:1415)
        at 
org.apache.hadoop.mapred.SequenceFileRecordReader.(SequenceFileRecordReader.java:43)
        at 
org.apache.hadoop.mapred.SequenceFileInputFormat.getRecordReader(SequenceFileInputFormat.java:54)
        at org.apache.hadoop.mapred.MapTask.run(MapTask.java:219)
        at 
org.apache.hadoop.mapred.TaskTracker$Child.main(TaskTracker.java:2206)

It runs fine if killed datanodes are again brought up. 

> Not able to run randomwriter/sort on hdfs if all the nodes of same rack are 
> killed.
> -----------------------------------------------------------------------------------
>
>                 Key: HADOOP-5334
>                 URL: https://issues.apache.org/jira/browse/HADOOP-5334
>             Project: Hadoop Core
>          Issue Type: Bug
>          Components: dfs
>    Affects Versions: 0.18.3
>            Reporter: Suman Sehgal
>
> Not able to run randomwriter if all the datanodes of any one of the racks are 
> killed. (replication factor : 3)
> Randomwriter job gets failed and following error message is displayed in log:
> java.net.ConnectException: Connection refused
>       at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
>       at 
> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:574)
>       at sun.nio.ch.SocketAdaptor.connect(SocketAdaptor.java:100)
>       at 
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.createBlockOutputStream(DFSClient.java:2398)
>       at 
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2354)
>       at 
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.access$1800(DFSClient.java:1744)
>       at 
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:1927)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to