[ 
https://issues.apache.org/jira/browse/HDFS-1787?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13037996#comment-13037996
 ] 

Jonathan Hsieh commented on HDFS-1787:
--------------------------------------

Actually, I assumed that the test focused on the patch in the 'changes' section 
of the jenkins result of build 608.  This actually ran the newly added test 
case from the HDFS-1787 patch..  

The 
org.apache.hadoop.hdfs.server.datanode.TestFiDataTransferProtocol2.pipeline_Fi_30
  test seems to be intermittently failing.   It also isn't reported by hudson.  
 Is there a reason why?  

> "Not enough xcievers" error should propagate to client
> ------------------------------------------------------
>
>                 Key: HDFS-1787
>                 URL: https://issues.apache.org/jira/browse/HDFS-1787
>             Project: Hadoop HDFS
>          Issue Type: Improvement
>          Components: data-node
>    Affects Versions: 0.23.0
>            Reporter: Todd Lipcon
>            Assignee: Jonathan Hsieh
>              Labels: newbie
>             Fix For: 0.23.0
>
>         Attachments: hdfs-1787.2.patch, hdfs-1787.3.patch, hdfs-1787.3.patch, 
> hdfs-1787.patch
>
>
> We find that users often run into the default transceiver limits in the DN. 
> Putting aside the inherent issues with xceiver threads, it would be nice if 
> the "xceiver limit exceeded" error propagated to the client. Currently, 
> clients simply see an EOFException which is hard to interpret, and have to go 
> slogging through DN logs to find the underlying issue.
> The data transfer protocol should be extended to either have a special error 
> code for "not enough xceivers" or should have some error code for generic 
> errors with which a string can be attached and propagated.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira

Reply via email to