[ https://issues.apache.org/jira/browse/HDFS-7005?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14126894#comment-14126894 ]
Hudson commented on HDFS-7005: ------------------------------ FAILURE: Integrated in Hadoop-Yarn-trunk #675 (See [https://builds.apache.org/job/Hadoop-Yarn-trunk/675/]) HDFS-7005. DFS input streams do not timeout. Contributed by Daryn Sharp. (kihwal: rev 6a84f88c1190a8fecadd81deb6e7b8a69675fa91) * hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDistributedFileSystem.java * hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSClient.java > DFS input streams do not timeout > -------------------------------- > > Key: HDFS-7005 > URL: https://issues.apache.org/jira/browse/HDFS-7005 > Project: Hadoop HDFS > Issue Type: Bug > Components: hdfs-client > Affects Versions: 3.0.0, 2.5.0 > Reporter: Daryn Sharp > Assignee: Daryn Sharp > Priority: Critical > Fix For: 2.6.0 > > Attachments: HDFS-7005.patch > > > Input streams lost their timeout. The problem appears to be > {{DFSClient#newConnectedPeer}} does not set the read timeout. During a > temporary network interruption the server will close the socket, unbeknownst > to the client host, which blocks on a read forever. > The results are dire. Services such as the RM, JHS, NMs, oozie servers, etc > all need to be restarted to recover - unless you want to wait many hours for > the tcp stack keepalive to detect the broken socket. -- This message was sent by Atlassian JIRA (v6.3.4#6332)