[ 
https://issues.apache.org/jira/browse/HDFS-11711?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brahma Reddy Battula updated HDFS-11711:
----------------------------------------
       Resolution: Fixed
     Hadoop Flags: Reviewed
    Fix Version/s: 2.8.2
                   2.8.1
                   3.0.0-alpha4
                   2.9.0
           Status: Resolved  (was: Patch Available)

Committed to {{trunk}},{{branch-2}},{{branch-2.8}} and {{branch-2.8.1}}. Test 
failures are unrelated

Thanks to all for your great reviews.

bq.Must throw File not found would be better updated to Must throw 
FileNotFoundException,
bq.Should throw too many open would be better updated to Should throw too many 
open files

Addressed these minor nits while committing.

bq. But I wish there's a more portable way to check for Too many open files 
error.

do you mean, it's already there..?

> DN should not delete the block On "Too many open files" Exception
> -----------------------------------------------------------------
>
>                 Key: HDFS-11711
>                 URL: https://issues.apache.org/jira/browse/HDFS-11711
>             Project: Hadoop HDFS
>          Issue Type: Bug
>          Components: datanode
>            Reporter: Brahma Reddy Battula
>            Assignee: Brahma Reddy Battula
>            Priority: Critical
>             Fix For: 2.9.0, 3.0.0-alpha4, 2.8.1, 2.8.2
>
>         Attachments: HDFS-11711-002.patch, HDFS-11711-003.patch, 
> HDFS-11711-004.patch, HDFS-11711-branch-2-002.patch, 
> HDFS-11711-branch-2-003.patch, HDFS-11711.patch
>
>
>  *Seen the following scenario in one of our customer environment* 
> * while jobclient writing {{"job.xml"}} there are pipeline failures and 
> written to only one DN.
> * when mapper reading the {{"job.xml"}}, DN got {{"Too many open files"}} (as 
> system exceed limit) and block got deleted. Hence mapper failed to read and 
> job got failed.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to