[ https://issues.apache.org/jira/browse/HADOOP-2757?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
dhruba borthakur updated HADOOP-2757: ------------------------------------- Attachment: softMount3.patch A configuration parameter ipc.client.rpctimeout that makes an RPC to not hang for more than this period while waiting for a response from the remote end. A configuration parameter dfs.softmount.timeout that allows a dfs writer to exit gracefully if the dfsclient is unable to talk to the namenode for this period of time. > Should DFS outputstream's close wait forever? > --------------------------------------------- > > Key: HADOOP-2757 > URL: https://issues.apache.org/jira/browse/HADOOP-2757 > Project: Hadoop Core > Issue Type: Improvement > Components: dfs > Reporter: Raghu Angadi > Assignee: dhruba borthakur > Attachments: softMount1.patch, softMount1.patch, softMount2.patch, > softMount3.patch > > > Currently {{DFSOutputStream.close()}} waits for ever if Namenode keeps > throwing {{NotYetReplicated}} exception, for whatever reason. Its pretty > annoying for a user. Shoud the loop inside close have a timeout? If so how > much? It could probably something like 10 minutes. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.