Allow configurable timeouts when connecting to HDFS via java FileSystem API
---------------------------------------------------------------------------
Key: HADOOP-7397
URL: https://issues.apache.org/jira/browse/HADOOP-7397
Project: Hadoop Common
Issue Type: Improvement
Components: ipc
Affects Versions: 0.20.2
Environment: Any
Reporter: Scott Fines
Priority: Minor
If the NameNode is not available (in, for example, a network partition event
separating the client from the NameNode), and an attempt is made to connect,
then the FileSystem api will *eventually* timeout and throw an error. However,
that timeout is currently hardcoded to be 20 seconds to connect, with 45
retries, for a total of a 15 minute wait before failure. While in many
circumstances this is fine, there are also many circumstances (such as booting
a service) where both the connection timeout and the number of retries should
be significantly less, so as not to harm availability of other services.
Investigating Client.java, I see that there are two fields in Connection:
maxRetries and rpcTimeout. I propose either re-using those fields for
initiating the connection as well; alternatively, using the already existing
dfs.socket.timeout parameter to set the connection timeout on initialization,
and potentially adding a new field such as dfs.connection.retries with a
default of 45 to replicate current behaviors.
--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira