[ 
https://issues.apache.org/jira/browse/HDFS-2819?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13201265#comment-13201265
 ] 

Suresh Srinivas commented on HDFS-2819:
---------------------------------------

Eli, thanks for documenting the config.

Some comments:
# Why are we changing the name to DFS_HA_NAMENODES_KEY_PREFIX?
# I am not sure what you mean by "prefix for given nameservice". Also please 
mention value contains list of comma-separated namenodes.
{noformat}
+<property>
+  <name>dfs.ha.namenodes.EXAMPLENAMESERVICE</name>
+  <value></value>
+  <description>
+    The prefix for a given nameservice, contains a comma-separated
+    list of namenodes for a given nameservice (eg EXAMPLENAMESERVICE).
+  </description>
+</property>
{noformat}
# Does the existing code that finds nameservice ID work with this config added? 
 Is this not equivalent to adding an empty nameservice ID? Not sure how our 
config behaves for <value></value>. 
{noformat}
+<property>
+  <name>dfs.federation.nameservice.id</name>
+  <value></value>
+  <description>
+    The ID of this nameservice. If the nameservice ID is not
+    configured or more than one nameservice is configured for
+    dfs.federation.nameservices it is determined automatically by
+    matching the local node's address with the configured address.
+  </description>
+</property>
{noformat}
# Please comment out the properties dfs.federation.nameservices and 
dfs.federation.nameservice.id.
# Can you please describe what 0 means for users? Same for 
dfs.client.failover.connection.retries. Users many not understand what failover 
IPC client means.
{noformat}
+<property>
+  <name>dfs.client.failover.connection.retries.on.timeouts</name>
+  <value>0</value>
+  <description>
+    Expert only. The number of retry attempts a failover IPC client
+    will make on socket timeout when establishing a server connection.
+  </description>
+</property>
{noformat}
# For client.failover, are the default retry attempts, timeouts correct? I am 
sure I understand the rationale for these timeouts. The failover attempts are 
made for ~ 165 seconds in following sequence ~(0.5 1 2 4 8 15 15 15 15 15 15 15 
15 15 15)
# From config perspective client.failover config items are confusing. There 
configs related to failover.max.attempts and failover.connection.retries. Not 
sure if the description helps understand the difference.

                
> Document new HA-related configs in hdfs-default.xml
> ---------------------------------------------------
>
>                 Key: HDFS-2819
>                 URL: https://issues.apache.org/jira/browse/HDFS-2819
>             Project: Hadoop HDFS
>          Issue Type: Sub-task
>          Components: documentation, ha
>    Affects Versions: HA branch (HDFS-1623)
>            Reporter: Todd Lipcon
>            Assignee: Eli Collins
>         Attachments: hdfs-2819.txt, hdfs-2819.txt, hdfs-2819.txt
>
>
> We've added a few configs, like shared edits dir, dfs.ha.namenodes, etc - we 
> should probably add these to hdfs-default.xml so they get documented.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira

        

Reply via email to