[
https://issues.apache.org/jira/browse/HDFS-2819?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13201265#comment-13201265
]
Suresh Srinivas commented on HDFS-2819:
---------------------------------------
Eli, thanks for documenting the config.
Some comments:
# Why are we changing the name to DFS_HA_NAMENODES_KEY_PREFIX?
# I am not sure what you mean by "prefix for given nameservice". Also please
mention value contains list of comma-separated namenodes.
{noformat}
+<property>
+ <name>dfs.ha.namenodes.EXAMPLENAMESERVICE</name>
+ <value></value>
+ <description>
+ The prefix for a given nameservice, contains a comma-separated
+ list of namenodes for a given nameservice (eg EXAMPLENAMESERVICE).
+ </description>
+</property>
{noformat}
# Does the existing code that finds nameservice ID work with this config added?
Is this not equivalent to adding an empty nameservice ID? Not sure how our
config behaves for <value></value>.
{noformat}
+<property>
+ <name>dfs.federation.nameservice.id</name>
+ <value></value>
+ <description>
+ The ID of this nameservice. If the nameservice ID is not
+ configured or more than one nameservice is configured for
+ dfs.federation.nameservices it is determined automatically by
+ matching the local node's address with the configured address.
+ </description>
+</property>
{noformat}
# Please comment out the properties dfs.federation.nameservices and
dfs.federation.nameservice.id.
# Can you please describe what 0 means for users? Same for
dfs.client.failover.connection.retries. Users many not understand what failover
IPC client means.
{noformat}
+<property>
+ <name>dfs.client.failover.connection.retries.on.timeouts</name>
+ <value>0</value>
+ <description>
+ Expert only. The number of retry attempts a failover IPC client
+ will make on socket timeout when establishing a server connection.
+ </description>
+</property>
{noformat}
# For client.failover, are the default retry attempts, timeouts correct? I am
sure I understand the rationale for these timeouts. The failover attempts are
made for ~ 165 seconds in following sequence ~(0.5 1 2 4 8 15 15 15 15 15 15 15
15 15 15)
# From config perspective client.failover config items are confusing. There
configs related to failover.max.attempts and failover.connection.retries. Not
sure if the description helps understand the difference.
> Document new HA-related configs in hdfs-default.xml
> ---------------------------------------------------
>
> Key: HDFS-2819
> URL: https://issues.apache.org/jira/browse/HDFS-2819
> Project: Hadoop HDFS
> Issue Type: Sub-task
> Components: documentation, ha
> Affects Versions: HA branch (HDFS-1623)
> Reporter: Todd Lipcon
> Assignee: Eli Collins
> Attachments: hdfs-2819.txt, hdfs-2819.txt, hdfs-2819.txt
>
>
> We've added a few configs, like shared edits dir, dfs.ha.namenodes, etc - we
> should probably add these to hdfs-default.xml so they get documented.
--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators:
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira