zhuxiangyi commented on a change in pull request #2981: URL: https://github.com/apache/hadoop/pull/2981#discussion_r628667273
########## File path: hadoop-hdfs-project/hadoop-hdfs-rbf/src/site/markdown/HDFSRouterFederation.md ########## @@ -241,6 +241,22 @@ Mount table permission can be set by following command: The option mode is UNIX-style permissions for the mount table. Permissions are specified in octal, e.g. 0755. By default, this is set to 0755. +#### Init ViewFs To Router +Router supports initializing the ViewFS mount point to the Router. The mapping directory protocol of ViewFS must be HDFS, and the initializer only supports one-to-one mapping. + +For example, use the following viewfs to configure the initial mount table to the router. + + <configuration> + <property> + <name>fs.viewfs.mounttable.ClusterX.link./data</name> + <value>hdfs://nn1-clusterx.example.com:8020/data</value> + </property> + </configuration> + +The ViewFS mount table can be initialized to the Router by using the following command: + + [hdfs]$ $HADOOP_HOME/bin/hdfs dfsrouteradmin -initViewFsToMountTable ClusterX Review comment: Do you mean that there are mount points for multiple clusters, can the user import all of them without specifying the cluster name? For example, the following configuration. <configuration> <property> <name>fs.viewfs.mounttable.ClusterX.link./data</name> <value>hdfs://nn1-clusterx.example.com:8020/data</value> </property> <property> <name>fs.viewfs.mounttable.ClusterY.link./project</name> <value>hdfs://nn2-clusterx.example.com:8020/project</value> </property> </configuration> -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org --------------------------------------------------------------------- To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org