Hi,

Following are not required I guess. I am able to connect to cluster without
these. Is there any reason to include them?

dfs.namenode.http-address.${dfs.nameservices}.nn1

dfs.namenode.http-address.${dfs.nameservices}.nn2

Regards

Pushparaj



On Wed, Oct 12, 2016 at 6:39 AM, 권병창 <magnu...@navercorp.com> wrote:

> Hi.
>
>
>
> 1. minimal configuration to connect HA namenode is below properties.
>
> zookeeper information does not necessary.
>
>
>
> dfs.nameservices
>
> dfs.ha.namenodes.${dfs.nameservices}
>
> dfs.namenode.rpc-address.${dfs.nameservices}.nn1
>
> dfs.namenode.rpc-address.${dfs.nameservices}.nn2
>
> dfs.namenode.http-address.${dfs.nameservices}.nn1
>
> dfs.namenode.http-address.${dfs.nameservices}.nn2
> dfs.client.failover.proxy.provider.c3=org.apache.hadoop.
> hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider
>
>
>
>
>
> 2. client use round robin manner for selecting active namenode.
>
>
>
>
>
> -----Original Message-----
> *From:* "Pushparaj Motamari"<pushpara...@gmail.com>
> *To:* <user@hadoop.apache.org>;
> *Cc:*
> *Sent:* 2016-10-12 (수) 03:20:53
> *Subject:* Connecting Hadoop HA cluster via java client
>
> Hi,
>
> I have two questions pertaining to accessing the hadoop ha cluster from
> java client.
>
> 1. Is  it necessary to supply
>
> conf.set("dfs.ha.automatic-failover.enabled",true);
>
> and
>
> conf.set("ha.zookeeper.quorum","zk1.example.com:2181,zk2.example.com:2181,zk3.example.com:2181");
>
> in addition to the other properties set in the code below?
>
> private Configuration initHAConf(URI journalURI, Configuration conf) {
>   conf.set(DFSConfigKeys.DFS_NAMENODE_SHARED_EDITS_DIR_KEY,
>       journalURI.toString());
>
>   String address1 = "127.0.0.1:" + NN1_IPC_PORT;
>   String address2 = "127.0.0.1:" + NN2_IPC_PORT;
>   conf.set(DFSUtil.addKeySuffixes(DFS_NAMENODE_RPC_ADDRESS_KEY,
>       NAMESERVICE, NN1), address1);
>   conf.set(DFSUtil.addKeySuffixes(DFS_NAMENODE_RPC_ADDRESS_KEY,
>       NAMESERVICE, NN2), address2);
>   conf.set(DFSConfigKeys.DFS_NAMESERVICES, NAMESERVICE);
>   conf.set(DFSUtil.addKeySuffixes(DFS_HA_NAMENODES_KEY_PREFIX, NAMESERVICE),
>       NN1 + "," + NN2);
>   conf.set(DFS_CLIENT_FAILOVER_PROXY_PROVIDER_KEY_PREFIX + "." + NAMESERVICE,
>       ConfiguredFailoverProxyProvider.class.getName());
>   conf.set("fs.defaultFS", "hdfs://" + NAMESERVICE);
>
>   return conf;}
>
> 2. If we supply zookeeper configuration details as mentioned in the question 
> 1 is it necessary to set the primary and secondary namenode addresses as 
> mentioned in the code above? Since we have
> given zookeeper connection details the client should be able to figure out 
> the active namenode connection details.
>
>
> Regards
>
> Pushparaj
>
>

Reply via email to