Hadoop has lots of different configurations in core-site.xml, hdfs-site.xml, ... all of which eventually get loaded into the Configuration object used to create a FileSystem instance. There are so many different configurations related to security, HA, etc. that it is almost impossible for me to guess exactly which ones you need to have set correctly to make this work. Typically what we do for storm to be able to talk to HDFS is to package the complete set of configs that appear on a Hadoop Gateway with the topology jar when it is shipped. This guarantees that the config is the same as on the gateway and should behave the same way. You can also grab them from the name node or any of the hadoop compute nodes. This will work for the HdfsBolt that loads default configurations from the classpath before overriding them with any custom configurations you set for that bolt.
- Bobby On Thursday, February 19, 2015 10:42 AM, clay teahouse <clayteaho...@gmail.com> wrote: Bobby,What do you mean by client here? In this context, do you consider hdfsbolt a client? If yes, then which configuration you are referring to? I've seen the following, but I am not sure if I follow. - dfs.client.failover.proxy.provider.[nameservice ID] - the Java class that HDFS clients use to contact the Active NameNodeConfigure the name of the Java class which will be used by the DFS Client to determine which NameNode is the current Active, and therefore which NameNode is currently serving client requests. The only implementation which currently ships with Hadoop is the ConfiguredFailoverProxyProvider, so use this unless you are using a custom one. For example: <property> <name>dfs.client.failover.proxy.provider.mycluster</name> <value>org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider</value> </property> thanks,Clay On Thu, Feb 19, 2015 at 8:38 AM, Bobby Evans <ev...@yahoo-inc.com.invalid> wrote: HDFS HA provides fail-over for the name node and the client determines which name node is the active one but should be completely transparent to you if the client is configured correctly. - Bobby On Thursday, February 19, 2015 6:47 AM, clay teahouse <clayteaho...@gmail.com> wrote: Hi All, Has anyone used HdfsBolt with hdfs in HA mode? How would you determine which hdfs node is the active node? thanks Clay