Hi All

I have 5 node hadoop cluster with 2 HA name nodes runs in active / standby
mode .And im going to setup Hbase cluster on top of that .

So my issue is how can point Hmaster and Hregion servers to active namenode
from hbase config  ? ( When namenode failover happen Hreagion or Hmaster
servers should automatically point to the active namenode )

is it like below


*hbase-site.xml----------------------*

<configuration>
  <property>
    <name>hbase.rootdir</name>
    <value>hdfs://ha-cluster/hbase</value>
  </property>
  <property>
    <name>hbase.cluster.distributed</name>
    <value>true</value>
  </property>
  <property>
    <name>hbase.zookeeper.quorum</name>
    <value>dn1.cluster.com,nn1.cluster.com,nn2.cluster.com</value>
  </property>
</configuration>


but when i made the above config and runs *./start-hbase.sh ,  Hreagion*
servers throws the below error .and *Hmaster* runs on one of the namenode
servers and it starts fine without any error


aused by: java.lang.IllegalArgumentException:
java.net.UnknownHostException: ha-cluster





Hadoop configuration
================


*hdfs-site.xml --------------------*

<configuration>


<property>
 <name>dfs.namenode.name.dir</name>
 <value>/usr/local/hadoop/data/namenode</value>
 </property>
 <property>
 <name>dfs.replication</name>
 <value>3</value>
 </property>
 <property>
 <name>dfs.permissions</name>
 <value>false</value>
 </property>
 <property>
 <name>dfs.nameservices</name>
 <value>ha-cluster</value>
 </property>
 <property>
 <name>dfs.ha.namenodes.ha-cluster</name>
 <value>nn1,nn2</value>
 </property>
 <property>
 <name>dfs.namenode.rpc-address.ha-cluster.nn1</name>
 <value>nn1.cluster.com:9000</value>
 </property>
 <property>
 <name>dfs.namenode.rpc-address.ha-cluster.nn2</name>
 <value>nn2.cluster.com:9000</value>
 </property>
 <property>
 <name>dfs.namenode.http-address.ha-cluster.nn1</name>
 <value>nn1.cluster.com:50070</value>
 </property>
 <property>
 <name>dfs.namenode.http-address.ha-cluster.nn2</name>
 <value>nn2.cluster.com:50070</value>
 </property>
 <property>
 <name>dfs.namenode.shared.edits.dir</name>
 <value>qjournal://nn1.cluster.com:8485;nn2.cluster.com:8485;
dn1.cluster.com:8485/ha-cluster</value>
 </property>
 <property>
 <name>dfs.client.failover.proxy.provider.ha-cluster</name>
 
<value>org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider</value>
 </property>
 <property>
 <name>dfs.ha.automatic-failover.enabled</name>
 <value>true</value>
 </property>
 <property>
 <name>ha.zookeeper.quorum</name>
 <value>nn1.cluster.com:2181,nn2.cluster.com:2181,dn1.cluster.com:2181
</value>
 </property>

 <property>
 <name>dfs.ha.fencing.methods</name>
 <value>sshfence</value>
 </property>

 <property>
 <name>dfs.ha.fencing.ssh.private-key-files</name>
 <value>/home/hadoop/.ssh/id_rsa</value>
 </property>

<property>
<name>dfs.hosts.exclude</name>
<value>/usr/local/hadoop/current/etc/hadoop/exclude</value>
</property>

</configuration>


*Core-site.xml*
--------------------

<configuration>

<property>
<name>fs.defaultFS</name>
<value>hdfs://ha-cluster</value>
</property>

<property>
<name>dfs.journalnode.edits.dir</name>
<value>/usr/local/hadoop/data/jn</value>
</property>

</configuration>



Thank You
Dhanushka

Reply via email to