If the master doesn't shut down, it means it's waiting on something...
you looked at the logs?

You say you ran ./jps ... did you install that in the local directory?
Also what do you mean "it didn't work as well"? What didn't work? The
command didn't return anything or the HMaster process wasn't listed?

Also did you check the zookeeper logs like Patrick said? You should
see in there when the master tries to connect, and you should see why
it wasn't able to do so.

To help you I need more data about your problem.

J-D

On Thu, Apr 1, 2010 at 11:39 AM, jayavelu jaisenthilkumar
<joysent...@gmail.com> wrote:
> Hi Daniel,
>                   I removed the property tags from the hbase-site.xml.
>
> Same error occurs.
>
> Also one strange behaviour,  If i give ./stop-hbase.sh , the terminal says
> stopping master ....................................
> and never stopped.
>
> I couldnt able to ./jps to check the java in this scenario, it didnt work
> aswell.  So I killed the Hmaster start ( ps -ef | grep java)
>
> Also manually need to kill Hregionserver both on master, slave1 and slave2.
>
> Any suggestions please.......
>
> Regs,
> senthil
> On 31 March 2010 19:15, Jean-Daniel Cryans <jdcry...@apache.org> wrote:
>
>> You set the tick time like this:
>>
>>  <property>
>>   <name>hbase.zookeeper.property.tickTime</name>
>>   <value>1</value>
>>   <description>Property from ZooKeeper's config zoo.cfg.
>>   The number of milliseconds of each tick.  See
>>   zookeeper.session.timeout description.
>>   </description>
>>
>> 1 means HBase has to report to zookeeper every 1 millisecond and if
>> for any reason it doesn't after 20ms, the session is expired (!!). I
>> recommend using the default value.
>>
>> Also you should keep the same config on every node, rsync can do wonders.
>>
>> J-D
>>
>> On Wed, Mar 31, 2010 at 9:24 AM, jayavelu jaisenthilkumar
>> <joysent...@gmail.com> wrote:
>> > Hi,
>> >            I am using 1 master and 2 slaves one has password for ssh.
>> >
>> > I am using hadoop0.20.1 and hbase0.20.3(direct one not upgraded one)
>> >
>> > 1)The slave one with password is could not be disabled, i removed the
>> whole
>> > .ssh directory try to ssh-keygen with passwordless phrase, still i am
>> asked
>> > for the password  when i
>> > "ssh localhost"
>> >
>> > 2) I am able to run hadoop and successfuly run the Mapreduce in the
>> hadoop
>> > environment as per the Running Hadoop On Ubuntu Linux (Multi-Node
>> Cluster)
>> > by noel
>> >
>> > 3) I am now following the tutorial hbase: overview HBase 0.20.3 API
>> >
>> > Its not clearly stated as the mulitnode cluster hadoop for the
>> distributed
>> > mode hbase.
>> >
>> > I ran the hdfs and the hbase using start-dfs.sh and start-hbase.sh
>> > respectively.
>> >
>> > The master log indicates connection loss on the /hbase :  ( is this hbase
>> is
>> > created by Hbase or should we do to create it again
>> >
>> > 2010-03-31 16:45:57,850 INFO org.apache.zookeeper.
>> > ClientCnxn: Attempting connection to server Hadoopserver/
>> 192.168.1.65:2222
>> > 2010-03-31 16:45:57,858 INFO org.apache.zookeeper.ClientCnxn: Priming
>> > connection to java.nio.channels.SocketChannel[connected local=/
>> > 192.168.1.65:43017 remote=Hadoopserver/192.168.1.65:2222]
>> > 2010-03-31 16:45:57,881 INFO org.apache.zookeeper.ClientCnxn: Server
>> > connection successful
>> > 2010-03-31 16:45:57,883 WARN org.apache.zookeeper.ClientCnxn: Exception
>> > closing session 0x0 to sun.nio.ch.selectionkeyi...@11c2b67
>> > java.io.IOException: Read error rc = -1 java.nio.DirectByteBuffer[pos=0
>> > lim=4 cap=4]
>> >    at
>> org.apache.zookeeper.ClientCnxn$SendThread.doIO(ClientCnxn.java:701)
>> >    at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:945)
>> > 2010-03-31 16:45:57,885 WARN org.apache.zookeeper.ClientCnxn: Ignoring
>> > exception during shutdown input
>> > java.net.SocketException: Transport endpoint is not connected
>> >    at sun.nio.ch.SocketChannelImpl.shutdown(Native Method)
>> >    at
>> > sun.nio.ch.SocketChannelImpl.shutdownInput(SocketChannelImpl.java:640)
>> >    at sun.nio.ch.SocketAdaptor.shutdownInput(SocketAdaptor.java:360)
>> >    at
>> > org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:999)
>> >    at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970)
>> > 2010-03-31 16:45:57,885 WARN org.apache.zookeeper.ClientCnxn: Ignoring
>> > exception during shutdown output
>> > java.net.SocketException: Transport endpoint is not connected
>> >    at sun.nio.ch.SocketChannelImpl.shutdown(Native Method)
>> >    at
>> > sun.nio.ch.SocketChannelImpl.shutdownOutput(SocketChannelImpl.java:651)
>> >    at sun.nio.ch.SocketAdaptor.shutdownOutput(SocketAdaptor.java:368)
>> >    at
>> > org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:1004)
>> >    at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970)
>> > 2010-03-31 16:45:57,933 INFO
>> org.apache.hadoop.hbase.master.RegionManager:
>> > -ROOT- region unset (but not set to be reassigned)
>> > 2010-03-31 16:45:57,934 INFO
>> org.apache.hadoop.hbase.master.RegionManager:
>> > ROOT inserted into regionsInTransition
>> > 2010-03-31 16:45:58,024 DEBUG
>> > org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper: Failed to read:
>> > org.apache.zookeeper.KeeperException$ConnectionLossException:
>> > KeeperErrorCode = ConnectionLoss for /hbase/master
>> > 2010-03-31 16:45:58,422 INFO org.apache.zookeeper.ClientCnxn: Attempting
>> > connection to server Hadoopclient1/192.168.1.2:2222
>> > 2010-03-31 16:45:58,423 INFO org.apache.zookeeper.ClientCnxn: Priming
>> > connection to java.nio.channels.SocketChannel[connected local=/
>> > 192.168.1.65:51219 remote=Hadoopclient1/192.168.1.2:2222]
>> > 2010-03-31 16:45:58,423 INFO org.apache.zookeeper.ClientCnxn: Server
>> > connection successful
>> > 2010-03-31 16:45:58,436 WARN org.apache.zookeeper.ClientCnxn: Exception
>> > closing session 0x0 to sun.nio.ch.selectionkeyi...@17b6643
>> > java.io.IOException: Read error rc = -1 java.nio.DirectByteBuffer[pos=0
>> > lim=4 cap=4]
>> >    at
>> org.apache.zookeeper.ClientCnxn$SendThread.doIO(ClientCnxn.java:701)
>> >    at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:945)
>> > 2010-03-31 16:45:58,437 WARN org.apache.zookeeper.ClientCnxn: Ignoring
>> > exception during shutdown input
>> > java.net.SocketException: Transport endpoint is not connected
>> >    at sun.nio.ch.SocketChannelImpl.shutdown(Native Method)
>> >    at
>> > sun.nio.ch.SocketChannelImpl.shutdownInput(SocketChannelImpl.java:640)
>> >    at sun.nio.ch.SocketAdaptor.shutdownInput(SocketAdaptor.java:360)
>> >    at
>> > org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:999)
>> >    at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970)
>> > 2010-03-31 16:45:58,437 WARN org.apache.zookeeper.ClientCnxn: Ignoring
>> > exception during shutdown output
>> > java.net.SocketException: Transport endpoint is not connected
>> >    at sun.nio.ch.SocketChannelImpl.shutdown(Native Method)
>> >    at
>> > sun.nio.ch.SocketChannelImpl.shutdownOutput(SocketChannelImpl.java:651)
>> >    at sun.nio.ch.SocketAdaptor.shutdownOutput(SocketAdaptor.java:368)
>> >    at
>> > org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:1004)
>> >    at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970)
>> > 2010-03-31 16:45:58,537 WARN
>> > org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper: Failed to create
>> /hbase
>> > -- check quorum servers,
>> > currently=Hadoopclient1:2222,Hadoopclient:2222,Hadoopserver:2222
>> > org.apache.zookeeper.KeeperException$ConnectionLossException:
>> > KeeperErrorCode = ConnectionLoss for /hbase
>> >    at
>> org.apache.zookeeper.KeeperException.create(KeeperException.java:90)
>> >    at
>> org.apache.zookeeper.KeeperException.create(KeeperException.java:42)
>> >    at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:608)
>> >    at
>> >
>> org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper.ensureExists(ZooKeeperWrapper.java:405)
>> >    at
>> >
>> org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper.ensureParentExists(ZooKeeperWrapper.java:428)
>> >    at
>> >
>> org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper.writeMasterAddress(ZooKeeperWrapper.java:516)
>> >    at
>> >
>> org.apache.hadoop.hbase.master.HMaster.writeAddressToZooKeeper(HMaster.java:263)
>> >    at org.apache.hadoop.hbase.master.HMaster.<init>(HMaster.java:245)
>> >    at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
>> Method)
>> >    at
>> >
>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
>> >    at
>> >
>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
>> >    at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
>> >    at org.apache.hadoop.hbase.master.HMaster.doMain(HMaster.java:1241)
>> >    at org.apache.hadoop.hbase.master.HMaster.main(HMaster.java:1282)
>> > 2010-03-31 16:45:58,549 INFO org.apache.zookeeper.ClientCnxn: Attempting
>> > connection to server Hadoopclient/192.168.1.3:2222
>> > 2010-03-31 16:45:58,550 INFO org.apache.zookeeper.ClientCnxn: Priming
>> > connection to java.nio.channels.SocketChannel[connected local=/
>> > 192.168.1.65:56142 remote=Hadoopclient/192.168.1.3:2222]
>> > 2010-03-31 16:45:58,550 INFO org.apache.zookeeper.ClientCnxn: Server
>> > connection successful
>> > 2010-03-31 16:45:58,577 WARN org.apache.zookeeper.ClientCnxn: Exception
>> > closing session 0x0 to sun.nio.ch.selectionkeyi...@76e8a7
>> > java.io.IOException: Read error rc = -1 java.nio.DirectByteBuffer[pos=0
>> > lim=4 cap=4]
>> >    at
>> org.apache.zookeeper.ClientCnxn$SendThread.doIO(ClientCnxn.java:701)
>> >    at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:945)
>> > 2010-03-31 16:45:58,577 WARN org.apache.zookeeper.ClientCnxn: Ignoring
>> > exception during shutdown input
>> > java.net.SocketException: Transport endpoint is not connected
>> >    at sun.nio.ch.SocketChannelImpl.shutdown(Native Method)
>> >    at
>> > sun.nio.ch.SocketChannelImpl.shutdownInput(SocketChannelImpl.java:640)
>> >    at sun.nio.ch.SocketAdaptor.shutdownInput(SocketAdaptor.java:360)
>> >    at
>> > org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:999)
>> >    at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970)
>> > 2010-03-31 16:45:58,577 WARN org.apache.zookeeper.ClientCnxn: Ignoring
>> > exception during shutdown output
>> > java.net.SocketException: Transport endpoint is not connected
>> >    at sun.nio.ch.SocketChannelImpl.shutdown(Native Method)
>> >    at
>> > sun.nio.ch.SocketChannelImpl.shutdownOutput(SocketChannelImpl.java:651)
>> >    at sun.nio.ch.SocketAdaptor.shutdownOutput(SocketAdaptor.java:368)
>> >    at
>> > org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:1004)
>> >    at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970)
>> > 2010-03-31 16:45:58,678 DEBUG
>> > org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper: Failed to read:
>> > org.apache.zookeeper.KeeperException$ConnectionLossException:
>> > KeeperErrorCode = ConnectionLoss for /hbase/master
>> >
>> > my hbase-site.xml ( Master)
>> > <configuration>
>> >  <property>
>> >    <name>hbase.rootdir</name>
>> >    <value>hdfs://Hadoopserver:54310/hbase</value>
>> >    <description>The directory shared by region servers.
>> >    Should be fully-qualified to include the filesystem to use.
>> >    E.g: hdfs://NAMENODE_SERVER:PORT/HBASE_ROOTDIR
>> >    </description>
>> >  </property>
>> >  <property>
>> >    <name>hbase.cluster.distributed</name>
>> >    <value>true</value>
>> >    <description>The mode the cluster will be in. Possible values are
>> >      false: standalone and pseudo-distributed setups with managed
>> Zookeeper
>> >      true: fully-distributed with unmanaged Zookeeper Quorum (see
>> > hbase-env.sh)
>> >    </description>
>> >  </property>
>> >  <property>
>> >    <name>hbase.zookeeper.quorum</name>
>> >    <value>Hadoopserver,Hadoopclient1,Hadoopclient</value>
>> >    <description>Comma separated list of servers in the ZooKeeper Quorum.
>> >    For example, "host1.mydomain.com,host2.mydomain.com,
>> host3.mydomain.com".
>> >    By default this is set to localhost for local and pseudo-distributed
>> > modes
>> >    of operation. For a fully-distributed setup, this should be set to a
>> > full
>> >    list of ZooKeeper quorum servers. If HBASE_MANAGES_ZK is set in
>> > hbase-env.sh
>> >    this is the list of servers which we will start/stop ZooKeeper on.
>> >    </description>
>> >  </property>
>> >  <property>
>> >    <name>hbase.zookeeper.property.tickTime</name>
>> >    <value>1</value>
>> >    <description>Property from ZooKeeper's config zoo.cfg.
>> >    The number of milliseconds of each tick.  See
>> >    zookeeper.session.timeout description.
>> >    </description>
>> >  </property>
>> > <property>
>> >    <name>zookeeper.retries</name>
>> >    <value>5</value>
>> >    <description>How many times to retry connections to ZooKeeper. Used
>> for
>> >      reading/writing root region location, checking/writing out of safe
>> > mode.
>> >      Used together with ${zookeeper.pause} in an exponential backoff
>> > fashion
>> >      when making queries to ZooKeeper.
>> >    </description>
>> >  </property>
>> > <property>
>> > <name>hbase.zookeeper.property.clientPort</name>
>> > <value>2222</value>
>> > <description>Property from ZooKeeper's config zoo.cfg.
>> > the port at which the clients will connect.
>> > </description>
>> > </property>
>> >
>> > <property>
>> >  <name>dfs.replication</name>
>> > <value>3</value>
>> > <description>Default block replication.
>> > The actual number of replications can be specified when the file is
>> created.
>> > The default is used if replication is not specified in create time.
>> > </description>
>> > </property>
>> > </configuration>
>> >
>> >
>> > The hbase-site.xml in slave1, slave2
>> > <configuration>
>> >  <property>
>> >    <name>hbase.rootdir</name>
>> >    <value>hdfs://Hadoopserver:54310/hbase</value>
>> >    <description>The directory shared by region servers.
>> >    Should be fully-qualified to include the filesystem to use.
>> >    E.g: hdfs://NAMENODE_SERVER:PORT/HBASE_ROOTDIR
>> >    </description>
>> >  </property>
>> >  <property>
>> >    <name>hbase.cluster.distributed</name>
>> >    <value>true</value>
>> >    <description>The mode the cluster will be in. Possible values are
>> >      false: standalone and pseudo-distributed setups with managed
>> Zookeeper
>> >      true: fully-distributed with unmanaged Zookeeper Quorum (see
>> > hbase-env.sh)
>> >    </description>
>> >  </property>
>> >  <property>
>> >    <name>hbase.zookeeper.quorum</name>
>> >    <value>Hadoopserver,Hadoopclient1,Hadoopclient</value>
>> >    <description>Comma separated list of servers in the ZooKeeper Quorum.
>> >    For example, "host1.mydomain.com,host2.mydomain.com,
>> host3.mydomain.com".
>> >    By default this is set to localhost for local and pseudo-distributed
>> > modes
>> >    of operation. For a fully-distributed setup, this should be set to a
>> > full
>> >    list of ZooKeeper quorum servers. If HBASE_MANAGES_ZK is set in
>> > hbase-env.sh
>> >    this is the list of servers which we will start/stop ZooKeeper on.
>> >    </description>
>> >  </property>
>> >
>> > <property>
>> > <name>hbase.zookeeper.property.clientPort</name>
>> > <value>2222</value>
>> > <description>Property from ZooKeeper's config zoo.cfg.
>> > the port at which the clients will connect.
>> > </description>
>> > </property>
>> >
>> > <property>
>> >  <name>dfs.replication</name>
>> > <value>3</value>
>> > <description>Default block replication.
>> > The actual number of replications can be specified when the file is
>> created.
>> > The default is used if replication is not specified in create time.
>> > </description>
>> > </property>
>> > </configuration>
>> >
>> > Regionservers ( only on Master in my case Hadoopserver)
>> >
>> > Hadoopserver
>> > Hadoopclient1
>> > Hadoopclient
>> >
>> >
>> > Regionservers ( on slaves)
>> > localhost
>> >
>> > Blocked with this error for the past one week. Googled so much didnt get
>> any
>> > solution.
>> >
>> >
>> > REgs.
>> > senthil
>> >
>>
>

Reply via email to