Please check the following entry in hbase-env.sh: hbase-env.sh:# The directory where pid files are stored. /tmp by default. hbase-env.sh:# export HBASE_PID_DIR=/var/hadoop/pids
If pid file is stored under /tmp, it might have been cleaned up. On Thu, Apr 1, 2010 at 11:44 AM, Jean-Daniel Cryans <jdcry...@apache.org>wrote: > If the master doesn't shut down, it means it's waiting on something... > you looked at the logs? > > You say you ran ./jps ... did you install that in the local directory? > Also what do you mean "it didn't work as well"? What didn't work? The > command didn't return anything or the HMaster process wasn't listed? > > Also did you check the zookeeper logs like Patrick said? You should > see in there when the master tries to connect, and you should see why > it wasn't able to do so. > > To help you I need more data about your problem. > > J-D > > On Thu, Apr 1, 2010 at 11:39 AM, jayavelu jaisenthilkumar > <joysent...@gmail.com> wrote: > > Hi Daniel, > > I removed the property tags from the hbase-site.xml. > > > > Same error occurs. > > > > Also one strange behaviour, If i give ./stop-hbase.sh , the terminal > says > > stopping master .................................... > > and never stopped. > > > > I couldnt able to ./jps to check the java in this scenario, it didnt work > > aswell. So I killed the Hmaster start ( ps -ef | grep java) > > > > Also manually need to kill Hregionserver both on master, slave1 and > slave2. > > > > Any suggestions please....... > > > > Regs, > > senthil > > On 31 March 2010 19:15, Jean-Daniel Cryans <jdcry...@apache.org> wrote: > > > >> You set the tick time like this: > >> > >> <property> > >> <name>hbase.zookeeper.property.tickTime</name> > >> <value>1</value> > >> <description>Property from ZooKeeper's config zoo.cfg. > >> The number of milliseconds of each tick. See > >> zookeeper.session.timeout description. > >> </description> > >> > >> 1 means HBase has to report to zookeeper every 1 millisecond and if > >> for any reason it doesn't after 20ms, the session is expired (!!). I > >> recommend using the default value. > >> > >> Also you should keep the same config on every node, rsync can do > wonders. > >> > >> J-D > >> > >> On Wed, Mar 31, 2010 at 9:24 AM, jayavelu jaisenthilkumar > >> <joysent...@gmail.com> wrote: > >> > Hi, > >> > I am using 1 master and 2 slaves one has password for ssh. > >> > > >> > I am using hadoop0.20.1 and hbase0.20.3(direct one not upgraded one) > >> > > >> > 1)The slave one with password is could not be disabled, i removed the > >> whole > >> > .ssh directory try to ssh-keygen with passwordless phrase, still i am > >> asked > >> > for the password when i > >> > "ssh localhost" > >> > > >> > 2) I am able to run hadoop and successfuly run the Mapreduce in the > >> hadoop > >> > environment as per the Running Hadoop On Ubuntu Linux (Multi-Node > >> Cluster) > >> > by noel > >> > > >> > 3) I am now following the tutorial hbase: overview HBase 0.20.3 API > >> > > >> > Its not clearly stated as the mulitnode cluster hadoop for the > >> distributed > >> > mode hbase. > >> > > >> > I ran the hdfs and the hbase using start-dfs.sh and start-hbase.sh > >> > respectively. > >> > > >> > The master log indicates connection loss on the /hbase : ( is this > hbase > >> is > >> > created by Hbase or should we do to create it again > >> > > >> > 2010-03-31 16:45:57,850 INFO org.apache.zookeeper. > >> > ClientCnxn: Attempting connection to server Hadoopserver/ > >> 192.168.1.65:2222 > >> > 2010-03-31 16:45:57,858 INFO org.apache.zookeeper.ClientCnxn: Priming > >> > connection to java.nio.channels.SocketChannel[connected local=/ > >> > 192.168.1.65:43017 remote=Hadoopserver/192.168.1.65:2222] > >> > 2010-03-31 16:45:57,881 INFO org.apache.zookeeper.ClientCnxn: Server > >> > connection successful > >> > 2010-03-31 16:45:57,883 WARN org.apache.zookeeper.ClientCnxn: > Exception > >> > closing session 0x0 to sun.nio.ch.selectionkeyi...@11c2b67 > >> > java.io.IOException: Read error rc = -1 > java.nio.DirectByteBuffer[pos=0 > >> > lim=4 cap=4] > >> > at > >> org.apache.zookeeper.ClientCnxn$SendThread.doIO(ClientCnxn.java:701) > >> > at > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:945) > >> > 2010-03-31 16:45:57,885 WARN org.apache.zookeeper.ClientCnxn: Ignoring > >> > exception during shutdown input > >> > java.net.SocketException: Transport endpoint is not connected > >> > at sun.nio.ch.SocketChannelImpl.shutdown(Native Method) > >> > at > >> > sun.nio.ch.SocketChannelImpl.shutdownInput(SocketChannelImpl.java:640) > >> > at sun.nio.ch.SocketAdaptor.shutdownInput(SocketAdaptor.java:360) > >> > at > >> > > org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:999) > >> > at > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970) > >> > 2010-03-31 16:45:57,885 WARN org.apache.zookeeper.ClientCnxn: Ignoring > >> > exception during shutdown output > >> > java.net.SocketException: Transport endpoint is not connected > >> > at sun.nio.ch.SocketChannelImpl.shutdown(Native Method) > >> > at > >> > > sun.nio.ch.SocketChannelImpl.shutdownOutput(SocketChannelImpl.java:651) > >> > at sun.nio.ch.SocketAdaptor.shutdownOutput(SocketAdaptor.java:368) > >> > at > >> > > org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:1004) > >> > at > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970) > >> > 2010-03-31 16:45:57,933 INFO > >> org.apache.hadoop.hbase.master.RegionManager: > >> > -ROOT- region unset (but not set to be reassigned) > >> > 2010-03-31 16:45:57,934 INFO > >> org.apache.hadoop.hbase.master.RegionManager: > >> > ROOT inserted into regionsInTransition > >> > 2010-03-31 16:45:58,024 DEBUG > >> > org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper: Failed to read: > >> > org.apache.zookeeper.KeeperException$ConnectionLossException: > >> > KeeperErrorCode = ConnectionLoss for /hbase/master > >> > 2010-03-31 16:45:58,422 INFO org.apache.zookeeper.ClientCnxn: > Attempting > >> > connection to server Hadoopclient1/192.168.1.2:2222 > >> > 2010-03-31 16:45:58,423 INFO org.apache.zookeeper.ClientCnxn: Priming > >> > connection to java.nio.channels.SocketChannel[connected local=/ > >> > 192.168.1.65:51219 remote=Hadoopclient1/192.168.1.2:2222] > >> > 2010-03-31 16:45:58,423 INFO org.apache.zookeeper.ClientCnxn: Server > >> > connection successful > >> > 2010-03-31 16:45:58,436 WARN org.apache.zookeeper.ClientCnxn: > Exception > >> > closing session 0x0 to sun.nio.ch.selectionkeyi...@17b6643 > >> > java.io.IOException: Read error rc = -1 > java.nio.DirectByteBuffer[pos=0 > >> > lim=4 cap=4] > >> > at > >> org.apache.zookeeper.ClientCnxn$SendThread.doIO(ClientCnxn.java:701) > >> > at > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:945) > >> > 2010-03-31 16:45:58,437 WARN org.apache.zookeeper.ClientCnxn: Ignoring > >> > exception during shutdown input > >> > java.net.SocketException: Transport endpoint is not connected > >> > at sun.nio.ch.SocketChannelImpl.shutdown(Native Method) > >> > at > >> > sun.nio.ch.SocketChannelImpl.shutdownInput(SocketChannelImpl.java:640) > >> > at sun.nio.ch.SocketAdaptor.shutdownInput(SocketAdaptor.java:360) > >> > at > >> > > org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:999) > >> > at > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970) > >> > 2010-03-31 16:45:58,437 WARN org.apache.zookeeper.ClientCnxn: Ignoring > >> > exception during shutdown output > >> > java.net.SocketException: Transport endpoint is not connected > >> > at sun.nio.ch.SocketChannelImpl.shutdown(Native Method) > >> > at > >> > > sun.nio.ch.SocketChannelImpl.shutdownOutput(SocketChannelImpl.java:651) > >> > at sun.nio.ch.SocketAdaptor.shutdownOutput(SocketAdaptor.java:368) > >> > at > >> > > org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:1004) > >> > at > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970) > >> > 2010-03-31 16:45:58,537 WARN > >> > org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper: Failed to create > >> /hbase > >> > -- check quorum servers, > >> > currently=Hadoopclient1:2222,Hadoopclient:2222,Hadoopserver:2222 > >> > org.apache.zookeeper.KeeperException$ConnectionLossException: > >> > KeeperErrorCode = ConnectionLoss for /hbase > >> > at > >> org.apache.zookeeper.KeeperException.create(KeeperException.java:90) > >> > at > >> org.apache.zookeeper.KeeperException.create(KeeperException.java:42) > >> > at org.apache.zookeeper.ZooKeeper.create(ZooKeeper.java:608) > >> > at > >> > > >> > org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper.ensureExists(ZooKeeperWrapper.java:405) > >> > at > >> > > >> > org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper.ensureParentExists(ZooKeeperWrapper.java:428) > >> > at > >> > > >> > org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper.writeMasterAddress(ZooKeeperWrapper.java:516) > >> > at > >> > > >> > org.apache.hadoop.hbase.master.HMaster.writeAddressToZooKeeper(HMaster.java:263) > >> > at org.apache.hadoop.hbase.master.HMaster.<init>(HMaster.java:245) > >> > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native > >> Method) > >> > at > >> > > >> > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39) > >> > at > >> > > >> > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27) > >> > at java.lang.reflect.Constructor.newInstance(Constructor.java:513) > >> > at org.apache.hadoop.hbase.master.HMaster.doMain(HMaster.java:1241) > >> > at org.apache.hadoop.hbase.master.HMaster.main(HMaster.java:1282) > >> > 2010-03-31 16:45:58,549 INFO org.apache.zookeeper.ClientCnxn: > Attempting > >> > connection to server Hadoopclient/192.168.1.3:2222 > >> > 2010-03-31 16:45:58,550 INFO org.apache.zookeeper.ClientCnxn: Priming > >> > connection to java.nio.channels.SocketChannel[connected local=/ > >> > 192.168.1.65:56142 remote=Hadoopclient/192.168.1.3:2222] > >> > 2010-03-31 16:45:58,550 INFO org.apache.zookeeper.ClientCnxn: Server > >> > connection successful > >> > 2010-03-31 16:45:58,577 WARN org.apache.zookeeper.ClientCnxn: > Exception > >> > closing session 0x0 to sun.nio.ch.selectionkeyi...@76e8a7 > >> > java.io.IOException: Read error rc = -1 > java.nio.DirectByteBuffer[pos=0 > >> > lim=4 cap=4] > >> > at > >> org.apache.zookeeper.ClientCnxn$SendThread.doIO(ClientCnxn.java:701) > >> > at > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:945) > >> > 2010-03-31 16:45:58,577 WARN org.apache.zookeeper.ClientCnxn: Ignoring > >> > exception during shutdown input > >> > java.net.SocketException: Transport endpoint is not connected > >> > at sun.nio.ch.SocketChannelImpl.shutdown(Native Method) > >> > at > >> > sun.nio.ch.SocketChannelImpl.shutdownInput(SocketChannelImpl.java:640) > >> > at sun.nio.ch.SocketAdaptor.shutdownInput(SocketAdaptor.java:360) > >> > at > >> > > org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:999) > >> > at > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970) > >> > 2010-03-31 16:45:58,577 WARN org.apache.zookeeper.ClientCnxn: Ignoring > >> > exception during shutdown output > >> > java.net.SocketException: Transport endpoint is not connected > >> > at sun.nio.ch.SocketChannelImpl.shutdown(Native Method) > >> > at > >> > > sun.nio.ch.SocketChannelImpl.shutdownOutput(SocketChannelImpl.java:651) > >> > at sun.nio.ch.SocketAdaptor.shutdownOutput(SocketAdaptor.java:368) > >> > at > >> > > org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:1004) > >> > at > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970) > >> > 2010-03-31 16:45:58,678 DEBUG > >> > org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper: Failed to read: > >> > org.apache.zookeeper.KeeperException$ConnectionLossException: > >> > KeeperErrorCode = ConnectionLoss for /hbase/master > >> > > >> > my hbase-site.xml ( Master) > >> > <configuration> > >> > <property> > >> > <name>hbase.rootdir</name> > >> > <value>hdfs://Hadoopserver:54310/hbase</value> > >> > <description>The directory shared by region servers. > >> > Should be fully-qualified to include the filesystem to use. > >> > E.g: hdfs://NAMENODE_SERVER:PORT/HBASE_ROOTDIR > >> > </description> > >> > </property> > >> > <property> > >> > <name>hbase.cluster.distributed</name> > >> > <value>true</value> > >> > <description>The mode the cluster will be in. Possible values are > >> > false: standalone and pseudo-distributed setups with managed > >> Zookeeper > >> > true: fully-distributed with unmanaged Zookeeper Quorum (see > >> > hbase-env.sh) > >> > </description> > >> > </property> > >> > <property> > >> > <name>hbase.zookeeper.quorum</name> > >> > <value>Hadoopserver,Hadoopclient1,Hadoopclient</value> > >> > <description>Comma separated list of servers in the ZooKeeper > Quorum. > >> > For example, "host1.mydomain.com,host2.mydomain.com, > >> host3.mydomain.com". > >> > By default this is set to localhost for local and > pseudo-distributed > >> > modes > >> > of operation. For a fully-distributed setup, this should be set to > a > >> > full > >> > list of ZooKeeper quorum servers. If HBASE_MANAGES_ZK is set in > >> > hbase-env.sh > >> > this is the list of servers which we will start/stop ZooKeeper on. > >> > </description> > >> > </property> > >> > <property> > >> > <name>hbase.zookeeper.property.tickTime</name> > >> > <value>1</value> > >> > <description>Property from ZooKeeper's config zoo.cfg. > >> > The number of milliseconds of each tick. See > >> > zookeeper.session.timeout description. > >> > </description> > >> > </property> > >> > <property> > >> > <name>zookeeper.retries</name> > >> > <value>5</value> > >> > <description>How many times to retry connections to ZooKeeper. Used > >> for > >> > reading/writing root region location, checking/writing out of > safe > >> > mode. > >> > Used together with ${zookeeper.pause} in an exponential backoff > >> > fashion > >> > when making queries to ZooKeeper. > >> > </description> > >> > </property> > >> > <property> > >> > <name>hbase.zookeeper.property.clientPort</name> > >> > <value>2222</value> > >> > <description>Property from ZooKeeper's config zoo.cfg. > >> > the port at which the clients will connect. > >> > </description> > >> > </property> > >> > > >> > <property> > >> > <name>dfs.replication</name> > >> > <value>3</value> > >> > <description>Default block replication. > >> > The actual number of replications can be specified when the file is > >> created. > >> > The default is used if replication is not specified in create time. > >> > </description> > >> > </property> > >> > </configuration> > >> > > >> > > >> > The hbase-site.xml in slave1, slave2 > >> > <configuration> > >> > <property> > >> > <name>hbase.rootdir</name> > >> > <value>hdfs://Hadoopserver:54310/hbase</value> > >> > <description>The directory shared by region servers. > >> > Should be fully-qualified to include the filesystem to use. > >> > E.g: hdfs://NAMENODE_SERVER:PORT/HBASE_ROOTDIR > >> > </description> > >> > </property> > >> > <property> > >> > <name>hbase.cluster.distributed</name> > >> > <value>true</value> > >> > <description>The mode the cluster will be in. Possible values are > >> > false: standalone and pseudo-distributed setups with managed > >> Zookeeper > >> > true: fully-distributed with unmanaged Zookeeper Quorum (see > >> > hbase-env.sh) > >> > </description> > >> > </property> > >> > <property> > >> > <name>hbase.zookeeper.quorum</name> > >> > <value>Hadoopserver,Hadoopclient1,Hadoopclient</value> > >> > <description>Comma separated list of servers in the ZooKeeper > Quorum. > >> > For example, "host1.mydomain.com,host2.mydomain.com, > >> host3.mydomain.com". > >> > By default this is set to localhost for local and > pseudo-distributed > >> > modes > >> > of operation. For a fully-distributed setup, this should be set to > a > >> > full > >> > list of ZooKeeper quorum servers. If HBASE_MANAGES_ZK is set in > >> > hbase-env.sh > >> > this is the list of servers which we will start/stop ZooKeeper on. > >> > </description> > >> > </property> > >> > > >> > <property> > >> > <name>hbase.zookeeper.property.clientPort</name> > >> > <value>2222</value> > >> > <description>Property from ZooKeeper's config zoo.cfg. > >> > the port at which the clients will connect. > >> > </description> > >> > </property> > >> > > >> > <property> > >> > <name>dfs.replication</name> > >> > <value>3</value> > >> > <description>Default block replication. > >> > The actual number of replications can be specified when the file is > >> created. > >> > The default is used if replication is not specified in create time. > >> > </description> > >> > </property> > >> > </configuration> > >> > > >> > Regionservers ( only on Master in my case Hadoopserver) > >> > > >> > Hadoopserver > >> > Hadoopclient1 > >> > Hadoopclient > >> > > >> > > >> > Regionservers ( on slaves) > >> > localhost > >> > > >> > Blocked with this error for the past one week. Googled so much didnt > get > >> any > >> > solution. > >> > > >> > > >> > REgs. > >> > senthil > >> > > >> > > >