Dear Jimmy, After changing my Linux from RedHat 9 to Ubuntu 10, all the problems are solved.
Maybe RedHat 9 is too old to fit HBase and Hadoop? Thanks, Bing On Tue, Feb 14, 2012 at 5:23 AM, Bing Li <lbl...@gmail.com> wrote: > Dear Jimmy, > > I noticed that my name node was not started. That might be the reason? I > still tried to figure out why the name node was not started. > > Thanks so much! > > Bing > > > On Tue, Feb 14, 2012 at 5:11 AM, Jimmy Xiang <jxi...@cloudera.com> wrote: > >> Which HDFS/Hadoop are you using? The name node configuration for >> fs.default.name >> should be hdfs://localhost:9000 if you want your hbase.rootdir to be >> hdfs://localhost:9000/ >> hbase. They need to match. >> >> >> >> >> On Mon, Feb 13, 2012 at 11:58 AM, Bing Li <lbl...@gmail.com> wrote: >> >>> Dear Jimmy, >>> >>> Thanks so much for your instant reply! >>> >>> My hbase-site.xml is like the following. >>> >>> <property> >>> <name>hbase.rootdir</name> >>> <value>hdfs://localhost:9000/hbase</value> >>> </property> >>> <property> >>> <name>dfs.replication</name> >>> <value>1</value> >>> </property> >>> <property> >>> <name>hbase.master</name> >>> <value>localhost:60000</value> >>> </property> >>> <property> >>> <name>hbase.cluster.distributed</name> >>> <value>true</value> >>> </property> >>> <property> >>> <name>hbase.zookeeper.quorum</name> >>> <value>localhost</value> >>> </property> >>> >>> When I run hadoop fs -ls /, the directories and files under the linux >>> root are displayed. >>> >>> Best, >>> Bing >>> >>> On Tue, Feb 14, 2012 at 3:48 AM, Jimmy Xiang <jxi...@cloudera.com>wrote: >>> >>>> Which port does your HDFS listen to? It is not 9000, right? >>>> >>>> <name>hbase.rootdir</name> >>>> <value>hdfs://localhost:9000/hbase</value> >>>> >>>> You need to fix this and make sure your HDFS is working, for example, >>>> the following command should work for you. >>>> >>>> hadoop fs -ls / >>>> >>>> >>>> >>>> On Mon, Feb 13, 2012 at 11:44 AM, Bing Li <lbl...@gmail.com> wrote: >>>> >>>>> Dear Jimmy, >>>>> >>>>> I configured the standalone mode successfully. But I wonder why the >>>>> pseudo distributed one does work. >>>>> >>>>> I checked in logs and got the following exceptions. Does the >>>>> information give you some hints? >>>>> >>>>> Thanks so much for your help again! >>>>> >>>>> Best, >>>>> Bing >>>>> >>>>> 2012-02-13 18:25:49,782 FATAL org.apache.hadoop.hbase.master.HMaster: >>>>> Unhandled exception. Starting shutdown. >>>>> java.net.ConnectException: Call to localhost/127.0.0.1:9000 failed on >>>>> connection exception: java.net.ConnectException: Connection refuse >>>>> d >>>>> at org.apache.hadoop.ipc.Client.wrapException(Client.java:1095) >>>>> at org.apache.hadoop.ipc.Client.call(Client.java:1071) >>>>> at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225) >>>>> at $Proxy10.getProtocolVersion(Unknown Source) >>>>> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:396) >>>>> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:379) >>>>> at >>>>> org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:119) >>>>> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:238) >>>>> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:203) >>>>> at >>>>> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:89) >>>>> at >>>>> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1386) >>>>> at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66) >>>>> at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1404) >>>>> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:254) >>>>> at org.apache.hadoop.fs.Path.getFileSystem(Path.java:187) >>>>> at org.apache.hadoop.hbase.util.FSUtils.getRootDir(FSUtils.java:471) >>>>> at >>>>> org.apache.hadoop.hbase.master.MasterFileSystem.<init>(MasterFileSystem.java:94) >>>>> at >>>>> org.apache.hadoop.hbase.master.HMaster.finishInitialization(HMaster.java:448) >>>>> at org.apache.hadoop.hbase.master.HMaster.run(HMaster.java:326) >>>>> at java.lang.Thread.run(Thread.java:662) >>>>> Caused by: java.net.ConnectException: Connection refused >>>>> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) >>>>> at >>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567) >>>>> at >>>>> org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206) >>>>> at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:656) >>>>> at >>>>> org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:434) >>>>> at >>>>> org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:560) >>>>> at org.apache.hadoop.ipc.Client$Connection.access$2000(Client.java:184) >>>>> at org.apache.hadoop.ipc.Client.getConnection(Client.java:1202) >>>>> at org.apache.hadoop.ipc.Client.call(Client.java:1046) >>>>> ... 18 more >>>>> 2012-02-13 18:25:49,787 INFO org.apache.hadoop.hbase.master.HMaster: >>>>> Aborting >>>>> 2012-02-13 18:25:49,787 DEBUG org.apache.hadoop.hbase.master.HMaster: >>>>> Stopping service threads >>>>> >>>>> >>>>> Thanks so much! >>>>> Bing >>>>> >>>>> >>>>> On Tue, Feb 14, 2012 at 3:35 AM, Jimmy Xiang <jxi...@cloudera.com>wrote: >>>>> >>>>>> In this case, you may just use the standalone mode. You can follow >>>>>> the quick start step by step. >>>>>> >>>>>> The default zookeeper port is 2181, you don't need to configure it. >>>>>> >>>>>> >>>>>> >>>>>> On Mon, Feb 13, 2012 at 11:28 AM, Bing Li <lbl...@gmail.com> wrote: >>>>>> >>>>>>> Dear Jimmy, >>>>>>> >>>>>>> I am a new user of HBase. My experiences in HBase and Hadoop is very >>>>>>> limited. I just tried to follow some books, such as Hadoop/HBase the >>>>>>> Definitive Guide. However, I still got some problems. >>>>>>> >>>>>>> What I am trying to do is just to set up a pseudo distributed HBase >>>>>>> environment on a single node. After that, I will start my system >>>>>>> programming in Java. I hope I could deploy the system in fully >>>>>>> distributed >>>>>>> mode when my system is done. >>>>>>> >>>>>>> So what I am configuring is very simple. Do I need to set up the >>>>>>> zookeeper port in hbase-site.xml? >>>>>>> >>>>>>> Thanks so much! >>>>>>> >>>>>>> Best, >>>>>>> Bing >>>>>>> >>>>>>> >>>>>>> On Tue, Feb 14, 2012 at 3:16 AM, Jimmy Xiang <jxi...@cloudera.com>wrote: >>>>>>> >>>>>>>> Have you restarted your HBase after the change? What's the >>>>>>>> zookeeper port does your HMaster use? >>>>>>>> >>>>>>>> Can you run the following to checkout where is your HMaster as >>>>>>>> below? >>>>>>>> >>>>>>>> hbase zkcli >>>>>>>> then: get /hbase/master >>>>>>>> It should show you master location. >>>>>>>> >>>>>>>> It seems you have a distributed installation. How many >>>>>>>> regionservers do you have? Can you check your >>>>>>>> master web UI to make sure all look fine. >>>>>>>> >>>>>>>> Thanks, >>>>>>>> Jimmy >>>>>>>> >>>>>>>> >>>>>>>> On Mon, Feb 13, 2012 at 10:51 AM, Bing Li <lbl...@gmail.com> wrote: >>>>>>>> >>>>>>>>> Dear Jimmy, >>>>>>>>> >>>>>>>>> Thanks so much for your reply! >>>>>>>>> >>>>>>>>> I didn't set up the zookeeper.quorom. After getting your email, I >>>>>>>>> made a change. Now my hbase-site.xml is as follows. >>>>>>>>> >>>>>>>>> <configuration> >>>>>>>>> <property> >>>>>>>>> <name>hbase.rootdir</name> >>>>>>>>> <value>hdfs://localhost:9000/hbase</value> >>>>>>>>> </property> >>>>>>>>> <property> >>>>>>>>> <name>dfs.replication</name> >>>>>>>>> <value>1</value> >>>>>>>>> </property> >>>>>>>>> <property> >>>>>>>>> <name>hbase.cluster.distributed</name> >>>>>>>>> <value>true</value> >>>>>>>>> </property> >>>>>>>>> <property> >>>>>>>>> <name>hbase.zookeeper.quorum</name> >>>>>>>>> <value>localhost</value> >>>>>>>>> </property> >>>>>>>>> </configuration> >>>>>>>>> >>>>>>>>> The previous error is still existed. I feel weird why HBase >>>>>>>>> developers cannot provide a reliable description about their work. >>>>>>>>> >>>>>>>>> Best, >>>>>>>>> Bing >>>>>>>>> >>>>>>>>> >>>>>>>>> On Tue, Feb 14, 2012 at 2:16 AM, Jimmy Xiang >>>>>>>>> <jxi...@cloudera.com>wrote: >>>>>>>>> >>>>>>>>>> What's your hbase.zookeeper.quorom configuration? You can check >>>>>>>>>> out this quick start guide: >>>>>>>>>> >>>>>>>>>> http://hbase.apache.org/book/quickstart.html >>>>>>>>>> >>>>>>>>>> Thanks, >>>>>>>>>> Jimmy >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Mon, Feb 13, 2012 at 10:09 AM, Bing Li <lbl...@gmail.com>wrote: >>>>>>>>>> >>>>>>>>>>> Dear all, >>>>>>>>>>> >>>>>>>>>>> After searching on the Web and asking for help from friends, I >>>>>>>>>>> noticed that >>>>>>>>>>> the pseudo distributed configuration in the book, HBase the >>>>>>>>>>> Definitive >>>>>>>>>>> Guide, was not complete. Now the ZooKeeper related exception is >>>>>>>>>>> fixed. >>>>>>>>>>> However, I got another error when typing "status" in the HBase >>>>>>>>>>> shell. >>>>>>>>>>> >>>>>>>>>>> ERROR: org.apache.hadoop.hbase.MasterNotRunningException: >>>>>>>>>>> Retried 7 >>>>>>>>>>> Times >>>>>>>>>>> >>>>>>>>>>> I am trying to fix it myself. Your help is highly appreciated. >>>>>>>>>>> >>>>>>>>>>> Thanks so much! >>>>>>>>>>> Bing Li >>>>>>>>>>> >>>>>>>>>>> On Mon, Feb 13, 2012 at 5:00 AM, Bing Li <lbl...@gmail.com> >>>>>>>>>>> wrote: >>>>>>>>>>> >>>>>>>>>>> > Dear all, >>>>>>>>>>> > >>>>>>>>>>> > I am a new learner of HBase. I tried to set up my HBase on a >>>>>>>>>>> > pseudo-distributed HDFS. >>>>>>>>>>> > >>>>>>>>>>> > After starting HDFS by running ./start-dfs.sh and >>>>>>>>>>> ./start-hbase.sh, I >>>>>>>>>>> > started the HBase shell. >>>>>>>>>>> > >>>>>>>>>>> > ./hbase shell >>>>>>>>>>> > >>>>>>>>>>> > It was started properly. However, when I typed the command, >>>>>>>>>>> status, as >>>>>>>>>>> > follows. >>>>>>>>>>> > >>>>>>>>>>> > hbase(main):001:0> status >>>>>>>>>>> > >>>>>>>>>>> > It got the following exception. Since I had very limited >>>>>>>>>>> experiences to >>>>>>>>>>> > use HBase, I could not figure out what the problem was. >>>>>>>>>>> > >>>>>>>>>>> > SLF4J: Class path contains multiple SLF4J bindings. >>>>>>>>>>> > SLF4J: Found binding in >>>>>>>>>>> > >>>>>>>>>>> [jar:file:/opt/hbase-0.92.0/lib/slf4j-log4j12-1.5.8.jar!/org/slf4j/impl/StaticLoggerBinder.class] >>>>>>>>>>> > SLF4J: Found binding in >>>>>>>>>>> > >>>>>>>>>>> [jar:file:/opt/hadoop-1.0.0/lib/slf4j-log4j12-1.4.3.jar!/org/slf4j/impl/StaticLoggerBinder.class] >>>>>>>>>>> > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindingsfor an >>>>>>>>>>> > explanation. >>>>>>>>>>> > 12/02/13 04:34:01 ERROR zookeeper.RecoverableZooKeeper: >>>>>>>>>>> ZooKeeper exists >>>>>>>>>>> > failed after 3 retries >>>>>>>>>>> > 12/02/13 04:34:01 WARN zookeeper.ZKUtil: hconnection Unable to >>>>>>>>>>> set watcher >>>>>>>>>>> > on znode /hbase/master >>>>>>>>>>> > org.apache.zookeeper.KeeperException$ConnectionLossException: >>>>>>>>>>> > KeeperErrorCode = ConnectionLoss for /hbase/master >>>>>>>>>>> > at >>>>>>>>>>> > >>>>>>>>>>> org.apache.zookeeper.KeeperException.create(KeeperException.java:99) >>>>>>>>>>> > at >>>>>>>>>>> > >>>>>>>>>>> org.apache.zookeeper.KeeperException.create(KeeperException.java:51) >>>>>>>>>>> > at >>>>>>>>>>> org.apache.zookeeper.ZooKeeper.exists(ZooKeeper.java:1003) >>>>>>>>>>> > at >>>>>>>>>>> > >>>>>>>>>>> org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.exists(RecoverableZooKeeper.java:154) >>>>>>>>>>> > at >>>>>>>>>>> > >>>>>>>>>>> org.apache.hadoop.hbase.zookeeper.ZKUtil.watchAndCheckExists(ZKUtil.java:226) >>>>>>>>>>> > at >>>>>>>>>>> > >>>>>>>>>>> org.apache.hadoop.hbase.zookeeper.ZooKeeperNodeTracker.start(ZooKeeperNodeTracker.java:76) >>>>>>>>>>> > at >>>>>>>>>>> > >>>>>>>>>>> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.setupZookeeperTrackers(HConnectionManager.java:580) >>>>>>>>>>> > at >>>>>>>>>>> > >>>>>>>>>>> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.<init>(HConnectionManager.java:569) >>>>>>>>>>> > at >>>>>>>>>>> > >>>>>>>>>>> org.apache.hadoop.hbase.client.HConnectionManager.getConnection(HConnectionManager.java:186) >>>>>>>>>>> > at >>>>>>>>>>> > >>>>>>>>>>> org.apache.hadoop.hbase.client.HBaseAdmin.<init>(HBaseAdmin.java:98) >>>>>>>>>>> > at >>>>>>>>>>> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native >>>>>>>>>>> > Method) >>>>>>>>>>> > at >>>>>>>>>>> > >>>>>>>>>>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39) >>>>>>>>>>> > at >>>>>>>>>>> > >>>>>>>>>>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27) >>>>>>>>>>> > at >>>>>>>>>>> java.lang.reflect.Constructor.newInstance(Constructor.java:513) >>>>>>>>>>> > at >>>>>>>>>>> > >>>>>>>>>>> org.jruby.javasupport.JavaConstructor.newInstanceDirect(JavaConstructor.java:275) >>>>>>>>>>> > at >>>>>>>>>>> > >>>>>>>>>>> org.jruby.java.invokers.ConstructorInvoker.call(ConstructorInvoker.java:91) >>>>>>>>>>> > at >>>>>>>>>>> > >>>>>>>>>>> org.jruby.java.invokers.ConstructorInvoker.call(ConstructorInvoker.java:178) >>>>>>>>>>> > at >>>>>>>>>>> > >>>>>>>>>>> org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:322) >>>>>>>>>>> > at >>>>>>>>>>> > >>>>>>>>>>> org.jruby.runtime.callsite.CachingCallSite.callBlock(CachingCallSite.java:178) >>>>>>>>>>> > at >>>>>>>>>>> > >>>>>>>>>>> org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:182) >>>>>>>>>>> > at >>>>>>>>>>> > >>>>>>>>>>> org.jruby.java.proxies.ConcreteJavaProxy$2.call(ConcreteJavaProxy.java:47) >>>>>>>>>>> > at >>>>>>>>>>> > >>>>>>>>>>> org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:322) >>>>>>>>>>> > >>>>>>>>>>> > Could you please give me a hand? Thanks so much! >>>>>>>>>>> > >>>>>>>>>>> > Best regards, >>>>>>>>>>> > Bing >>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >