There seems to be a mismatch between the hbase versions you are using. In particular, there is a known bug when using hbase 0.20.0 with 0.20.1 and 0.20.2. The best is to just upgrade to 0.20.2
J-D On Thu, Jan 14, 2010 at 12:11 AM, Muhammad Mudassar <mudassa...@gmail.com> wrote: > Basically I am trying to create table in Hbase by using *hbaseAdmin* by > using a java programe but i am getting trouble however table is created but > it does not store anything in it when i use *batchUpdate.put* to insert > anything in it the exception shown in ide is > > Exception in thread "main" java.lang.reflect.UndeclaredThrowableException > at $Proxy1.getRegionInfo(Unknown Source) > at > org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRootRegion(HConnectionManager.java:795) > at > org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegion(HConnectionManager.java:465) > at > org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegion(HConnectionManager.java:440) > at > org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegionInMeta(HConnectionManager.java:515) > at > org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegion(HConnectionManager.java:474) > at > org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegion(HConnectionManager.java:440) > at > org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegionInMeta(HConnectionManager.java:515) > at > org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegion(HConnectionManager.java:478) > at > org.apache.hadoop.hbase.client.HConnectionManager$TableServers.locateRegion(HConnectionManager.java:440) > at > org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:159) > > Caused by: org.apache.hadoop.ipc.RemoteException: java.io.IOException: > java.lang.NoSuchMethodException: > org.apache.hadoop.hbase.regionserver.HRegionServer.getRow([B) > at java.lang.Class.getMethod(Class.java:1605) > at > org.apache.hadoop.hbase.ipc.HBaseRPC$Server.call(HBaseRPC.java:627) > at > org.apache.hadoop.hbase.ipc.HBaseServer$Handler.run(HBaseServer.java:912) > > at > org.apache.hadoop.hbase.ipc.HBaseClient.call(HBaseClient.java:701) > at > org.apache.hadoop.hbase.ipc.HBaseRPC$Invoker.invoke(HBaseRPC.java:321) > ... 14 more > Java Result: 1 > > when i checked logs of hbase master > > On Wed, Jan 13, 2010 at 10:37 PM, Jean-Daniel Cryans > <jdcry...@apache.org>wrote: > >> This is probably a question better for common-user rather than hbase. >> >> But to answer your problem, your JobTracker is able to talk to your >> Namenode but there's something wrong with the Datanode, your should >> grep its log for any exception. >> >> J-D >> >> On Wed, Jan 13, 2010 at 3:11 AM, Muhammad Mudassar <mudassa...@gmail.com> >> wrote: >> > hi i am running hadoop 0.20.1 on single node and i am getting some >> problem >> > My hdfs-site configurations are >> > <configuration> >> > <property> >> > <name>dfs.replication</name> >> > <value>1</value> >> > </property> >> > <property> >> > <name>hadoop.tmp.dir</name> >> > <value>/home/hadoop/Desktop/hadoop-store/hadoop-$hadoop</value> >> > <description>A base for other temporary directories.</description> >> > </property> >> > </configuration> >> > >> > >> > and core site configurations are >> > <configuration> >> > <property> >> > <name>fs.default.name</name> >> > <value>hdfs://localhost:54310</value> >> > </property> >> > <property> >> > <name>hadoop.tmp.dir</name> >> > <value>/home/hadoop/Desktop/hadoop-store/hadoop-$hadoop</value> >> > <description>A base for other temporary directories.</description> >> > </property> >> > </configuration> >> > >> > >> > the problem is with jobtracker log file says that >> > >> > 2010-01-13 16:00:33,015 INFO org.apache.hadoop.mapred.JobTracker: >> Scheduler >> > configured with (memSizeForMapSlotOnJT, memSizeForReduceSlotOnJT, >> > limitMaxMemForMapTasks, limitMaxMemForReduceTasks) (-1, -1, -1, -1) >> > 2010-01-13 16:00:33,043 INFO org.apache.hadoop.ipc.metrics.RpcMetrics: >> > Initializing RPC Metrics with hostName=JobTracker, port=54311 >> > 2010-01-13 16:00:38,309 INFO org.mortbay.log: Logging to >> > org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via >> > org.mortbay.log.Slf4jLog >> > 2010-01-13 16:00:38,407 INFO org.apache.hadoop.http.HttpServer: Port >> > returned by webServer.getConnectors()[0].getLocalPort() before open() is >> -1. >> > Opening the listener on 50030 >> > 2010-01-13 16:00:38,408 INFO org.apache.hadoop.http.HttpServer: >> > listener.getLocalPort() returned 50030 >> > webServer.getConnectors()[0].getLocalPort() returned 50030 >> > 2010-01-13 16:00:38,408 INFO org.apache.hadoop.http.HttpServer: Jetty >> bound >> > to port 50030 >> > 2010-01-13 16:00:38,408 INFO org.mortbay.log: jetty-6.1.14 >> > 2010-01-13 16:00:51,429 INFO org.mortbay.log: Started >> > selectchannelconnec...@0.0.0.0:50030 >> > 2010-01-13 16:00:51,430 INFO org.apache.hadoop.metrics.jvm.JvmMetrics: >> > Initializing JVM Metrics with processName=JobTracker, sessionId= >> > 2010-01-13 16:00:51,431 INFO org.apache.hadoop.mapred.JobTracker: >> JobTracker >> > up at: 54311 >> > 2010-01-13 16:00:51,431 INFO org.apache.hadoop.mapred.JobTracker: >> JobTracker >> > webserver: 50030 >> > 2010-01-13 16:00:51,574 INFO org.apache.hadoop.mapred.JobTracker: >> Cleaning >> > up the system directory >> > 2010-01-13 16:00:51,643 INFO >> > org.apache.hadoop.mapred.CompletedJobStatusStore: Completed job store is >> > inactive >> > 2010-01-13 16:00:51,674 WARN org.apache.hadoop.hdfs.DFSClient: >> DataStreamer >> > Exception: org.apache.hadoop.ipc.RemoteException: java.io.IOException: >> File >> > /home/hadoop/Desktop/hadoop-store/hadoop-$hadoop/mapred/system/ >> > jobtracker.info could only be replicated to 0 nodes, instead of 1 >> > at >> > >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1267) >> > at >> > >> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:422) >> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> > at >> > >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) >> > at >> > >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) >> > at java.lang.reflect.Method.invoke(Method.java:597) >> > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508) >> > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959) >> > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955) >> > at java.security.AccessController.doPrivileged(Native Method) >> > at javax.security.auth.Subject.doAs(Subject.java:396) >> > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953) >> > >> > at org.apache.hadoop.ipc.Client.call(Client.java:739) >> > at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220) >> > at $Proxy4.addBlock(Unknown Source) >> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> > at >> > >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) >> > at >> > >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) >> > at java.lang.reflect.Method.invoke(Method.java:597) >> > at >> > >> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82) >> > at >> > >> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59) >> > at $Proxy4.addBlock(Unknown Source) >> > at >> > >> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.locateFollowingBlock(DFSClient.java:2904) >> > at >> > >> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2786) >> > at >> > >> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:2076) >> > at >> > >> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2262) >> > >> > 2*010-01-13 16:00:51,674 WARN org.apache.hadoop.hdfs.DFSClient: Error >> > Recovery for block null bad datanode[0] nodes == null >> > 2010-01-13 16:00:51,674 WARN org.apache.hadoop.hdfs.DFSClient: Could not >> get >> > block locations. Source file >> > "/home/hadoop/Desktop/hadoop-store/hadoop-$hadoop/mapred/system/ >> > jobtracker.info" - Aborting... >> > 2010-01-13 16:00:51,674 WARN org.apache.hadoop.mapred.JobTracker: Writing >> to >> > file >> > >> hdfs://localhost:54310/home/hadoop/Desktop/hadoop-store/hadoop-$hadoop/mapred/system/ >> > jobtracker.info failed! >> > 2010-01-13 16:00:51,674 WARN org.apache.hadoop.mapred.JobTracker: >> FileSystem >> > is not ready yet! >> > 2010-01-13 16:00:51,679 WARN org.apache.hadoop.mapred.JobTracker: Failed >> to >> > initialize recovery manager. >> > org.apache.hadoop.ipc.RemoteException: java.io.IOException: File >> > /home/hadoop/Desktop/hadoop-store/hadoop-$hadoop/mapred/system/ >> > jobtracker.info could only be replicated to 0 nodes, instead of 1* >> > at >> > >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1267) >> > at >> > >> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:422) >> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> > at >> > >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) >> > at >> > >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) >> > at java.lang.reflect.Method.invoke(Method.java:597) >> > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508) >> > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959) >> > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955) >> > at java.security.AccessController.doPrivileged(Native Method) >> > at javax.security.auth.Subject.doAs(Subject.java:396) >> > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953) >> > >> > at org.apache.hadoop.ipc.Client.call(Client.java:739) >> > at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220) >> > at $Proxy4.addBlock(Unknown Source) >> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> > at >> > >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) >> > at >> > >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) >> > at java.lang.reflect.Method.invoke(Method.java:597) >> > at >> > >> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82) >> > at >> > >> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59) >> > at $Proxy4.addBlock(Unknown Source) >> > at >> > >> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.locateFollowingBlock(DFSClient.java:2904) >> > at >> > >> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2786) >> > at >> > >> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:2076) >> > at >> > >> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2262) >> > >> > >> > >> > i checked with jps it says that processes are running >> > >> > 15030 SecondaryNameNode >> > 14904 DataNode >> > 15129 JobTracker >> > 15231 TaskTracker >> > 14787 NameNode >> > >> > but log file has errors can any one tell what the problem is >> > >> >