The error indicates incompatible protobuf versions. Please take a look at 4.1.1 under http://hbase.apache.org/book.html#basic.prerequisites
Cheers On Thu, Apr 30, 2015 at 3:49 AM, Saurabh Gupta <saurabh.gu...@semusi.com> wrote: > Now able to solve the issue by setting > > SparkConf sconf = *new* SparkConf().setAppName(“App").setMaster("local") > and > > conf.set(“zookeeper.znode.parent”, “/hbase-unsecure”) > > Standalone hbase has a table 'test' > hbase(main):001:0> scan 'test' > ROW COLUMN+CELL > row1 column=cf:a, timestamp=1430234895637, > value=value1 > row2 column=cf:b, timestamp=1430234907537, > value=value2 > row3 column=cf:c, timestamp=1430234918284, > value=value3 > > Now facing this issue: > > ERROR TableInputFormat: java.io.IOException: > java.lang.reflect.InvocationTargetException > at > org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:416) > at > org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:393) > at > org.apache.hadoop.hbase.client.HConnectionManager.getConnection(HConnectionManager.java:274) > at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:194) > at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:156) > at > org.apache.hadoop.hbase.mapreduce.TableInputFormat.setConf(TableInputFormat.java:101) > at org.apache.spark.rdd.NewHadoopRDD.getPartitions(NewHadoopRDD.scala:91) > at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:219) > at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:217) > at scala.Option.getOrElse(Option.scala:120) > at org.apache.spark.rdd.RDD.partitions(RDD.scala:217) > at org.apache.spark.SparkContext.runJob(SparkContext.scala:1632) > at org.apache.spark.rdd.RDD.count(RDD.scala:1012) > at org.apache.spark.examples.HBaseTest$.main(HBaseTest.scala:58) > at org.apache.spark.examples.HBaseTest.main(HBaseTest.scala) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:606) > at > org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:607) > at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:167) > at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:190) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:111) > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > Caused by: java.lang.reflect.InvocationTargetException > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) > at > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:526) > at > org.apache.hadoop.hbase.client.HConnectionManager.createConnection(HConnectionManager.java:414) > ... 23 more > Caused by: java.lang.VerifyError: class > org.apache.hadoop.hbase.protobuf.generated.ClientProtos$Result overrides > final method getUnknownFields.()Lcom/google/protobuf/UnknownFieldSet; > at java.lang.ClassLoader.defineClass1(Native Method) > at java.lang.ClassLoader.defineClass(ClassLoader.java:800) > at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) > at java.net.URLClassLoader.defineClass(URLClassLoader.java:449) > at java.net.URLClassLoader.access$100(URLClassLoader.java:71) > at java.net.URLClassLoader$1.run(URLClassLoader.java:361) > at java.net.URLClassLoader$1.run(URLClassLoader.java:355) > at java.security.AccessController.doPrivileged(Native Method) > at java.net.URLClassLoader.findClass(URLClassLoader.java:354) > at java.lang.ClassLoader.loadClass(ClassLoader.java:425) > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) > at java.lang.ClassLoader.loadClass(ClassLoader.java:358) > at > org.apache.hadoop.hbase.protobuf.ProtobufUtil.<clinit>(ProtobufUtil.java:176) > at org.apache.hadoop.hbase.ClusterId.parseFrom(ClusterId.java:64) > at > org.apache.hadoop.hbase.zookeeper.ZKClusterId.readClusterIdZNode(ZKClusterId.java:69) > at > org.apache.hadoop.hbase.client.ZooKeeperRegistry.getClusterId(ZooKeeperRegistry.java:83) > at > org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.retrieveClusterId(HConnectionManager.java:857) > at > org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.<init>(HConnectionManager.java:662) > ... 28 more > > Exception in thread "main" java.io.IOException: No table was provided. > at > org.apache.hadoop.hbase.mapreduce.TableInputFormatBase.getSplits(TableInputFormatBase.java:154) > at org.apache.spark.rdd.NewHadoopRDD.getPartitions(NewHadoopRDD.scala:95) > at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:219) > at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:217) > at scala.Option.getOrElse(Option.scala:120) > at org.apache.spark.rdd.RDD.partitions(RDD.scala:217) > at org.apache.spark.SparkContext.runJob(SparkContext.scala:1632) > at org.apache.spark.rdd.RDD.count(RDD.scala:1012) > at org.apache.spark.examples.HBaseTest$.main(HBaseTest.scala:58) > at org.apache.spark.examples.HBaseTest.main(HBaseTest.scala) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:606) > at > org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:607) > at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:167) > at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:190) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:111) > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > > On Thu, Apr 30, 2015 at 1:25 PM, Saurabh Gupta <saurabh.gu...@semusi.com> > wrote: > >> I am using hbase -0.94.8. >> >> On Wed, Apr 29, 2015 at 11:56 PM, Ted Yu <yuzhih...@gmail.com> wrote: >> >>> Can you enable HBase DEBUG logging in log4j.properties so that we can >>> have more clue ? >>> >>> What hbase release are you using ? >>> >>> Cheers >>> >>> On Wed, Apr 29, 2015 at 4:27 AM, Saurabh Gupta <saurabh.gu...@semusi.com >>> > wrote: >>> >>>> Hi, >>>> >>>> I am working with standalone HBase. And I want to execute >>>> HBaseTest.scala (in scala examples) . >>>> >>>> I have created a test table with three rows and I just want to get the >>>> count using HBaseTest.scala >>>> >>>> I am getting this issue: >>>> >>>> 15/04/29 11:17:10 INFO BlockManagerMaster: Registered BlockManager >>>> 15/04/29 11:17:11 INFO ZooKeeper: Client >>>> environment:zookeeper.version=3.4.5-1392090, built on 09/30/2012 17:52 GMT >>>> 15/04/29 11:17:11 INFO ZooKeeper: Client environment:host.name >>>> =ip-10-144-185-113 >>>> 15/04/29 11:17:11 INFO ZooKeeper: Client >>>> environment:java.version=1.7.0_79 >>>> 15/04/29 11:17:11 INFO ZooKeeper: Client environment:java.vendor=Oracle >>>> Corporation >>>> 15/04/29 11:17:11 INFO ZooKeeper: Client >>>> environment:java.home=/usr/lib/jvm/java-7-openjdk-amd64/jre >>>> 15/04/29 11:17:11 INFO ZooKeeper: Client >>>> environment:java.class.path=/home/ubuntu/sparkfolder/conf/:/home/ubuntu/sparkfolder/assembly/target/scala-2.10/spark-assembly-1.4.0-SNAPSHOT-hadoop2.2.0.jar:/home/ubuntu/sparkfolder/lib_managed/jars/datanucleus-core-3.2.10.jar:/home/ubuntu/sparkfolder/lib_managed/jars/datanucleus-api-jdo-3.2.6.jar:/home/ubuntu/sparkfolder/lib_managed/jars/datanucleus-rdbms-3.2.9.jar >>>> 15/04/29 11:17:11 INFO ZooKeeper: Client >>>> environment:java.library.path=/usr/java/packages/lib/amd64:/usr/lib/x86_64-linux-gnu/jni:/lib/x86_64-linux-gnu:/usr/lib/x86_64-linux-gnu:/usr/lib/jni:/lib:/usr/lib >>>> 15/04/29 11:17:11 INFO ZooKeeper: Client environment:java.io.tmpdir=/tmp >>>> 15/04/29 11:17:11 INFO ZooKeeper: Client environment:java.compiler=<NA> >>>> 15/04/29 11:17:11 INFO ZooKeeper: Client environment:os.name=Linux >>>> 15/04/29 11:17:11 INFO ZooKeeper: Client environment:os.arch=amd64 >>>> 15/04/29 11:17:11 INFO ZooKeeper: Client >>>> environment:os.version=3.13.0-49-generic >>>> 15/04/29 11:17:11 INFO ZooKeeper: Client environment:user.name=root >>>> 15/04/29 11:17:11 INFO ZooKeeper: Client environment:user.home=/root >>>> 15/04/29 11:17:11 INFO ZooKeeper: Client >>>> environment:user.dir=/home/ubuntu/sparkfolder >>>> 15/04/29 11:17:11 INFO ZooKeeper: Initiating client connection, >>>> connectString=localhost:2181 sessionTimeout=90000 >>>> watcher=hconnection-0x2711025f, quorum=localhost:2181, baseZNode=/hbase >>>> 15/04/29 11:17:11 INFO RecoverableZooKeeper: Process >>>> identifier=hconnection-0x2711025f connecting to ZooKeeper >>>> ensemble=localhost:2181 >>>> 15/04/29 11:17:11 INFO ClientCnxn: Opening socket connection to server >>>> ip-10-144-185-113/10.144.185.113:2181. Will not attempt to >>>> authenticate using SASL (unknown error) >>>> 15/04/29 11:17:11 INFO ClientCnxn: Socket connection established to >>>> ip-10-144-185-113/10.144.185.113:2181, initiating session >>>> 15/04/29 11:17:11 INFO ClientCnxn: Session establishment complete on >>>> server ip-10-144-185-113/10.144.185.113:2181, sessionid = >>>> 0x14d04d506da0005, negotiated timeout = 40000 >>>> 15/04/29 11:17:11 INFO ZooKeeperRegistry: ClusterId read in ZooKeeper >>>> is null >>>> >>>> Its just stuck Not showing any error. There is no Hadoop on my machine. >>>> What could be the issue? >>>> >>>> here is hbase-site.xml: >>>> >>>> <configuration> >>>> <property> >>>> <name>hbase.zookeeper.quorum</name> >>>> <value>localhost</value> >>>> </property> >>>> >>>> <property> >>>> <name>hbase.zookeeper.property.clientPort</name> >>>> <value>2181</value> >>>> </property> >>>> <property> >>>> <name>zookeeper.znode.parent</name> >>>> <value>/hbase</value> >>>> </property> >>>> </configuration> >>>> >>>> >>> >> >