Hi , I think you can just install phoenix from cloudera manager, it's very easy. You can read instruction here, http://www.cloudera.com/content/cloudera/en/developers/home/cloudera-labs/apache-phoenix/install-apache-phoenix-cloudera-labs.pdf
http://blog.cloudera.com/blog/2015/05/apache-phoenix-joins-cloudera-labs/ 2015-05-11 13:56 GMT+08:00 Anirudha Khanna <[email protected]>: > I first tried the "server" jar but kept getting an exception like "No > coprocessor $SomeCoprocessor available." So then I went and checked which > "server" jar had the coprocessor classes and based on this picked the > "runnable" jar. > > $jar tf phoenix-server-4.4.0-HBase-1.0-runnable.jar | grep coprocessor | > wc -l > Picked up JAVA_TOOL_OPTIONS: -Dfile.encoding=UTF-8 > 490 > $jar tf phoenix-server-client-4.4.0-HBase-1.0.jar | grep coprocessor | wc > -l > Picked up JAVA_TOOL_OPTIONS: -Dfile.encoding=UTF-8 > 0 > $jar tf phoenix-server-4.4.0-HBase-1.0.jar | grep coprocessor | wc -l > Picked up JAVA_TOOL_OPTIONS: -Dfile.encoding=UTF-8 > 0 > > Is this just a case of wrong packaging or is this expected? > > > -- Anirudha > > On Fri, May 8, 2015 at 8:50 PM, Nick Dimiduk <[email protected]> wrote: > >> Phoenix-runnable is actually the uberjar for the query server. Instead, >> you'll need the "server" jar. Sorry, these names are a bit confusing. >> >> >> On Friday, May 8, 2015, Anirudha Khanna <[email protected]> >> wrote: >> >>> Hi All, >>> >>> I am trying to deploy Phoenix-4.4.0 with HBase-1.0.0 from CDH5.4.0. From >>> the phoenix build, I copied the >>> jar, phoenix-server-4.4.0-HBase-1.0-runnable.jar over to the HBase lib >>> directory and was able to successfully start the HBase cluster. >>> >>> But when I tried to connect to the cluster using sqlline.py >>> $zookeeperQuorum I got the following exception, >>> >>> Error: org.apache.hadoop.hbase.DoNotRetryIOException: SYSTEM.CATALOG: >>> org.apache.hadoop.hbase.client.Scan.setRaw(Z)Lorg/apache/hadoop/hbase/client/Scan; >>> at >>> org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:84) >>> at >>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:1148) >>> at >>> org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:10515) >>> at >>> org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:7054) >>> at >>> org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:1740) >>> at >>> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:1722) >>> at >>> org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31309) >>> at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2035) >>> at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:107) >>> at >>> org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130) >>> at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107) >>> at java.lang.Thread.run(Thread.java:745) >>> Caused by: java.lang.NoSuchMethodError: >>> org.apache.hadoop.hbase.client.Scan.setRaw(Z)Lorg/apache/hadoop/hbase/client/Scan; >>> at >>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.buildDeletedTable(MetaDataEndpointImpl.java:925) >>> at >>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.loadTable(MetaDataEndpointImpl.java:1001) >>> at >>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:1097) >>> ... 10 more (state=08000,code=101) >>> >>> Looking into the HBase master logs they looked clean, (no expcetions), >>> but the RegionServer had a similar exception, >>> >>> 2015-05-08 15:43:20,039 INFO >>> [PostOpenDeployTasks:f629ad7161c1b03ba7b0aa4459a85a60] >>> hbase.MetaTableAccessor: Updated row >>> SYSTEM.CATALOG,,1431079592560.f629ad7161c1b03ba7b0aa4459a85a60. with >>> server=localhost,16201,1431079934872 >>> 2015-05-08 15:43:23,138 ERROR >>> [B.defaultRpcServer.handler=4,queue=1,port=16201] >>> coprocessor.MetaDataEndpointImpl: createTable failed >>> java.lang.NoSuchMethodError: >>> org.apache.hadoop.hbase.client.Scan.setRaw(Z)Lorg/apache/hadoop/hbase/client/Scan; >>> at >>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.buildDeletedTable(MetaDataEndpointImpl.java:925) >>> at >>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.loadTable(MetaDataEndpointImpl.java:1001) >>> at >>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:1097) >>> at >>> org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:10515) >>> at >>> org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:7054) >>> at >>> org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:1740) >>> at >>> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:1722) >>> at >>> org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31309) >>> at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2035) >>> at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:107) >>> at >>> org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130) >>> at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107) >>> at java.lang.Thread.run(Thread.java:745) >>> >>> >>> Any help in helping resolve this is greatly appreciated. >>> >>> Cheers, >>> Anirudha >>> >> >
