That's really helpful do you know if this parcel contains phoenix-spark support?
Anirudha On Mon, May 11, 2015 at 12:16 PM, Biyuhao <[email protected]> wrote: > Hi , > > I think you can just install phoenix from cloudera manager, it's very easy. > You can read instruction here, > http://www.cloudera.com/content/cloudera/en/developers/home/cloudera-labs/apache-phoenix/install-apache-phoenix-cloudera-labs.pdf > > http://blog.cloudera.com/blog/2015/05/apache-phoenix-joins-cloudera-labs/ > > 2015-05-11 13:56 GMT+08:00 Anirudha Khanna <[email protected]>: > >> I first tried the "server" jar but kept getting an exception like "No >> coprocessor $SomeCoprocessor available." So then I went and checked which >> "server" jar had the coprocessor classes and based on this picked the >> "runnable" jar. >> >> $jar tf phoenix-server-4.4.0-HBase-1.0-runnable.jar | grep coprocessor | >> wc -l >> Picked up JAVA_TOOL_OPTIONS: -Dfile.encoding=UTF-8 >> 490 >> $jar tf phoenix-server-client-4.4.0-HBase-1.0.jar | grep coprocessor | wc >> -l >> Picked up JAVA_TOOL_OPTIONS: -Dfile.encoding=UTF-8 >> 0 >> $jar tf phoenix-server-4.4.0-HBase-1.0.jar | grep coprocessor | wc -l >> Picked up JAVA_TOOL_OPTIONS: -Dfile.encoding=UTF-8 >> 0 >> >> Is this just a case of wrong packaging or is this expected? >> >> >> -- Anirudha >> >> On Fri, May 8, 2015 at 8:50 PM, Nick Dimiduk <[email protected]> wrote: >> >>> Phoenix-runnable is actually the uberjar for the query server. Instead, >>> you'll need the "server" jar. Sorry, these names are a bit confusing. >>> >>> >>> On Friday, May 8, 2015, Anirudha Khanna <[email protected]> >>> wrote: >>> >>>> Hi All, >>>> >>>> I am trying to deploy Phoenix-4.4.0 with HBase-1.0.0 from CDH5.4.0. >>>> From the phoenix build, I copied the >>>> jar, phoenix-server-4.4.0-HBase-1.0-runnable.jar over to the HBase lib >>>> directory and was able to successfully start the HBase cluster. >>>> >>>> But when I tried to connect to the cluster using sqlline.py >>>> $zookeeperQuorum I got the following exception, >>>> >>>> Error: org.apache.hadoop.hbase.DoNotRetryIOException: SYSTEM.CATALOG: >>>> org.apache.hadoop.hbase.client.Scan.setRaw(Z)Lorg/apache/hadoop/hbase/client/Scan; >>>> at >>>> org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:84) >>>> at >>>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:1148) >>>> at >>>> org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:10515) >>>> at >>>> org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:7054) >>>> at >>>> org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:1740) >>>> at >>>> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:1722) >>>> at >>>> org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31309) >>>> at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2035) >>>> at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:107) >>>> at >>>> org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130) >>>> at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107) >>>> at java.lang.Thread.run(Thread.java:745) >>>> Caused by: java.lang.NoSuchMethodError: >>>> org.apache.hadoop.hbase.client.Scan.setRaw(Z)Lorg/apache/hadoop/hbase/client/Scan; >>>> at >>>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.buildDeletedTable(MetaDataEndpointImpl.java:925) >>>> at >>>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.loadTable(MetaDataEndpointImpl.java:1001) >>>> at >>>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:1097) >>>> ... 10 more (state=08000,code=101) >>>> >>>> Looking into the HBase master logs they looked clean, (no expcetions), >>>> but the RegionServer had a similar exception, >>>> >>>> 2015-05-08 15:43:20,039 INFO >>>> [PostOpenDeployTasks:f629ad7161c1b03ba7b0aa4459a85a60] >>>> hbase.MetaTableAccessor: Updated row >>>> SYSTEM.CATALOG,,1431079592560.f629ad7161c1b03ba7b0aa4459a85a60. with >>>> server=localhost,16201,1431079934872 >>>> 2015-05-08 15:43:23,138 ERROR >>>> [B.defaultRpcServer.handler=4,queue=1,port=16201] >>>> coprocessor.MetaDataEndpointImpl: createTable failed >>>> java.lang.NoSuchMethodError: >>>> org.apache.hadoop.hbase.client.Scan.setRaw(Z)Lorg/apache/hadoop/hbase/client/Scan; >>>> at >>>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.buildDeletedTable(MetaDataEndpointImpl.java:925) >>>> at >>>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.loadTable(MetaDataEndpointImpl.java:1001) >>>> at >>>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:1097) >>>> at >>>> org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:10515) >>>> at >>>> org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:7054) >>>> at >>>> org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:1740) >>>> at >>>> org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:1722) >>>> at >>>> org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31309) >>>> at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2035) >>>> at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:107) >>>> at >>>> org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130) >>>> at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107) >>>> at java.lang.Thread.run(Thread.java:745) >>>> >>>> >>>> Any help in helping resolve this is greatly appreciated. >>>> >>>> Cheers, >>>> Anirudha >>>> >>> >> >
