On the first connection to the cluster when you've installed Phoenix
2.2.3 and were previously using Phoenix 2.2.2, Phoenix will upgrade
your Phoenix tables to use the new coprocessor names
(org.apache.phoenix.*) instead of the old coprocessor names
(com.salesforce.phoenix.*).

Thanks,
James

On Thu, Aug 14, 2014 at 8:38 AM, Ted Yu <[email protected]> wrote:
> Adding Phoenix dev@
>
>
>
> On Thu, Aug 14, 2014 at 8:05 AM, Kristoffer Sjögren <[email protected]>
> wrote:
>
>> It seems that the region servers are complaining about wrong phoenix
>> classes for some reason. We are running 2.2.0 which is the version before
>> phoenix was moved to apache.
>>
>> But looking at the regionserver logs are stuck complaining about
>> "org.apache.phoenix.coprocessor.MetaDataEndpointImpl" which IS the apache
>> version? We might have connected with a newer client - but how could this
>> trigger this?
>>
>>
>> 2014-08-14 17:01:40,052 DEBUG
>> org.apache.hadoop.hbase.coprocessor.CoprocessorHost: Loading coprocessor
>> class org.apache.phoenix.coprocessor.ServerCachingEndpointImpl with path
>> null and priority 1
>> 2014-08-14 17:01:40,053 WARN
>> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost: attribute
>> 'coprocessor$12' has invalid coprocessor specification
>> '|org.apache.phoenix.coprocessor.ServerCachingEndpointImpl|1|'
>> 2014-08-14 17:01:40,053 WARN
>> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost:
>> java.io.IOException: No jar path specified for
>> org.apache.phoenix.coprocessor.ServerCachingEndpointImpl
>> at
>>
>> org.apache.hadoop.hbase.coprocessor.CoprocessorHost.load(CoprocessorHost.java:183)
>> at
>>
>> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.loadTableCoprocessors(RegionCoprocessorHost.java:190)
>> at
>>
>> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.<init>(RegionCoprocessorHost.java:154)
>> at org.apache.hadoop.hbase.regionserver.HRegion.<init>(HRegion.java:474)
>> at sun.reflect.GeneratedConstructorAccessor13.newInstance(Unknown Source)
>> at
>>
>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
>> at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
>> at
>> org.apache.hadoop.hbase.regionserver.HRegion.newHRegion(HRegion.java:4084)
>> at
>> org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:4267)
>> at
>>
>> org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.openRegion(OpenRegionHandler.java:329)
>> at
>>
>> org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.process(OpenRegionHandler.java:100)
>> at org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:175)
>> at
>>
>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>> at
>>
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>> at java.lang.Thread.run(Thread.java:662)
>>
>> 2014-08-14 17:01:40,053 DEBUG
>> org.apache.hadoop.hbase.coprocessor.CoprocessorHost: Loading coprocessor
>> class org.apache.phoenix.coprocessor.MetaDataEndpointImpl with path null
>> and priority 1
>> 2014-08-14 17:01:40,053 WARN
>> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost: attribute
>> 'coprocessor$13' has invalid coprocessor specification
>> '|org.apache.phoenix.coprocessor.MetaDataEndpointImpl|1|'
>> 2014-08-14 17:01:40,053 WARN
>> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost:
>> java.io.IOException: No jar path specified for
>> org.apache.phoenix.coprocessor.MetaDataEndpointImpl
>> at
>>
>> org.apache.hadoop.hbase.coprocessor.CoprocessorHost.load(CoprocessorHost.java:183)
>> at
>>
>> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.loadTableCoprocessors(RegionCoprocessorHost.java:190)
>> at
>>
>> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.<init>(RegionCoprocessorHost.java:154)
>> at org.apache.hadoop.hbase.regionserver.HRegion.<init>(HRegion.java:474)
>> at sun.reflect.GeneratedConstructorAccessor13.newInstance(Unknown Source)
>> at
>>
>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
>> at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
>> at
>> org.apache.hadoop.hbase.regionserver.HRegion.newHRegion(HRegion.java:4084)
>> at
>> org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:4267)
>> at
>>
>> org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.openRegion(OpenRegionHandler.java:329)
>> at
>>
>> org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.process(OpenRegionHandler.java:100)
>> at org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:175)
>> at
>>
>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>> at
>>
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>> at java.lang.Thread.run(Thread.java:662)
>>
>>
>>
>> On Thu, Aug 14, 2014 at 4:31 PM, Kristoffer Sjögren <[email protected]>
>> wrote:
>>
>> > Hi
>> >
>> > We are running hbase 0.94.6 cdh 4.4 and have a problem with one table not
>> > being assigned to any region. This is the SYSTEM.TABLE in Phoenix so all
>> > tables are basically non functional at the moment.
>> >
>> > When running hbck repair we get the following...
>> >
>> > ERROR: Region { meta =>
>> > SYSTEM.TABLE,,1385477659546.b739cf1ef14dd664ae873bf5b16e7a35., hdfs =>
>> > hdfs://primecluster/hbase/SYSTEM.TABLE/b739cf1ef14dd664ae873bf5b16e7a35,
>> > deployed =>  } not deployed on any region server.
>> >
>> > And when hbck -repair (hbck_repair.txt attachment) it never assigns a
>> > region and everything seems stuck.
>> >
>> > Master log keeps spewing out the following at very rapid pace (master.log
>> > attachement)...
>> >
>> > Any pointers?
>> >
>> > Cheers,
>> > -Kristoffer
>> >
>> >
>>

Reply via email to