It seems that the region servers are complaining about wrong phoenix
classes for some reason. We are running 2.2.0 which is the version before
phoenix was moved to apache.

But looking at the regionserver logs are stuck complaining about
"org.apache.phoenix.coprocessor.MetaDataEndpointImpl" which IS the apache
version? We might have connected with a newer client - but how could this
trigger this?


2014-08-14 17:01:40,052 DEBUG
org.apache.hadoop.hbase.coprocessor.CoprocessorHost: Loading coprocessor
class org.apache.phoenix.coprocessor.ServerCachingEndpointImpl with path
null and priority 1
2014-08-14 17:01:40,053 WARN
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost: attribute
'coprocessor$12' has invalid coprocessor specification
'|org.apache.phoenix.coprocessor.ServerCachingEndpointImpl|1|'
2014-08-14 17:01:40,053 WARN
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost:
java.io.IOException: No jar path specified for
org.apache.phoenix.coprocessor.ServerCachingEndpointImpl
at
org.apache.hadoop.hbase.coprocessor.CoprocessorHost.load(CoprocessorHost.java:183)
at
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.loadTableCoprocessors(RegionCoprocessorHost.java:190)
at
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.<init>(RegionCoprocessorHost.java:154)
at org.apache.hadoop.hbase.regionserver.HRegion.<init>(HRegion.java:474)
at sun.reflect.GeneratedConstructorAccessor13.newInstance(Unknown Source)
at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
at
org.apache.hadoop.hbase.regionserver.HRegion.newHRegion(HRegion.java:4084)
at
org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:4267)
at
org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.openRegion(OpenRegionHandler.java:329)
at
org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.process(OpenRegionHandler.java:100)
at org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:175)
at
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:662)

2014-08-14 17:01:40,053 DEBUG
org.apache.hadoop.hbase.coprocessor.CoprocessorHost: Loading coprocessor
class org.apache.phoenix.coprocessor.MetaDataEndpointImpl with path null
and priority 1
2014-08-14 17:01:40,053 WARN
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost: attribute
'coprocessor$13' has invalid coprocessor specification
'|org.apache.phoenix.coprocessor.MetaDataEndpointImpl|1|'
2014-08-14 17:01:40,053 WARN
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost:
java.io.IOException: No jar path specified for
org.apache.phoenix.coprocessor.MetaDataEndpointImpl
at
org.apache.hadoop.hbase.coprocessor.CoprocessorHost.load(CoprocessorHost.java:183)
at
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.loadTableCoprocessors(RegionCoprocessorHost.java:190)
at
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.<init>(RegionCoprocessorHost.java:154)
at org.apache.hadoop.hbase.regionserver.HRegion.<init>(HRegion.java:474)
at sun.reflect.GeneratedConstructorAccessor13.newInstance(Unknown Source)
at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
at
org.apache.hadoop.hbase.regionserver.HRegion.newHRegion(HRegion.java:4084)
at
org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:4267)
at
org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.openRegion(OpenRegionHandler.java:329)
at
org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.process(OpenRegionHandler.java:100)
at org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:175)
at
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:662)



On Thu, Aug 14, 2014 at 4:31 PM, Kristoffer Sjögren <sto...@gmail.com>
wrote:

> Hi
>
> We are running hbase 0.94.6 cdh 4.4 and have a problem with one table not
> being assigned to any region. This is the SYSTEM.TABLE in Phoenix so all
> tables are basically non functional at the moment.
>
> When running hbck repair we get the following...
>
> ERROR: Region { meta =>
> SYSTEM.TABLE,,1385477659546.b739cf1ef14dd664ae873bf5b16e7a35., hdfs =>
> hdfs://primecluster/hbase/SYSTEM.TABLE/b739cf1ef14dd664ae873bf5b16e7a35,
> deployed =>  } not deployed on any region server.
>
> And when hbck -repair (hbck_repair.txt attachment) it never assigns a
> region and everything seems stuck.
>
> Master log keeps spewing out the following at very rapid pace (master.log
> attachement)...
>
> Any pointers?
>
> Cheers,
> -Kristoffer
>
>

Reply via email to