Hi Ted, In morning today, I downloaded 4.1 from the link you provided. The problem is that i was unable to find 4.0.0-incubating release artifacts. So, i thought to use 4.1(thinking 4.1 will be a minor & compatible upgrade to 4.0) as my client. IMO, we should also have 4.0.0-incubating artifacts since its the compatible version with HDP2.1.5(6 month old release of HDP)
Thanks, Anil Gupta On Thu, Mar 5, 2015 at 9:17 PM, Ted Yu <[email protected]> wrote: > Ani: > You can find Phoenix release artifacts here: > http://archive.apache.org/dist/phoenix/ > > e.g. for 4.1.0: > http://archive.apache.org/dist/phoenix/phoenix-4.1.0/bin/ > > Cheers > > On Thu, Mar 5, 2015 at 5:26 PM, anil gupta <[email protected]> wrote: > > > @James: Could you point me to a place where i can find tar file of > > Phoenix-4.0.0-incubating release? All the links on this page are broken: > > http://www.apache.org/dyn/closer.cgi/incubator/phoenix/ > > > > On Thu, Mar 5, 2015 at 5:04 PM, anil gupta <[email protected]> > wrote: > > > > > I have tried to disable the table but since none of the RS are coming > up. > > > I am unable to do it. Am i missing something? > > > On the server side, we were using the "4.0.0-incubating". It seems like > > my > > > only option is to upgrade the server to 4.1. At-least, the HBase > cluster > > > to be UP. I just want my cluster to come and then i will disable the > > table > > > that has a Phoenix view. > > > What would be the possible side effects of using Phoenix 4.1 with > > > HDP2.1.5. > > > Even after updating to Phoenix4.1, if the problem is not fixed. What is > > > the next alternative? > > > > > > > > > On Thu, Mar 5, 2015 at 4:54 PM, Nick Dimiduk <[email protected]> > wrote: > > > > > >> Hi Anil, > > >> > > >> HDP-2.1.5 ships with Phoenix [0]. Are you using the version shipped, > or > > >> trying out a newer version? As James says, the upgrade must be servers > > >> first, then client. Also, Phoenix versions tend to be picky about > their > > >> underlying HBase version. > > >> > > >> You can also try altering the now-broken phoenix tables via HBase > shell, > > >> removing the phoenix coprocessor. I've tried this in the past with > other > > >> coprocessor-loading woes and had mixed results. Try: disable table, > > alter > > >> table, enable table. There's still sharp edges around > coprocessor-based > > >> deployment. > > >> > > >> Keep us posted, and sorry for the mess. > > >> > > >> -n > > >> > > >> [0]: > > >> > > > http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.1.7/bk_releasenotes_hdp_2.1/content/ch_relnotes-hdp-2.1.5-product.html > > >> > > >> On Thu, Mar 5, 2015 at 4:34 PM, anil gupta <[email protected]> > > wrote: > > >> > > >>> Unfortunately, we ran out of luck on this one because we are not > > running > > >>> the latest version of HBase. This property was introduced recently: > > >>> https://issues.apache.org/jira/browse/HBASE-13044 :( > > >>> Thanks, Vladimir. > > >>> > > >>> On Thu, Mar 5, 2015 at 3:44 PM, Vladimir Rodionov < > > >>> [email protected]> wrote: > > >>> > > >>>> Try the following: > > >>>> > > >>>> Update hbase-site.xml config, set > > >>>> > > >>>> hbase.coprocessor.enabed=false > > >>>> > > >>>> or: > > >>>> > > >>>> hbase.coprocessor.user.enabed=false > > >>>> > > >>>> sync config across cluster. > > >>>> > > >>>> restart the cluster > > >>>> > > >>>> than update your table's settings in hbase shell > > >>>> > > >>>> -Vlad > > >>>> > > >>>> > > >>>> On Thu, Mar 5, 2015 at 3:32 PM, anil gupta <[email protected]> > > >>>> wrote: > > >>>> > > >>>>> Hi All, > > >>>>> > > >>>>> I am using HDP2.1.5, Phoenix4-0.0 was installed on RS. I was > running > > >>>>> Phoenix4.1 client because i could not find tar file for > > >>>>> "Phoenix4-0.0-incubating". > > >>>>> I tried to create a view on existing table and then my entire > cluster > > >>>>> went down(all the RS went down. MAster is still up). > > >>>>> > > >>>>> > > >>>>> This is the exception i am seeing: > > >>>>> > > >>>>> 2015-03-05 14:30:53,296 FATAL [RS_OPEN_REGION-hdpslave8:60020-2] > > regionserver.HRegionServer: ABORTING region server bigdatabox.com > ,60020,1423589420136: > > The coprocessor org.apache.hadoop.hbase.regionserver.LocalIndexSplitter > > threw an unexpected exception > > >>>>> java.io.IOException: No jar path specified for > > org.apache.hadoop.hbase.regionserver.LocalIndexSplitter > > >>>>> at > > > org.apache.hadoop.hbase.coprocessor.CoprocessorHost.load(CoprocessorHost.java:177) > > >>>>> at > > > org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.loadTableCoprocessors(RegionCoprocessorHost.java:207) > > >>>>> at > > > org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.<init>(RegionCoprocessorHost.java:163) > > >>>>> at > > org.apache.hadoop.hbase.regionserver.HRegion.<init>(HRegion.java:555) > > >>>>> at > > org.apache.hadoop.hbase.regionserver.HRegion.<init>(HRegion.java:462) > > >>>>> at > > sun.reflect.GeneratedConstructorAccessor33.newInstance(Unknown Source) > > >>>>> at > > > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) > > >>>>> at > > java.lang.reflect.Constructor.newInstance(Constructor.java:526) > > >>>>> at > > > org.apache.hadoop.hbase.regionserver.HRegion.newHRegion(HRegion.java:4119) > > >>>>> at > > > org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:4430) > > >>>>> at > > > org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:4403) > > >>>>> at > > > org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:4359) > > >>>>> at > > > org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:4310) > > >>>>> at > > > org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.openRegion(OpenRegionHandler.java:465) > > >>>>> at > > > org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.process(OpenRegionHandler.java:139) > > >>>>> at > > org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:128) > > >>>>> at > > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > > >>>>> at > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > > >>>>> at java.lang.Thread.run(Thread.java:744) > > >>>>> > > >>>>> > > >>>>> We tried to restart the cluster. It died again. It seems, its > stucks > > at this point looking for > > >>>>> > > >>>>> LocalIndexSplitter class. How can i resolve this error? We cant do > > anything in the cluster until we fix it. > > >>>>> > > >>>>> I was thinking of disabling those tables but none of the RS is > > coming up. Can anyone suggest me how can i bail out of this BAD > situation. > > >>>>> > > >>>>> > > >>>>> -- > > >>>>> Thanks & Regards, > > >>>>> Anil Gupta > > >>>>> > > >>>> > > >>>> > > >>> > > >>> > > >>> -- > > >>> Thanks & Regards, > > >>> Anil Gupta > > >>> > > >> > > >> > > > > > > > > > -- > > > Thanks & Regards, > > > Anil Gupta > > > > > > > > > > > -- > > Thanks & Regards, > > Anil Gupta > > > -- Thanks & Regards, Anil Gupta
