np, thanks for the update Anil. On 8 March 2015 at 00:00, anil gupta <[email protected]> wrote:
> Hi Chandu, > > Unfortunately, Its a company private event so i wont me able to make it > public. > > Thanks, > Anil Gupta > > On Sat, Mar 7, 2015 at 2:03 AM, Chandu <[email protected]> wrote: > >> Hi Anil, >> >> Is it a webinar? How can I join the meeting? >> >> Thanks, >> Chandu. >> >> On 7 March 2015 at 02:20, anil gupta <[email protected]> wrote: >> >>> Hi James/Mujtaba, >>> >>> I am giving a tech talk of HBase on Monday morning. I wanted to demo >>> Phoenix as part of that. Installation of 4.0.0 jars can only be done in >>> office hours because i am dependent on other team to do it. If i can get >>> the jar in 1-2 hours. I would really appreciate it. >>> >>> Thanks, >>> Anil Gupta >>> >>> >>> On Thu, Mar 5, 2015 at 10:10 PM, James Taylor <[email protected]> >>> wrote: >>> >>>> Mujtaba - do you know where our 4.0.0-incubating artifacts are? >>>> >>>> On Thu, Mar 5, 2015 at 9:58 PM, anil gupta <[email protected]> >>>> wrote: >>>> > Hi Ted, >>>> > >>>> > In morning today, I downloaded 4.1 from the link you provided. The >>>> problem >>>> > is that i was unable to find 4.0.0-incubating release artifacts. So, i >>>> > thought to use 4.1(thinking 4.1 will be a minor & compatible upgrade >>>> to 4.0) >>>> > as my client. >>>> > IMO, we should also have 4.0.0-incubating artifacts since its the >>>> compatible >>>> > version with HDP2.1.5(6 month old release of HDP) >>>> > >>>> > Thanks, >>>> > Anil Gupta >>>> > >>>> > On Thu, Mar 5, 2015 at 9:17 PM, Ted Yu <[email protected]> wrote: >>>> >> >>>> >> Ani: >>>> >> You can find Phoenix release artifacts here: >>>> >> http://archive.apache.org/dist/phoenix/ >>>> >> >>>> >> e.g. for 4.1.0: >>>> >> http://archive.apache.org/dist/phoenix/phoenix-4.1.0/bin/ >>>> >> >>>> >> Cheers >>>> >> >>>> >> On Thu, Mar 5, 2015 at 5:26 PM, anil gupta <[email protected]> >>>> wrote: >>>> >> >>>> >> > @James: Could you point me to a place where i can find tar file of >>>> >> > Phoenix-4.0.0-incubating release? All the links on this page are >>>> broken: >>>> >> > http://www.apache.org/dyn/closer.cgi/incubator/phoenix/ >>>> >> > >>>> >> > On Thu, Mar 5, 2015 at 5:04 PM, anil gupta <[email protected]> >>>> >> > wrote: >>>> >> > >>>> >> > > I have tried to disable the table but since none of the RS are >>>> coming >>>> >> > > up. >>>> >> > > I am unable to do it. Am i missing something? >>>> >> > > On the server side, we were using the "4.0.0-incubating". It >>>> seems >>>> >> > > like >>>> >> > my >>>> >> > > only option is to upgrade the server to 4.1. At-least, the HBase >>>> >> > > cluster >>>> >> > > to be UP. I just want my cluster to come and then i will disable >>>> the >>>> >> > table >>>> >> > > that has a Phoenix view. >>>> >> > > What would be the possible side effects of using Phoenix 4.1 with >>>> >> > > HDP2.1.5. >>>> >> > > Even after updating to Phoenix4.1, if the problem is not fixed. >>>> What >>>> >> > > is >>>> >> > > the next alternative? >>>> >> > > >>>> >> > > >>>> >> > > On Thu, Mar 5, 2015 at 4:54 PM, Nick Dimiduk <[email protected] >>>> > >>>> >> > > wrote: >>>> >> > > >>>> >> > >> Hi Anil, >>>> >> > >> >>>> >> > >> HDP-2.1.5 ships with Phoenix [0]. Are you using the version >>>> shipped, >>>> >> > >> or >>>> >> > >> trying out a newer version? As James says, the upgrade must be >>>> >> > >> servers >>>> >> > >> first, then client. Also, Phoenix versions tend to be picky >>>> about >>>> >> > >> their >>>> >> > >> underlying HBase version. >>>> >> > >> >>>> >> > >> You can also try altering the now-broken phoenix tables via >>>> HBase >>>> >> > >> shell, >>>> >> > >> removing the phoenix coprocessor. I've tried this in the past >>>> with >>>> >> > >> other >>>> >> > >> coprocessor-loading woes and had mixed results. Try: disable >>>> table, >>>> >> > alter >>>> >> > >> table, enable table. There's still sharp edges around >>>> >> > >> coprocessor-based >>>> >> > >> deployment. >>>> >> > >> >>>> >> > >> Keep us posted, and sorry for the mess. >>>> >> > >> >>>> >> > >> -n >>>> >> > >> >>>> >> > >> [0]: >>>> >> > >> >>>> >> > >>>> >> > >>>> http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.1.7/bk_releasenotes_hdp_2.1/content/ch_relnotes-hdp-2.1.5-product.html >>>> >> > >> >>>> >> > >> On Thu, Mar 5, 2015 at 4:34 PM, anil gupta < >>>> [email protected]> >>>> >> > wrote: >>>> >> > >> >>>> >> > >>> Unfortunately, we ran out of luck on this one because we are >>>> not >>>> >> > running >>>> >> > >>> the latest version of HBase. This property was introduced >>>> recently: >>>> >> > >>> https://issues.apache.org/jira/browse/HBASE-13044 :( >>>> >> > >>> Thanks, Vladimir. >>>> >> > >>> >>>> >> > >>> On Thu, Mar 5, 2015 at 3:44 PM, Vladimir Rodionov < >>>> >> > >>> [email protected]> wrote: >>>> >> > >>> >>>> >> > >>>> Try the following: >>>> >> > >>>> >>>> >> > >>>> Update hbase-site.xml config, set >>>> >> > >>>> >>>> >> > >>>> hbase.coprocessor.enabed=false >>>> >> > >>>> >>>> >> > >>>> or: >>>> >> > >>>> >>>> >> > >>>> hbase.coprocessor.user.enabed=false >>>> >> > >>>> >>>> >> > >>>> sync config across cluster. >>>> >> > >>>> >>>> >> > >>>> restart the cluster >>>> >> > >>>> >>>> >> > >>>> than update your table's settings in hbase shell >>>> >> > >>>> >>>> >> > >>>> -Vlad >>>> >> > >>>> >>>> >> > >>>> >>>> >> > >>>> On Thu, Mar 5, 2015 at 3:32 PM, anil gupta < >>>> [email protected]> >>>> >> > >>>> wrote: >>>> >> > >>>> >>>> >> > >>>>> Hi All, >>>> >> > >>>>> >>>> >> > >>>>> I am using HDP2.1.5, Phoenix4-0.0 was installed on RS. I was >>>> >> > >>>>> running >>>> >> > >>>>> Phoenix4.1 client because i could not find tar file for >>>> >> > >>>>> "Phoenix4-0.0-incubating". >>>> >> > >>>>> I tried to create a view on existing table and then my entire >>>> >> > >>>>> cluster >>>> >> > >>>>> went down(all the RS went down. MAster is still up). >>>> >> > >>>>> >>>> >> > >>>>> >>>> >> > >>>>> This is the exception i am seeing: >>>> >> > >>>>> >>>> >> > >>>>> 2015-03-05 14:30:53,296 FATAL >>>> [RS_OPEN_REGION-hdpslave8:60020-2] >>>> >> > regionserver.HRegionServer: ABORTING region server >>>> >> > bigdatabox.com,60020,1423589420136: >>>> >> > The coprocessor >>>> org.apache.hadoop.hbase.regionserver.LocalIndexSplitter >>>> >> > threw an unexpected exception >>>> >> > >>>>> java.io.IOException: No jar path specified for >>>> >> > org.apache.hadoop.hbase.regionserver.LocalIndexSplitter >>>> >> > >>>>> at >>>> >> > >>>> >> > >>>> org.apache.hadoop.hbase.coprocessor.CoprocessorHost.load(CoprocessorHost.java:177) >>>> >> > >>>>> at >>>> >> > >>>> >> > >>>> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.loadTableCoprocessors(RegionCoprocessorHost.java:207) >>>> >> > >>>>> at >>>> >> > >>>> >> > >>>> org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.<init>(RegionCoprocessorHost.java:163) >>>> >> > >>>>> at >>>> >> > >>>> org.apache.hadoop.hbase.regionserver.HRegion.<init>(HRegion.java:555) >>>> >> > >>>>> at >>>> >> > >>>> org.apache.hadoop.hbase.regionserver.HRegion.<init>(HRegion.java:462) >>>> >> > >>>>> at >>>> >> > sun.reflect.GeneratedConstructorAccessor33.newInstance(Unknown >>>> Source) >>>> >> > >>>>> at >>>> >> > >>>> >> > >>>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) >>>> >> > >>>>> at >>>> >> > java.lang.reflect.Constructor.newInstance(Constructor.java:526) >>>> >> > >>>>> at >>>> >> > >>>> >> > >>>> org.apache.hadoop.hbase.regionserver.HRegion.newHRegion(HRegion.java:4119) >>>> >> > >>>>> at >>>> >> > >>>> >> > >>>> org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:4430) >>>> >> > >>>>> at >>>> >> > >>>> >> > >>>> org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:4403) >>>> >> > >>>>> at >>>> >> > >>>> >> > >>>> org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:4359) >>>> >> > >>>>> at >>>> >> > >>>> >> > >>>> org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:4310) >>>> >> > >>>>> at >>>> >> > >>>> >> > >>>> org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.openRegion(OpenRegionHandler.java:465) >>>> >> > >>>>> at >>>> >> > >>>> >> > >>>> org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.process(OpenRegionHandler.java:139) >>>> >> > >>>>> at >>>> >> > >>>> org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:128) >>>> >> > >>>>> at >>>> >> > >>>> >> > >>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>>> >> > >>>>> at >>>> >> > >>>> >> > >>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>>> >> > >>>>> at java.lang.Thread.run(Thread.java:744) >>>> >> > >>>>> >>>> >> > >>>>> >>>> >> > >>>>> We tried to restart the cluster. It died again. It seems, its >>>> >> > >>>>> stucks >>>> >> > at this point looking for >>>> >> > >>>>> >>>> >> > >>>>> LocalIndexSplitter class. How can i resolve this error? We >>>> cant do >>>> >> > anything in the cluster until we fix it. >>>> >> > >>>>> >>>> >> > >>>>> I was thinking of disabling those tables but none of the RS >>>> is >>>> >> > coming up. Can anyone suggest me how can i bail out of this BAD >>>> >> > situation. >>>> >> > >>>>> >>>> >> > >>>>> >>>> >> > >>>>> -- >>>> >> > >>>>> Thanks & Regards, >>>> >> > >>>>> Anil Gupta >>>> >> > >>>>> >>>> >> > >>>> >>>> >> > >>>> >>>> >> > >>> >>>> >> > >>> >>>> >> > >>> -- >>>> >> > >>> Thanks & Regards, >>>> >> > >>> Anil Gupta >>>> >> > >>> >>>> >> > >> >>>> >> > >> >>>> >> > > >>>> >> > > >>>> >> > > -- >>>> >> > > Thanks & Regards, >>>> >> > > Anil Gupta >>>> >> > > >>>> >> > >>>> >> > >>>> >> > >>>> >> > -- >>>> >> > Thanks & Regards, >>>> >> > Anil Gupta >>>> >> > >>>> > >>>> > >>>> > >>>> > >>>> > -- >>>> > Thanks & Regards, >>>> > Anil Gupta >>>> >>> >>> >>> >>> -- >>> Thanks & Regards, >>> Anil Gupta >>> >> >> >> >> -- >> Cheers, >> Chandu. >> > > > > -- > Thanks & Regards, > Anil Gupta > -- Cheers, Chandu.
