Thanks, in that case it sounds like it is more related to Hadoop classpath mixups, rather than class loading.
On Mon, Mar 26, 2018 at 3:03 PM, ashish pok <ashish...@yahoo.com> wrote: > Stephan, we are in 1.4.2. > > Thanks, > > -- Ashish > > On Mon, Mar 26, 2018 at 7:38 AM, Stephan Ewen > <se...@apache.org> wrote: > If you are on Flink 1.4.0 or 1.4.1, please check if you accidentally have > Hadoop in your application jar. That can mess up things with child-first > classloading. 1.4.2 should handle Hadoop properly in any case. > > On Sun, Mar 25, 2018 at 3:26 PM, Ashish Pokharel <ashish...@yahoo.com> > wrote: > > Hi Ken, > > Yes - we are on 1.4. Thanks for that link - it certainly now explains how > things are working :) > > We currently don’t have HADOOP_CLASSPATH env var setup and “hadoop class > path” command basically points to HDP2.6 locations (HDP = Hortonworks Data > Platform). Best guess I have for this right now is HDP2.6 back ported some > 2.9 changes into their distro. This is on my list to get to the bottom of > (hopefully no hiccups till prod) - we double checked our Salt Orchestration > packages which were used to built the cluster but couldn’t find a reference > to hadoop 2.9. For now, we are moving on with our testing to prepare for > deployment with hadoop free version which is using hadoop classpath as > described in FLINK-7477. > > Thanks, Ashish > > On Mar 23, 2018, at 12:31 AM, Ken Krugler <kkrugler_li...@transpac.com> > wrote: > > Hi Ashish, > > Are you using Flink 1.4? If so, what does the “hadoop classpath” command > return from the command line where you’re trying to start the job? > > Asking because I’d run into issues with https://issues.apache. > org/jira/browse/FLINK-7477 > <https://issues.apache.org/jira/browse/FLINK-7477>, where I had a old > version of Hadoop being referenced by the “hadoop" command. > > — Ken > > > On Mar 22, 2018, at 7:05 PM, Ashish Pokharel <ashish...@yahoo.com> wrote: > > Hi All, > > Looks like we are out of the woods for now (so we think) - we went with > Hadoop free version and relied on client libraries on edge node. > > However, I am still not very confident as I started digging into that > stack as well and realized what Till pointed out (traces leads to a class > that is part of 2.9). I did dig around env variables and nothing was set. > This is a brand new clustered installed a week back and our team is > literally the first hands on deck. I will fish around and see if > Hortonworks back-ported something for HDP (dots are still not completely > connected but nonetheless, we have a test session and app running in our > brand new Prod) > > Thanks, Ashish > > On Mar 22, 2018, at 4:47 AM, Till Rohrmann <trohrm...@apache.org> wrote: > > Hi Ashish, > > the class ` RequestHedgingRMFailoverProxyP rovider` was only introduced > with Hadoop 2.9.0. My suspicion is thus that you start the client with some > Hadoop 2.9.0 dependencies on the class path. Could you please check the > logs of the client what's on its class path? Maybe you could also share the > logs with us. Please also check whether HADOOP_CLASSPATH is set to > something suspicious. > > Thanks a lot! > > Cheers, > Till > > On Wed, Mar 21, 2018 at 6:25 PM, ashish pok <ashish...@yahoo.com> wrote: > > Hi Piotrek, > > At this point we are simply trying to start a YARN session. > > BTW, we are on Hortonworks HDP 2.6 which is on 2.7 Hadoop if anyone has > experienced similar issues. > > We actually pulled 2.6 binaries for the heck of it and ran into same > issues. > > I guess we are left with getting non-hadoop binaries and set > HADOOP_CLASSPATH then? > > -- Ashish > > On Wed, Mar 21, 2018 at 12:03 PM, Piotr Nowojski > <pi...@data-artisans.com> wrote: > Hi, > > > Does some simple word count example works on the cluster after the > upgrade? > > If not, maybe your job is pulling some dependency that’s causing this > version conflict? > > Piotrek > > On 21 Mar 2018, at 16:52, ashish pok <ashish...@yahoo.com> wrote: > > Hi Piotrek, > > Yes, this is a brand new Prod environment. 2.6 was in our lab. > > Thanks, > > -- Ashish > > On Wed, Mar 21, 2018 at 11:39 AM, Piotr Nowojski > <pi...@data-artisans.com> wrote: > Hi, > > Have you replaced all of your old Flink binaries with freshly downloaded > <https://flink.apache.org/downloads.html> Hadoop 2.7 versions? Are you > sure that something hasn't mix in the process? > > Does some simple word count example works on the cluster after the upgrade? > > Piotrek > > On 21 Mar 2018, at 16:11, ashish pok <ashish...@yahoo.com> wrote: > > Hi All, > > We ran into a roadblock in our new Hadoop environment, migrating from 2.6 > to 2.7. It was supposed to be an easy lift to get a YARN session but doesnt > seem like :) We definitely are using 2.7 binaries but it looks like there > is a call here to a private methos which screams runtime incompatibility. > > Anyone has seen this and have pointers? > > Thanks, Ashish > > Exception in thread "main" java.lang.IllegalAccessError: tried to access > method org.apache.hadoop.yarn.client. ConfiguredRMFailoverProxyProvi > der.getProxyInternal()Ljava/la ng/Object; from class > org.apache.hadoop.yarn.client. RequestHedgingRMFailoverProxyP rovider > at org.apache.hadoop.yarn.client. > RequestHedgingRMFailoverProxyP rovider.init(RequestHedgingRMF > ailoverProxyProvider.java:75) > at org.apache.hadoop.yarn.client. > RMProxy.createRMFailoverProxyP rovider(RMProxy.java:163) > at org.apache.hadoop.yarn.client. > RMProxy.createRMProxy(RMProxy. java:94) > at org.apache.hadoop.yarn.client. > ClientRMProxy.createRMProxy(Cl ientRMProxy.java:72) > at org.apache.hadoop.yarn.client. > api.impl.YarnClientImpl.servic eStart(YarnClientImpl.java: 187) > at org.apache.hadoop.service.Abst > ractService.start(AbstractServ ice.java:193) > at org.apache.flink.yarn.Abstract > YarnClusterDescriptor.getYarnC lient(AbstractYarnClusterDescr > iptor.java:314) > at org.apache.flink.yarn.Abstract > YarnClusterDescriptor.deployIn ternal(AbstractYarnClusterDesc > riptor.java:417) > at org.apache.flink.yarn.Abstract > YarnClusterDescriptor.deploySe ssionCluster(AbstractYarnClust > erDescriptor.java:367) > at org.apache.flink.yarn.cli.Flin > kYarnSessionCli.run(FlinkYarnS essionCli.java:679) > at org.apache.flink.yarn.cli.Flin > kYarnSessionCli$1.call(FlinkYa rnSessionCli.java:514) > at org.apache.flink.yarn.cli.Flin > kYarnSessionCli$1.call(FlinkYa rnSessionCli.java:511) > at java.security.AccessController .doPrivileged(Native Method) > at javax.security.auth.Subject.do As(Subject.java:422) > at org.apache.hadoop.security.Use > rGroupInformation.doAs(UserGro upInformation.java:1698) > at org.apache.flink.runtime.secur > ity.HadoopSecurityContext.runS ecured(HadoopSecurityContext. java:41) > at org.apache.flink.yarn.cli.Flin > kYarnSessionCli.main(FlinkYarn SessionCli.java:511) > > > > > > > ------------------------------ -------------- > http://about.me/kkrugler > +1 530-210-6378 > > > >