Any help on this?

On Mon, Jun 18, 2018 at 11:31 PM Garvit Sharma <garvit...@gmail.com> wrote:

> Yes, it is.
>
> On Mon, Jun 18, 2018 at 7:54 PM Till Rohrmann <trohrm...@apache.org>
> wrote:
>
>> Is `/usr/hdp/2.6.3.0-235/hadoop/client/xercesImpl.jar` a link to `
>> /usr/hdp/2.6.3.0-235/hadoop/client/xercesImpl-2.9.1.jar`?
>>
>> On Mon, Jun 18, 2018 at 4:02 PM Garvit Sharma <garvit...@gmail.com>
>> wrote:
>>
>>> I don't think I can access core-default as it comes with Hadoop jar
>>>
>>> On Mon, 18 Jun 2018 at 7:30 PM, Till Rohrmann <trohrm...@apache.org>
>>> wrote:
>>>
>>>> Hmm, could you check whether core-default.xml contains any suspicious
>>>> entries? Apparently xerces:2.9.1 cannot read it.
>>>>
>>>> On Mon, Jun 18, 2018 at 3:40 PM Garvit Sharma <garvit...@gmail.com>
>>>> wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> After putting the following log in my code, I can see that the Xerces
>>>>> version is - Xerces version : Xerces-J 2.9.1
>>>>>
>>>>> log.info("Xerces version : {}", 
>>>>> org.apache.xerces.impl.Version.getVersion());
>>>>>
>>>>> Also, following is the response of *$* *locate xerces* command on the
>>>>> server -
>>>>>
>>>>>
>>>>> /usr/hdp/2.6.1.0-129/falcon/client/lib/xercesImpl-2.10.0.jar
>>>>>
>>>>> /usr/hdp/2.6.1.0-129/hadoop/client/xercesImpl-2.9.1.jar
>>>>>
>>>>> /usr/hdp/2.6.1.0-129/hadoop/client/xercesImpl.jar
>>>>>
>>>>> /usr/hdp/2.6.1.0-129/hadoop-hdfs/lib/xercesImpl-2.9.1.jar
>>>>>
>>>>> /usr/hdp/2.6.1.0-129/hbase/lib/xercesImpl-2.9.1.jar
>>>>>
>>>>>
>>>>> /usr/hdp/2.6.1.0-129/hive-hcatalog/share/webhcat/svr/lib/xercesImpl-2.9.1.jar
>>>>>
>>>>> /usr/hdp/2.6.1.0-129/livy/jars/xercesImpl-2.9.1.jar
>>>>>
>>>>> /usr/hdp/2.6.1.0-129/livy2/jars/xercesImpl-2.9.1.jar
>>>>>
>>>>> /usr/hdp/2.6.1.0-129/oozie/lib/xercesImpl-2.10.0.jar
>>>>>
>>>>> /usr/hdp/2.6.1.0-129/oozie/libserver/xercesImpl-2.10.0.jar
>>>>>
>>>>> /usr/hdp/2.6.1.0-129/oozie/libtools/xercesImpl-2.10.0.jar
>>>>>
>>>>> /usr/hdp/2.6.1.0-129/slider/lib/xercesImpl-2.9.1.jar
>>>>>
>>>>> /usr/hdp/2.6.1.0-129/spark2/jars/xercesImpl-2.9.1.jar
>>>>>
>>>>> /usr/hdp/2.6.1.0-129/storm/contrib/storm-autocreds/xercesImpl-2.9.1.jar
>>>>>
>>>>> /usr/hdp/2.6.1.0-129/zookeeper/lib/xercesMinimal-1.9.6.2.jar
>>>>>
>>>>> /usr/hdp/2.6.3.0-235/falcon/client/lib/xercesImpl-2.10.0.jar
>>>>>
>>>>> /usr/hdp/2.6.3.0-235/hadoop/client/xercesImpl-2.9.1.jar
>>>>>
>>>>> /usr/hdp/2.6.3.0-235/hadoop/client/xercesImpl.jar
>>>>>
>>>>> /usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/xercesImpl-2.9.1.jar
>>>>>
>>>>> /usr/hdp/2.6.3.0-235/hbase/lib/xercesImpl-2.9.1.jar
>>>>>
>>>>>
>>>>> /usr/hdp/2.6.3.0-235/hive-hcatalog/share/webhcat/svr/lib/xercesImpl-2.9.1.jar
>>>>>
>>>>> /usr/hdp/2.6.3.0-235/livy/jars/xercesImpl-2.9.1.jar
>>>>>
>>>>> /usr/hdp/2.6.3.0-235/livy2/jars/xercesImpl-2.9.1.jar
>>>>>
>>>>> /usr/hdp/2.6.3.0-235/oozie/lib/xercesImpl-2.10.0.jar
>>>>>
>>>>> /usr/hdp/2.6.3.0-235/oozie/libserver/xercesImpl-2.10.0.jar
>>>>>
>>>>> /usr/hdp/2.6.3.0-235/oozie/libtools/xercesImpl-2.10.0.jar
>>>>>
>>>>>
>>>>> /usr/hdp/2.6.3.0-235/ranger-admin/ews/webapp/WEB-INF/lib/xercesImpl-2.9.1.jar
>>>>>
>>>>> /usr/hdp/2.6.3.0-235/slider/lib/xercesImpl-2.9.1.jar
>>>>>
>>>>> /usr/hdp/2.6.3.0-235/spark2/jars/xercesImpl-2.9.1.jar
>>>>>
>>>>> /usr/hdp/2.6.3.0-235/storm/contrib/storm-autocreds/xercesImpl-2.9.1.jar
>>>>>
>>>>> /usr/hdp/2.6.3.0-235/zookeeper/lib/xercesMinimal-1.9.6.2.jar
>>>>>
>>>>> /usr/hdp/share/hst/hst-common/lib/xercesImpl-2.9.1.jar
>>>>>
>>>>> Now, I can say that the version of xerces are same.
>>>>>
>>>>>
>>>>> So, what is causing this issue if Xerces version is in sync?
>>>>>
>>>>>
>>>>> I am very excited to discover the issue :)
>>>>>
>>>>>
>>>>> Thanks,
>>>>>
>>>>> On Mon, Jun 18, 2018 at 6:27 PM Till Rohrmann <trohrm...@apache.org>
>>>>> wrote:
>>>>>
>>>>>> Could you check which xerces version you have on your classpath?
>>>>>> Apparently, it cannot read core-default.xml as Ted pointed out. This 
>>>>>> might
>>>>>> be the root cause for the failure.
>>>>>>
>>>>>> Cheers,
>>>>>> Till
>>>>>>
>>>>>> On Mon, Jun 18, 2018 at 1:31 PM Garvit Sharma <garvit...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Hi,
>>>>>>>
>>>>>>> Sorry for the confusion, but the yarn is running on Hadoop version
>>>>>>> 2.7 only and hence I am using Flink 1.5 Hadoop 2.7 binary.
>>>>>>>
>>>>>>> Below are the details provided by Yarn version command :
>>>>>>>
>>>>>>> Hadoop 2.7.3.2.6.3.0-235
>>>>>>> Subversion g...@github.com:hortonworks/hadoop.git -r
>>>>>>> 45bfd33bba8acadfa0e6024c80981c023b28d454
>>>>>>> Compiled by jenkins on 2017-10-30T02:31Z
>>>>>>> Compiled with protoc 2.5.0
>>>>>>> From source with checksum cd1a4a466ef450f547c279989f3aa3
>>>>>>> This command was run using
>>>>>>> /usr/hdp/2.6.3.0-235/hadoop/hadoop-common-2.7.3.2.6.3.0-235.jar
>>>>>>>
>>>>>>> Please let me know if you have found the resolution to my issue :)
>>>>>>>
>>>>>>> Thanks,
>>>>>>>
>>>>>>>
>>>>>>> On Mon, Jun 18, 2018 at 4:50 PM Till Rohrmann <trohrm...@apache.org>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Which Hadoop version have you installed? It looks as if Flink has
>>>>>>>> been build with Hadoop 2.7 but I see /usr/hdp/2.6.3.0-235 in the class
>>>>>>>> path. If you want to run Flink on Hadoop 2.6, then try to use the 
>>>>>>>> Hadoop
>>>>>>>> free Flink binaries or the one built for Hadoop 2.6.
>>>>>>>>
>>>>>>>> Cheers,
>>>>>>>> Till
>>>>>>>>
>>>>>>>> On Mon, Jun 18, 2018 at 10:48 AM Garvit Sharma <garvit...@gmail.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Ok, I have attached the log file.
>>>>>>>>>
>>>>>>>>> Please check and let me know.
>>>>>>>>>
>>>>>>>>> Thanks,
>>>>>>>>>
>>>>>>>>> On Mon, Jun 18, 2018 at 2:07 PM Amit Jain <aj201...@gmail.com>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> Hi Gravit,
>>>>>>>>>>
>>>>>>>>>> I think Till is interested to know about classpath details
>>>>>>>>>> present at the start of JM and TM logs e.g. following logs provide
>>>>>>>>>> classpath details used by TM in our case.
>>>>>>>>>>
>>>>>>>>>> 2018-06-17 19:01:30,656 INFO
>>>>>>>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -
>>>>>>>>>> --------------------------------------------------------------------------------
>>>>>>>>>> 2018-06-17 19:01:30,658 INFO
>>>>>>>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -  
>>>>>>>>>> Starting
>>>>>>>>>> YARN TaskExecutor runner (Version: 1.5.0, Rev:c61b108, 
>>>>>>>>>> Date:24.05.2018 @
>>>>>>>>>> 14:54:44 UTC)
>>>>>>>>>> 2018-06-17 19:01:30,659 INFO
>>>>>>>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -  OS
>>>>>>>>>> current user: yarn
>>>>>>>>>> 2018-06-17 19:01:31,662 INFO
>>>>>>>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -  
>>>>>>>>>> Current
>>>>>>>>>> Hadoop/Kerberos user: hadoop
>>>>>>>>>> 2018-06-17 19:01:31,663 INFO
>>>>>>>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -  
>>>>>>>>>> JVM:
>>>>>>>>>> OpenJDK 64-Bit Server VM - Oracle Corporation - 1.8/25.171-b10
>>>>>>>>>> 2018-06-17 19:01:31,663 INFO
>>>>>>>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -  
>>>>>>>>>> Maximum
>>>>>>>>>> heap size: 6647 MiBytes
>>>>>>>>>> 2018-06-17 19:01:31,663 INFO
>>>>>>>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -
>>>>>>>>>>  JAVA_HOME: /usr/lib/jvm/java-openjdk
>>>>>>>>>> 2018-06-17 19:01:31,664 INFO
>>>>>>>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -  
>>>>>>>>>> Hadoop
>>>>>>>>>> version: 2.8.3
>>>>>>>>>> 2018-06-17 19:01:31,664 INFO
>>>>>>>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -  JVM
>>>>>>>>>> Options:
>>>>>>>>>> 2018-06-17 19:01:31,665 INFO
>>>>>>>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -
>>>>>>>>>> -Xms6936m
>>>>>>>>>> 2018-06-17 19:01:31,665 INFO
>>>>>>>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -
>>>>>>>>>> -Xmx6936m
>>>>>>>>>> 2018-06-17 19:01:31,665 INFO
>>>>>>>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -
>>>>>>>>>> -XX:MaxDirectMemorySize=4072m
>>>>>>>>>> 2018-06-17 19:01:31,665 INFO
>>>>>>>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -
>>>>>>>>>> -Dlog.file=/var/log/hadoop-yarn/containers/application_1528342246614_0002/container_1528342246614_0002_01_282649/taskmanager.log
>>>>>>>>>> 2018-06-17 19:01:31,665 INFO
>>>>>>>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -
>>>>>>>>>> -Dlogback.configurationFile=file:./logback.xml
>>>>>>>>>> 2018-06-17 19:01:31,665 INFO
>>>>>>>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -
>>>>>>>>>> -Dlog4j.configuration=file:./log4j.properties
>>>>>>>>>> 2018-06-17 19:01:31,665 INFO
>>>>>>>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -  
>>>>>>>>>> Program
>>>>>>>>>> Arguments:
>>>>>>>>>> 2018-06-17 19:01:31,665 INFO
>>>>>>>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -
>>>>>>>>>> --configDir
>>>>>>>>>> 2018-06-17 19:01:31,665 INFO
>>>>>>>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -     
>>>>>>>>>> .
>>>>>>>>>> *2018-06-17 19:01:31,666 INFO
>>>>>>>>>>  org.apache.flink.yarn.YarnTaskExecutorRunner                  -
>>>>>>>>>>  Classpath:
>>>>>>>>>> lib/flink-dist_2.11-1.5.0.jar:lib/flink-python_2.11-1.5.0.jar:lib/flink-shaded-hadoop2-uber-1.5.0.jar:lib/flink-shaded-include-yarn-0.9.1.jar:lib/guava-18.0.jar:lib/log4j-1.2.17.jar:lib/slf4j-log4j12-1.7.7.jar:log4j.properties:logback.xml:flink.jar:flink-conf.yaml::/etc/hadoop/conf:/usr/lib/hadoop/hadoop-common-2.8.3-amzn-0.jar:/usr/lib/hadoop/hadoop-archive-logs.jar:/usr/lib/hadoop/hadoop-auth.jar:/usr/lib/hadoop/hadoop-archives-2.8.3-amzn-0.jar:/usr/lib/hadoop/hadoop-archive-logs-2.8.3-amzn-0.jar:/usr/lib/hadoop/hadoop-azure-datalake-2.8.3-amzn-0.jar.........*
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> Thanks,
>>>>>>>>>> Amit
>>>>>>>>>>
>>>>>>>>>> On Mon, Jun 18, 2018 at 2:00 PM, Garvit Sharma <
>>>>>>>>>> garvit...@gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi,
>>>>>>>>>>>
>>>>>>>>>>> Please refer to my previous mail for complete logs.
>>>>>>>>>>>
>>>>>>>>>>> Thanks,
>>>>>>>>>>>
>>>>>>>>>>> On Mon, Jun 18, 2018 at 1:17 PM Till Rohrmann <
>>>>>>>>>>> trohrm...@apache.org> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Could you also please share the complete log file with us.
>>>>>>>>>>>>
>>>>>>>>>>>> Cheers,
>>>>>>>>>>>> Till
>>>>>>>>>>>>
>>>>>>>>>>>> On Sat, Jun 16, 2018 at 5:22 PM Ted Yu <yuzhih...@gmail.com>
>>>>>>>>>>>> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> The error for core-default.xml is interesting.
>>>>>>>>>>>>>
>>>>>>>>>>>>> Flink doesn't have this file. Probably it came with Yarn.
>>>>>>>>>>>>> Please check the hadoop version Flink was built with versus the 
>>>>>>>>>>>>> hadoop
>>>>>>>>>>>>> version in your cluster.
>>>>>>>>>>>>>
>>>>>>>>>>>>> Thanks
>>>>>>>>>>>>>
>>>>>>>>>>>>> -------- Original message --------
>>>>>>>>>>>>> From: Garvit Sharma <garvit...@gmail.com>
>>>>>>>>>>>>> Date: 6/16/18 7:23 AM (GMT-08:00)
>>>>>>>>>>>>> To: trohrm...@apache.org
>>>>>>>>>>>>> Cc: Chesnay Schepler <ches...@apache.org>,
>>>>>>>>>>>>> user@flink.apache.org
>>>>>>>>>>>>> Subject: Re: Exception while submitting jobs through Yarn
>>>>>>>>>>>>>
>>>>>>>>>>>>> I am not able to figure out, got stuck badly in this since
>>>>>>>>>>>>> last 1 week. Any little help would be appreciated.
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> 2018-06-16 19:25:10,523 DEBUG
>>>>>>>>>>>>> org.apache.flink.streaming.api.graph.StreamingJobGraphGenerator
>>>>>>>>>>>>> - Parallelism set: 1 for 8
>>>>>>>>>>>>>
>>>>>>>>>>>>> 2018-06-16 19:25:10,578 DEBUG
>>>>>>>>>>>>> org.apache.flink.streaming.api.graph.StreamingJobGraphGenerator
>>>>>>>>>>>>> - Parallelism set: 1 for 1
>>>>>>>>>>>>>
>>>>>>>>>>>>> 2018-06-16 19:25:10,588 DEBUG
>>>>>>>>>>>>> org.apache.flink.streaming.api.graph.StreamingJobGraphGenerator
>>>>>>>>>>>>> - CONNECTED: KeyGroupStreamPartitioner - 1 -> 8
>>>>>>>>>>>>>
>>>>>>>>>>>>> 2018-06-16 19:25:10,591 DEBUG
>>>>>>>>>>>>> org.apache.flink.streaming.api.graph.StreamingJobGraphGenerator
>>>>>>>>>>>>> - Parallelism set: 1 for 5
>>>>>>>>>>>>>
>>>>>>>>>>>>> 2018-06-16 19:25:10,597 DEBUG
>>>>>>>>>>>>> org.apache.flink.streaming.api.graph.StreamingJobGraphGenerator
>>>>>>>>>>>>> - CONNECTED: KeyGroupStreamPartitioner - 5 -> 8
>>>>>>>>>>>>>
>>>>>>>>>>>>> 2018-06-16 19:25:10,618 FATAL
>>>>>>>>>>>>> org.apache.hadoop.conf.Configuration                          -
>>>>>>>>>>>>> error parsing conf core-default.xml
>>>>>>>>>>>>>
>>>>>>>>>>>>> javax.xml.parsers.ParserConfigurationException: Feature '
>>>>>>>>>>>>> http://apache.org/xml/features/xinclude' is not recognized.
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.xerces.jaxp.DocumentBuilderFactoryImpl.newDocumentBuilder(Unknown
>>>>>>>>>>>>> Source)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2482)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2444)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2361)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.hadoop.conf.Configuration.get(Configuration.java:1188)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.hadoop.yarn.factory.providers.RecordFactoryProvider.getRecordFactory(RecordFactoryProvider.java:49)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.hadoop.yarn.util.Records.<clinit>(Records.java:32)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getQueueInfoRequest(YarnClientImpl.java:495)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getAllQueues(YarnClientImpl.java:525)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.flink.yarn.AbstractYarnClusterDescriptor.checkYarnQueues(AbstractYarnClusterDescriptor.java:658)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.flink.yarn.AbstractYarnClusterDescriptor.deployInternal(AbstractYarnClusterDescriptor.java:486)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.flink.yarn.YarnClusterDescriptor.deployJobCluster(YarnClusterDescriptor.java:75)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.flink.client.cli.CliFrontend.runProgram(CliFrontend.java:235)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:210)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:1020)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.flink.client.cli.CliFrontend.lambda$main$9(CliFrontend.java:1096)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at java.security.AccessController.doPrivileged(Native Method)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at javax.security.auth.Subject.doAs(Subject.java:422)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1692)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.flink.runtime.security.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:41)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1096)
>>>>>>>>>>>>>
>>>>>>>>>>>>> 2018-06-16 19:25:10,620 WARN  
>>>>>>>>>>>>> org.apache.flink.yarn.AbstractYarnClusterDescriptor
>>>>>>>>>>>>>           - Error while getting queue information from YARN:
>>>>>>>>>>>>> null
>>>>>>>>>>>>>
>>>>>>>>>>>>> 2018-06-16 19:25:10,621 DEBUG
>>>>>>>>>>>>> org.apache.flink.yarn.AbstractYarnClusterDescriptor           -
>>>>>>>>>>>>> Error details
>>>>>>>>>>>>>
>>>>>>>>>>>>> java.lang.ExceptionInInitializerError
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getQueueInfoRequest(YarnClientImpl.java:495)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getAllQueues(YarnClientImpl.java:525)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.flink.yarn.AbstractYarnClusterDescriptor.checkYarnQueues(AbstractYarnClusterDescriptor.java:658)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.flink.yarn.AbstractYarnClusterDescriptor.deployInternal(AbstractYarnClusterDescriptor.java:486)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.flink.yarn.YarnClusterDescriptor.deployJobCluster(YarnClusterDescriptor.java:75)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.flink.client.cli.CliFrontend.runProgram(CliFrontend.java:235)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:210)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:1020)
>>>>>>>>>>>>>
>>>>>>>>>>>>> at
>>>>>>>>>>>>> org.apache.flink.client.cli.CliFrontend.lambda$main$9(CliFrontend.java:1096)
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>>
>>>>>>>>>>> Garvit Sharma
>>>>>>>>>>> github.com/garvitlnmiit/
>>>>>>>>>>>
>>>>>>>>>>> No Body is a Scholar by birth, its only hard work and strong
>>>>>>>>>>> determination that makes him master.
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>>
>>>>>>>>> Garvit Sharma
>>>>>>>>> github.com/garvitlnmiit/
>>>>>>>>>
>>>>>>>>> No Body is a Scholar by birth, its only hard work and strong
>>>>>>>>> determination that makes him master.
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>>
>>>>>>> Garvit Sharma
>>>>>>> github.com/garvitlnmiit/
>>>>>>>
>>>>>>> No Body is a Scholar by birth, its only hard work and strong
>>>>>>> determination that makes him master.
>>>>>>>
>>>>>>
>>>>>
>>>>> --
>>>>>
>>>>> Garvit Sharma
>>>>> github.com/garvitlnmiit/
>>>>>
>>>>> No Body is a Scholar by birth, its only hard work and strong
>>>>> determination that makes him master.
>>>>>
>>>> --
>>>
>>> Garvit Sharma
>>> github.com/garvitlnmiit/
>>>
>>> No Body is a Scholar by birth, its only hard work and strong
>>> determination that makes him master.
>>>
>>
>
> --
>
> Garvit Sharma
> github.com/garvitlnmiit/
>
> No Body is a Scholar by birth, its only hard work and strong determination
> that makes him master.
>


-- 

Garvit Sharma
github.com/garvitlnmiit/

No Body is a Scholar by birth, its only hard work and strong determination
that makes him master.

Reply via email to