[
https://issues.apache.org/jira/browse/FLINK-25253?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Chesnay Schepler updated FLINK-25253:
-------------------------------------
Priority: Minor (was: Blocker)
> An ClassNotFoundException of missing Hadoop class occurred when submitting to
> yarn
> ----------------------------------------------------------------------------------
>
> Key: FLINK-25253
> URL: https://issues.apache.org/jira/browse/FLINK-25253
> Project: Flink
> Issue Type: Bug
> Components: Connectors / Hive, Deployment / YARN
> Affects Versions: 1.14.0
> Environment: Environment version:
> Hadoop 3.1.1
> Hive 3.1.1
> Flink 1.4.0
> kafka 2.6.1
>
> Reporter: ghost synth
> Priority: Minor
> Attachments: FlinkPlaySubmit.scala, flink lib.png,
> original-TropicalaLink-1.0-SNAPSHOT.jar, pom.xml, submit_log.log
>
> Original Estimate: 96h
> Remaining Estimate: 96h
>
> I use the Hive Table connector to write hive from Kafka and submit to yarn
> successfully, but it will always report during execution
> *Caused by: java.lang.ClassNotFoundException:
> org.apache.hadoop.mapred.JobConf*
> Before submitting, I executed "export HADOOP_CLASSPATH=`hadoop classpath`" to
> import hadoop dependencies.
> I found in the JM log that the classpath already contains Hadoop
> dependencies, but an exception still occurs
> The original jar that I submitted only contains code and does not contain
> dependencies,The program loads dependencies from the hadoop path and lib
> under the flink directory
> The attachment contains flink lib, code, jar and JM log
> Thanks
--
This message was sent by Atlassian Jira
(v8.20.1#820001)