I'm unable to figure out which dependency to add in order for the ByteArrayDeserializer class to get included in the jar. I have added all the dependencies as per the documentation still unable to figure out the cause.
On Fri, Sep 10, 2021 at 12:17 AM Robert Metzger <rmetz...@apache.org> wrote: > Does the jar file you are trying to submit contain > the org/apache/kafka/common/serialization/ByteArrayDeserializer class? > > On Thu, Sep 9, 2021 at 2:10 PM Harshvardhan Shinde < > harshvardhan.shi...@oyorooms.com> wrote: > >> Here's the complete stack trace: >> >> Server Response:org.apache.flink.runtime.rest.handler.RestHandlerException: >> Could not execute application. at >> org.apache.flink.runtime.webmonitor.handlers.JarRunHandler.lambda$handleRequest$1(JarRunHandler.java:108) >> at >> java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:836) >> at >> java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:811) >> at >> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488) >> at >> java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1609) >> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) >> at java.util.concurrent.FutureTask.run(FutureTask.java:266) at >> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) >> at >> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) >> at >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) >> at >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) >> at java.lang.Thread.run(Thread.java:748) Caused by: >> java.util.concurrent.CompletionException: java.lang.NoClassDefFoundError: >> org/apache/kafka/common/serialization/ByteArrayDeserializer at >> java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:273) >> at >> java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:280) >> at >> java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1606) >> ... 7 more Caused by: java.lang.NoClassDefFoundError: >> org/apache/kafka/common/serialization/ByteArrayDeserializer at >> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer.setDeserializer(FlinkKafkaConsumer.java:322) >> at >> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer.<init>(FlinkKafkaConsumer.java:223) >> at >> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer.<init>(FlinkKafkaConsumer.java:154) >> at >> org.apache.flink.streaming.connectors.kafka.table.KafkaDynamicSource.createKafkaConsumer(KafkaDynamicSource.java:383) >> at >> org.apache.flink.streaming.connectors.kafka.table.KafkaDynamicSource.getScanRuntimeProvider(KafkaDynamicSource.java:205) >> at >> org.apache.flink.table.planner.connectors.DynamicSourceUtils.validateScanSource(DynamicSourceUtils.java:453) >> at >> org.apache.flink.table.planner.connectors.DynamicSourceUtils.prepareDynamicSource(DynamicSourceUtils.java:161) >> at >> org.apache.flink.table.planner.connectors.DynamicSourceUtils.convertSourceToRel(DynamicSourceUtils.java:119) >> at >> org.apache.flink.table.planner.plan.schema.CatalogSourceTable.toRel(CatalogSourceTable.java:85) >> at >> org.apache.calcite.sql2rel.SqlToRelConverter.toRel(SqlToRelConverter.java:3585) >> at >> org.apache.calcite.sql2rel.SqlToRelConverter.convertIdentifier(SqlToRelConverter.java:2507) >> at >> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2144) >> at >> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2093) >> at >> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2050) >> at >> org.apache.calcite.sql2rel.SqlToRelConverter.convertSelectImpl(SqlToRelConverter.java:663) >> at >> org.apache.calcite.sql2rel.SqlToRelConverter.convertSelect(SqlToRelConverter.java:644) >> at >> org.apache.calcite.sql2rel.SqlToRelConverter.convertQueryRecursive(SqlToRelConverter.java:3438) >> at >> org.apache.calcite.sql2rel.SqlToRelConverter.convertQuery(SqlToRelConverter.java:570) >> at >> org.apache.flink.table.planner.calcite.FlinkPlannerImpl.org$apache$flink$table$planner$calcite$FlinkPlannerImpl$$rel(FlinkPlannerImpl.scala:169) >> at >> org.apache.flink.table.planner.calcite.FlinkPlannerImpl.rel(FlinkPlannerImpl.scala:161) >> at >> org.apache.flink.table.planner.operations.SqlToOperationConverter.toQueryOperation(SqlToOperationConverter.java:989) >> at >> org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlQuery(SqlToOperationConverter.java:958) >> at >> org.apache.flink.table.planner.operations.SqlToOperationConverter.convert(SqlToOperationConverter.java:283) >> at >> org.apache.flink.table.planner.delegation.ParserImpl.parse(ParserImpl.java:101) >> at >> org.apache.flink.table.api.internal.TableEnvironmentImpl.sqlQuery(TableEnvironmentImpl.java:704) >> at com.harsh.test.StreamingJob.main(StreamingJob.java:106) at >> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:498) at >> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:355) >> at >> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:222) >> at org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:114) >> at >> org.apache.flink.client.deployment.application.DetachedApplicationRunner.tryExecuteJobs(DetachedApplicationRunner.java:84) >> at >> org.apache.flink.client.deployment.application.DetachedApplicationRunner.run(DetachedApplicationRunner.java:70) >> at >> org.apache.flink.runtime.webmonitor.handlers.JarRunHandler.lambda$handleRequest$0(JarRunHandler.java:102) >> at >> java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1604) >> ... 7 more >> >> On Thu, Sep 9, 2021 at 5:36 PM Robert Metzger <rmetz...@apache.org> >> wrote: >> >>> Can you share the full stack trace, not just a part of it? >>> >>> On Thu, Sep 9, 2021 at 1:43 PM Harshvardhan Shinde < >>> harshvardhan.shi...@oyorooms.com> wrote: >>> >>>> Hi, >>>> >>>> I added the dependencies while trying to resolve the same >>>> issue, thought I was missing them. >>>> >>>> Thanks >>>> >>>> On Thu, Sep 9, 2021 at 4:26 PM Robert Metzger <rmetz...@apache.org> >>>> wrote: >>>> >>>>> Hey, >>>>> >>>>> Why do you have these dependencies in your pom? >>>>> >>>>> <!-- >>>>> https://mvnrepository.com/artifact/org.apache.kafka/kafka-clients --> >>>>> <dependency> >>>>> <groupId>org.apache.kafka</groupId> >>>>> <artifactId>kafka-clients</artifactId> >>>>> <version>2.8.0</version> >>>>> </dependency> >>>>> >>>>> <dependency> >>>>> <groupId>org.apache.kafka</groupId> >>>>> <artifactId>kafka_2.12</artifactId> >>>>> <version>2.8.0</version> >>>>> </dependency> >>>>> >>>>> >>>>> They are not needed for using the Kafka connector of Flink (the flink >>>>> kafka connector dependencies pulls the required dependencies) >>>>> >>>>> >>>>> On Thu, Sep 9, 2021 at 12:02 PM Harshvardhan Shinde < >>>>> harshvardhan.shi...@oyorooms.com> wrote: >>>>> >>>>>> Hi, >>>>>> >>>>>> I'm trying a simple flink job that reads data from a kafka topic and >>>>>> creates a Hive table. >>>>>> >>>>>> I'm following the steps from here >>>>>> <https://ci.apache.org/projects/flink/flink-docs-release-1.13/docs/connectors/table/hive/overview/#connecting-to-hive> >>>>>> . >>>>>> >>>>>> Here's my code: >>>>>> >>>>>> import org.apache.flink.table.api.EnvironmentSettings; >>>>>> import org.apache.flink.table.api.Table; >>>>>> import org.apache.flink.table.api.TableEnvironment; >>>>>> import org.apache.flink.table.catalog.hive.HiveCatalog; >>>>>> >>>>>> EnvironmentSettings settings = >>>>>> EnvironmentSettings.newInstance().useBlinkPlanner().inStreamingMode().build(); >>>>>> TableEnvironment tableEnv = TableEnvironment.create(settings); >>>>>> >>>>>> String name = "myhive"; >>>>>> String defaultDatabase = "harsh_test"; >>>>>> String hiveConfDir = "/etc/hive/conf"; >>>>>> >>>>>> HiveCatalog hive = new HiveCatalog(name, defaultDatabase, hiveConfDir); >>>>>> tableEnv.registerCatalog(name, hive); >>>>>> >>>>>> // set the HiveCatalog as the current catalog of the session >>>>>> tableEnv.useCatalog(name); >>>>>> >>>>>> tableEnv.executeSql("CREATE TABLE IF NOT EXISTS transactions (\n" + >>>>>> " `created_at` TIMESTAMP(3) METADATA FROM 'timestamp',\n" + >>>>>> " `partition` BIGINT METADATA VIRTUAL,\n" + >>>>>> " `offset` BIGINT METADATA VIRTUAL,\n" + >>>>>> " account_id BIGINT,\n" + >>>>>> " amount BIGINT,\n" + >>>>>> " transaction_time TIMESTAMP(3),\n" + >>>>>> " WATERMARK FOR transaction_time AS transaction_time - INTERVAL >>>>>> '5' SECOND\n" + >>>>>> ") WITH (\n" + >>>>>> " 'connector' = 'kafka',\n" + >>>>>> " 'topic' = 'flink-stream-table',\n" + >>>>>> " 'properties.bootstrap.servers' = '<BROKER_ADDRESS>:9092',\n" + >>>>>> " 'scan.startup.mode' = 'earliest-offset',\n" + >>>>>> " 'format' = 'csv'\n" + >>>>>> ")"); >>>>>> >>>>>> Table table = tableEnv.sqlQuery("Select * from transactions"); >>>>>> table.execute().print(); >>>>>> >>>>>> The code builds successfully, but I'm getting the following runtime >>>>>> error: >>>>>> >>>>>> Caused by: java.util.concurrent.CompletionException: >>>>>> java.lang.NoClassDefFoundError: >>>>>> org/apache/kafka/common/serialization/ByteArrayDeserializer at >>>>>> java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:273) >>>>>> at >>>>>> java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:280) >>>>>> at >>>>>> java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1606) >>>>>> .. >>>>>> >>>>>> Here are my pom.xml file contents: >>>>>> >>>>>> <!-- >>>>>> Licensed to the Apache Software Foundation (ASF) under one >>>>>> or more contributor license agreements. See the NOTICE file >>>>>> distributed with this work for additional information >>>>>> regarding copyright ownership. The ASF licenses this file >>>>>> to you under the Apache License, Version 2.0 (the >>>>>> "License"); you may not use this file except in compliance >>>>>> with the License. You may obtain a copy of the License at >>>>>> >>>>>> http://www.apache.org/licenses/LICENSE-2.0 >>>>>> >>>>>> Unless required by applicable law or agreed to in writing, >>>>>> software distributed under the License is distributed on an >>>>>> "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY >>>>>> KIND, either express or implied. See the License for the >>>>>> specific language governing permissions and limitations >>>>>> under the License. >>>>>> --> >>>>>> <project xmlns="http://maven.apache.org/POM/4.0.0" >>>>>> xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" >>>>>> xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 >>>>>> http://maven.apache.org/xsd/maven-4.0.0.xsd"> >>>>>> <modelVersion>4.0.0</modelVersion> >>>>>> >>>>>> <groupId>com.harsh.test</groupId> >>>>>> <artifactId>harsh-flink-test</artifactId> >>>>>> <version>1.0-SNAPSHOT</version> >>>>>> <packaging>jar</packaging> >>>>>> >>>>>> <name>Flink Quickstart Job</name> >>>>>> <url>http://www.myorganization.org</url> >>>>>> >>>>>> <properties> >>>>>> >>>>>> <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding> >>>>>> <flink.version>1.13.2</flink.version> >>>>>> <java.version>1.8</java.version> >>>>>> <hive.version>2.3.6</hive.version> >>>>>> <scala.binary.version>2.12</scala.binary.version> >>>>>> <maven.compiler.source>${java.version}</maven.compiler.source> >>>>>> <maven.compiler.target>${java.version}</maven.compiler.target> >>>>>> </properties> >>>>>> >>>>>> <repositories> >>>>>> <repository> >>>>>> <id>apache.snapshots</id> >>>>>> <name>Apache Development Snapshot Repository</name> >>>>>> >>>>>> <url>https://repository.apache.org/content/repositories/snapshots/</url> >>>>>> <releases> >>>>>> <enabled>false</enabled> >>>>>> </releases> >>>>>> <snapshots> >>>>>> <enabled>true</enabled> >>>>>> </snapshots> >>>>>> </repository> >>>>>> </repositories> >>>>>> >>>>>> <dependencies> >>>>>> <!-- Apache Flink dependencies --> >>>>>> <!-- These dependencies are provided, because they should not be >>>>>> packaged into the JAR file. --> >>>>>> <dependency> >>>>>> <groupId>org.apache.flink</groupId> >>>>>> <artifactId>flink-java</artifactId> >>>>>> <version>${flink.version}</version> >>>>>> </dependency> >>>>>> <dependency> >>>>>> <groupId>org.apache.flink</groupId> >>>>>> >>>>>> <artifactId>flink-streaming-java_${scala.binary.version}</artifactId> >>>>>> <version>${flink.version}</version> >>>>>> </dependency> >>>>>> >>>>>> <!-- Add connector dependencies here. They must be in the >>>>>> default scope (compile). --> >>>>>> >>>>>> <!-- Example: >>>>>> >>>>>> <dependency> >>>>>> <groupId>org.apache.flink</groupId> >>>>>> >>>>>> <artifactId>flink-connector-kafka-0.10_${scala.binary.version}</artifactId> >>>>>> <version>${flink.version}</version> >>>>>> </dependency> >>>>>> --> >>>>>> >>>>>> <!-- >>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-connector-kafka >>>>>> --> >>>>>> <dependency> >>>>>> <groupId>org.apache.flink</groupId> >>>>>> >>>>>> <artifactId>flink-connector-kafka_${scala.binary.version}</artifactId> >>>>>> <version>${flink.version}</version> >>>>>> </dependency> >>>>>> >>>>>> <dependency> >>>>>> <groupId>org.apache.flink</groupId> >>>>>> >>>>>> <artifactId>flink-table-api-java-bridge_${scala.binary.version}</artifactId> >>>>>> <version>${flink.version}</version> >>>>>> </dependency> >>>>>> >>>>>> <dependency> >>>>>> <groupId>org.apache.flink</groupId> >>>>>> >>>>>> <artifactId>flink-table-planner-blink_${scala.binary.version}</artifactId> >>>>>> <version>${flink.version}</version> >>>>>> </dependency> >>>>>> >>>>>> <!-- >>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-table-planner >>>>>> --> >>>>>> <dependency> >>>>>> <groupId>org.apache.flink</groupId> >>>>>> <artifactId>flink-table-planner_2.12</artifactId> >>>>>> <version>1.13.2</version> >>>>>> </dependency> >>>>>> >>>>>> >>>>>> <!-- Flink Dependency --> >>>>>> <dependency> >>>>>> <groupId>org.apache.flink</groupId> >>>>>> >>>>>> <artifactId>flink-connector-hive_${scala.binary.version}</artifactId> >>>>>> <version>${flink.version}</version> >>>>>> </dependency> >>>>>> >>>>>> <!-- Hive Dependency --> >>>>>> <dependency> >>>>>> <groupId>org.apache.hive</groupId> >>>>>> <artifactId>hive-exec</artifactId> >>>>>> <version>${hive.version}</version> >>>>>> </dependency> >>>>>> >>>>>> <!-- >>>>>> https://mvnrepository.com/artifact/javax.servlet/javax.servlet-api --> >>>>>> <dependency> >>>>>> <groupId>javax.servlet</groupId> >>>>>> <artifactId>javax.servlet-api</artifactId> >>>>>> <version>3.1.0</version> >>>>>> <scope>provided</scope> >>>>>> </dependency> >>>>>> >>>>>> <!-- >>>>>> https://mvnrepository.com/artifact/org.apache.htrace/htrace-core4 --> >>>>>> <dependency> >>>>>> <groupId>org.apache.htrace</groupId> >>>>>> <artifactId>htrace-core4</artifactId> >>>>>> <version>4.0.1-incubating</version> >>>>>> </dependency> >>>>>> >>>>>> <!-- >>>>>> https://mvnrepository.com/artifact/commons-configuration/commons-configuration >>>>>> --> >>>>>> <dependency> >>>>>> <groupId>commons-configuration</groupId> >>>>>> <artifactId>commons-configuration</artifactId> >>>>>> <version>1.10</version> >>>>>> </dependency> >>>>>> >>>>>> <!-- >>>>>> https://mvnrepository.com/artifact/commons-logging/commons-logging --> >>>>>> <dependency> >>>>>> <groupId>commons-logging</groupId> >>>>>> <artifactId>commons-logging</artifactId> >>>>>> <version>1.2</version> >>>>>> </dependency> >>>>>> >>>>>> <!-- >>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-shaded-hadoop-2 >>>>>> --> >>>>>> <dependency> >>>>>> <groupId>org.apache.flink</groupId> >>>>>> <artifactId>flink-shaded-hadoop-2</artifactId> >>>>>> <version>2.8.3-10.0</version> >>>>>> </dependency> >>>>>> >>>>>> >>>>>> <!-- >>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-hadoop-compatibility >>>>>> --> >>>>>> <dependency> >>>>>> <groupId>org.apache.flink</groupId> >>>>>> <artifactId>flink-hadoop-compatibility_2.12</artifactId> >>>>>> <version>1.13.2</version> >>>>>> </dependency> >>>>>> >>>>>> <!-- >>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-hadoop-fs --> >>>>>> <dependency> >>>>>> <groupId>org.apache.flink</groupId> >>>>>> <artifactId>flink-hadoop-fs</artifactId> >>>>>> <version>1.13.2</version> >>>>>> </dependency> >>>>>> >>>>>> <!-- >>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-csv --> >>>>>> <dependency> >>>>>> <groupId>org.apache.flink</groupId> >>>>>> <artifactId>flink-csv</artifactId> >>>>>> <version>1.13.2</version> >>>>>> </dependency> >>>>>> >>>>>> <!-- >>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-json --> >>>>>> <dependency> >>>>>> <groupId>org.apache.flink</groupId> >>>>>> <artifactId>flink-json</artifactId> >>>>>> <version>1.13.2</version> >>>>>> </dependency> >>>>>> >>>>>> <!-- >>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-sql-connector-hive-1.2.2 >>>>>> --> >>>>>> <dependency> >>>>>> <groupId>org.apache.flink</groupId> >>>>>> <artifactId>flink-sql-connector-hive-2.3.6_2.12</artifactId> >>>>>> <version>1.13.2</version> >>>>>> </dependency> >>>>>> >>>>>> <!-- >>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-core --> >>>>>> <dependency> >>>>>> <groupId>org.apache.flink</groupId> >>>>>> <artifactId>flink-core</artifactId> >>>>>> <version>1.13.2</version> >>>>>> </dependency> >>>>>> >>>>>> <!-- >>>>>> https://mvnrepository.com/artifact/org.apache.kafka/kafka-clients --> >>>>>> <dependency> >>>>>> <groupId>org.apache.kafka</groupId> >>>>>> <artifactId>kafka-clients</artifactId> >>>>>> <version>2.8.0</version> >>>>>> </dependency> >>>>>> >>>>>> <dependency> >>>>>> <groupId>org.apache.kafka</groupId> >>>>>> <artifactId>kafka_2.12</artifactId> >>>>>> <version>2.8.0</version> >>>>>> </dependency> >>>>>> >>>>>> >>>>>> <!-- Add logging framework, to produce console output when >>>>>> running in the IDE. --> >>>>>> <!-- These dependencies are excluded from the application JAR by >>>>>> default. --> >>>>>> <dependency> >>>>>> <groupId>org.slf4j</groupId> >>>>>> <artifactId>slf4j-log4j12</artifactId> >>>>>> <version>1.7.7</version> >>>>>> <scope>runtime</scope> >>>>>> </dependency> >>>>>> <dependency> >>>>>> <groupId>log4j</groupId> >>>>>> <artifactId>log4j</artifactId> >>>>>> <version>1.2.17</version> >>>>>> <scope>runtime</scope> >>>>>> </dependency> >>>>>> </dependencies> >>>>>> >>>>>> <build> >>>>>> <plugins> >>>>>> >>>>>> <!-- Java Compiler --> >>>>>> <plugin> >>>>>> <groupId>org.apache.maven.plugins</groupId> >>>>>> <artifactId>maven-compiler-plugin</artifactId> >>>>>> <version>3.1</version> >>>>>> <configuration> >>>>>> <source>${java.version}</source> >>>>>> <target>${java.version}</target> >>>>>> </configuration> >>>>>> </plugin> >>>>>> >>>>>> <!-- We use the maven-shade plugin to create a fat jar that >>>>>> contains all necessary dependencies. --> >>>>>> <!-- Change the value of <mainClass>...</mainClass> if your >>>>>> program entry point changes. --> >>>>>> <plugin> >>>>>> <groupId>org.apache.maven.plugins</groupId> >>>>>> <artifactId>maven-shade-plugin</artifactId> >>>>>> <version>3.0.0</version> >>>>>> <executions> >>>>>> <!-- Run shade goal on package phase --> >>>>>> <execution> >>>>>> <phase>package</phase> >>>>>> <goals> >>>>>> <goal>shade</goal> >>>>>> </goals> >>>>>> <configuration> >>>>>> <artifactSet> >>>>>> <excludes> >>>>>> >>>>>> <exclude>org.apache.flink:force-shading</exclude> >>>>>> >>>>>> <exclude>com.google.code.findbugs:jsr305</exclude> >>>>>> <exclude>org.slf4j:*</exclude> >>>>>> <exclude>log4j:*</exclude> >>>>>> </excludes> >>>>>> </artifactSet> >>>>>> <filters> >>>>>> <filter> >>>>>> <!-- Do not copy the signatures in >>>>>> the META-INF folder. >>>>>> Otherwise, this might cause >>>>>> SecurityExceptions when using the JAR. --> >>>>>> <artifact>*:*</artifact> >>>>>> <excludes> >>>>>> <exclude>META-INF/*.SF</exclude> >>>>>> <exclude>META-INF/*.DSA</exclude> >>>>>> <exclude>META-INF/*.RSA</exclude> >>>>>> </excludes> >>>>>> </filter> >>>>>> </filters> >>>>>> <transformers> >>>>>> <transformer >>>>>> implementation="org.apache.maven.plugins.shade.resource.ManifestResourceTransformer"> >>>>>> >>>>>> <mainClass>com.harsh.test.StreamingJob</mainClass> >>>>>> </transformer> >>>>>> </transformers> >>>>>> </configuration> >>>>>> </execution> >>>>>> </executions> >>>>>> </plugin> >>>>>> </plugins> >>>>>> >>>>>> <pluginManagement> >>>>>> <plugins> >>>>>> >>>>>> <!-- This improves the out-of-the-box experience in >>>>>> Eclipse by resolving some warnings. --> >>>>>> <plugin> >>>>>> <groupId>org.eclipse.m2e</groupId> >>>>>> <artifactId>lifecycle-mapping</artifactId> >>>>>> <version>1.0.0</version> >>>>>> <configuration> >>>>>> <lifecycleMappingMetadata> >>>>>> <pluginExecutions> >>>>>> <pluginExecution> >>>>>> <pluginExecutionFilter> >>>>>> >>>>>> <groupId>org.apache.maven.plugins</groupId> >>>>>> >>>>>> <artifactId>maven-shade-plugin</artifactId> >>>>>> >>>>>> <versionRange>[3.0.0,)</versionRange> >>>>>> <goals> >>>>>> <goal>shade</goal> >>>>>> </goals> >>>>>> </pluginExecutionFilter> >>>>>> <action> >>>>>> <ignore/> >>>>>> </action> >>>>>> </pluginExecution> >>>>>> <pluginExecution> >>>>>> <pluginExecutionFilter> >>>>>> >>>>>> <groupId>org.apache.maven.plugins</groupId> >>>>>> >>>>>> <artifactId>maven-compiler-plugin</artifactId> >>>>>> >>>>>> <versionRange>[3.1,)</versionRange> >>>>>> <goals> >>>>>> <goal>testCompile</goal> >>>>>> <goal>compile</goal> >>>>>> </goals> >>>>>> </pluginExecutionFilter> >>>>>> <action> >>>>>> <ignore/> >>>>>> </action> >>>>>> </pluginExecution> >>>>>> </pluginExecutions> >>>>>> </lifecycleMappingMetadata> >>>>>> </configuration> >>>>>> </plugin> >>>>>> </plugins> >>>>>> </pluginManagement> >>>>>> </build> >>>>>> >>>>>> <!-- This profile helps to make things run out of the box in >>>>>> IntelliJ --> >>>>>> <!-- Its adds Flink's core classes to the runtime class path. --> >>>>>> <!-- Otherwise they are missing in IntelliJ, because the dependency >>>>>> is 'provided' --> >>>>>> <profiles> >>>>>> <profile> >>>>>> <id>add-dependencies-for-IDEA</id> >>>>>> >>>>>> <activation> >>>>>> <property> >>>>>> <name>idea.version</name> >>>>>> </property> >>>>>> </activation> >>>>>> >>>>>> <dependencies> >>>>>> <dependency> >>>>>> <groupId>org.apache.flink</groupId> >>>>>> <artifactId>flink-java</artifactId> >>>>>> <version>${flink.version}</version> >>>>>> <scope>compile</scope> >>>>>> </dependency> >>>>>> <dependency> >>>>>> <groupId>org.apache.flink</groupId> >>>>>> >>>>>> <artifactId>flink-streaming-java_${scala.binary.version}</artifactId> >>>>>> <version>${flink.version}</version> >>>>>> <scope>compile</scope> >>>>>> </dependency> >>>>>> </dependencies> >>>>>> </profile> >>>>>> </profiles> >>>>>> >>>>>> </project> >>>>>> >>>>>> >>>>>> Please help me resolve the issue. >>>>>> >>>>>> Thanks >>>>>> >>>>>> >>>>>> >>>>>> >>>> >>>> -- >>>> Thanks and Regards, >>>> Harshvardhan >>>> Data Platform >>>> >>> >> >> -- >> Thanks and Regards, >> Harshvardhan >> Data Platform >> > -- Thanks and Regards, Harshvardhan Data Platform