It seems that your Kafka clients dependency is not in your JAR file.


ByteArrayDeserializer is a symptom that seems to occur often. At least, I can find a similar question on Stackoverflow:

https://stackoverflow.com/questions/58408494/flink-1-3-class-not-found-org-apache-kafka-common-bytearraydeserializer

Maybe this helps.

Btw for Table API and SQL, you could also try out the fat JAR that we provide for the SQL Client:

https://ci.apache.org/projects/flink/flink-docs-release-1.13/docs/connectors/table/kafka/

links to

https://repo.maven.apache.org/maven2/org/apache/flink/flink-sql-connector-kafka_2.11/1.13.2/flink-sql-connector-kafka_2.11-1.13.2.jar

Here we try to do the dependency management for users.

Regards,
Timo

On 10.09.21 11:20, Harshvardhan Shinde wrote:
I'm unable to figure out which dependency to add in order for the ByteArrayDeserializer class to get included in the jar. I have added all the dependencies as per the documentation still unable to figure out the cause.

On Fri, Sep 10, 2021 at 12:17 AM Robert Metzger <rmetz...@apache.org <mailto:rmetz...@apache.org>> wrote:

    Does the jar file you are trying to submit contain
    the org/apache/kafka/common/serialization/ByteArrayDeserializer class?

    On Thu, Sep 9, 2021 at 2:10 PM Harshvardhan Shinde
    <harshvardhan.shi...@oyorooms.com
    <mailto:harshvardhan.shi...@oyorooms.com>> wrote:

        Here's the complete stack trace:

        Server
        Response:org.apache.flink.runtime.rest.handler.RestHandlerException:
        Could not execute application. at
        
org.apache.flink.runtime.webmonitor.handlers.JarRunHandler.lambda$handleRequest$1(JarRunHandler.java:108)
        at
        
java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:836)
        at
        
java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:811)
        at
        
java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
        at
        
java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1609)
        at
        java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
        at java.util.concurrent.FutureTask.run(FutureTask.java:266) at
        
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
        at
        
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
        at
        
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at
        
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748) Caused by:
        java.util.concurrent.CompletionException:
        java.lang.NoClassDefFoundError:
        org/apache/kafka/common/serialization/ByteArrayDeserializer at
        
java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:273)
        at
        
java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:280)
        at
        
java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1606)
        ... 7 more Caused by: java.lang.NoClassDefFoundError:
        org/apache/kafka/common/serialization/ByteArrayDeserializer at
        
org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer.setDeserializer(FlinkKafkaConsumer.java:322)
        at
        
org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer.<init>(FlinkKafkaConsumer.java:223)
        at
        
org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer.<init>(FlinkKafkaConsumer.java:154)
        at
        
org.apache.flink.streaming.connectors.kafka.table.KafkaDynamicSource.createKafkaConsumer(KafkaDynamicSource.java:383)
        at
        
org.apache.flink.streaming.connectors.kafka.table.KafkaDynamicSource.getScanRuntimeProvider(KafkaDynamicSource.java:205)
        at
        
org.apache.flink.table.planner.connectors.DynamicSourceUtils.validateScanSource(DynamicSourceUtils.java:453)
        at
        
org.apache.flink.table.planner.connectors.DynamicSourceUtils.prepareDynamicSource(DynamicSourceUtils.java:161)
        at
        
org.apache.flink.table.planner.connectors.DynamicSourceUtils.convertSourceToRel(DynamicSourceUtils.java:119)
        at
        
org.apache.flink.table.planner.plan.schema.CatalogSourceTable.toRel(CatalogSourceTable.java:85)
        at
        
org.apache.calcite.sql2rel.SqlToRelConverter.toRel(SqlToRelConverter.java:3585)
        at
        
org.apache.calcite.sql2rel.SqlToRelConverter.convertIdentifier(SqlToRelConverter.java:2507)
        at
        
org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2144)
        at
        
org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2093)
        at
        
org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2050)
        at
        
org.apache.calcite.sql2rel.SqlToRelConverter.convertSelectImpl(SqlToRelConverter.java:663)
        at
        
org.apache.calcite.sql2rel.SqlToRelConverter.convertSelect(SqlToRelConverter.java:644)
        at
        
org.apache.calcite.sql2rel.SqlToRelConverter.convertQueryRecursive(SqlToRelConverter.java:3438)
        at
        
org.apache.calcite.sql2rel.SqlToRelConverter.convertQuery(SqlToRelConverter.java:570)
        at org.apache.flink.table.planner.calcite.FlinkPlannerImpl.org
        
<http://org.apache.flink.table.planner.calcite.FlinkPlannerImpl.org>$apache$flink$table$planner$calcite$FlinkPlannerImpl$$rel(FlinkPlannerImpl.scala:169)
        at
        
org.apache.flink.table.planner.calcite.FlinkPlannerImpl.rel(FlinkPlannerImpl.scala:161)
        at
        
org.apache.flink.table.planner.operations.SqlToOperationConverter.toQueryOperation(SqlToOperationConverter.java:989)
        at
        
org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlQuery(SqlToOperationConverter.java:958)
        at
        
org.apache.flink.table.planner.operations.SqlToOperationConverter.convert(SqlToOperationConverter.java:283)
        at
        
org.apache.flink.table.planner.delegation.ParserImpl.parse(ParserImpl.java:101)
        at
        
org.apache.flink.table.api.internal.TableEnvironmentImpl.sqlQuery(TableEnvironmentImpl.java:704)
        at com.harsh.test.StreamingJob.main(StreamingJob.java:106) at
        sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at
        
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at
        
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:498) at
        
org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:355)
        at
        
org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:222)
        at
        org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:114)
        at
        
org.apache.flink.client.deployment.application.DetachedApplicationRunner.tryExecuteJobs(DetachedApplicationRunner.java:84)
        at
        
org.apache.flink.client.deployment.application.DetachedApplicationRunner.run(DetachedApplicationRunner.java:70)
        at
        
org.apache.flink.runtime.webmonitor.handlers.JarRunHandler.lambda$handleRequest$0(JarRunHandler.java:102)
        at
        
java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1604)
        ... 7 more

        On Thu, Sep 9, 2021 at 5:36 PM Robert Metzger
        <rmetz...@apache.org <mailto:rmetz...@apache.org>> wrote:

            Can you share the full stack trace, not just a part of it?

            On Thu, Sep 9, 2021 at 1:43 PM Harshvardhan Shinde
            <harshvardhan.shi...@oyorooms.com
            <mailto:harshvardhan.shi...@oyorooms.com>> wrote:

                Hi,

                I added the dependencies while trying to resolve the
                same issue, thought I was missing them.

                Thanks

                On Thu, Sep 9, 2021 at 4:26 PM Robert Metzger
                <rmetz...@apache.org <mailto:rmetz...@apache.org>> wrote:

                    Hey,

                    Why do you have these dependencies in your pom?

                    <!--
                    
https://mvnrepository.com/artifact/org.apache.kafka/kafka-clients
                    
<https://mvnrepository.com/artifact/org.apache.kafka/kafka-clients>
                    -->
                    <dependency>
                    <groupId>org.apache.kafka</groupId>
                    <artifactId>kafka-clients</artifactId>
                    <version>2.8.0</version>
                    </dependency>

                    <dependency>
                    <groupId>org.apache.kafka</groupId>
                    <artifactId>kafka_2.12</artifactId>
                    <version>2.8.0</version>
                    </dependency>


                    They are not needed for using the Kafka connector of
                    Flink (the flink kafka connector dependencies pulls
                    the required dependencies)


                    On Thu, Sep 9, 2021 at 12:02 PM Harshvardhan Shinde
                    <harshvardhan.shi...@oyorooms.com
                    <mailto:harshvardhan.shi...@oyorooms.com>> wrote:

                        Hi,

                        I'm trying a simple flink job that reads data
                        from a kafka topic and creates a Hive table.

                        I'm following the steps from here
                        
<https://ci.apache.org/projects/flink/flink-docs-release-1.13/docs/connectors/table/hive/overview/#connecting-to-hive>.

                        Here's my code:

                        import org.apache.flink.table.api.EnvironmentSettings;
                        import org.apache.flink.table.api.Table;
                        import org.apache.flink.table.api.TableEnvironment;
                        import org.apache.flink.table.catalog.hive.HiveCatalog;

                        EnvironmentSettings settings = 
EnvironmentSettings.newInstance().useBlinkPlanner().inStreamingMode().build();
                        TableEnvironment tableEnv = 
TableEnvironment.create(settings);

                        String name            ="myhive";
                        String defaultDatabase ="harsh_test";
                        String hiveConfDir     ="/etc/hive/conf";

                        HiveCatalog hive =new HiveCatalog(name, 
defaultDatabase, hiveConfDir);
                        tableEnv.registerCatalog(name, hive);

                        // set the HiveCatalog as the current catalog of
                        the session
                        tableEnv.useCatalog(name);

                        tableEnv.executeSql("CREATE TABLE IF NOT EXISTS transactions 
(\n" +
                               " `created_at` TIMESTAMP(3) METADATA FROM
                        'timestamp',\n" +
                               " `partition` BIGINT METADATA VIRTUAL,\n" +
                               " `offset` BIGINT METADATA VIRTUAL,\n" +
                               " account_id BIGINT,\n" +
                               " amount BIGINT,\n" +
                               " transaction_time TIMESTAMP(3),\n" +
                               " WATERMARK FOR transaction_time AS
                        transaction_time - INTERVAL '5' SECOND\n" +
                               ") WITH (\n" +
                               " 'connector' = 'kafka',\n" +
                               " 'topic' = 'flink-stream-table',\n" +
                               " 'properties.bootstrap.servers' =
                        '<BROKER_ADDRESS>:9092',\n" +
                               " 'scan.startup.mode' = 'earliest-offset',\n" +
                               " 'format' = 'csv'\n" +
                               ")");

                        Table table = tableEnv.sqlQuery("Select * from 
transactions");
                        table.execute().print();

                        The code builds successfully, but I'm getting
                        the following runtime error:

                        Caused by:
                        java.util.concurrent.CompletionException:
                        java.lang.NoClassDefFoundError:
                        
org/apache/kafka/common/serialization/ByteArrayDeserializer
                        at
                        
java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:273)
                        at
                        
java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:280)
                        at
                        
java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1606)
                        ..

                        Here are my pom.xml file contents:

                        <!--
                        Licensed to the Apache Software Foundation (ASF)
                        under one
                        or more contributor license agreements. See the
                        NOTICE file
                        distributed with this work for additional
                        information
                        regarding copyright ownership. The ASF licenses
                        this file
                        to you under the Apache License, Version 2.0 (the
                        "License"); you may not use this file except in
                        compliance
                        with the License. You may obtain a copy of the
                        License at

                        http://www.apache.org/licenses/LICENSE-2.0
                        <http://www.apache.org/licenses/LICENSE-2.0>

                        Unless required by applicable law or agreed to
                        in writing,
                        software distributed under the License is
                        distributed on an
                        "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS
                        OF ANY
                        KIND, either express or implied. See the License
                        for the
                        specific language governing permissions and
                        limitations
                        under the License.
                        -->
                        <project
                        xmlns="http://maven.apache.org/POM/4.0.0
                        <http://maven.apache.org/POM/4.0.0>"
                        xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance
                        <http://www.w3.org/2001/XMLSchema-instance>"
                        xsi:schemaLocation="http://maven.apache.org/POM/4.0.0
                        <http://maven.apache.org/POM/4.0.0>
                        http://maven.apache.org/xsd/maven-4.0.0.xsd
                        <http://maven.apache.org/xsd/maven-4.0.0.xsd>">
                        <modelVersion>4.0.0</modelVersion>

                        <groupId>com.harsh.test</groupId>
                        <artifactId>harsh-flink-test</artifactId>
                        <version>1.0-SNAPSHOT</version>
                        <packaging>jar</packaging>

                        <name>Flink Quickstart Job</name>
                        <url>http://www.myorganization.org  
<http://www.myorganization.org></url>

                        <properties>
                        
<project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
                        <flink.version>1.13.2</flink.version>
                        <java.version>1.8</java.version>
                        <hive.version>2.3.6</hive.version>
                        <scala.binary.version>2.12</scala.binary.version>
                        
<maven.compiler.source>${java.version}</maven.compiler.source>
                        
<maven.compiler.target>${java.version}</maven.compiler.target>
                        </properties>

                        <repositories>
                        <repository>
                        <id>apache.snapshots</id>
                        <name>Apache Development Snapshot Repository</name>
                        <url>https://repository.apache.org/content/repositories/snapshots/  
<https://repository.apache.org/content/repositories/snapshots/></url>
                        <releases>
                        <enabled>false</enabled>
                        </releases>
                        <snapshots>
                        <enabled>true</enabled>
                        </snapshots>
                        </repository>
                        </repositories>

                        <dependencies>
                        <!-- Apache Flink dependencies -->
                        <!-- These dependencies are provided, because
                        they should not be packaged into the JAR file. -->
                        <dependency>
                        <groupId>org.apache.flink</groupId>
                        <artifactId>flink-java</artifactId>
                        <version>${flink.version}</version>
                        </dependency>
                        <dependency>
                        <groupId>org.apache.flink</groupId>
                        
<artifactId>flink-streaming-java_${scala.binary.version}</artifactId>
                        <version>${flink.version}</version>
                        </dependency>

                        <!-- Add connector dependencies here. They must
                        be in the default scope (compile). -->

                        <!-- Example:

                        <dependency>
                        <groupId>org.apache.flink</groupId>
                        
<artifactId>flink-connector-kafka-0.10_${scala.binary.version}</artifactId>
                        <version>${flink.version}</version>
                        </dependency>
                        -->

                        <!--
                        
https://mvnrepository.com/artifact/org.apache.flink/flink-connector-kafka
                        
<https://mvnrepository.com/artifact/org.apache.flink/flink-connector-kafka>
                        -->
                        <dependency>
                        <groupId>org.apache.flink</groupId>
                        
<artifactId>flink-connector-kafka_${scala.binary.version}</artifactId>
                        <version>${flink.version}</version>
                        </dependency>

                        <dependency>
                        <groupId>org.apache.flink</groupId>
                        
<artifactId>flink-table-api-java-bridge_${scala.binary.version}</artifactId>
                        <version>${flink.version}</version>
                        </dependency>

                        <dependency>
                        <groupId>org.apache.flink</groupId>
                        
<artifactId>flink-table-planner-blink_${scala.binary.version}</artifactId>
                        <version>${flink.version}</version>
                        </dependency>

                        <!--
                        
https://mvnrepository.com/artifact/org.apache.flink/flink-table-planner
                        
<https://mvnrepository.com/artifact/org.apache.flink/flink-table-planner>
                        -->
                        <dependency>
                        <groupId>org.apache.flink</groupId>
                        <artifactId>flink-table-planner_2.12</artifactId>
                        <version>1.13.2</version>
                        </dependency>


                        <!-- Flink Dependency -->
                        <dependency>
                        <groupId>org.apache.flink</groupId>
                        
<artifactId>flink-connector-hive_${scala.binary.version}</artifactId>
                        <version>${flink.version}</version>
                        </dependency>

                        <!-- Hive Dependency -->
                        <dependency>
                        <groupId>org.apache.hive</groupId>
                        <artifactId>hive-exec</artifactId>
                        <version>${hive.version}</version>
                        </dependency>

                        <!--
                        
https://mvnrepository.com/artifact/javax.servlet/javax.servlet-api
                        
<https://mvnrepository.com/artifact/javax.servlet/javax.servlet-api>
                        -->
                        <dependency>
                        <groupId>javax.servlet</groupId>
                        <artifactId>javax.servlet-api</artifactId>
                        <version>3.1.0</version>
                        <scope>provided</scope>
                        </dependency>

                        <!--
                        
https://mvnrepository.com/artifact/org.apache.htrace/htrace-core4
                        
<https://mvnrepository.com/artifact/org.apache.htrace/htrace-core4>
                        -->
                        <dependency>
                        <groupId>org.apache.htrace</groupId>
                        <artifactId>htrace-core4</artifactId>
                        <version>4.0.1-incubating</version>
                        </dependency>

                        <!--
                        
https://mvnrepository.com/artifact/commons-configuration/commons-configuration
                        
<https://mvnrepository.com/artifact/commons-configuration/commons-configuration>
                        -->
                        <dependency>
                        <groupId>commons-configuration</groupId>
                        <artifactId>commons-configuration</artifactId>
                        <version>1.10</version>
                        </dependency>

                        <!--
                        
https://mvnrepository.com/artifact/commons-logging/commons-logging
                        
<https://mvnrepository.com/artifact/commons-logging/commons-logging>
                        -->
                        <dependency>
                        <groupId>commons-logging</groupId>
                        <artifactId>commons-logging</artifactId>
                        <version>1.2</version>
                        </dependency>

                        <!--
                        
https://mvnrepository.com/artifact/org.apache.flink/flink-shaded-hadoop-2
                        
<https://mvnrepository.com/artifact/org.apache.flink/flink-shaded-hadoop-2>
                        -->
                        <dependency>
                        <groupId>org.apache.flink</groupId>
                        <artifactId>flink-shaded-hadoop-2</artifactId>
                        <version>2.8.3-10.0</version>
                        </dependency>


                        <!--
                        
https://mvnrepository.com/artifact/org.apache.flink/flink-hadoop-compatibility
                        
<https://mvnrepository.com/artifact/org.apache.flink/flink-hadoop-compatibility>
                        -->
                        <dependency>
                        <groupId>org.apache.flink</groupId>
                        <artifactId>flink-hadoop-compatibility_2.12</artifactId>
                        <version>1.13.2</version>
                        </dependency>

                        <!--
                        
https://mvnrepository.com/artifact/org.apache.flink/flink-hadoop-fs
                        
<https://mvnrepository.com/artifact/org.apache.flink/flink-hadoop-fs>
                        -->
                        <dependency>
                        <groupId>org.apache.flink</groupId>
                        <artifactId>flink-hadoop-fs</artifactId>
                        <version>1.13.2</version>
                        </dependency>

                        <!--
                        
https://mvnrepository.com/artifact/org.apache.flink/flink-csv
                        
<https://mvnrepository.com/artifact/org.apache.flink/flink-csv>
                        -->
                        <dependency>
                        <groupId>org.apache.flink</groupId>
                        <artifactId>flink-csv</artifactId>
                        <version>1.13.2</version>
                        </dependency>

                        <!--
                        
https://mvnrepository.com/artifact/org.apache.flink/flink-json
                        
<https://mvnrepository.com/artifact/org.apache.flink/flink-json>
                        -->
                        <dependency>
                        <groupId>org.apache.flink</groupId>
                        <artifactId>flink-json</artifactId>
                        <version>1.13.2</version>
                        </dependency>

                        <!--
                        
https://mvnrepository.com/artifact/org.apache.flink/flink-sql-connector-hive-1.2.2
                        
<https://mvnrepository.com/artifact/org.apache.flink/flink-sql-connector-hive-1.2.2>
                        -->
                        <dependency>
                        <groupId>org.apache.flink</groupId>
                        
<artifactId>flink-sql-connector-hive-2.3.6_2.12</artifactId>
                        <version>1.13.2</version>
                        </dependency>

                        <!--
                        
https://mvnrepository.com/artifact/org.apache.flink/flink-core
                        
<https://mvnrepository.com/artifact/org.apache.flink/flink-core>
                        -->
                        <dependency>
                        <groupId>org.apache.flink</groupId>
                        <artifactId>flink-core</artifactId>
                        <version>1.13.2</version>
                        </dependency>

                        <!--
                        
https://mvnrepository.com/artifact/org.apache.kafka/kafka-clients
                        
<https://mvnrepository.com/artifact/org.apache.kafka/kafka-clients>
                        -->
                        <dependency>
                        <groupId>org.apache.kafka</groupId>
                        <artifactId>kafka-clients</artifactId>
                        <version>2.8.0</version>
                        </dependency>

                        <dependency>
                        <groupId>org.apache.kafka</groupId>
                        <artifactId>kafka_2.12</artifactId>
                        <version>2.8.0</version>
                        </dependency>


                        <!-- Add logging framework, to produce console
                        output when running in the IDE. -->
                        <!-- These dependencies are excluded from the
                        application JAR by default. -->
                        <dependency>
                        <groupId>org.slf4j</groupId>
                        <artifactId>slf4j-log4j12</artifactId>
                        <version>1.7.7</version>
                        <scope>runtime</scope>
                        </dependency>
                        <dependency>
                        <groupId>log4j</groupId>
                        <artifactId>log4j</artifactId>
                        <version>1.2.17</version>
                        <scope>runtime</scope>
                        </dependency>
                        </dependencies>

                        <build>
                        <plugins>

                        <!-- Java Compiler -->
                        <plugin>
                        <groupId>org.apache.maven.plugins</groupId>
                        <artifactId>maven-compiler-plugin</artifactId>
                        <version>3.1</version>
                        <configuration>
                        <source>${java.version}</source>
                        <target>${java.version}</target>
                        </configuration>
                        </plugin>

                        <!-- We use the maven-shade plugin to create a
                        fat jar that contains all necessary
                        dependencies. -->
                        <!-- Change the value of
                        <mainClass>...</mainClass> if your program entry
                        point changes. -->
                        <plugin>
                        <groupId>org.apache.maven.plugins</groupId>
                        <artifactId>maven-shade-plugin</artifactId>
                        <version>3.0.0</version>
                        <executions>
                        <!-- Run shade goal on package phase -->
                        <execution>
                        <phase>package</phase>
                        <goals>
                        <goal>shade</goal>
                        </goals>
                        <configuration>
                        <artifactSet>
                        <excludes>
                        <exclude>org.apache.flink:force-shading</exclude>
                        <exclude>com.google.code.findbugs:jsr305</exclude>
                        <exclude>org.slf4j:*</exclude>
                        <exclude>log4j:*</exclude>
                        </excludes>
                        </artifactSet>
                        <filters>
                        <filter>
                        <!-- Do not copy the signatures in the META-INF
                        folder.
                        Otherwise, this might cause SecurityExceptions
                        when using the JAR. -->
                        <artifact>*:*</artifact>
                        <excludes>
                        <exclude>META-INF/*.SF</exclude>
                        <exclude>META-INF/*.DSA</exclude>
                        <exclude>META-INF/*.RSA</exclude>
                        </excludes>
                        </filter>
                        </filters>
                        <transformers>
                        <transformer
                        
implementation="org.apache.maven.plugins.shade.resource.ManifestResourceTransformer">
                        <mainClass>com.harsh.test.StreamingJob</mainClass>
                        </transformer>
                        </transformers>
                        </configuration>
                        </execution>
                        </executions>
                        </plugin>
                        </plugins>

                        <pluginManagement>
                        <plugins>

                        <!-- This improves the out-of-the-box experience
                        in Eclipse by resolving some warnings. -->
                        <plugin>
                        <groupId>org.eclipse.m2e</groupId>
                        <artifactId>lifecycle-mapping</artifactId>
                        <version>1.0.0</version>
                        <configuration>
                        <lifecycleMappingMetadata>
                        <pluginExecutions>
                        <pluginExecution>
                        <pluginExecutionFilter>
                        <groupId>org.apache.maven.plugins</groupId>
                        <artifactId>maven-shade-plugin</artifactId>
                        <versionRange>[3.0.0,)</versionRange>
                        <goals>
                        <goal>shade</goal>
                        </goals>
                        </pluginExecutionFilter>
                        <action>
                        <ignore/>
                        </action>
                        </pluginExecution>
                        <pluginExecution>
                        <pluginExecutionFilter>
                        <groupId>org.apache.maven.plugins</groupId>
                        <artifactId>maven-compiler-plugin</artifactId>
                        <versionRange>[3.1,)</versionRange>
                        <goals>
                        <goal>testCompile</goal>
                        <goal>compile</goal>
                        </goals>
                        </pluginExecutionFilter>
                        <action>
                        <ignore/>
                        </action>
                        </pluginExecution>
                        </pluginExecutions>
                        </lifecycleMappingMetadata>
                        </configuration>
                        </plugin>
                        </plugins>
                        </pluginManagement>
                        </build>

                        <!-- This profile helps to make things run out
                        of the box in IntelliJ -->
                        <!-- Its adds Flink's core classes to the
                        runtime class path. -->
                        <!-- Otherwise they are missing in IntelliJ,
                        because the dependency is 'provided' -->
                        <profiles>
                        <profile>
                        <id>add-dependencies-for-IDEA</id>

                        <activation>
                        <property>
                        <name>idea.version</name>
                        </property>
                        </activation>

                        <dependencies>
                        <dependency>
                        <groupId>org.apache.flink</groupId>
                        <artifactId>flink-java</artifactId>
                        <version>${flink.version}</version>
                        <scope>compile</scope>
                        </dependency>
                        <dependency>
                        <groupId>org.apache.flink</groupId>
                        
<artifactId>flink-streaming-java_${scala.binary.version}</artifactId>
                        <version>${flink.version}</version>
                        <scope>compile</scope>
                        </dependency>
                        </dependencies>
                        </profile>
                        </profiles>

                        </project>


                        Please help me resolve the issue.

                        Thanks





-- Thanks and Regards,
                Harshvardhan
                Data Platform



-- Thanks and Regards,
        Harshvardhan
        Data Platform



--
Thanks and Regards,
Harshvardhan
Data Platform

Reply via email to