I'm unable to figure out which dependency to add in order for the
ByteArrayDeserializer class to get included in the jar. I have added all
the dependencies as per the documentation still unable to figure out the
cause.

On Fri, Sep 10, 2021 at 12:17 AM Robert Metzger <rmetz...@apache.org> wrote:

> Does the jar file you are trying to submit contain
> the org/apache/kafka/common/serialization/ByteArrayDeserializer class?
>
> On Thu, Sep 9, 2021 at 2:10 PM Harshvardhan Shinde <
> harshvardhan.shi...@oyorooms.com> wrote:
>
>> Here's the complete stack trace:
>>
>> Server Response:org.apache.flink.runtime.rest.handler.RestHandlerException:
>> Could not execute application. at
>> org.apache.flink.runtime.webmonitor.handlers.JarRunHandler.lambda$handleRequest$1(JarRunHandler.java:108)
>> at
>> java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:836)
>> at
>> java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:811)
>> at
>> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
>> at
>> java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1609)
>> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>> at java.util.concurrent.FutureTask.run(FutureTask.java:266) at
>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
>> at
>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
>> at
>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>> at
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>> at java.lang.Thread.run(Thread.java:748) Caused by:
>> java.util.concurrent.CompletionException: java.lang.NoClassDefFoundError:
>> org/apache/kafka/common/serialization/ByteArrayDeserializer at
>> java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:273)
>> at
>> java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:280)
>> at
>> java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1606)
>> ... 7 more Caused by: java.lang.NoClassDefFoundError:
>> org/apache/kafka/common/serialization/ByteArrayDeserializer at
>> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer.setDeserializer(FlinkKafkaConsumer.java:322)
>> at
>> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer.<init>(FlinkKafkaConsumer.java:223)
>> at
>> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer.<init>(FlinkKafkaConsumer.java:154)
>> at
>> org.apache.flink.streaming.connectors.kafka.table.KafkaDynamicSource.createKafkaConsumer(KafkaDynamicSource.java:383)
>> at
>> org.apache.flink.streaming.connectors.kafka.table.KafkaDynamicSource.getScanRuntimeProvider(KafkaDynamicSource.java:205)
>> at
>> org.apache.flink.table.planner.connectors.DynamicSourceUtils.validateScanSource(DynamicSourceUtils.java:453)
>> at
>> org.apache.flink.table.planner.connectors.DynamicSourceUtils.prepareDynamicSource(DynamicSourceUtils.java:161)
>> at
>> org.apache.flink.table.planner.connectors.DynamicSourceUtils.convertSourceToRel(DynamicSourceUtils.java:119)
>> at
>> org.apache.flink.table.planner.plan.schema.CatalogSourceTable.toRel(CatalogSourceTable.java:85)
>> at
>> org.apache.calcite.sql2rel.SqlToRelConverter.toRel(SqlToRelConverter.java:3585)
>> at
>> org.apache.calcite.sql2rel.SqlToRelConverter.convertIdentifier(SqlToRelConverter.java:2507)
>> at
>> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2144)
>> at
>> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2093)
>> at
>> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2050)
>> at
>> org.apache.calcite.sql2rel.SqlToRelConverter.convertSelectImpl(SqlToRelConverter.java:663)
>> at
>> org.apache.calcite.sql2rel.SqlToRelConverter.convertSelect(SqlToRelConverter.java:644)
>> at
>> org.apache.calcite.sql2rel.SqlToRelConverter.convertQueryRecursive(SqlToRelConverter.java:3438)
>> at
>> org.apache.calcite.sql2rel.SqlToRelConverter.convertQuery(SqlToRelConverter.java:570)
>> at 
>> org.apache.flink.table.planner.calcite.FlinkPlannerImpl.org$apache$flink$table$planner$calcite$FlinkPlannerImpl$$rel(FlinkPlannerImpl.scala:169)
>> at
>> org.apache.flink.table.planner.calcite.FlinkPlannerImpl.rel(FlinkPlannerImpl.scala:161)
>> at
>> org.apache.flink.table.planner.operations.SqlToOperationConverter.toQueryOperation(SqlToOperationConverter.java:989)
>> at
>> org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlQuery(SqlToOperationConverter.java:958)
>> at
>> org.apache.flink.table.planner.operations.SqlToOperationConverter.convert(SqlToOperationConverter.java:283)
>> at
>> org.apache.flink.table.planner.delegation.ParserImpl.parse(ParserImpl.java:101)
>> at
>> org.apache.flink.table.api.internal.TableEnvironmentImpl.sqlQuery(TableEnvironmentImpl.java:704)
>> at com.harsh.test.StreamingJob.main(StreamingJob.java:106) at
>> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>> at java.lang.reflect.Method.invoke(Method.java:498) at
>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:355)
>> at
>> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:222)
>> at org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:114)
>> at
>> org.apache.flink.client.deployment.application.DetachedApplicationRunner.tryExecuteJobs(DetachedApplicationRunner.java:84)
>> at
>> org.apache.flink.client.deployment.application.DetachedApplicationRunner.run(DetachedApplicationRunner.java:70)
>> at
>> org.apache.flink.runtime.webmonitor.handlers.JarRunHandler.lambda$handleRequest$0(JarRunHandler.java:102)
>> at
>> java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1604)
>> ... 7 more
>>
>> On Thu, Sep 9, 2021 at 5:36 PM Robert Metzger <rmetz...@apache.org>
>> wrote:
>>
>>> Can you share the full stack trace, not just a part of it?
>>>
>>> On Thu, Sep 9, 2021 at 1:43 PM Harshvardhan Shinde <
>>> harshvardhan.shi...@oyorooms.com> wrote:
>>>
>>>> Hi,
>>>>
>>>> I added the dependencies while trying to resolve the same
>>>> issue, thought I was missing them.
>>>>
>>>> Thanks
>>>>
>>>> On Thu, Sep 9, 2021 at 4:26 PM Robert Metzger <rmetz...@apache.org>
>>>> wrote:
>>>>
>>>>> Hey,
>>>>>
>>>>> Why do you have these dependencies in your pom?
>>>>>
>>>>>         <!-- 
>>>>> https://mvnrepository.com/artifact/org.apache.kafka/kafka-clients -->
>>>>>         <dependency>
>>>>>             <groupId>org.apache.kafka</groupId>
>>>>>             <artifactId>kafka-clients</artifactId>
>>>>>             <version>2.8.0</version>
>>>>>         </dependency>
>>>>>
>>>>>         <dependency>
>>>>>             <groupId>org.apache.kafka</groupId>
>>>>>             <artifactId>kafka_2.12</artifactId>
>>>>>             <version>2.8.0</version>
>>>>>         </dependency>
>>>>>
>>>>>
>>>>> They are not needed for using the Kafka connector of Flink (the flink
>>>>> kafka connector dependencies pulls the required dependencies)
>>>>>
>>>>>
>>>>> On Thu, Sep 9, 2021 at 12:02 PM Harshvardhan Shinde <
>>>>> harshvardhan.shi...@oyorooms.com> wrote:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> I'm trying a simple flink job that reads data from a kafka topic and
>>>>>> creates a Hive table.
>>>>>>
>>>>>> I'm following the steps from here
>>>>>> <https://ci.apache.org/projects/flink/flink-docs-release-1.13/docs/connectors/table/hive/overview/#connecting-to-hive>
>>>>>> .
>>>>>>
>>>>>> Here's my code:
>>>>>>
>>>>>> import org.apache.flink.table.api.EnvironmentSettings;
>>>>>> import org.apache.flink.table.api.Table;
>>>>>> import org.apache.flink.table.api.TableEnvironment;
>>>>>> import org.apache.flink.table.catalog.hive.HiveCatalog;
>>>>>>
>>>>>> EnvironmentSettings settings = 
>>>>>> EnvironmentSettings.newInstance().useBlinkPlanner().inStreamingMode().build();
>>>>>> TableEnvironment tableEnv = TableEnvironment.create(settings);
>>>>>>
>>>>>> String name            = "myhive";
>>>>>> String defaultDatabase = "harsh_test";
>>>>>> String hiveConfDir     = "/etc/hive/conf";
>>>>>>
>>>>>> HiveCatalog hive = new HiveCatalog(name, defaultDatabase, hiveConfDir);
>>>>>> tableEnv.registerCatalog(name, hive);
>>>>>>
>>>>>> // set the HiveCatalog as the current catalog of the session
>>>>>> tableEnv.useCatalog(name);
>>>>>>
>>>>>> tableEnv.executeSql("CREATE TABLE IF NOT EXISTS transactions (\n" +
>>>>>>       "  `created_at` TIMESTAMP(3) METADATA FROM 'timestamp',\n" +
>>>>>>       "      `partition` BIGINT METADATA VIRTUAL,\n" +
>>>>>>       "      `offset` BIGINT METADATA VIRTUAL,\n" +
>>>>>>       "    account_id  BIGINT,\n" +
>>>>>>       "    amount      BIGINT,\n" +
>>>>>>       "    transaction_time TIMESTAMP(3),\n" +
>>>>>>       "    WATERMARK FOR transaction_time AS transaction_time - INTERVAL 
>>>>>> '5' SECOND\n" +
>>>>>>       ") WITH (\n" +
>>>>>>       "    'connector' = 'kafka',\n" +
>>>>>>       "    'topic'     = 'flink-stream-table',\n" +
>>>>>>       "    'properties.bootstrap.servers' = '<BROKER_ADDRESS>:9092',\n" +
>>>>>>       "   'scan.startup.mode' = 'earliest-offset',\n" +
>>>>>>       "    'format'    = 'csv'\n" +
>>>>>>       ")");
>>>>>>
>>>>>> Table table = tableEnv.sqlQuery("Select * from transactions");
>>>>>> table.execute().print();
>>>>>>
>>>>>> The code builds successfully, but I'm getting the following runtime
>>>>>> error:
>>>>>>
>>>>>> Caused by: java.util.concurrent.CompletionException:
>>>>>> java.lang.NoClassDefFoundError:
>>>>>> org/apache/kafka/common/serialization/ByteArrayDeserializer at
>>>>>> java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:273)
>>>>>> at
>>>>>> java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:280)
>>>>>> at
>>>>>> java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1606)
>>>>>> ..
>>>>>>
>>>>>> Here are my pom.xml file contents:
>>>>>>
>>>>>> <!--
>>>>>> Licensed to the Apache Software Foundation (ASF) under one
>>>>>> or more contributor license agreements.  See the NOTICE file
>>>>>> distributed with this work for additional information
>>>>>> regarding copyright ownership.  The ASF licenses this file
>>>>>> to you under the Apache License, Version 2.0 (the
>>>>>> "License"); you may not use this file except in compliance
>>>>>> with the License.  You may obtain a copy of the License at
>>>>>>
>>>>>>   http://www.apache.org/licenses/LICENSE-2.0
>>>>>>
>>>>>> Unless required by applicable law or agreed to in writing,
>>>>>> software distributed under the License is distributed on an
>>>>>> "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
>>>>>> KIND, either express or implied.  See the License for the
>>>>>> specific language governing permissions and limitations
>>>>>> under the License.
>>>>>> -->
>>>>>> <project xmlns="http://maven.apache.org/POM/4.0.0"; 
>>>>>> xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance";
>>>>>>          xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 
>>>>>> http://maven.apache.org/xsd/maven-4.0.0.xsd";>
>>>>>>     <modelVersion>4.0.0</modelVersion>
>>>>>>
>>>>>>     <groupId>com.harsh.test</groupId>
>>>>>>     <artifactId>harsh-flink-test</artifactId>
>>>>>>     <version>1.0-SNAPSHOT</version>
>>>>>>     <packaging>jar</packaging>
>>>>>>
>>>>>>     <name>Flink Quickstart Job</name>
>>>>>>     <url>http://www.myorganization.org</url>
>>>>>>
>>>>>>     <properties>
>>>>>>         
>>>>>> <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
>>>>>>         <flink.version>1.13.2</flink.version>
>>>>>>         <java.version>1.8</java.version>
>>>>>>         <hive.version>2.3.6</hive.version>
>>>>>>         <scala.binary.version>2.12</scala.binary.version>
>>>>>>         <maven.compiler.source>${java.version}</maven.compiler.source>
>>>>>>         <maven.compiler.target>${java.version}</maven.compiler.target>
>>>>>>     </properties>
>>>>>>
>>>>>>     <repositories>
>>>>>>         <repository>
>>>>>>             <id>apache.snapshots</id>
>>>>>>             <name>Apache Development Snapshot Repository</name>
>>>>>>             
>>>>>> <url>https://repository.apache.org/content/repositories/snapshots/</url>
>>>>>>             <releases>
>>>>>>                 <enabled>false</enabled>
>>>>>>             </releases>
>>>>>>             <snapshots>
>>>>>>                 <enabled>true</enabled>
>>>>>>             </snapshots>
>>>>>>         </repository>
>>>>>>     </repositories>
>>>>>>
>>>>>>     <dependencies>
>>>>>>         <!-- Apache Flink dependencies -->
>>>>>>         <!-- These dependencies are provided, because they should not be 
>>>>>> packaged into the JAR file. -->
>>>>>>         <dependency>
>>>>>>             <groupId>org.apache.flink</groupId>
>>>>>>             <artifactId>flink-java</artifactId>
>>>>>>             <version>${flink.version}</version>
>>>>>>         </dependency>
>>>>>>         <dependency>
>>>>>>             <groupId>org.apache.flink</groupId>
>>>>>>             
>>>>>> <artifactId>flink-streaming-java_${scala.binary.version}</artifactId>
>>>>>>             <version>${flink.version}</version>
>>>>>>         </dependency>
>>>>>>
>>>>>>         <!-- Add connector dependencies here. They must be in the 
>>>>>> default scope (compile). -->
>>>>>>
>>>>>>         <!-- Example:
>>>>>>
>>>>>>         <dependency>
>>>>>>             <groupId>org.apache.flink</groupId>
>>>>>>             
>>>>>> <artifactId>flink-connector-kafka-0.10_${scala.binary.version}</artifactId>
>>>>>>             <version>${flink.version}</version>
>>>>>>         </dependency>
>>>>>>         -->
>>>>>>
>>>>>>         <!-- 
>>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-connector-kafka
>>>>>>  -->
>>>>>>         <dependency>
>>>>>>             <groupId>org.apache.flink</groupId>
>>>>>>             
>>>>>> <artifactId>flink-connector-kafka_${scala.binary.version}</artifactId>
>>>>>>             <version>${flink.version}</version>
>>>>>>         </dependency>
>>>>>>
>>>>>>         <dependency>
>>>>>>             <groupId>org.apache.flink</groupId>
>>>>>>             
>>>>>> <artifactId>flink-table-api-java-bridge_${scala.binary.version}</artifactId>
>>>>>>             <version>${flink.version}</version>
>>>>>>         </dependency>
>>>>>>
>>>>>>         <dependency>
>>>>>>             <groupId>org.apache.flink</groupId>
>>>>>>             
>>>>>> <artifactId>flink-table-planner-blink_${scala.binary.version}</artifactId>
>>>>>>             <version>${flink.version}</version>
>>>>>>         </dependency>
>>>>>>
>>>>>>         <!-- 
>>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-table-planner 
>>>>>> -->
>>>>>>         <dependency>
>>>>>>             <groupId>org.apache.flink</groupId>
>>>>>>             <artifactId>flink-table-planner_2.12</artifactId>
>>>>>>             <version>1.13.2</version>
>>>>>>         </dependency>
>>>>>>
>>>>>>
>>>>>>         <!-- Flink Dependency -->
>>>>>>         <dependency>
>>>>>>             <groupId>org.apache.flink</groupId>
>>>>>>             
>>>>>> <artifactId>flink-connector-hive_${scala.binary.version}</artifactId>
>>>>>>             <version>${flink.version}</version>
>>>>>>         </dependency>
>>>>>>
>>>>>>         <!-- Hive Dependency -->
>>>>>>         <dependency>
>>>>>>             <groupId>org.apache.hive</groupId>
>>>>>>             <artifactId>hive-exec</artifactId>
>>>>>>             <version>${hive.version}</version>
>>>>>>         </dependency>
>>>>>>
>>>>>>         <!-- 
>>>>>> https://mvnrepository.com/artifact/javax.servlet/javax.servlet-api -->
>>>>>>         <dependency>
>>>>>>             <groupId>javax.servlet</groupId>
>>>>>>             <artifactId>javax.servlet-api</artifactId>
>>>>>>             <version>3.1.0</version>
>>>>>>             <scope>provided</scope>
>>>>>>         </dependency>
>>>>>>
>>>>>>         <!-- 
>>>>>> https://mvnrepository.com/artifact/org.apache.htrace/htrace-core4 -->
>>>>>>         <dependency>
>>>>>>             <groupId>org.apache.htrace</groupId>
>>>>>>             <artifactId>htrace-core4</artifactId>
>>>>>>             <version>4.0.1-incubating</version>
>>>>>>         </dependency>
>>>>>>
>>>>>>         <!-- 
>>>>>> https://mvnrepository.com/artifact/commons-configuration/commons-configuration
>>>>>>  -->
>>>>>>         <dependency>
>>>>>>             <groupId>commons-configuration</groupId>
>>>>>>             <artifactId>commons-configuration</artifactId>
>>>>>>             <version>1.10</version>
>>>>>>         </dependency>
>>>>>>
>>>>>>         <!-- 
>>>>>> https://mvnrepository.com/artifact/commons-logging/commons-logging -->
>>>>>>         <dependency>
>>>>>>             <groupId>commons-logging</groupId>
>>>>>>             <artifactId>commons-logging</artifactId>
>>>>>>             <version>1.2</version>
>>>>>>         </dependency>
>>>>>>
>>>>>>         <!-- 
>>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-shaded-hadoop-2
>>>>>>  -->
>>>>>>         <dependency>
>>>>>>             <groupId>org.apache.flink</groupId>
>>>>>>             <artifactId>flink-shaded-hadoop-2</artifactId>
>>>>>>             <version>2.8.3-10.0</version>
>>>>>>         </dependency>
>>>>>>
>>>>>>
>>>>>>         <!-- 
>>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-hadoop-compatibility
>>>>>>  -->
>>>>>>         <dependency>
>>>>>>             <groupId>org.apache.flink</groupId>
>>>>>>             <artifactId>flink-hadoop-compatibility_2.12</artifactId>
>>>>>>             <version>1.13.2</version>
>>>>>>         </dependency>
>>>>>>
>>>>>>         <!-- 
>>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-hadoop-fs -->
>>>>>>         <dependency>
>>>>>>             <groupId>org.apache.flink</groupId>
>>>>>>             <artifactId>flink-hadoop-fs</artifactId>
>>>>>>             <version>1.13.2</version>
>>>>>>         </dependency>
>>>>>>
>>>>>>         <!-- 
>>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-csv -->
>>>>>>         <dependency>
>>>>>>             <groupId>org.apache.flink</groupId>
>>>>>>             <artifactId>flink-csv</artifactId>
>>>>>>             <version>1.13.2</version>
>>>>>>         </dependency>
>>>>>>
>>>>>>         <!-- 
>>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-json -->
>>>>>>         <dependency>
>>>>>>             <groupId>org.apache.flink</groupId>
>>>>>>             <artifactId>flink-json</artifactId>
>>>>>>             <version>1.13.2</version>
>>>>>>         </dependency>
>>>>>>
>>>>>>         <!-- 
>>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-sql-connector-hive-1.2.2
>>>>>>  -->
>>>>>>         <dependency>
>>>>>>             <groupId>org.apache.flink</groupId>
>>>>>>             <artifactId>flink-sql-connector-hive-2.3.6_2.12</artifactId>
>>>>>>             <version>1.13.2</version>
>>>>>>         </dependency>
>>>>>>
>>>>>>         <!-- 
>>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-core -->
>>>>>>         <dependency>
>>>>>>             <groupId>org.apache.flink</groupId>
>>>>>>             <artifactId>flink-core</artifactId>
>>>>>>             <version>1.13.2</version>
>>>>>>         </dependency>
>>>>>>
>>>>>>         <!-- 
>>>>>> https://mvnrepository.com/artifact/org.apache.kafka/kafka-clients -->
>>>>>>         <dependency>
>>>>>>             <groupId>org.apache.kafka</groupId>
>>>>>>             <artifactId>kafka-clients</artifactId>
>>>>>>             <version>2.8.0</version>
>>>>>>         </dependency>
>>>>>>
>>>>>>         <dependency>
>>>>>>             <groupId>org.apache.kafka</groupId>
>>>>>>             <artifactId>kafka_2.12</artifactId>
>>>>>>             <version>2.8.0</version>
>>>>>>         </dependency>
>>>>>>
>>>>>>
>>>>>>         <!-- Add logging framework, to produce console output when 
>>>>>> running in the IDE. -->
>>>>>>         <!-- These dependencies are excluded from the application JAR by 
>>>>>> default. -->
>>>>>>         <dependency>
>>>>>>             <groupId>org.slf4j</groupId>
>>>>>>             <artifactId>slf4j-log4j12</artifactId>
>>>>>>             <version>1.7.7</version>
>>>>>>             <scope>runtime</scope>
>>>>>>         </dependency>
>>>>>>         <dependency>
>>>>>>             <groupId>log4j</groupId>
>>>>>>             <artifactId>log4j</artifactId>
>>>>>>             <version>1.2.17</version>
>>>>>>             <scope>runtime</scope>
>>>>>>         </dependency>
>>>>>>     </dependencies>
>>>>>>
>>>>>>     <build>
>>>>>>         <plugins>
>>>>>>
>>>>>>             <!-- Java Compiler -->
>>>>>>             <plugin>
>>>>>>                 <groupId>org.apache.maven.plugins</groupId>
>>>>>>                 <artifactId>maven-compiler-plugin</artifactId>
>>>>>>                 <version>3.1</version>
>>>>>>                 <configuration>
>>>>>>                     <source>${java.version}</source>
>>>>>>                     <target>${java.version}</target>
>>>>>>                 </configuration>
>>>>>>             </plugin>
>>>>>>
>>>>>>             <!-- We use the maven-shade plugin to create a fat jar that 
>>>>>> contains all necessary dependencies. -->
>>>>>>             <!-- Change the value of <mainClass>...</mainClass> if your 
>>>>>> program entry point changes. -->
>>>>>>             <plugin>
>>>>>>                 <groupId>org.apache.maven.plugins</groupId>
>>>>>>                 <artifactId>maven-shade-plugin</artifactId>
>>>>>>                 <version>3.0.0</version>
>>>>>>                 <executions>
>>>>>>                     <!-- Run shade goal on package phase -->
>>>>>>                     <execution>
>>>>>>                         <phase>package</phase>
>>>>>>                         <goals>
>>>>>>                             <goal>shade</goal>
>>>>>>                         </goals>
>>>>>>                         <configuration>
>>>>>>                             <artifactSet>
>>>>>>                                 <excludes>
>>>>>>                                     
>>>>>> <exclude>org.apache.flink:force-shading</exclude>
>>>>>>                                     
>>>>>> <exclude>com.google.code.findbugs:jsr305</exclude>
>>>>>>                                     <exclude>org.slf4j:*</exclude>
>>>>>>                                     <exclude>log4j:*</exclude>
>>>>>>                                 </excludes>
>>>>>>                             </artifactSet>
>>>>>>                             <filters>
>>>>>>                                 <filter>
>>>>>>                                     <!-- Do not copy the signatures in 
>>>>>> the META-INF folder.
>>>>>>                                     Otherwise, this might cause 
>>>>>> SecurityExceptions when using the JAR. -->
>>>>>>                                     <artifact>*:*</artifact>
>>>>>>                                     <excludes>
>>>>>>                                         <exclude>META-INF/*.SF</exclude>
>>>>>>                                         <exclude>META-INF/*.DSA</exclude>
>>>>>>                                         <exclude>META-INF/*.RSA</exclude>
>>>>>>                                     </excludes>
>>>>>>                                 </filter>
>>>>>>                             </filters>
>>>>>>                             <transformers>
>>>>>>                                 <transformer 
>>>>>> implementation="org.apache.maven.plugins.shade.resource.ManifestResourceTransformer">
>>>>>>                                     
>>>>>> <mainClass>com.harsh.test.StreamingJob</mainClass>
>>>>>>                                 </transformer>
>>>>>>                             </transformers>
>>>>>>                         </configuration>
>>>>>>                     </execution>
>>>>>>                 </executions>
>>>>>>             </plugin>
>>>>>>         </plugins>
>>>>>>
>>>>>>         <pluginManagement>
>>>>>>             <plugins>
>>>>>>
>>>>>>                 <!-- This improves the out-of-the-box experience in 
>>>>>> Eclipse by resolving some warnings. -->
>>>>>>                 <plugin>
>>>>>>                     <groupId>org.eclipse.m2e</groupId>
>>>>>>                     <artifactId>lifecycle-mapping</artifactId>
>>>>>>                     <version>1.0.0</version>
>>>>>>                     <configuration>
>>>>>>                         <lifecycleMappingMetadata>
>>>>>>                             <pluginExecutions>
>>>>>>                                 <pluginExecution>
>>>>>>                                     <pluginExecutionFilter>
>>>>>>                                         
>>>>>> <groupId>org.apache.maven.plugins</groupId>
>>>>>>                                         
>>>>>> <artifactId>maven-shade-plugin</artifactId>
>>>>>>                                         
>>>>>> <versionRange>[3.0.0,)</versionRange>
>>>>>>                                         <goals>
>>>>>>                                             <goal>shade</goal>
>>>>>>                                         </goals>
>>>>>>                                     </pluginExecutionFilter>
>>>>>>                                     <action>
>>>>>>                                         <ignore/>
>>>>>>                                     </action>
>>>>>>                                 </pluginExecution>
>>>>>>                                 <pluginExecution>
>>>>>>                                     <pluginExecutionFilter>
>>>>>>                                         
>>>>>> <groupId>org.apache.maven.plugins</groupId>
>>>>>>                                         
>>>>>> <artifactId>maven-compiler-plugin</artifactId>
>>>>>>                                         
>>>>>> <versionRange>[3.1,)</versionRange>
>>>>>>                                         <goals>
>>>>>>                                             <goal>testCompile</goal>
>>>>>>                                             <goal>compile</goal>
>>>>>>                                         </goals>
>>>>>>                                     </pluginExecutionFilter>
>>>>>>                                     <action>
>>>>>>                                         <ignore/>
>>>>>>                                     </action>
>>>>>>                                 </pluginExecution>
>>>>>>                             </pluginExecutions>
>>>>>>                         </lifecycleMappingMetadata>
>>>>>>                     </configuration>
>>>>>>                 </plugin>
>>>>>>             </plugins>
>>>>>>         </pluginManagement>
>>>>>>     </build>
>>>>>>
>>>>>>     <!-- This profile helps to make things run out of the box in 
>>>>>> IntelliJ -->
>>>>>>     <!-- Its adds Flink's core classes to the runtime class path. -->
>>>>>>     <!-- Otherwise they are missing in IntelliJ, because the dependency 
>>>>>> is 'provided' -->
>>>>>>     <profiles>
>>>>>>         <profile>
>>>>>>             <id>add-dependencies-for-IDEA</id>
>>>>>>
>>>>>>             <activation>
>>>>>>                 <property>
>>>>>>                     <name>idea.version</name>
>>>>>>                 </property>
>>>>>>             </activation>
>>>>>>
>>>>>>             <dependencies>
>>>>>>                 <dependency>
>>>>>>                     <groupId>org.apache.flink</groupId>
>>>>>>                     <artifactId>flink-java</artifactId>
>>>>>>                     <version>${flink.version}</version>
>>>>>>                     <scope>compile</scope>
>>>>>>                 </dependency>
>>>>>>                 <dependency>
>>>>>>                     <groupId>org.apache.flink</groupId>
>>>>>>                     
>>>>>> <artifactId>flink-streaming-java_${scala.binary.version}</artifactId>
>>>>>>                     <version>${flink.version}</version>
>>>>>>                     <scope>compile</scope>
>>>>>>                 </dependency>
>>>>>>             </dependencies>
>>>>>>         </profile>
>>>>>>     </profiles>
>>>>>>
>>>>>> </project>
>>>>>>
>>>>>>
>>>>>> Please help me resolve the issue.
>>>>>>
>>>>>> Thanks
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>
>>>> --
>>>> Thanks and Regards,
>>>> Harshvardhan
>>>> Data Platform
>>>>
>>>
>>
>> --
>> Thanks and Regards,
>> Harshvardhan
>> Data Platform
>>
>

-- 
Thanks and Regards,
Harshvardhan
Data Platform

Reply via email to