+1 on the mentioned change, Spark uses the following kafka-clients library:
<kafka.version>2.4.1</kafka.version> G On Mon, Dec 7, 2020 at 9:30 AM German Schiavon <gschiavonsp...@gmail.com> wrote: > Hi, > > I think the issue is that you are overriding the kafka-clients that comes > with <artifactId>spark-sql-kafka-0-10_2.12</artifactId> > > > I'd try removing the kafka-clients and see if it works > > > On Sun, 6 Dec 2020 at 08:01, Amit Joshi <mailtojoshia...@gmail.com> wrote: > >> Hi All, >> >> I am running the Spark Structured Streaming along with Kafka. >> Below is the pom.xml >> >> <properties> >> <maven.compiler.source>1.8</maven.compiler.source> >> <maven.compiler.target>1.8</maven.compiler.target> >> <encoding>UTF-8</encoding> >> <!-- Put the Scala version of the cluster --> >> <scalaVersion>2.12.10</scalaVersion> >> <sparkVersion>3.0.1</sparkVersion> >> </properties> >> >> <dependency> >> <groupId>org.apache.kafka</groupId> >> <artifactId>kafka-clients</artifactId> >> <version>2.1.0</version> >> </dependency> >> >> <dependency> >> <groupId>org.apache.spark</groupId> >> <artifactId>spark-core_2.12</artifactId> >> <version>${sparkVersion}</version> >> <scope>provided</scope> >> </dependency> >> <!-- https://mvnrepository.com/artifact/org.apache.spark/spark-sql --> >> <dependency> >> <groupId>org.apache.spark</groupId> >> <artifactId>spark-sql_2.12</artifactId> >> <version>${sparkVersion}</version> >> <scope>provided</scope> >> </dependency> >> <!-- >> https://mvnrepository.com/artifact/org.apache.spark/spark-sql-kafka-0-10 --> >> <dependency> >> <groupId>org.apache.spark</groupId> >> <artifactId>spark-sql-kafka-0-10_2.12</artifactId> >> <version>${sparkVersion}</version> >> </dependency> >> >> Building the fat jar with shade plugin. The jar is running as expected in my >> local setup with the command >> >> *spark-submit --master local[*] --class com.stream.Main --num-executors 3 >> --driver-memory 2g --executor-cores 2 --executor-memory 3g >> prism-event-synch-rta.jar* >> >> But when I am trying to run same jar in spark cluster using yarn with >> command: >> >> *spark-submit --master yarn --deploy-mode cluster --class com.stream.Main >> --num-executors 4 --driver-memory 2g --executor-cores 1 --executor-memory 4g >> gs://jars/prism-event-synch-rta.jar* >> >> Getting the this exception: >> >> >> >> >> *at org.apache.spark.sql.execution.streaming.StreamExecution.org >> <http://org.apache.spark.sql.execution.streaming.StreamExecution.org>$apache$spark$sql$execution$streaming$StreamExecution$$runStream(StreamExecution.scala:355) >> at >> org.apache.spark.sql.execution.streaming.StreamExecution$$anon$1.run(StreamExecution.scala:245)Caused >> by: org.apache.kafka.common.config.ConfigException: Missing required >> configuration "partition.assignment.strategy" which has no default value. at >> org.apache.kafka.common.config.ConfigDef.parse(ConfigDef.java:124)* >> >> I have tried setting up the "partition.assignment.strategy", then also its >> not working. >> >> Please help. >> >> >> Regards >> >> Amit Joshi >> >>