Hi,

I have a problem that's bothering me for a few days, and I'm pretty out of
ideas.

I built a Spark docker container where Spark runs in standalone mode. Both
master and worker are started there.

Now I tried to deploy my Spark Scala App in a separate container(same
machine) where I pass the Spark master URL and other stuff I need to connect
to Spark. Connection is seamless.

First problem I encountered was:

Exception in thread "main" org.apache.spark.SparkException: Job aborted due
to stage failure: Task 0 in stage 0.0 failed 4 times, most recent failure:
Lost task 0.3 in stage 0.0 (TID 3, 10.1.0.4):
java.lang.ClassNotFoundException:
org.apache.spark.streaming.kafka.KafkaRDDPartition

Then I made a folder of my dependencies except Spark, put them in a folder
alongside my app JAR file and added them to SparkConf using
SparkConf.setJars,

Now the strange thing happens:
Exception in thread "main" org.apache.spark.SparkException: Job aborted due
to stage failure: Task 0 in stage 0.0 failed 4 times, most recent failure:
Lost task 0.3 in stage 0.0 (TID 3, 10.1.0.4): java.lang.ClassCastException:
cannot assign instance of scala.collection.immutable.List$SerializationProxy
to field org.apache.spark.rdd.RDD.org$apache$spark$rdd$RDD$$dependencies_ of
type scala.collection.Seq in instance of
org.apache.spark.rdd.MapPartitionsRDD

More than this, if I just run the scala app from local using java -cp
<dependencies(including spark jars) cp> myApp.jar it works perfectly, jobs
run ok.

I don't have any SPARK_HOME locally and the setJars basically takes an empty
list, as If I won't use it.

I guess it uses jars provided in the classpath when I run my app and I don't
need to provide anything else.

If any of you guys have any ideas I would be grateful, I really can't
explain myself why this doesn't work and I didn't do any Spark deployments
until now. I mostly ran in embedded Spark.

Spark is same version in my app dependencies (2.0.0) as the one running in
the docker container.
I used Scala 2.11.7 for my app
Java 1.8 on both containers(app, spark)



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/Spark-Docker-Container-Jars-problem-when-deploying-my-app-tp27878.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

---------------------------------------------------------------------
To unsubscribe e-mail: user-unsubscr...@spark.apache.org

Reply via email to