Did you try this way?

/usr/local/spark/bin/spark-submit --master mesos://mesos.master:5050 --conf
spark.mesos.executor.docker.image=docker.repo/spark:latest --class
org.apache.spark.examples.SparkPi *--jars hdfs://hdfs1/tmp/spark-*
*examples-1.4.1-hadoop2.6.0-**cdh5.4.4.jar* 100

Thanks
Best Regards

On Fri, Aug 7, 2015 at 5:51 AM, Alan Braithwaite <a...@cloudflare.com>
wrote:

> Hi All,
>
> We're trying to run spark with mesos and docker in client mode (since
> mesos doesn't support cluster mode) and load the application Jar from
> HDFS.  The following is the command we're running:
>
> /usr/local/spark/bin/spark-submit --master mesos://mesos.master:5050
> --conf spark.mesos.executor.docker.image=docker.repo/spark:latest --class
> org.apache.spark.examples.SparkPi
> hdfs://hdfs1/tmp/spark-examples-1.4.1-hadoop2.6.0-cdh5.4.4.jar 100
>
> We're getting the following warning before an exception from that command:
>
> Warning: Skip remote jar
> hdfs://hdfs1/tmp/spark-examples-1.4.1-hadoop2.6.0-cdh5.4.4.jar.
> java.lang.ClassNotFoundException: org.apache.spark.examples.SparkPi
>
> Before I debug further, is this even supported?  I started reading the
> code and it wasn't clear that it's possible to load a remote jar in client
> mode at all.  I did see a related issue in [2] but it didn't quite clarify
> everything I was looking for.
>
> Thanks,
> - Alan
>
> [1] https://spark.apache.org/docs/latest/submitting-applications.html
>
> [2]
> http://apache-spark-user-list.1001560.n3.nabble.com/Spark-submit-not-working-when-application-jar-is-in-hdfs-td21840.html
>

Reply via email to