class java.io.IOException: Cannot run program
"/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh"
(in directory "."): error=2, No such file or directory

By any chance, are your SPARK_HOME directories different on the
machine where you're submitting from and the cluster? I'm on an older
drop so not sure about the finer points of spark-submit but do
remember a very similar issue when trying to run a Spark driver on a
windows machine against a Spark Master on Ubuntu cluster (the
SPARK_HOME directories were obviously different)

On Wed, Jul 9, 2014 at 7:18 PM, Aris Vlasakakis <a...@vlasakakis.com> wrote:
> Hello everybody,
>
> I am trying to figure out how to submit a Spark application from one
> separate physical machine to a Spark stand alone cluster. I have an
> application that I wrote in Python that works if I am on the 1-Node Spark
> server itself, and from that spark installation I run bin/spark-submit with
> 1)  MASTER=local[*] or if 2) MASTER=spark://localhost:7077.
>
> However, I want to be on a separate machine that submits a job to Spark. Am
> I doing something wrong here? I think something is wrong because I am
> working from two different spark "installations" -- as in, on the big server
> I have one spark installation and I am running sbin/start-all.sh to run the
> standalone server (and that works), and then on a separate laptop I have a
> different installation of spark-1.0.0, but I am using the laptop's
> bin/spark-submit script to submit to the remote Spark server (using
> MASTER=spark://<remote-spark-master>:7077
>
> This "submit-to-remote cluster" does not work, even for the Scala examples
> like SparkPi.
>
> Concrete Example: I want to do submit the example SparkPi to the cluster,
> from my laptop.
>
> Server is 10.20.10.152, running master and slave, I can look at the Master
> web UI at http://10.20.10.152:8080. Great.
>
> From laptop (10.20.10.154), I try the following, using bin/run-example from
> a locally built version of spark 1.0.0 (so that I have the script
> spark-submit!):
>
> bin/spark-submit --verbose --class org.apache.spark.examples.SparkPi
> --master spark://10.20.10.152:7077
> examples/target/scala-2.10/spark-examples-1.0.0-hadoop1.0.4.jar
>
>
> This fails, with the errors at the bottom of this email.
>
> Am I doing something wrong? How can I submit to  a remote cluster? I get the
> same problem with bin/spark-submit.
>
>
>  bin/spark-submit --verbose --class org.apache.spark.examples.SparkPi
> --master spark://10.20.10.152:7077
> examples/target/scala-2.10/spark-examples-1.0.0-hadoop1.0.4.jar
> Using properties file: null
> Using properties file: null
> Parsed arguments:
>   master                  spark://10.20.10.152:7077
>   deployMode              null
>   executorMemory          null
>   executorCores           null
>   totalExecutorCores      null
>   propertiesFile          null
>   driverMemory            null
>   driverCores             null
>   driverExtraClassPath    null
>   driverExtraLibraryPath  null
>   driverExtraJavaOptions  null
>   supervise               false
>   queue                   null
>   numExecutors            null
>   files                   null
>   pyFiles                 null
>   archives                null
>   mainClass               org.apache.spark.examples.SparkPi
>   primaryResource
> file:/Users/aris.vlasakakis/Documents/spark-1.0.0/examples/target/scala-2.10/spark-examples-1.0.0-hadoop1.0.4.jar
>   name                    org.apache.spark.examples.SparkPi
>   childArgs               []
>   jars                    null
>   verbose                 true
>
> Default properties from null:
>
>
>
> Using properties file: null
> Main class:
> org.apache.spark.examples.SparkPi
> Arguments:
>
> System properties:
> SPARK_SUBMIT -> true
> spark.app.name -> org.apache.spark.examples.SparkPi
> spark.jars ->
> file:/Users/aris.vlasakakis/Documents/spark-1.0.0/examples/target/scala-2.10/spark-examples-1.0.0-hadoop1.0.4.jar
> spark.master -> spark://10.20.10.152:7077
> Classpath elements:
> file:/Users/aris.vlasakakis/Documents/spark-1.0.0/examples/target/scala-2.10/spark-examples-1.0.0-hadoop1.0.4.jar
>
>
> 14/07/09 16:16:08 INFO SecurityManager: Using Spark's default log4j profile:
> org/apache/spark/log4j-defaults.properties
> 14/07/09 16:16:08 INFO SecurityManager: Changing view acls to:
> aris.vlasakakis
> 14/07/09 16:16:08 INFO SecurityManager: SecurityManager: authentication
> disabled; ui acls disabled; users with view permissions:
> Set(aris.vlasakakis)
> 14/07/09 16:16:08 INFO Slf4jLogger: Slf4jLogger started
> 14/07/09 16:16:08 INFO Remoting: Starting remoting
> 14/07/09 16:16:08 INFO Remoting: Remoting started; listening on addresses
> :[akka.tcp://spark@10.20.10.154:50478]
> 14/07/09 16:16:08 INFO Remoting: Remoting now listens on addresses:
> [akka.tcp://spark@10.20.10.154:50478]
> 14/07/09 16:16:08 INFO SparkEnv: Registering MapOutputTracker
> 14/07/09 16:16:08 INFO SparkEnv: Registering BlockManagerMaster
> 14/07/09 16:16:08 INFO DiskBlockManager: Created local directory at
> /var/folders/ch/yfyhs7px5h90505g4n21n8d5k3svt3/T/spark-local-20140709161608-0531
> 14/07/09 16:16:08 INFO MemoryStore: MemoryStore started with capacity 5.8
> GB.
> 14/07/09 16:16:08 INFO ConnectionManager: Bound socket to port 50479 with id
> = ConnectionManagerId(10.20.10.154,50479)
> 14/07/09 16:16:08 INFO BlockManagerMaster: Trying to register BlockManager
> 14/07/09 16:16:08 INFO BlockManagerInfo: Registering block manager
> 10.20.10.154:50479 with 5.8 GB RAM
> 14/07/09 16:16:08 INFO BlockManagerMaster: Registered BlockManager
> 14/07/09 16:16:08 INFO HttpServer: Starting HTTP Server
> 14/07/09 16:16:09 INFO HttpBroadcast: Broadcast server started at
> http://10.20.10.154:50480
> 14/07/09 16:16:09 INFO HttpFileServer: HTTP File server directory is
> /var/folders/ch/yfyhs7px5h90505g4n21n8d5k3svt3/T/spark-edd787f4-f606-473c-965b-9f3b131cfb43
> 14/07/09 16:16:09 INFO HttpServer: Starting HTTP Server
> 14/07/09 16:16:09 INFO SparkUI: Started SparkUI at http://10.20.10.154:4040
> 2014-07-09 16:16:09.439 java[37388:d17] Unable to load realm mapping info
> from SCDynamicStore
> 14/07/09 16:16:09 INFO SparkContext: Added JAR
> file:/Users/aris.vlasakakis/Documents/spark-1.0.0/examples/target/scala-2.10/spark-examples-1.0.0-hadoop1.0.4.jar
> at http://10.20.10.154:50481/jars/spark-examples-1.0.0-hadoop1.0.4.jar with
> timestamp 1404947769853
> 14/07/09 16:16:09 INFO AppClient$ClientActor: Connecting to master
> spark://10.20.10.152:7077...
> 14/07/09 16:16:09 INFO SparkContext: Starting job: reduce at
> SparkPi.scala:35
> 14/07/09 16:16:09 INFO DAGScheduler: Got job 0 (reduce at SparkPi.scala:35)
> with 2 output partitions (allowLocal=false)
> 14/07/09 16:16:09 INFO DAGScheduler: Final stage: Stage 0(reduce at
> SparkPi.scala:35)
> 14/07/09 16:16:10 INFO DAGScheduler: Parents of final stage: List()
> 14/07/09 16:16:10 INFO DAGScheduler: Missing parents: List()
> 14/07/09 16:16:10 INFO DAGScheduler: Submitting Stage 0 (MappedRDD[1] at map
> at SparkPi.scala:31), which has no missing parents
> 14/07/09 16:16:10 INFO DAGScheduler: Submitting 2 missing tasks from Stage 0
> (MappedRDD[1] at map at SparkPi.scala:31)
> 14/07/09 16:16:10 INFO TaskSchedulerImpl: Adding task set 0.0 with 2 tasks
> 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Connected to Spark
> cluster with app ID app-20140709161541-0001
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor added:
> app-20140709161541-0001/0 on worker-20140709160420-10.20.10.152-57674
> (10.20.10.152:57674) with 4 cores
> 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Granted executor ID
> app-20140709161541-0001/0 on hostPort 10.20.10.152:57674 with 4 cores, 4.0
> GB RAM
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
> app-20140709161541-0001/0 is now RUNNING
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
> app-20140709161541-0001/0 is now FAILED (class java.io.IOException: Cannot
> run program
> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
> directory "."): error=2, No such file or directory)
> 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Executor
> app-20140709161541-0001/0 removed: class java.io.IOException: Cannot run
> program
> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
> directory "."): error=2, No such file or directory
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor added:
> app-20140709161541-0001/1 on worker-20140709160420-10.20.10.152-57674
> (10.20.10.152:57674) with 4 cores
> 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Granted executor ID
> app-20140709161541-0001/1 on hostPort 10.20.10.152:57674 with 4 cores, 4.0
> GB RAM
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
> app-20140709161541-0001/1 is now RUNNING
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
> app-20140709161541-0001/1 is now FAILED (class java.io.IOException: Cannot
> run program
> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
> directory "."): error=2, No such file or directory)
> 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Executor
> app-20140709161541-0001/1 removed: class java.io.IOException: Cannot run
> program
> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
> directory "."): error=2, No such file or directory
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor added:
> app-20140709161541-0001/2 on worker-20140709160420-10.20.10.152-57674
> (10.20.10.152:57674) with 4 cores
> 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Granted executor ID
> app-20140709161541-0001/2 on hostPort 10.20.10.152:57674 with 4 cores, 4.0
> GB RAM
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
> app-20140709161541-0001/2 is now RUNNING
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
> app-20140709161541-0001/2 is now FAILED (class java.io.IOException: Cannot
> run program
> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
> directory "."): error=2, No such file or directory)
> 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Executor
> app-20140709161541-0001/2 removed: class java.io.IOException: Cannot run
> program
> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
> directory "."): error=2, No such file or directory
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor added:
> app-20140709161541-0001/3 on worker-20140709160420-10.20.10.152-57674
> (10.20.10.152:57674) with 4 cores
> 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Granted executor ID
> app-20140709161541-0001/3 on hostPort 10.20.10.152:57674 with 4 cores, 4.0
> GB RAM
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
> app-20140709161541-0001/3 is now RUNNING
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
> app-20140709161541-0001/3 is now FAILED (class java.io.IOException: Cannot
> run program
> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
> directory "."): error=2, No such file or directory)
> 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Executor
> app-20140709161541-0001/3 removed: class java.io.IOException: Cannot run
> program
> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
> directory "."): error=2, No such file or directory
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor added:
> app-20140709161541-0001/4 on worker-20140709160420-10.20.10.152-57674
> (10.20.10.152:57674) with 4 cores
> 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Granted executor ID
> app-20140709161541-0001/4 on hostPort 10.20.10.152:57674 with 4 cores, 4.0
> GB RAM
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
> app-20140709161541-0001/4 is now RUNNING
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
> app-20140709161541-0001/4 is now FAILED (class java.io.IOException: Cannot
> run program
> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
> directory "."): error=2, No such file or directory)
> 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Executor
> app-20140709161541-0001/4 removed: class java.io.IOException: Cannot run
> program
> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
> directory "."): error=2, No such file or directory
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor added:
> app-20140709161541-0001/5 on worker-20140709160420-10.20.10.152-57674
> (10.20.10.152:57674) with 4 cores
> 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Granted executor ID
> app-20140709161541-0001/5 on hostPort 10.20.10.152:57674 with 4 cores, 4.0
> GB RAM
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
> app-20140709161541-0001/5 is now RUNNING
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
> app-20140709161541-0001/5 is now FAILED (class java.io.IOException: Cannot
> run program
> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
> directory "."): error=2, No such file or directory)
> 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Executor
> app-20140709161541-0001/5 removed: class java.io.IOException: Cannot run
> program
> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
> directory "."): error=2, No such file or directory
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor added:
> app-20140709161541-0001/6 on worker-20140709160420-10.20.10.152-57674
> (10.20.10.152:57674) with 4 cores
> 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Granted executor ID
> app-20140709161541-0001/6 on hostPort 10.20.10.152:57674 with 4 cores, 4.0
> GB RAM
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
> app-20140709161541-0001/6 is now RUNNING
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
> app-20140709161541-0001/6 is now FAILED (class java.io.IOException: Cannot
> run program
> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
> directory "."): error=2, No such file or directory)
> 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Executor
> app-20140709161541-0001/6 removed: class java.io.IOException: Cannot run
> program
> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
> directory "."): error=2, No such file or directory
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor added:
> app-20140709161541-0001/7 on worker-20140709160420-10.20.10.152-57674
> (10.20.10.152:57674) with 4 cores
> 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Granted executor ID
> app-20140709161541-0001/7 on hostPort 10.20.10.152:57674 with 4 cores, 4.0
> GB RAM
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
> app-20140709161541-0001/7 is now RUNNING
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
> app-20140709161541-0001/7 is now FAILED (class java.io.IOException: Cannot
> run program
> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
> directory "."): error=2, No such file or directory)
> 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Executor
> app-20140709161541-0001/7 removed: class java.io.IOException: Cannot run
> program
> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
> directory "."): error=2, No such file or directory
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor added:
> app-20140709161541-0001/8 on worker-20140709160420-10.20.10.152-57674
> (10.20.10.152:57674) with 4 cores
> 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Granted executor ID
> app-20140709161541-0001/8 on hostPort 10.20.10.152:57674 with 4 cores, 4.0
> GB RAM
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
> app-20140709161541-0001/8 is now RUNNING
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
> app-20140709161541-0001/8 is now FAILED (class java.io.IOException: Cannot
> run program
> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
> directory "."): error=2, No such file or directory)
> 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Executor
> app-20140709161541-0001/8 removed: class java.io.IOException: Cannot run
> program
> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
> directory "."): error=2, No such file or directory
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor added:
> app-20140709161541-0001/9 on worker-20140709160420-10.20.10.152-57674
> (10.20.10.152:57674) with 4 cores
> 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Granted executor ID
> app-20140709161541-0001/9 on hostPort 10.20.10.152:57674 with 4 cores, 4.0
> GB RAM
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
> app-20140709161541-0001/9 is now RUNNING
> 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
> app-20140709161541-0001/9 is now FAILED (class java.io.IOException: Cannot
> run program
> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
> directory "."): error=2, No such file or directory)
> 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Executor
> app-20140709161541-0001/9 removed: class java.io.IOException: Cannot run
> program
> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
> directory "."): error=2, No such file or directory
> 14/07/09 16:16:10 ERROR SparkDeploySchedulerBackend: Application has been
> killed. Reason: Master removed our application: FAILED
> 14/07/09 16:16:10 INFO TaskSchedulerImpl: Removed TaskSet 0.0, whose tasks
> have all completed, from pool
> 14/07/09 16:16:10 INFO DAGScheduler: Failed to run reduce at
> SparkPi.scala:35
> Exception in thread "main" 14/07/09 16:16:10 INFO TaskSchedulerImpl:
> Cancelling stage 0
> org.apache.spark.SparkException: Job aborted due to stage failure: Master
> removed our application: FAILED
>         at
> org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1033)
>         at
> org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1017)
>         at
> org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1015)
>         at
> scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
>         at
> scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
>         at
> org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1015)
>         at
> org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:633)
>         at
> org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:633)
>         at scala.Option.foreach(Option.scala:236)
>         at
> org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:633)
>         at
> org.apache.spark.scheduler.DAGSchedulerEventProcessActor$$anonfun$receive$2.applyOrElse(DAGScheduler.scala:1207)
>         at akka.actor.ActorCell.receiveMessage(ActorCell.scala:498)
>         at akka.actor.ActorCell.invoke(ActorCell.scala:456)
>         at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:237)
>         at akka.dispatch.Mailbox.run(Mailbox.scala:219)
>         at
> akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386)
>         at
> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>         at
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
>         at
> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
>         at
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> --
> Άρης Βλασακάκης
> Aris Vlasakakis

Reply via email to