Andrew, thank you so much! That worked! I had to manually set the
spark.home configuration in the SparkConf object using
.set("spark.home","/cluster/path/to/spark/"), and then I was able to submit
from my laptop to the cluster!

Aris


On Thu, Jul 10, 2014 at 11:41 AM, Andrew Or <and...@databricks.com> wrote:

> Setting SPARK_HOME is not super effective, because it is overridden very
> quickly by bin/spark-submit here
> <https://github.com/apache/spark/blob/88006a62377d2b7c9886ba49ceef158737bc1b97/bin/spark-submit#L20>.
> Instead you should set the config "spark.home". Here's why:
>
> Each of your executors inherits its spark home from the application
> description, and this is created by your SparkContext on your local
> machine. By default, as you noticed, this uses your local spark home that
> is not applicable to your remote executors. There are two ways of
> controlling the spark home set in your application description: through the
> "spark.home" config or "SPARK_HOME" environment variable, with the former
> taking priority over the latter
> <https://github.com/apache/spark/blob/88006a62377d2b7c9886ba49ceef158737bc1b97/core/src/main/scala/org/apache/spark/SparkContext.scala#L1017>.
> However, since spark-submit overwrites whatever value you set SPARK_HOME
> to, there is really only one way: by setting "spark.home". Note that this
> config is only used to launch the executors, meaning the driver has already
> started by the time this config is consumed, so this does not create any
> ordering issue.
>
> Does that make sense?
>
> Andrew
>
>
> 2014-07-10 10:17 GMT-07:00 Aris Vlasakakis <a...@vlasakakis.com>:
>
> Thank you very much Yana for replying!
>>
>> So right now the set up is a single-node machine which is my "cluster",
>> and YES you are right my submitting laptop has a different path to the
>> spark-1.0.0 installation than the "cluster" machine.
>>
>> I tried to set SPARK_HOME on my submittor laptop using the actual path of
>> the CLUSTER's directories...hoping this would help. It didn't change
>> anything. Let me show you the command from the submission side on my laptop
>> - keep in mind I set SPARK_HOME as the path for the cluster.
>>
>> SPARK_HOME=/home/data/Documents/spark-1.0.0/
>>  ~/Documents/spark-1.0.0/bin/spark-submit --verbose --class
>> org.apache.spark.examples.SparkPi --master spark://10.20.10.152:7077
>> ~/Documents/spark-1.0.0/examples/target/scala-2.10/spark-examples-1.0.0-hadoop1.0.4.jar
>>
>> I get the same problems as originally. I am actually confused here - if I
>> set the SPARK_HOME on the same line as the spark-submit command, I don't
>> get how the submittor script did not get all confused just running
>> bin/spark-submit...but what do I know.
>>
>> Am I setting SPARK_HOME correctly?
>>
>>
>> On Wed, Jul 9, 2014 at 7:09 PM, Yana Kadiyska <yana.kadiy...@gmail.com>
>> wrote:
>>
>>>  class java.io.IOException: Cannot run program
>>> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh"
>>> (in directory "."): error=2, No such file or directory
>>>
>>> By any chance, are your SPARK_HOME directories different on the
>>> machine where you're submitting from and the cluster? I'm on an older
>>> drop so not sure about the finer points of spark-submit but do
>>> remember a very similar issue when trying to run a Spark driver on a
>>> windows machine against a Spark Master on Ubuntu cluster (the
>>> SPARK_HOME directories were obviously different)
>>>
>>> On Wed, Jul 9, 2014 at 7:18 PM, Aris Vlasakakis <a...@vlasakakis.com>
>>> wrote:
>>> > Hello everybody,
>>> >
>>> > I am trying to figure out how to submit a Spark application from one
>>> > separate physical machine to a Spark stand alone cluster. I have an
>>> > application that I wrote in Python that works if I am on the 1-Node
>>> Spark
>>> > server itself, and from that spark installation I run bin/spark-submit
>>> with
>>> > 1)  MASTER=local[*] or if 2) MASTER=spark://localhost:7077.
>>> >
>>> > However, I want to be on a separate machine that submits a job to
>>> Spark. Am
>>> > I doing something wrong here? I think something is wrong because I am
>>> > working from two different spark "installations" -- as in, on the big
>>> server
>>> > I have one spark installation and I am running sbin/start-all.sh to
>>> run the
>>> > standalone server (and that works), and then on a separate laptop I
>>> have a
>>> > different installation of spark-1.0.0, but I am using the laptop's
>>> > bin/spark-submit script to submit to the remote Spark server (using
>>> > MASTER=spark://<remote-spark-master>:7077
>>> >
>>> > This "submit-to-remote cluster" does not work, even for the Scala
>>> examples
>>> > like SparkPi.
>>> >
>>> > Concrete Example: I want to do submit the example SparkPi to the
>>> cluster,
>>> > from my laptop.
>>> >
>>> > Server is 10.20.10.152, running master and slave, I can look at the
>>> Master
>>> > web UI at http://10.20.10.152:8080. Great.
>>> >
>>> > From laptop (10.20.10.154), I try the following, using bin/run-example
>>> from
>>> > a locally built version of spark 1.0.0 (so that I have the script
>>> > spark-submit!):
>>> >
>>> > bin/spark-submit --verbose --class org.apache.spark.examples.SparkPi
>>> > --master spark://10.20.10.152:7077
>>> > examples/target/scala-2.10/spark-examples-1.0.0-hadoop1.0.4.jar
>>> >
>>> >
>>> > This fails, with the errors at the bottom of this email.
>>> >
>>> > Am I doing something wrong? How can I submit to  a remote cluster? I
>>> get the
>>> > same problem with bin/spark-submit.
>>> >
>>> >
>>> >  bin/spark-submit --verbose --class org.apache.spark.examples.SparkPi
>>> > --master spark://10.20.10.152:7077
>>> > examples/target/scala-2.10/spark-examples-1.0.0-hadoop1.0.4.jar
>>> > Using properties file: null
>>> > Using properties file: null
>>> > Parsed arguments:
>>> >   master                  spark://10.20.10.152:7077
>>> >   deployMode              null
>>> >   executorMemory          null
>>> >   executorCores           null
>>> >   totalExecutorCores      null
>>> >   propertiesFile          null
>>> >   driverMemory            null
>>> >   driverCores             null
>>> >   driverExtraClassPath    null
>>> >   driverExtraLibraryPath  null
>>> >   driverExtraJavaOptions  null
>>> >   supervise               false
>>> >   queue                   null
>>> >   numExecutors            null
>>> >   files                   null
>>> >   pyFiles                 null
>>> >   archives                null
>>> >   mainClass               org.apache.spark.examples.SparkPi
>>> >   primaryResource
>>> >
>>> file:/Users/aris.vlasakakis/Documents/spark-1.0.0/examples/target/scala-2.10/spark-examples-1.0.0-hadoop1.0.4.jar
>>> >   name                    org.apache.spark.examples.SparkPi
>>> >   childArgs               []
>>> >   jars                    null
>>> >   verbose                 true
>>> >
>>> > Default properties from null:
>>> >
>>> >
>>> >
>>> > Using properties file: null
>>> > Main class:
>>> > org.apache.spark.examples.SparkPi
>>> > Arguments:
>>> >
>>> > System properties:
>>> > SPARK_SUBMIT -> true
>>> > spark.app.name -> org.apache.spark.examples.SparkPi
>>> > spark.jars ->
>>> >
>>> file:/Users/aris.vlasakakis/Documents/spark-1.0.0/examples/target/scala-2.10/spark-examples-1.0.0-hadoop1.0.4.jar
>>> > spark.master -> spark://10.20.10.152:7077
>>> > Classpath elements:
>>> >
>>> file:/Users/aris.vlasakakis/Documents/spark-1.0.0/examples/target/scala-2.10/spark-examples-1.0.0-hadoop1.0.4.jar
>>> >
>>> >
>>> > 14/07/09 16:16:08 INFO SecurityManager: Using Spark's default log4j
>>> profile:
>>> > org/apache/spark/log4j-defaults.properties
>>> > 14/07/09 16:16:08 INFO SecurityManager: Changing view acls to:
>>> > aris.vlasakakis
>>> > 14/07/09 16:16:08 INFO SecurityManager: SecurityManager: authentication
>>> > disabled; ui acls disabled; users with view permissions:
>>> > Set(aris.vlasakakis)
>>> > 14/07/09 16:16:08 INFO Slf4jLogger: Slf4jLogger started
>>> > 14/07/09 16:16:08 INFO Remoting: Starting remoting
>>> > 14/07/09 16:16:08 INFO Remoting: Remoting started; listening on
>>> addresses
>>> > :[akka.tcp://spark@10.20.10.154:50478]
>>> > 14/07/09 16:16:08 INFO Remoting: Remoting now listens on addresses:
>>> > [akka.tcp://spark@10.20.10.154:50478]
>>> > 14/07/09 16:16:08 INFO SparkEnv: Registering MapOutputTracker
>>> > 14/07/09 16:16:08 INFO SparkEnv: Registering BlockManagerMaster
>>> > 14/07/09 16:16:08 INFO DiskBlockManager: Created local directory at
>>> >
>>> /var/folders/ch/yfyhs7px5h90505g4n21n8d5k3svt3/T/spark-local-20140709161608-0531
>>> > 14/07/09 16:16:08 INFO MemoryStore: MemoryStore started with capacity
>>> 5.8
>>> > GB.
>>> > 14/07/09 16:16:08 INFO ConnectionManager: Bound socket to port 50479
>>> with id
>>> > = ConnectionManagerId(10.20.10.154,50479)
>>> > 14/07/09 16:16:08 INFO BlockManagerMaster: Trying to register
>>> BlockManager
>>> > 14/07/09 16:16:08 INFO BlockManagerInfo: Registering block manager
>>> > 10.20.10.154:50479 with 5.8 GB RAM
>>> > 14/07/09 16:16:08 INFO BlockManagerMaster: Registered BlockManager
>>> > 14/07/09 16:16:08 INFO HttpServer: Starting HTTP Server
>>> > 14/07/09 16:16:09 INFO HttpBroadcast: Broadcast server started at
>>> > http://10.20.10.154:50480
>>> > 14/07/09 16:16:09 INFO HttpFileServer: HTTP File server directory is
>>> >
>>> /var/folders/ch/yfyhs7px5h90505g4n21n8d5k3svt3/T/spark-edd787f4-f606-473c-965b-9f3b131cfb43
>>> > 14/07/09 16:16:09 INFO HttpServer: Starting HTTP Server
>>> > 14/07/09 16:16:09 INFO SparkUI: Started SparkUI at
>>> http://10.20.10.154:4040
>>> > 2014-07-09 16:16:09.439 java[37388:d17] Unable to load realm mapping
>>> info
>>> > from SCDynamicStore
>>> > 14/07/09 16:16:09 INFO SparkContext: Added JAR
>>> >
>>> file:/Users/aris.vlasakakis/Documents/spark-1.0.0/examples/target/scala-2.10/spark-examples-1.0.0-hadoop1.0.4.jar
>>> > at http://10.20.10.154:50481/jars/spark-examples-1.0.0-hadoop1.0.4.jar
>>> with
>>> > timestamp 1404947769853
>>> > 14/07/09 16:16:09 INFO AppClient$ClientActor: Connecting to master
>>> > spark://10.20.10.152:7077...
>>> > 14/07/09 16:16:09 INFO SparkContext: Starting job: reduce at
>>> > SparkPi.scala:35
>>> > 14/07/09 16:16:09 INFO DAGScheduler: Got job 0 (reduce at
>>> SparkPi.scala:35)
>>> > with 2 output partitions (allowLocal=false)
>>> > 14/07/09 16:16:09 INFO DAGScheduler: Final stage: Stage 0(reduce at
>>> > SparkPi.scala:35)
>>> > 14/07/09 16:16:10 INFO DAGScheduler: Parents of final stage: List()
>>> > 14/07/09 16:16:10 INFO DAGScheduler: Missing parents: List()
>>> > 14/07/09 16:16:10 INFO DAGScheduler: Submitting Stage 0 (MappedRDD[1]
>>> at map
>>> > at SparkPi.scala:31), which has no missing parents
>>> > 14/07/09 16:16:10 INFO DAGScheduler: Submitting 2 missing tasks from
>>> Stage 0
>>> > (MappedRDD[1] at map at SparkPi.scala:31)
>>> > 14/07/09 16:16:10 INFO TaskSchedulerImpl: Adding task set 0.0 with 2
>>> tasks
>>> > 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Connected to Spark
>>> > cluster with app ID app-20140709161541-0001
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor added:
>>> > app-20140709161541-0001/0 on worker-20140709160420-10.20.10.152-57674
>>> > (10.20.10.152:57674) with 4 cores
>>> > 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Granted executor ID
>>> > app-20140709161541-0001/0 on hostPort 10.20.10.152:57674 with 4
>>> cores, 4.0
>>> > GB RAM
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
>>> > app-20140709161541-0001/0 is now RUNNING
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
>>> > app-20140709161541-0001/0 is now FAILED (class java.io.IOException:
>>> Cannot
>>> > run program
>>> >
>>> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
>>> > directory "."): error=2, No such file or directory)
>>> > 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Executor
>>> > app-20140709161541-0001/0 removed: class java.io.IOException: Cannot
>>> run
>>> > program
>>> >
>>> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
>>> > directory "."): error=2, No such file or directory
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor added:
>>> > app-20140709161541-0001/1 on worker-20140709160420-10.20.10.152-57674
>>> > (10.20.10.152:57674) with 4 cores
>>> > 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Granted executor ID
>>> > app-20140709161541-0001/1 on hostPort 10.20.10.152:57674 with 4
>>> cores, 4.0
>>> > GB RAM
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
>>> > app-20140709161541-0001/1 is now RUNNING
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
>>> > app-20140709161541-0001/1 is now FAILED (class java.io.IOException:
>>> Cannot
>>> > run program
>>> >
>>> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
>>> > directory "."): error=2, No such file or directory)
>>> > 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Executor
>>> > app-20140709161541-0001/1 removed: class java.io.IOException: Cannot
>>> run
>>> > program
>>> >
>>> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
>>> > directory "."): error=2, No such file or directory
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor added:
>>> > app-20140709161541-0001/2 on worker-20140709160420-10.20.10.152-57674
>>> > (10.20.10.152:57674) with 4 cores
>>> > 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Granted executor ID
>>> > app-20140709161541-0001/2 on hostPort 10.20.10.152:57674 with 4
>>> cores, 4.0
>>> > GB RAM
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
>>> > app-20140709161541-0001/2 is now RUNNING
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
>>> > app-20140709161541-0001/2 is now FAILED (class java.io.IOException:
>>> Cannot
>>> > run program
>>> >
>>> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
>>> > directory "."): error=2, No such file or directory)
>>> > 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Executor
>>> > app-20140709161541-0001/2 removed: class java.io.IOException: Cannot
>>> run
>>> > program
>>> >
>>> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
>>> > directory "."): error=2, No such file or directory
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor added:
>>> > app-20140709161541-0001/3 on worker-20140709160420-10.20.10.152-57674
>>> > (10.20.10.152:57674) with 4 cores
>>> > 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Granted executor ID
>>> > app-20140709161541-0001/3 on hostPort 10.20.10.152:57674 with 4
>>> cores, 4.0
>>> > GB RAM
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
>>> > app-20140709161541-0001/3 is now RUNNING
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
>>> > app-20140709161541-0001/3 is now FAILED (class java.io.IOException:
>>> Cannot
>>> > run program
>>> >
>>> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
>>> > directory "."): error=2, No such file or directory)
>>> > 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Executor
>>> > app-20140709161541-0001/3 removed: class java.io.IOException: Cannot
>>> run
>>> > program
>>> >
>>> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
>>> > directory "."): error=2, No such file or directory
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor added:
>>> > app-20140709161541-0001/4 on worker-20140709160420-10.20.10.152-57674
>>> > (10.20.10.152:57674) with 4 cores
>>> > 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Granted executor ID
>>> > app-20140709161541-0001/4 on hostPort 10.20.10.152:57674 with 4
>>> cores, 4.0
>>> > GB RAM
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
>>> > app-20140709161541-0001/4 is now RUNNING
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
>>> > app-20140709161541-0001/4 is now FAILED (class java.io.IOException:
>>> Cannot
>>> > run program
>>> >
>>> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
>>> > directory "."): error=2, No such file or directory)
>>> > 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Executor
>>> > app-20140709161541-0001/4 removed: class java.io.IOException: Cannot
>>> run
>>> > program
>>> >
>>> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
>>> > directory "."): error=2, No such file or directory
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor added:
>>> > app-20140709161541-0001/5 on worker-20140709160420-10.20.10.152-57674
>>> > (10.20.10.152:57674) with 4 cores
>>> > 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Granted executor ID
>>> > app-20140709161541-0001/5 on hostPort 10.20.10.152:57674 with 4
>>> cores, 4.0
>>> > GB RAM
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
>>> > app-20140709161541-0001/5 is now RUNNING
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
>>> > app-20140709161541-0001/5 is now FAILED (class java.io.IOException:
>>> Cannot
>>> > run program
>>> >
>>> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
>>> > directory "."): error=2, No such file or directory)
>>> > 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Executor
>>> > app-20140709161541-0001/5 removed: class java.io.IOException: Cannot
>>> run
>>> > program
>>> >
>>> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
>>> > directory "."): error=2, No such file or directory
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor added:
>>> > app-20140709161541-0001/6 on worker-20140709160420-10.20.10.152-57674
>>> > (10.20.10.152:57674) with 4 cores
>>> > 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Granted executor ID
>>> > app-20140709161541-0001/6 on hostPort 10.20.10.152:57674 with 4
>>> cores, 4.0
>>> > GB RAM
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
>>> > app-20140709161541-0001/6 is now RUNNING
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
>>> > app-20140709161541-0001/6 is now FAILED (class java.io.IOException:
>>> Cannot
>>> > run program
>>> >
>>> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
>>> > directory "."): error=2, No such file or directory)
>>> > 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Executor
>>> > app-20140709161541-0001/6 removed: class java.io.IOException: Cannot
>>> run
>>> > program
>>> >
>>> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
>>> > directory "."): error=2, No such file or directory
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor added:
>>> > app-20140709161541-0001/7 on worker-20140709160420-10.20.10.152-57674
>>> > (10.20.10.152:57674) with 4 cores
>>> > 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Granted executor ID
>>> > app-20140709161541-0001/7 on hostPort 10.20.10.152:57674 with 4
>>> cores, 4.0
>>> > GB RAM
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
>>> > app-20140709161541-0001/7 is now RUNNING
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
>>> > app-20140709161541-0001/7 is now FAILED (class java.io.IOException:
>>> Cannot
>>> > run program
>>> >
>>> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
>>> > directory "."): error=2, No such file or directory)
>>> > 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Executor
>>> > app-20140709161541-0001/7 removed: class java.io.IOException: Cannot
>>> run
>>> > program
>>> >
>>> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
>>> > directory "."): error=2, No such file or directory
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor added:
>>> > app-20140709161541-0001/8 on worker-20140709160420-10.20.10.152-57674
>>> > (10.20.10.152:57674) with 4 cores
>>> > 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Granted executor ID
>>> > app-20140709161541-0001/8 on hostPort 10.20.10.152:57674 with 4
>>> cores, 4.0
>>> > GB RAM
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
>>> > app-20140709161541-0001/8 is now RUNNING
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
>>> > app-20140709161541-0001/8 is now FAILED (class java.io.IOException:
>>> Cannot
>>> > run program
>>> >
>>> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
>>> > directory "."): error=2, No such file or directory)
>>> > 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Executor
>>> > app-20140709161541-0001/8 removed: class java.io.IOException: Cannot
>>> run
>>> > program
>>> >
>>> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
>>> > directory "."): error=2, No such file or directory
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor added:
>>> > app-20140709161541-0001/9 on worker-20140709160420-10.20.10.152-57674
>>> > (10.20.10.152:57674) with 4 cores
>>> > 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Granted executor ID
>>> > app-20140709161541-0001/9 on hostPort 10.20.10.152:57674 with 4
>>> cores, 4.0
>>> > GB RAM
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
>>> > app-20140709161541-0001/9 is now RUNNING
>>> > 14/07/09 16:16:10 INFO AppClient$ClientActor: Executor updated:
>>> > app-20140709161541-0001/9 is now FAILED (class java.io.IOException:
>>> Cannot
>>> > run program
>>> >
>>> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
>>> > directory "."): error=2, No such file or directory)
>>> > 14/07/09 16:16:10 INFO SparkDeploySchedulerBackend: Executor
>>> > app-20140709161541-0001/9 removed: class java.io.IOException: Cannot
>>> run
>>> > program
>>> >
>>> "/Users/aris.vlasakakis/Documents/spark-1.0.0/bin/compute-classpath.sh" (in
>>> > directory "."): error=2, No such file or directory
>>> > 14/07/09 16:16:10 ERROR SparkDeploySchedulerBackend: Application has
>>> been
>>> > killed. Reason: Master removed our application: FAILED
>>> > 14/07/09 16:16:10 INFO TaskSchedulerImpl: Removed TaskSet 0.0, whose
>>> tasks
>>> > have all completed, from pool
>>> > 14/07/09 16:16:10 INFO DAGScheduler: Failed to run reduce at
>>> > SparkPi.scala:35
>>> > Exception in thread "main" 14/07/09 16:16:10 INFO TaskSchedulerImpl:
>>> > Cancelling stage 0
>>> > org.apache.spark.SparkException: Job aborted due to stage failure:
>>> Master
>>> > removed our application: FAILED
>>> >         at
>>> > org.apache.spark.scheduler.DAGScheduler.org
>>> $apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1033)
>>> >         at
>>> >
>>> org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1017)
>>> >         at
>>> >
>>> org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1015)
>>> >         at
>>> >
>>> scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
>>> >         at
>>> > scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
>>> >         at
>>> >
>>> org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1015)
>>> >         at
>>> >
>>> org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:633)
>>> >         at
>>> >
>>> org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:633)
>>> >         at scala.Option.foreach(Option.scala:236)
>>> >         at
>>> >
>>> org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:633)
>>> >         at
>>> >
>>> org.apache.spark.scheduler.DAGSchedulerEventProcessActor$$anonfun$receive$2.applyOrElse(DAGScheduler.scala:1207)
>>> >         at akka.actor.ActorCell.receiveMessage(ActorCell.scala:498)
>>> >         at akka.actor.ActorCell.invoke(ActorCell.scala:456)
>>> >         at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:237)
>>> >         at akka.dispatch.Mailbox.run(Mailbox.scala:219)
>>> >         at
>>> >
>>> akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386)
>>> >         at
>>> > scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>>> >         at
>>> >
>>> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
>>> >         at
>>> >
>>> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
>>> >         at
>>> >
>>> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
>>> > --
>>> > Άρης Βλασακάκης
>>> > Aris Vlasakakis
>>>
>>
>>
>>
>> --
>> Άρης Βλασακάκης
>> Aris Vlasakakis
>>
>
>


-- 
Άρης Βλασακάκης
Aris Vlasakakis

Reply via email to