if I understand your issue correctly, there is an open PR to deal w/ this
issue:

https://github.com/apache/incubator-spark/pull/192

though I dunno if its gonna get merged or not ...

On Thu, Jan 23, 2014 at 6:16 PM, kyocum <[email protected]> wrote:

> Trying to run spark-shell from my laptop to a master node in a cluster.  It
> appears that if you've installed spark at loc A on a cluster and in loc B
> on
> your local machine, the current app framework uses loc B as the path for
> start up.  Any way to config my way around this?  TIA
>
> The shell connects to the master and can connect to the workers:
>
> Using Scala version 2.9.3 (Java HotSpot(TM) 64-Bit Server VM, Java
> 1.7.0_09)
> Initializing interpreter...
> Creating SparkContext...
> 14/01/23 18:07:36 INFO Slf4jEventHandler: Slf4jEventHandler started
> 14/01/23 18:07:36 INFO SparkEnv: Registering BlockManagerMaster
> 14/01/23 18:07:36 INFO DiskBlockManager: Created local directory at
>
> /var/folders/d1/9h0hs71d0h112s1tk6vcbx540000gp/T/spark-local-20140123180736-bcf9
> 14/01/23 18:07:36 INFO MemoryStore: MemoryStore started with capacity 647.7
> MB.
> 14/01/23 18:07:36 INFO ConnectionManager: Bound socket to port 59238 with
> id
> = ConnectionManagerId(ilmn-coe-2.net,59238)
> 14/01/23 18:07:36 INFO BlockManagerMaster: Trying to register BlockManager
> 14/01/23 18:07:36 INFO BlockManagerMasterActor$BlockManagerInfo:
> Registering
> block manager ilmn-coe-2.net:59238 with 647.7 MB RAM
> 14/01/23 18:07:36 INFO BlockManagerMaster: Registered BlockManager
> 14/01/23 18:07:36 INFO HttpBroadcast: Broadcast server started at
> http://10.12.195.116:59239
> 14/01/23 18:07:36 INFO SparkEnv: Registering MapOutputTracker
> 14/01/23 18:07:36 INFO HttpFileServer: HTTP File server directory is
>
> /var/folders/d1/9h0hs71d0h112s1tk6vcbx540000gp/T/spark-d683bfd4-333a-4610-be20-a7390aa8d0ba
> 14/01/23 18:07:36 INFO SparkUI: Started Spark Web UI at
> http://ilmn-coe-2.net:4040
> 14/01/23 18:07:36 INFO Client$ClientActor: Connecting to master
> spark://hnn05.net:7077...
> 2014-01-23 18:07:36.475 java[9793:6403] Unable to load realm info from
> SCDynamicStore
> Spark context available as sc.
> 14/01/23 18:07:36 INFO SparkDeploySchedulerBackend: Connected to Spark
> cluster with app ID app-20140123180739-0021
> 14/01/23 18:07:36 INFO Client$ClientActor: Executor added:
> app-20140123180739-0021/0 on worker-20140123164655-192.168.28.232-51898
> (192.168.28.232:7077) with 4 cores
> 14/01/23 18:07:36 INFO SparkDeploySchedulerBackend: Granted executor ID
> app-20140123180739-0021/0 on hostPort 192.168.28.232:7077 with 4 cores,
> 1024.0 MB RAM
> 14/01/23 18:07:36 INFO Client$ClientActor: Executor updated:
> app-20140123180739-0021/0 is now RUNNING
> 14/01/23 18:07:36 INFO Client$ClientActor: Executor updated:
> app-20140123180739-0021/0 is now FAILED (class java.io.IOException: Cannot
> run program
> "/Users/kyocum/spark/spark-0.8.1-incubating/bin/compute-classpath.sh" (in
> directory "."): java.io.IOException: error=2, No such file or directory)
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/executor-failed-cannot-find-compute-classpath-sh-tp859.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>

Reply via email to