Hi Denny, What do you see in the task log?
Thanks! Tim On Tue, Dec 30, 2014 at 8:25 AM, Denny Lee <[email protected]> wrote: > I've been working with Spark 1.2 and Mesos 0.21.0 and while I have set the > spark.executor.uri within spark-env.sh (and directly within bash as well), > the Mesos slaves do not seem to be able to access the spark tgz file via > HTTP or HDFS as per the message below. > > > 14/12/30 15:57:35 INFO SparkILoop: Created spark context.. > Spark context available as sc. > > scala> 14/12/30 15:57:38 INFO CoarseMesosSchedulerBackend: Mesos task 0 is > now TASK_FAILED > 14/12/30 15:57:38 INFO CoarseMesosSchedulerBackend: Mesos task 1 is now > TASK_FAILED > 14/12/30 15:57:39 INFO CoarseMesosSchedulerBackend: Mesos task 2 is now > TASK_FAILED > 14/12/30 15:57:41 INFO CoarseMesosSchedulerBackend: Mesos task 3 is now > TASK_FAILED > 14/12/30 15:57:41 INFO CoarseMesosSchedulerBackend: Blacklisting Mesos > slave value: "20141228-183059-3045950474-5050-2788-S1" > due to too many failures; is Spark installed on it? > > > I've verified that the Mesos slaves can access both the HTTP and HDFS > locations. I'll start digging into the Mesos logs but was wondering if > anyone had run into this issue before. I was able to get this to run > successfully on Spark 1.1 on GCP - my current environment that I'm > experimenting with is Digital Ocean - perhaps this is in play? > > Thanks! > Denny > >

