Also, make sure that the machine you're trying to launch from has Mesos-DNS as its DNS server :)
On Thu, Apr 14, 2016 at 3:33 AM Stefano Bianchi <[email protected]> wrote: > Im correctly running mesos-dns from marathon and it seems to work. > But when i launch: > > http://test.marathon.mesos > > (where test is a funning task on marathon) > > I get: > > curl: (7) Failed connect to test.marathon.mesos:80; Connection refused > > Where am i wrong? > Il 13/apr/2016 17:46, "June Taylor" <[email protected]> ha scritto: > >> We are running pyspark against our cluster in coarse-grained mode by >> specifying the --master mesos://host:5050 flag, which properly creates >> one task on each node. >> >> However, if the driver is shut down, it appears that these executors >> become orphaned_tasks, still consuming resources on the slave, but no >> longer being represented in the master's understanding of available >> resources. >> >> Examining the stdout/stderr shows it exited: >> >> Registered executor on node4 >> Starting task 0 >> sh -c 'cd spark-1*; ./bin/spark-class >> org.apache.spark.executor.CoarseGrainedExecutorBackend --driver-url spark:// >> [email protected]:41563 --executor-id >> aa1337b6-43b0-4236-b445-c8ccbfb60506-S2/0 --hostname node4 --cores 31 >> --app-id aa1337b6-43b0-4236-b445-c8ccbfb60506-0097' >> Forked command at 117620 >> Command exited with status 1 (pid: 117620) >> >> But, these executors are remaining on all the slaves. >> >> What can we do to clear them out? Stopping mesos-slave and removing the >> full work-dir is successful, but also destroys our other tasks. >> >> Thanks, >> June Taylor >> System Administrator, Minnesota Population Center >> University of Minnesota >> >

