You're using the proper Spark definition of "job", but I believe Richard
means "driver".

On Wed, Oct 5, 2016 at 2:17 PM, Mark Hamstra <m...@clearstorydata.com>
wrote:

> Yes and no.  Something that you need to be aware of is that a Job as such
> exists in the DAGScheduler as part of the Application running on the
> Driver.  When talking about stopping or killing a Job, however, what people
> often mean is not just stopping the DAGScheduler from telling the Executors
> to run more Tasks associated with the Job, but also to stop any associated
> Tasks that are already running on Executors.  That is something that Spark
> doesn't try to do by default, and changing that behavior has been an open
> issue for a long time -- cf. SPARK-17064
>
> On Wed, Oct 5, 2016 at 2:07 PM, Michael Gummelt <mgumm...@mesosphere.io>
> wrote:
>
>> If running in client mode, just kill the job.  If running in cluster
>> mode, the Spark Dispatcher exposes an HTTP API for killing jobs.  I don't
>> think this is externally documented, so you might have to check the code to
>> find this endpoint.  If you run in dcos, you can just run "dcos spark kill
>> <id>".
>>
>> You can also find which node is running the driver, ssh in, and kill the
>> process.
>>
>> On Wed, Oct 5, 2016 at 1:55 PM, Richard Siebeling <rsiebel...@gmail.com>
>> wrote:
>>
>>> Hi,
>>>
>>> how can I stop a long running job?
>>>
>>> We're having Spark running in Mesos Coarse-grained mode. Suppose the
>>> user start a long running job, makes a mistake, changes a transformation
>>> and runs the job again. In this case I'd like to cancel the first job and
>>> after that start the second job. It would be a waste of resources to finish
>>> the first job (which could possibly take several hours...)
>>>
>>> How can this be accomplished?
>>> thanks in advance,
>>> Richard
>>>
>>>
>>
>>
>> --
>> Michael Gummelt
>> Software Engineer
>> Mesosphere
>>
>
>


-- 
Michael Gummelt
Software Engineer
Mesosphere

Reply via email to