Thanks Mayur...

I need both...but to start with even application killer will help a lot...

Somehow that command did not work for me....I will try it again from the
spark main folder..


On Sun, Mar 16, 2014 at 1:43 PM, Mayur Rustagi <mayur.rust...@gmail.com>wrote:

> This is meant to kill the whole driver hosted inside the Master (new
> feature as of 0.9.0).
> I assume you are trying to kill a job/task/stage inside the Spark rather
> than the whole application.
> Regards
> Mayur
>
> Mayur Rustagi
> Ph: +1 (760) 203 3257
> http://www.sigmoidanalytics.com
> @mayur_rustagi <https://twitter.com/mayur_rustagi>
>
>
>
> On Sun, Mar 16, 2014 at 4:36 PM, Debasish Das <debasish.da...@gmail.com>wrote:
>
>> From
>> http://spark.incubator.apache.org/docs/latest/spark-standalone.html#launching-applications-inside-the-cluster
>>
>> ./bin/spark-class org.apache.spark.deploy.Client kill <driverId>
>>
>>
>> does not work / has bugs ?
>>
>>
>> On Sun, Mar 16, 2014 at 1:17 PM, Mayur Rustagi 
>> <mayur.rust...@gmail.com>wrote:
>>
>>> Thr is a no good way to kill jobs in Spark yet. The closest is
>>> cancelAllJobs & cancelJobGroup in spark context. I have had bugs using
>>> both. I am trying to test them out, typically you would start a different
>>> thread & call these functions on it when you wish to cancel a job.
>>> Regards
>>> Mayur
>>>
>>> Mayur Rustagi
>>> Ph: +1 (760) 203 3257
>>> http://www.sigmoidanalytics.com
>>>  @mayur_rustagi <https://twitter.com/mayur_rustagi>
>>>
>>>
>>>
>>> On Sun, Mar 16, 2014 at 2:59 PM, Debasish Das 
>>> <debasish.da...@gmail.com>wrote:
>>>
>>>> Are these the right options:
>>>>
>>>> 1. If there is a spark script, just do a ctrl-c from spark-shell and
>>>> the job will be killed property.
>>>>
>>>> 2. For spark application also ctrl c will kill the job property on the
>>>> cluster:
>>>>
>>>> Somehow the ctrl-c option did not work for us...
>>>>
>>>> Similar option works fine for scalding for example but we see lot of
>>>> dead nodes if too many jobs are killed abruptly.
>>>>
>>>> 3. Use the Client script...
>>>>
>>>> /bin/spark-class org.apache.spark.deploy.Client kill spark://
>>>> myspark.com:7077 app-20140316142129-0000
>>>> Runner java
>>>> Classpath
>>>> :/home/debasish/sag_spark/conf:/home/debasish/sag_spark/assembly/target/scala-2.10/spark-assembly-1.0.0-incubating-SNAPSHOT-hadoop2.0.0-mr1-cdh4.5.0.jar
>>>> Java opts  -Djava.library.path= -Xms512m -Xmx512m
>>>> Options -Dspark.cores.max=16
>>>> Sending kill command to spark://myspark.com:7077
>>>> Driver app-20140316142129-0000 has already finished or does not exist
>>>>
>>>> This option also did not kill the job. I can still see the job running
>>>> on spark webui...
>>>>
>>>> Thanks.
>>>> Deb
>>>>
>>>
>>>
>>
>

Reply via email to