[ 
https://issues.apache.org/jira/browse/SPARK-10410?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bryce Ageno updated SPARK-10410:
--------------------------------
    Shepherd:   (was: Bryce Ageno)

> spark 1.4.1 kill command does not work with streaming job.
> ----------------------------------------------------------
>
>                 Key: SPARK-10410
>                 URL: https://issues.apache.org/jira/browse/SPARK-10410
>             Project: Spark
>          Issue Type: Bug
>          Components: Deploy
>    Affects Versions: 1.4.1
>            Reporter: Bryce Ageno
>
> Our team recently upgraded a cluster to 1.4.1 from 1.3.1 and we discovered 
> that when you run the kill command for a driver (/usr/spark/bin/spark-submit 
> --master spark://$SPARK_MASTER_IP:6066 --kill $SPARK_DRIVER) it is not 
> removing the driver off of the sparkUI.  It is a streaming job and the kill 
> command "ends" the job but it does not free up the resources or remove it 
> from the spark master.
> We are running in cluster mode.  We have also noticed that with 1.4.1 
> multiple spark-submits all of the drivers ends up on a single worker.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to