[
https://issues.apache.org/jira/browse/SPARK-8941?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14623226#comment-14623226
]
Jesper Lundgren edited comment on SPARK-8941 at 7/11/15 4:31 AM:
-----------------------------------------------------------------
Maybe it is better to close this issue and open a new one for the API change
and the documentation issues.
I'll try to review some of the issues we have had with the stand alone cluster
to see if I should create JIRA tickets for some of them.
ex, when using supervised mode in HA cluster, there is not a well documented
procedure to force stop and disable restart of a driver (in case the driver
exits with the wrong exit code). I know of the "kill" command
bin/spark-class org.apache.spark.deploy.Client kill
But in my experience it does not always work.
was (Author: koudelka):
Maybe it is better to close this issue and open a new one for the API change
and the documentation issues.
I'll try to review some of the issues we had with the stand alone cluster and
see if I should create JIRA tickets for some of them.
ex, when using supervised mode in HA cluster, there is not a well documented
procedure to force stop and disable restart of a driver (in case the driver
exits with the wrong exit code). I know of the "kill" command
bin/spark-class org.apache.spark.deploy.Client kill
But in my experience it does not always work.
> Standalone cluster worker does not accept multiple masters on launch
> --------------------------------------------------------------------
>
> Key: SPARK-8941
> URL: https://issues.apache.org/jira/browse/SPARK-8941
> Project: Spark
> Issue Type: Bug
> Components: Deploy, Documentation
> Affects Versions: 1.4.0, 1.4.1
> Reporter: Jesper Lundgren
> Priority: Critical
>
> Before 1.4 it was possible to launch a worker node using a comma separated
> list of master nodes.
> ex:
> sbin/start-slave.sh 1 "spark://localhost:7077,localhost:7078"
> starting org.apache.spark.deploy.worker.Worker, logging to
> /Users/jesper/Downloads/spark-1.4.0-bin-cdh4/sbin/../logs/spark-jesper-org.apache.spark.deploy.worker.Worker-1-Jespers-MacBook-Air.local.out
> failed to launch org.apache.spark.deploy.worker.Worker:
> Default is conf/spark-defaults.conf.
> 15/07/09 12:33:06 INFO Utils: Shutdown hook called
> Spark 1.2 and 1.3.1 accepts multiple masters in this format.
> update: start-slave.sh only expects master lists in 1.4 (no instance number)
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]