Jesper Lundgren created SPARK-8941:
--------------------------------------

             Summary: Standalone cluster worker does not accept multiple 
masters on launch
                 Key: SPARK-8941
                 URL: https://issues.apache.org/jira/browse/SPARK-8941
             Project: Spark
          Issue Type: Bug
          Components: Deploy
    Affects Versions: 1.4.0
            Reporter: Jesper Lundgren


Before 1.4 it was possible to launch a worker node using a comma separated list 
of master nodes. 

ex:
sbin/start-slave.sh 1 "spark://localhost:7077,localhost:7078"
starting org.apache.spark.deploy.worker.Worker, logging to 
/Users/jesper/Downloads/spark-1.4.0-bin-cdh4/sbin/../logs/spark-jesper-org.apache.spark.deploy.worker.Worker-1-Jespers-MacBook-Air.local.out
failed to launch org.apache.spark.deploy.worker.Worker:
                             Default is conf/spark-defaults.conf.
  15/07/09 12:33:06 INFO Utils: Shutdown hook called

I tried in spark 1.2 and spark 1.3.1 those works so seems to an issue 
introduced with 1.4.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to