[ https://issues.apache.org/jira/browse/SPARK-693?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Albert Chu updated SPARK-693: ----------------------------- Attachment: SPARK-693.patch We required this support in our environment. Attached is my patch to implement this for Spark 1.0.0. Git pull request will be sent too. > Let deploy scripts set alternate conf, work directories > ------------------------------------------------------- > > Key: SPARK-693 > URL: https://issues.apache.org/jira/browse/SPARK-693 > Project: Spark > Issue Type: Improvement > Affects Versions: 0.6.2 > Reporter: David Chiang > Priority: Minor > Attachments: SPARK-693.patch > > > Currently SPARK_CONF_DIR is overridden in spark-config.sh, and > start-slaves.sh doesn't allow the user to pass a -d option in to set the work > directory. Allowing this is a small change and makes it possible to have > multiple clusters running at once. -- This message was sent by Atlassian JIRA (v6.2#6252)