Github user pwendell commented on the pull request:
https://github.com/apache/spark/pull/772#issuecomment-43175575
Hey @ScrapCodes - this looks like a good start. In terms of how to pass
options I think the nicest behavior would be to just make it identical to maven:
```
sbt/sbt -Pyarn -Dspark.hadoop.version=1.2.3 ~compile
```
Since we control the `sbt/sbt` script we could do this (I think) by just
capturing the arguments and then putting them in an environment variable or
passing them somehow so they can be read in the sbt build.
It would also be good, for backwards compatibility, to do some automatic
conversions:
```
if SPARK_YARN is set ==> set 'yarn' profile
if SPARK_GANGLIA_LGPL is set => set 'spark-ganglia-lgpl' profile
if SPARK_HADOOP_VERSION is set ==> spark.hadoop.version property
```
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---