Github user pwendell commented on a diff in the pull request:
https://github.com/apache/spark/pull/601#discussion_r12173709
--- Diff: docs/running-on-yarn.md ---
@@ -12,12 +12,14 @@ was added to Spark in version 0.6.0, and improved in
0.7.0 and 0.8.0.
We need a consolidated Spark JAR (which bundles all the required
dependencies) to run Spark jobs on a YARN cluster.
This can be built by setting the Hadoop version and `SPARK_YARN`
environment variable, as follows:
- SPARK_HADOOP_VERSION=2.0.5-alpha SPARK_YARN=true sbt/sbt assembly
+ mvn package -Pyarn -Dyarn.version=2.2.0 -Dhadoop.version=2.2.0
-DskipTests
--- End diff --
You no longer need to set `-Dyarn.version` in this case due a recent
improvement in the maven build.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---