Github user pwendell commented on a diff in the pull request:

    https://github.com/apache/spark/pull/2014#discussion_r16933038
  
    --- Diff: README.md ---
    @@ -66,78 +69,24 @@ Many of the example programs print usage help if no 
params are given.
     
     ## Running Tests
     
    -Testing first requires [building Spark](#building-spark). Once Spark is 
built, tests
    -can be run using:
    -
    -    ./dev/run-tests
    +Please see the guidance on how to 
    +[run all automated 
tests](https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark#ContributingtoSpark-AutomatedTesting)
     
     ## A Note About Hadoop Versions
     
     Spark uses the Hadoop core library to talk to HDFS and other 
Hadoop-supported
     storage systems. Because the protocols have changed in different versions 
of
     Hadoop, you must build Spark against the same version that your cluster 
runs.
    -You can change the version by setting `-Dhadoop.version` when building 
Spark.
    -
    -For Apache Hadoop versions 1.x, Cloudera CDH MRv1, and other Hadoop
    -versions without YARN, use:
    -
    -    # Apache Hadoop 1.2.1
    -    $ sbt/sbt -Dhadoop.version=1.2.1 assembly
    -
    -    # Cloudera CDH 4.2.0 with MapReduce v1
    -    $ sbt/sbt -Dhadoop.version=2.0.0-mr1-cdh4.2.0 assembly
    -
    -For Apache Hadoop 2.2.X, 2.1.X, 2.0.X, 0.23.x, Cloudera CDH MRv2, and 
other Hadoop versions
    -with YARN, also set `-Pyarn`:
    -
    -    # Apache Hadoop 2.0.5-alpha
    -    $ sbt/sbt -Dhadoop.version=2.0.5-alpha -Pyarn assembly
    -
    -    # Cloudera CDH 4.2.0 with MapReduce v2
    -    $ sbt/sbt -Dhadoop.version=2.0.0-cdh4.2.0 -Pyarn assembly
    -
    -    # Apache Hadoop 2.2.X and newer
    -    $ sbt/sbt -Dhadoop.version=2.2.0 -Pyarn assembly
    -
    -When developing a Spark application, specify the Hadoop version by adding 
the
    -"hadoop-client" artifact to your project's dependencies. For example, if 
you're
    -using Hadoop 1.2.1 and build your application using SBT, add this entry to
    -`libraryDependencies`:
    -
    -    "org.apache.hadoop" % "hadoop-client" % "1.2.1"
     
    -If your project is built with Maven, add this to your POM file's 
`<dependencies>` section:
    -
    -    <dependency>
    -      <groupId>org.apache.hadoop</groupId>
    -      <artifactId>hadoop-client</artifactId>
    -      <version>1.2.1</version>
    -    </dependency>
    -
    -
    -## A Note About Thrift JDBC server and CLI for Spark SQL
    -
    -Spark SQL supports Thrift JDBC server and CLI.
    -See sql-programming-guide.md for more information about using the JDBC 
server and CLI.
    -You can use those features by setting `-Phive` when building Spark as 
follows.
    -
    -    $ sbt/sbt -Phive  assembly
    +Please refer to the build documentation at
    +["Specifying the Hadoop 
Version"](http://spark.apache.org/docs/latest/building-with-maven.html#specifying-the-hadoop-version)
    --- End diff --
    
    Yeah maybe we should change the title of the doc to be called "Building 
Spark" and also it might be nice at the bottom to have a quick note that SBT is 
supported for developer builds but Maven is the reference build for all 
packaging. That might be the best way to do this overall.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to