Repository: spark
Updated Branches:
  refs/heads/master 4cea9da2a -> dc0a4c916


[SPARK-17445][DOCS] Reference an ASF page as the main place to find third-party 
packages

## What changes were proposed in this pull request?

Point references to spark-packages.org to 
https://cwiki.apache.org/confluence/display/SPARK/Third+Party+Projects

This will be accompanied by a parallel change to the spark-website repo, and 
additional changes to this wiki.

## How was this patch tested?

Jenkins tests.

Author: Sean Owen <[email protected]>

Closes #15075 from srowen/SPARK-17445.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/dc0a4c91
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/dc0a4c91
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/dc0a4c91

Branch: refs/heads/master
Commit: dc0a4c916151c795dc41b5714e9d23b4937f4636
Parents: 4cea9da
Author: Sean Owen <[email protected]>
Authored: Wed Sep 14 10:10:16 2016 +0100
Committer: Sean Owen <[email protected]>
Committed: Wed Sep 14 10:10:16 2016 +0100

----------------------------------------------------------------------
 CONTRIBUTING.md                                             | 2 +-
 R/pkg/R/sparkR.R                                            | 4 ++--
 docs/_layouts/global.html                                   | 2 +-
 docs/index.md                                               | 2 +-
 docs/sparkr.md                                              | 3 ++-
 docs/streaming-programming-guide.md                         | 2 +-
 .../apache/spark/sql/execution/datasources/DataSource.scala | 7 ++++---
 .../src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala | 9 +++------
 .../apache/spark/sql/sources/ResolvedDataSourceSuite.scala  | 6 +++---
 9 files changed, 18 insertions(+), 19 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/spark/blob/dc0a4c91/CONTRIBUTING.md
----------------------------------------------------------------------
diff --git a/CONTRIBUTING.md b/CONTRIBUTING.md
index f10d7e2..1a8206a 100644
--- a/CONTRIBUTING.md
+++ b/CONTRIBUTING.md
@@ -6,7 +6,7 @@ It lists steps that are required before creating a PR. In 
particular, consider:
 
 - Is the change important and ready enough to ask the community to spend time 
reviewing?
 - Have you searched for existing, related JIRAs and pull requests?
-- Is this a new feature that can stand alone as a package on 
http://spark-packages.org ?
+- Is this a new feature that can stand alone as a [third party 
project](https://cwiki.apache.org/confluence/display/SPARK/Third+Party+Projects)
 ?
 - Is the change being proposed clearly explained and motivated?
 
 When you contribute code, you affirm that the contribution is your original 
work and that you 

http://git-wip-us.apache.org/repos/asf/spark/blob/dc0a4c91/R/pkg/R/sparkR.R
----------------------------------------------------------------------
diff --git a/R/pkg/R/sparkR.R b/R/pkg/R/sparkR.R
index 15afe01..0601536 100644
--- a/R/pkg/R/sparkR.R
+++ b/R/pkg/R/sparkR.R
@@ -100,7 +100,7 @@ sparkR.stop <- function() {
 #' @param sparkEnvir Named list of environment variables to set on worker nodes
 #' @param sparkExecutorEnv Named list of environment variables to be used when 
launching executors
 #' @param sparkJars Character vector of jar files to pass to the worker nodes
-#' @param sparkPackages Character vector of packages from spark-packages.org
+#' @param sparkPackages Character vector of package coordinates
 #' @seealso \link{sparkR.session}
 #' @rdname sparkR.init-deprecated
 #' @export
@@ -327,7 +327,7 @@ sparkRHive.init <- function(jsc = NULL) {
 #' @param sparkHome Spark Home directory.
 #' @param sparkConfig named list of Spark configuration to set on worker nodes.
 #' @param sparkJars character vector of jar files to pass to the worker nodes.
-#' @param sparkPackages character vector of packages from spark-packages.org
+#' @param sparkPackages character vector of package coordinates
 #' @param enableHiveSupport enable support for Hive, fallback if not built 
with Hive support; once
 #'        set, this cannot be turned off on an existing session
 #' @param ... named Spark properties passed to the method.

http://git-wip-us.apache.org/repos/asf/spark/blob/dc0a4c91/docs/_layouts/global.html
----------------------------------------------------------------------
diff --git a/docs/_layouts/global.html b/docs/_layouts/global.html
index d3bf082..ad5b5c9 100755
--- a/docs/_layouts/global.html
+++ b/docs/_layouts/global.html
@@ -114,7 +114,7 @@
                                 <li class="divider"></li>
                                 <li><a href="building-spark.html">Building 
Spark</a></li>
                                 <li><a 
href="https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark";>Contributing
 to Spark</a></li>
-                                <li><a 
href="https://cwiki.apache.org/confluence/display/SPARK/Supplemental+Spark+Projects";>Supplemental
 Projects</a></li>
+                                <li><a 
href="https://cwiki.apache.org/confluence/display/SPARK/Third+Party+Projects";>Third
 Party Projects</a></li>
                             </ul>
                         </li>
                     </ul>

http://git-wip-us.apache.org/repos/asf/spark/blob/dc0a4c91/docs/index.md
----------------------------------------------------------------------
diff --git a/docs/index.md b/docs/index.md
index 0cb8803..a7a92f6 100644
--- a/docs/index.md
+++ b/docs/index.md
@@ -120,7 +120,7 @@ options for deployment:
   * [OpenStack Swift](storage-openstack-swift.html)
 * [Building Spark](building-spark.html): build Spark using the Maven system
 * [Contributing to 
Spark](https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark)
-* [Supplemental 
Projects](https://cwiki.apache.org/confluence/display/SPARK/Supplemental+Spark+Projects):
 related third party Spark projects
+* [Third Party 
Projects](https://cwiki.apache.org/confluence/display/SPARK/Third+Party+Projects):
 related third party Spark projects
 
 **External Resources:**
 

http://git-wip-us.apache.org/repos/asf/spark/blob/dc0a4c91/docs/sparkr.md
----------------------------------------------------------------------
diff --git a/docs/sparkr.md b/docs/sparkr.md
index 4bbc362..b881119 100644
--- a/docs/sparkr.md
+++ b/docs/sparkr.md
@@ -110,7 +110,8 @@ head(df)
 
 SparkR supports operating on a variety of data sources through the 
`SparkDataFrame` interface. This section describes the general methods for 
loading and saving data using Data Sources. You can check the Spark SQL 
programming guide for more [specific 
options](sql-programming-guide.html#manually-specifying-options) that are 
available for the built-in data sources.
 
-The general method for creating SparkDataFrames from data sources is 
`read.df`. This method takes in the path for the file to load and the type of 
data source, and the currently active SparkSession will be used automatically. 
SparkR supports reading JSON, CSV and Parquet files natively and through [Spark 
Packages](http://spark-packages.org/) you can find data source connectors for 
popular file formats like 
[Avro](http://spark-packages.org/package/databricks/spark-avro). These packages 
can either be added by
+The general method for creating SparkDataFrames from data sources is 
`read.df`. This method takes in the path for the file to load and the type of 
data source, and the currently active SparkSession will be used automatically.
+SparkR supports reading JSON, CSV and Parquet files natively, and through 
packages available from sources like [Third Party 
Projects](https://cwiki.apache.org/confluence/display/SPARK/Third+Party+Projects),
 you can find data source connectors for popular file formats like Avro. These 
packages can either be added by
 specifying `--packages` with `spark-submit` or `sparkR` commands, or if 
initializing SparkSession with `sparkPackages` parameter when in an interactive 
R shell or from RStudio.
 
 <div data-lang="r" markdown="1">

http://git-wip-us.apache.org/repos/asf/spark/blob/dc0a4c91/docs/streaming-programming-guide.md
----------------------------------------------------------------------
diff --git a/docs/streaming-programming-guide.md 
b/docs/streaming-programming-guide.md
index 5392b4a..43f1cf3 100644
--- a/docs/streaming-programming-guide.md
+++ b/docs/streaming-programming-guide.md
@@ -2382,7 +2382,7 @@ additional effort may be necessary to achieve 
exactly-once semantics. There are
     - [Kafka Integration Guide](streaming-kafka-integration.html)
     - [Kinesis Integration Guide](streaming-kinesis-integration.html)
     - [Custom Receiver Guide](streaming-custom-receivers.html)
-* Third-party DStream data sources can be found in [Spark 
Packages](https://spark-packages.org/)
+* Third-party DStream data sources can be found in [Third Party 
Projects](https://cwiki.apache.org/confluence/display/SPARK/Third+Party+Projects)
 * API documentation
   - Scala docs
     * 
[StreamingContext](api/scala/index.html#org.apache.spark.streaming.StreamingContext)
 and

http://git-wip-us.apache.org/repos/asf/spark/blob/dc0a4c91/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSource.scala
----------------------------------------------------------------------
diff --git 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSource.scala
 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSource.scala
index 71807b7..825c013 100644
--- 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSource.scala
+++ 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSource.scala
@@ -142,12 +142,13 @@ case class DataSource(
                 } else if (provider.toLowerCase == "avro" ||
                   provider == "com.databricks.spark.avro") {
                   throw new AnalysisException(
-                    s"Failed to find data source: ${provider.toLowerCase}. 
Please use Spark " +
-                      "package 
http://spark-packages.org/package/databricks/spark-avro";)
+                    s"Failed to find data source: ${provider.toLowerCase}. 
Please find an Avro " +
+                      "package at " +
+                      
"https://cwiki.apache.org/confluence/display/SPARK/Third+Party+Projects";)
                 } else {
                   throw new ClassNotFoundException(
                     s"Failed to find data source: $provider. Please find 
packages at " +
-                      "http://spark-packages.org";,
+                      
"https://cwiki.apache.org/confluence/display/SPARK/Third+Party+Projects";,
                     error)
                 }
             }

http://git-wip-us.apache.org/repos/asf/spark/blob/dc0a4c91/sql/core/src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala
----------------------------------------------------------------------
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala 
b/sql/core/src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala
index a2164f9..3cc3b31 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala
@@ -1645,21 +1645,18 @@ class SQLQuerySuite extends QueryTest with 
SharedSQLContext {
     e = intercept[AnalysisException] {
       sql(s"select id from `com.databricks.spark.avro`.`file_path`")
     }
-    assert(e.message.contains("Failed to find data source: 
com.databricks.spark.avro. " +
-      "Please use Spark package 
http://spark-packages.org/package/databricks/spark-avro";))
+    assert(e.message.contains("Failed to find data source: 
com.databricks.spark.avro."))
 
     // data source type is case insensitive
     e = intercept[AnalysisException] {
       sql(s"select id from Avro.`file_path`")
     }
-    assert(e.message.contains("Failed to find data source: avro. Please use 
Spark package " +
-      "http://spark-packages.org/package/databricks/spark-avro";))
+    assert(e.message.contains("Failed to find data source: avro."))
 
     e = intercept[AnalysisException] {
       sql(s"select id from avro.`file_path`")
     }
-    assert(e.message.contains("Failed to find data source: avro. Please use 
Spark package " +
-      "http://spark-packages.org/package/databricks/spark-avro";))
+    assert(e.message.contains("Failed to find data source: avro."))
 
     e = intercept[AnalysisException] {
       sql(s"select id from 
`org.apache.spark.sql.sources.HadoopFsRelationProvider`.`file_path`")

http://git-wip-us.apache.org/repos/asf/spark/blob/dc0a4c91/sql/core/src/test/scala/org/apache/spark/sql/sources/ResolvedDataSourceSuite.scala
----------------------------------------------------------------------
diff --git 
a/sql/core/src/test/scala/org/apache/spark/sql/sources/ResolvedDataSourceSuite.scala
 
b/sql/core/src/test/scala/org/apache/spark/sql/sources/ResolvedDataSourceSuite.scala
index 5ea1f32..76ffb94 100644
--- 
a/sql/core/src/test/scala/org/apache/spark/sql/sources/ResolvedDataSourceSuite.scala
+++ 
b/sql/core/src/test/scala/org/apache/spark/sql/sources/ResolvedDataSourceSuite.scala
@@ -74,16 +74,16 @@ class ResolvedDataSourceSuite extends SparkFunSuite {
     val error1 = intercept[AnalysisException] {
       getProvidingClass("avro")
     }
-    assert(error1.getMessage.contains("spark-packages"))
+    assert(error1.getMessage.contains("Failed to find data source: avro."))
 
     val error2 = intercept[AnalysisException] {
       getProvidingClass("com.databricks.spark.avro")
     }
-    assert(error2.getMessage.contains("spark-packages"))
+    assert(error2.getMessage.contains("Failed to find data source: 
com.databricks.spark.avro."))
 
     val error3 = intercept[ClassNotFoundException] {
       getProvidingClass("asfdwefasdfasdf")
     }
-    assert(error3.getMessage.contains("spark-packages"))
+    assert(error3.getMessage.contains("Failed to find data source: 
asfdwefasdfasdf."))
   }
 }


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to