[spark] Git Push Summary

2018-06-27 Thread tgraves
Repository: spark
Updated Tags:  refs/tags/v2.2.2-rc2 [created] fc28ba3db

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[2/2] spark git commit: Preparing development version 2.2.3-SNAPSHOT

2018-06-27 Thread tgraves
Preparing development version 2.2.3-SNAPSHOT


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/47958270
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/47958270
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/47958270

Branch: refs/heads/branch-2.2
Commit: 47958270f51d08a076e20e299bb2d86fa1927d26
Parents: fc28ba3
Author: Thomas Graves 
Authored: Wed Jun 27 13:55:20 2018 +
Committer: Thomas Graves 
Committed: Wed Jun 27 13:55:20 2018 +

--
 R/pkg/DESCRIPTION | 2 +-
 assembly/pom.xml  | 2 +-
 common/network-common/pom.xml | 2 +-
 common/network-shuffle/pom.xml| 2 +-
 common/network-yarn/pom.xml   | 2 +-
 common/sketch/pom.xml | 2 +-
 common/tags/pom.xml   | 2 +-
 common/unsafe/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 docs/_config.yml  | 4 ++--
 examples/pom.xml  | 2 +-
 external/docker-integration-tests/pom.xml | 2 +-
 external/flume-assembly/pom.xml   | 2 +-
 external/flume-sink/pom.xml   | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka-0-10-assembly/pom.xml  | 2 +-
 external/kafka-0-10-sql/pom.xml   | 2 +-
 external/kafka-0-10/pom.xml   | 2 +-
 external/kafka-0-8-assembly/pom.xml   | 2 +-
 external/kafka-0-8/pom.xml| 2 +-
 external/kinesis-asl-assembly/pom.xml | 2 +-
 external/kinesis-asl/pom.xml  | 2 +-
 external/spark-ganglia-lgpl/pom.xml   | 2 +-
 graphx/pom.xml| 2 +-
 launcher/pom.xml  | 2 +-
 mllib-local/pom.xml   | 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 2 +-
 python/pyspark/version.py | 2 +-
 repl/pom.xml  | 2 +-
 resource-managers/mesos/pom.xml   | 2 +-
 resource-managers/yarn/pom.xml| 2 +-
 sql/catalyst/pom.xml  | 2 +-
 sql/core/pom.xml  | 2 +-
 sql/hive-thriftserver/pom.xml | 2 +-
 sql/hive/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 38 files changed, 39 insertions(+), 39 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/47958270/R/pkg/DESCRIPTION
--
diff --git a/R/pkg/DESCRIPTION b/R/pkg/DESCRIPTION
index 380b3ef..ad72330 100644
--- a/R/pkg/DESCRIPTION
+++ b/R/pkg/DESCRIPTION
@@ -1,6 +1,6 @@
 Package: SparkR
 Type: Package
-Version: 2.2.2
+Version: 2.2.3
 Title: R Frontend for Apache Spark
 Description: Provides an R Frontend for Apache Spark.
 Authors@R: c(person("Shivaram", "Venkataraman", role = c("aut", "cre"),

http://git-wip-us.apache.org/repos/asf/spark/blob/47958270/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index 775d9f8..f9ec6e7 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   
 org.apache.spark
 spark-parent_2.11
-2.2.2
+2.2.3-SNAPSHOT
 ../pom.xml
   
 

http://git-wip-us.apache.org/repos/asf/spark/blob/47958270/common/network-common/pom.xml
--
diff --git a/common/network-common/pom.xml b/common/network-common/pom.xml
index 017f0f1..55d29d5 100644
--- a/common/network-common/pom.xml
+++ b/common/network-common/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.spark
 spark-parent_2.11
-2.2.2
+2.2.3-SNAPSHOT
 ../../pom.xml
   
 

http://git-wip-us.apache.org/repos/asf/spark/blob/47958270/common/network-shuffle/pom.xml
--
diff --git a/common/network-shuffle/pom.xml b/common/network-shuffle/pom.xml
index 17d5216..6d84766 100644
--- a/common/network-shuffle/pom.xml
+++ b/common/network-shuffle/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.spark
 spark-parent_2.11
-2.2.2
+2.2.3-SNAPSHOT
 ../../pom.xml
   
 

http://git-wip-us.apache.org/repos/asf/spark/blob/47958270/common/network-yarn/pom.xml
--
diff --git a/common/network-yarn/pom.xml b/common/network-yarn/pom.xml
index 8d03c86..6228be6 100644
--- a/common/network-yarn/pom.xml
+++ b/common/network-yarn/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.spark
 spark-parent_2.11
-2.2.2
+2.2.3-SNAPSHOT
 ../../pom.xml
   
 


[1/2] spark git commit: Preparing Spark release v2.2.2-rc2

2018-06-27 Thread tgraves
Repository: spark
Updated Branches:
  refs/heads/branch-2.2 72575d0bb -> 47958270f


Preparing Spark release v2.2.2-rc2


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/fc28ba3d
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/fc28ba3d
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/fc28ba3d

Branch: refs/heads/branch-2.2
Commit: fc28ba3db7185e84b6dbd02ad8ef8f1d06b9e3c6
Parents: 72575d0
Author: Thomas Graves 
Authored: Wed Jun 27 13:55:11 2018 +
Committer: Thomas Graves 
Committed: Wed Jun 27 13:55:11 2018 +

--
 R/pkg/DESCRIPTION | 2 +-
 assembly/pom.xml  | 2 +-
 common/network-common/pom.xml | 2 +-
 common/network-shuffle/pom.xml| 2 +-
 common/network-yarn/pom.xml   | 2 +-
 common/sketch/pom.xml | 2 +-
 common/tags/pom.xml   | 2 +-
 common/unsafe/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 docs/_config.yml  | 4 ++--
 examples/pom.xml  | 2 +-
 external/docker-integration-tests/pom.xml | 2 +-
 external/flume-assembly/pom.xml   | 2 +-
 external/flume-sink/pom.xml   | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka-0-10-assembly/pom.xml  | 2 +-
 external/kafka-0-10-sql/pom.xml   | 2 +-
 external/kafka-0-10/pom.xml   | 2 +-
 external/kafka-0-8-assembly/pom.xml   | 2 +-
 external/kafka-0-8/pom.xml| 2 +-
 external/kinesis-asl-assembly/pom.xml | 2 +-
 external/kinesis-asl/pom.xml  | 2 +-
 external/spark-ganglia-lgpl/pom.xml   | 2 +-
 graphx/pom.xml| 2 +-
 launcher/pom.xml  | 2 +-
 mllib-local/pom.xml   | 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 2 +-
 python/pyspark/version.py | 2 +-
 repl/pom.xml  | 2 +-
 resource-managers/mesos/pom.xml   | 2 +-
 resource-managers/yarn/pom.xml| 2 +-
 sql/catalyst/pom.xml  | 2 +-
 sql/core/pom.xml  | 2 +-
 sql/hive-thriftserver/pom.xml | 2 +-
 sql/hive/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 38 files changed, 39 insertions(+), 39 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/fc28ba3d/R/pkg/DESCRIPTION
--
diff --git a/R/pkg/DESCRIPTION b/R/pkg/DESCRIPTION
index ad72330..380b3ef 100644
--- a/R/pkg/DESCRIPTION
+++ b/R/pkg/DESCRIPTION
@@ -1,6 +1,6 @@
 Package: SparkR
 Type: Package
-Version: 2.2.3
+Version: 2.2.2
 Title: R Frontend for Apache Spark
 Description: Provides an R Frontend for Apache Spark.
 Authors@R: c(person("Shivaram", "Venkataraman", role = c("aut", "cre"),

http://git-wip-us.apache.org/repos/asf/spark/blob/fc28ba3d/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index f9ec6e7..775d9f8 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   
 org.apache.spark
 spark-parent_2.11
-2.2.3-SNAPSHOT
+2.2.2
 ../pom.xml
   
 

http://git-wip-us.apache.org/repos/asf/spark/blob/fc28ba3d/common/network-common/pom.xml
--
diff --git a/common/network-common/pom.xml b/common/network-common/pom.xml
index 55d29d5..017f0f1 100644
--- a/common/network-common/pom.xml
+++ b/common/network-common/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.spark
 spark-parent_2.11
-2.2.3-SNAPSHOT
+2.2.2
 ../../pom.xml
   
 

http://git-wip-us.apache.org/repos/asf/spark/blob/fc28ba3d/common/network-shuffle/pom.xml
--
diff --git a/common/network-shuffle/pom.xml b/common/network-shuffle/pom.xml
index 6d84766..17d5216 100644
--- a/common/network-shuffle/pom.xml
+++ b/common/network-shuffle/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.spark
 spark-parent_2.11
-2.2.3-SNAPSHOT
+2.2.2
 ../../pom.xml
   
 

http://git-wip-us.apache.org/repos/asf/spark/blob/fc28ba3d/common/network-yarn/pom.xml
--
diff --git a/common/network-yarn/pom.xml b/common/network-yarn/pom.xml
index 6228be6..8d03c86 100644
--- a/common/network-yarn/pom.xml
+++ b/common/network-yarn/pom.xml
@@ -22,7 +22,7 @@
   
 org.apache.spark
 spark-parent_2.11
-2.2.3-SNAPSHOT
+2.2.2
 ../../pom.xml
   

svn commit: r27768 - /dev/spark/v2.2.2-rc2-bin/

2018-06-27 Thread tgraves
Author: tgraves
Date: Wed Jun 27 17:16:26 2018
New Revision: 27768

Log:
Apache Spark v2.2.2-rc2

Added:
dev/spark/v2.2.2-rc2-bin/
dev/spark/v2.2.2-rc2-bin/SparkR_2.2.2.tar.gz   (with props)
dev/spark/v2.2.2-rc2-bin/SparkR_2.2.2.tar.gz.asc
dev/spark/v2.2.2-rc2-bin/SparkR_2.2.2.tar.gz.sha512
dev/spark/v2.2.2-rc2-bin/pyspark-2.2.2.tar.gz   (with props)
dev/spark/v2.2.2-rc2-bin/pyspark-2.2.2.tar.gz.asc
dev/spark/v2.2.2-rc2-bin/pyspark-2.2.2.tar.gz.sha512
dev/spark/v2.2.2-rc2-bin/spark-2.2.2-bin-hadoop2.6.tgz   (with props)
dev/spark/v2.2.2-rc2-bin/spark-2.2.2-bin-hadoop2.6.tgz.asc
dev/spark/v2.2.2-rc2-bin/spark-2.2.2-bin-hadoop2.6.tgz.sha512
dev/spark/v2.2.2-rc2-bin/spark-2.2.2-bin-hadoop2.7.tgz   (with props)
dev/spark/v2.2.2-rc2-bin/spark-2.2.2-bin-hadoop2.7.tgz.asc
dev/spark/v2.2.2-rc2-bin/spark-2.2.2-bin-hadoop2.7.tgz.sha512
dev/spark/v2.2.2-rc2-bin/spark-2.2.2-bin-without-hadoop.tgz   (with props)
dev/spark/v2.2.2-rc2-bin/spark-2.2.2-bin-without-hadoop.tgz.asc
dev/spark/v2.2.2-rc2-bin/spark-2.2.2-bin-without-hadoop.tgz.sha512
dev/spark/v2.2.2-rc2-bin/spark-2.2.2.tgz   (with props)
dev/spark/v2.2.2-rc2-bin/spark-2.2.2.tgz.asc
dev/spark/v2.2.2-rc2-bin/spark-2.2.2.tgz.sha512

Added: dev/spark/v2.2.2-rc2-bin/SparkR_2.2.2.tar.gz
==
Binary file - no diff available.

Propchange: dev/spark/v2.2.2-rc2-bin/SparkR_2.2.2.tar.gz
--
svn:mime-type = application/octet-stream

Added: dev/spark/v2.2.2-rc2-bin/SparkR_2.2.2.tar.gz.asc
==
--- dev/spark/v2.2.2-rc2-bin/SparkR_2.2.2.tar.gz.asc (added)
+++ dev/spark/v2.2.2-rc2-bin/SparkR_2.2.2.tar.gz.asc Wed Jun 27 17:16:26 2018
@@ -0,0 +1,17 @@
+-BEGIN PGP SIGNATURE-
+Version: GnuPG v1
+
+iQIcBAABAgAGBQJbM576AAoJEN5PvM2B5sdqBcQP/390/0Cyo/E2UYUs361AkaR2
+jCI+w0/+msaX4wd0TExllr2BG2Cwz2XHxL0nIOtd3h8MKAw/iBwO2eVd2imHa27i
+iZkFXrEBT8Gsompq6Nl/j4Cht/I5ay5qEaG4QHSTnJbLgtQIVVw5Rovw/KLVVm0/
+9r+YYFjahsFTw3JTp62LFUXbF81aTgJWwSw05b5TvAwfws4TZoZmES8yz/YKrRWe
+UBZFShAvehVEkSqP7qYm8Dd/fQnfyxaTTUfWXcJQX7fHNrQGXdPOECnlC+TMrmn2
+AqQaKTjHU3p8Wvv78tVLmal9By0NWcPqiSrQ2TlJFEygKseqrDNTw47p06xZHuuz
+bYmJu5k7JcQJyG7ThpQgnIFga0TjM4Ymm+9pN5OKsPyNIGhFmWHVp/zAkdvMHVUg
+D+5KIDJZPtOK3iL8hAjvgCM7W+Q0GcSOCtigNc7S3CaP+0xcI7RfOAy/Ns4wHPwD
+hJZA6u8lLntVwYfmamW4SCB6WHDvdHXl8H2QQ+mVqRpGdnK2fYBUpvcs0dijUAs9
+ZWIvDPuvCE+SnW3zShY0CGxwAhYFEO6vXVExQWsdInTlG7qXDtJ2S+9shm8PohYy
+CdDijjBW5sVR9OnLEFoMtxiOGXaGNo04kK4U8lzkq5reFtQc8I9fr+TUSvki9WqM
+Mf1RUFRr959InxcvafD8
+=WMWJ
+-END PGP SIGNATURE-

Added: dev/spark/v2.2.2-rc2-bin/SparkR_2.2.2.tar.gz.sha512
==
--- dev/spark/v2.2.2-rc2-bin/SparkR_2.2.2.tar.gz.sha512 (added)
+++ dev/spark/v2.2.2-rc2-bin/SparkR_2.2.2.tar.gz.sha512 Wed Jun 27 17:16:26 2018
@@ -0,0 +1,3 @@
+SparkR_2.2.2.tar.gz: FD8AB48E 7379B62C 85744DD8 D9CC0C85 C76EA843 460C3FEC
+ A65DBA30 A47B427C A388A8E2 99CF593A 0CE70FA6 F66BA6BF
+ 2280DF30 F3C4FA63 E451C4BD D49D124D

Added: dev/spark/v2.2.2-rc2-bin/pyspark-2.2.2.tar.gz
==
Binary file - no diff available.

Propchange: dev/spark/v2.2.2-rc2-bin/pyspark-2.2.2.tar.gz
--
svn:mime-type = application/octet-stream

Added: dev/spark/v2.2.2-rc2-bin/pyspark-2.2.2.tar.gz.asc
==
--- dev/spark/v2.2.2-rc2-bin/pyspark-2.2.2.tar.gz.asc (added)
+++ dev/spark/v2.2.2-rc2-bin/pyspark-2.2.2.tar.gz.asc Wed Jun 27 17:16:26 2018
@@ -0,0 +1,17 @@
+-BEGIN PGP SIGNATURE-
+Version: GnuPG v1
+
+iQIcBAABAgAGBQJbM6QPAAoJEN5PvM2B5sdqnhsP/0pwzvbnrycc6EgAr0/Qnzeu
+IDc56d+NIOFLX3KcqF+ggUpll5tj6+Llq8Qs+JNz4ufBT5FNd6/0sot9PfU86AzN
+A5+mlLAyOUTh1WSxY54Mo3Mc8nrLKQHw7QQFMxCfay9oTBrlURhyNp5bEwDfOurf
+LBlX36sTv3H60N9/RZQWt2OF9YnGHntLIe2Fsqk7GtnUMmkdAqdGGtq1w4cjVtci
+5/8xRqRwhVoaGcLLWzsdYMuOqnhBXq6GKtCgF4PLy25vAPVpVQmwcwrx2GVSY6Rr
+M5RNGw8cq2AdPHGojbygUtwg7h2N8BFqwM5EgYKIaKyWtw0N7RKRJ4YWl7QJorAe
+Q4xfFGCwmzmwqMpl3/ZontQMw56FedWLVuETmBoE30/lqPZKsJCU9eS0/8pdBo4t
+7UpQgao25kaBqy4RS2A48343oY9y4vreUQxxjxilV/RxCYI+yNalU3+oreHpaDAk
+k9SGprdmt5zXAkEd4wNmrLWHVB/IlyzjH8s7gKBlHA9GLpxnFwlMInZJjZkN7Jxt
+6A3SDOzhuuEkPHjb5rT1DyCFv7w5rN6TZ0iFxAh9sgR5MbMBaQFRr8krPiYVcG/2
+k9d8FMrDgAfttRVMe6fOg721bnAZaDwdCcV07DMn5/lR60bjhhHMrsD7JWy89RI9
+qrS3ckY1stERB3toI/Li
+=yKed
+-END PGP SIGNATURE-

Added: dev/spark/v2.2.2-rc2-bin/pyspark-2.2.2.tar.gz.sha512
==
--- dev/spark/v2.2.2-rc2-bin/pyspark-2.2.2.tar.gz.sha512 (added)
+++ dev/spark/v2.2.2-rc2-bin/pyspark-2.2.2.tar.gz.sha512 

spark git commit: [SPARK-24215][PYSPARK][FOLLOW UP] Implement eager evaluation for DataFrame APIs in PySpark

2018-06-27 Thread lixiao
Repository: spark
Updated Branches:
  refs/heads/master a1a64e358 -> 6a0b77a55


[SPARK-24215][PYSPARK][FOLLOW UP] Implement eager evaluation for DataFrame APIs 
in PySpark

## What changes were proposed in this pull request?

Address comments in #21370 and add more test.

## How was this patch tested?

Enhance test in pyspark/sql/test.py and DataFrameSuite

Author: Yuanjian Li 

Closes #21553 from xuanyuanking/SPARK-24215-follow.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/6a0b77a5
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/6a0b77a5
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/6a0b77a5

Branch: refs/heads/master
Commit: 6a0b77a55d53e74ac0a0892556c3a7a933474948
Parents: a1a64e3
Author: Yuanjian Li 
Authored: Wed Jun 27 10:43:06 2018 -0700
Committer: Xiao Li 
Committed: Wed Jun 27 10:43:06 2018 -0700

--
 docs/configuration.md   | 27 -
 python/pyspark/sql/dataframe.py |  3 +-
 python/pyspark/sql/tests.py | 46 ++-
 .../org/apache/spark/sql/internal/SQLConf.scala | 23 
 .../scala/org/apache/spark/sql/Dataset.scala| 11 ++--
 .../org/apache/spark/sql/DataFrameSuite.scala   | 59 
 6 files changed, 131 insertions(+), 38 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/6a0b77a5/docs/configuration.md
--
diff --git a/docs/configuration.md b/docs/configuration.md
index 6aa7878..0c7c447 100644
--- a/docs/configuration.md
+++ b/docs/configuration.md
@@ -457,33 +457,6 @@ Apart from these, the following properties are also 
available, and may be useful
   
 
 
-  spark.sql.repl.eagerEval.enabled
-  false
-  
-Enable eager evaluation or not. If true and the REPL you are using 
supports eager evaluation,
-Dataset will be ran automatically. The HTML table which generated by 
_repl_html_
-called by notebooks like Jupyter will feedback the queries user have 
defined. For plain Python
-REPL, the output will be shown like dataframe.show()
-(see https://issues.apache.org/jira/browse/SPARK-24215;>SPARK-24215 for 
more details).
-  
-
-
-  spark.sql.repl.eagerEval.maxNumRows
-  20
-  
-Default number of rows in eager evaluation output HTML table generated by 
_repr_html_ or plain text,
-this only take effect when spark.sql.repl.eagerEval.enabled 
is set to true.
-  
-
-
-  spark.sql.repl.eagerEval.truncate
-  20
-  
-Default number of truncate in eager evaluation output HTML table generated 
by _repr_html_ or
-plain text, this only take effect when 
spark.sql.repl.eagerEval.enabled set to true.
-  
-
-
   spark.files
   
   

http://git-wip-us.apache.org/repos/asf/spark/blob/6a0b77a5/python/pyspark/sql/dataframe.py
--
diff --git a/python/pyspark/sql/dataframe.py b/python/pyspark/sql/dataframe.py
index 1e6a1ac..cb3fe44 100644
--- a/python/pyspark/sql/dataframe.py
+++ b/python/pyspark/sql/dataframe.py
@@ -393,9 +393,8 @@ class DataFrame(object):
 self._support_repr_html = True
 if self._eager_eval:
 max_num_rows = max(self._max_num_rows, 0)
-vertical = False
 sock_info = self._jdf.getRowsToPython(
-max_num_rows, self._truncate, vertical)
+max_num_rows, self._truncate)
 rows = list(_load_from_socket(sock_info, 
BatchedSerializer(PickleSerializer(
 head = rows[0]
 row_data = rows[1:]

http://git-wip-us.apache.org/repos/asf/spark/blob/6a0b77a5/python/pyspark/sql/tests.py
--
diff --git a/python/pyspark/sql/tests.py b/python/pyspark/sql/tests.py
index 35a0636..8d73806 100644
--- a/python/pyspark/sql/tests.py
+++ b/python/pyspark/sql/tests.py
@@ -3351,11 +3351,41 @@ class SQLTests(ReusedSQLTestCase):
 finally:
 shutil.rmtree(path)
 
-def test_repr_html(self):
+def test_repr_behaviors(self):
 import re
 pattern = re.compile(r'^ *\|', re.MULTILINE)
 df = self.spark.createDataFrame([(1, "1"), (2, "2")], ("key", 
"value"))
-self.assertEquals(None, df._repr_html_())
+
+# test when eager evaluation is enabled and _repr_html_ will not be 
called
+with self.sql_conf({"spark.sql.repl.eagerEval.enabled": True}):
+expected1 = """+-+-+
+||  key|value|
+|+-+-+
+||1|1|
+||2|2|
+|+-+-+
+|"""
+self.assertEquals(re.sub(pattern, '', expected1), df.__repr__())
+with 

svn commit: r27772 - in /dev/spark/2.4.0-SNAPSHOT-2018_06_27_12_01-c04cb2d-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-06-27 Thread pwendell
Author: pwendell
Date: Wed Jun 27 19:15:40 2018
New Revision: 27772

Log:
Apache Spark 2.4.0-SNAPSHOT-2018_06_27_12_01-c04cb2d docs


[This commit notification would consist of 1468 parts, 
which exceeds the limit of 50 ones, so it was shortened to the summary.]

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



svn commit: r27769 - in /dev/spark/v2.2.2-rc2-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _site/api/java/org/apache/spark

2018-06-27 Thread tgraves
Author: tgraves
Date: Wed Jun 27 17:44:22 2018
New Revision: 27769

Log:
Apache Spark v2.2.2-rc2 docs


[This commit notification would consist of 1346 parts, 
which exceeds the limit of 50 ones, so it was shortened to the summary.]

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



spark git commit: [SPARK-21687][SQL] Spark SQL should set createTime for Hive partition

2018-06-27 Thread lixiao
Repository: spark
Updated Branches:
  refs/heads/master 78ecb6d45 -> c04cb2d1b


[SPARK-21687][SQL] Spark SQL should set createTime for Hive partition

## What changes were proposed in this pull request?

Set createTime for every hive partition created in Spark SQL, which could be 
used to manage data lifecycle in Hive warehouse. We found  that almost every 
partition modified by spark sql has not been set createTime.

```
mysql> select * from partitions where create_time=0 limit 1\G;
*** 1. row ***
 PART_ID: 1028584
 CREATE_TIME: 0
LAST_ACCESS_TIME: 1502203611
   PART_NAME: date=20170130
   SD_ID: 1543605
  TBL_ID: 211605
  LINK_TARGET_ID: NULL
1 row in set (0.27 sec)
```

## How was this patch tested?
 N/A

Author: debugger87 
Author: Chaozhong Yang 

Closes #18900 from debugger87/fix/set-create-time-for-hive-partition.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/c04cb2d1
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/c04cb2d1
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/c04cb2d1

Branch: refs/heads/master
Commit: c04cb2d1b72b1edaddf684755f5a9d6aaf00e03b
Parents: 78ecb6d
Author: debugger87 
Authored: Wed Jun 27 11:34:28 2018 -0700
Committer: Xiao Li 
Committed: Wed Jun 27 11:34:28 2018 -0700

--
 .../apache/spark/sql/catalyst/catalog/interface.scala |  6 ++
 .../sql/catalyst/catalog/SessionCatalogSuite.scala|  6 --
 .../results/describe-part-after-analyze.sql.out   | 14 ++
 .../test/resources/sql-tests/results/describe.sql.out |  4 
 .../resources/sql-tests/results/show-tables.sql.out   |  2 ++
 .../apache/spark/sql/hive/client/HiveClientImpl.scala |  4 
 6 files changed, 34 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/c04cb2d1/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/catalog/interface.scala
--
diff --git 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/catalog/interface.scala
 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/catalog/interface.scala
index f3e67dc..c6105c5 100644
--- 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/catalog/interface.scala
+++ 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/catalog/interface.scala
@@ -93,12 +93,16 @@ object CatalogStorageFormat {
  * @param spec partition spec values indexed by column name
  * @param storage storage format of the partition
  * @param parameters some parameters for the partition
+ * @param createTime creation time of the partition, in milliseconds
+ * @param lastAccessTime last access time, in milliseconds
  * @param stats optional statistics (number of rows, total size, etc.)
  */
 case class CatalogTablePartition(
 spec: CatalogTypes.TablePartitionSpec,
 storage: CatalogStorageFormat,
 parameters: Map[String, String] = Map.empty,
+createTime: Long = System.currentTimeMillis,
+lastAccessTime: Long = -1,
 stats: Option[CatalogStatistics] = None) {
 
   def toLinkedHashMap: mutable.LinkedHashMap[String, String] = {
@@ -109,6 +113,8 @@ case class CatalogTablePartition(
 if (parameters.nonEmpty) {
   map.put("Partition Parameters", s"{${parameters.map(p => p._1 + "=" + 
p._2).mkString(", ")}}")
 }
+map.put("Created Time", new Date(createTime).toString)
+map.put("Last Access", new Date(lastAccessTime).toString)
 stats.foreach(s => map.put("Partition Statistics", s.simpleString))
 map
   }

http://git-wip-us.apache.org/repos/asf/spark/blob/c04cb2d1/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/catalog/SessionCatalogSuite.scala
--
diff --git 
a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/catalog/SessionCatalogSuite.scala
 
b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/catalog/SessionCatalogSuite.scala
index 6abab00..6a7375e 100644
--- 
a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/catalog/SessionCatalogSuite.scala
+++ 
b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/catalog/SessionCatalogSuite.scala
@@ -1114,11 +1114,13 @@ abstract class SessionCatalogSuite extends AnalysisTest 
{
 // And for hive serde table, hive metastore will set some 
values(e.g.transient_lastDdlTime)
 // in table's parameters and storage's properties, here we also ignore 
them.
 val actualPartsNormalize = actualParts.map(p =>
-  p.copy(parameters = Map.empty, storage = p.storage.copy(
+  p.copy(parameters = Map.empty, createTime = -1, lastAccessTime = -1,
+storage = p.storage.copy(
 properties = Map.empty, locationUri = None, serde = 

svn commit: r27773 - /release/spark/KEYS

2018-06-27 Thread tgraves
Author: tgraves
Date: Wed Jun 27 20:04:43 2018
New Revision: 27773

Log:
Update KEYS

Modified:
release/spark/KEYS

Modified: release/spark/KEYS
==
--- release/spark/KEYS (original)
+++ release/spark/KEYS Wed Jun 27 20:04:43 2018
@@ -589,3 +589,60 @@ pp4VlK8tOgCZ3o6bWPR95GvVBvALnpoOJjS+cGwO
 7fva6arUhIPxFUb+7G0SqJkpib8s4IZ+os3wh4Rv5ZfDAavB4ftC298=
 =DFmT
 -END PGP PUBLIC KEY BLOCK-
+
+pub   4096R/81E6C76A 2018-06-13
+uid  Thomas Graves (CODE SIGNING KEY) 
+sub   4096R/FC8689C4 2018-06-13
+
+-BEGIN PGP PUBLIC KEY BLOCK-
+Version: GnuPG v2.0.22 (GNU/Linux)
+
+mQINBFsgdxABEACwfMYBgIW9v73TIIQJnpHWE7SgGHfAs60FR128uLzzcAJcAly6
+/lq4nRU6inNqiATM3f/HXKhKsv1koFtqPXz3yT6SYTgbMLP0zCU41nPi00r0AOY0
+KrRmM76bbx1hYVjw3L0gTz2froeQxcYoTUMPNqGFjGfEYHfWo04O2Ev5nf+P5kEU
+btk+QhzetcmIKRHVg1DSivPSknGYFRjVREWJTsfz8dZ/2MkN3gqPDcrIIgHJrGpM
+cAESFz1E7RLX3i+cJGH90BCDXv2DgcRtVAHJ/9b9IbfZ24pWqwk7nt7tpcGMwDPy
+9caQ2H5q4H+ECd2Rj2Tzh+ptRz6csLWUAONlS/+rP208wVTNqNCONlw5euYh8wJe
+DVrahHUlsI/SB5edBKQY6/nW+54wphHdbF1iE7GV0XFyG4t0RARkvJlaO+5SMngw
+3nxMd/CoL8r0j+0oncUkb8ZAPzJaVs0YHtEKOf/k7PpaRZW5LYCxiZzZWcjyTxtT
+kqrPZ11MwEqLVFiEMnEg4JVCX14WqP9AdWnrw1pOf9iAgs7LwyOEWxaR9CN70CLK
+Qao56vILC5/B9FN4jyee5y5U8yujqIXjU1ptKnMmYDEZn/fvns0kOix//lWSdHg5
+yt6/95Q/QXxOo4tenlKvLfORVkY4RNmWUOErZ+hbu0cysS8YToOif4+F1QARAQAB
+tDVUaG9tYXMgR3JhdmVzIChDT0RFIFNJR05JTkcgS0VZKSA8dGdyYXZlc0BhcGFj
+aGUub3JnPokCNwQTAQoAIQUCWyB3EAIbAwULCQgHAwUVCgkICwUWAgMBAAIeAQIX
+gAAKCRDeT7zNgebHam5pEACHgEA+xEsL9TYKj7lk5pwa7tVSFRIjiZqu7fausssS
+sAnCDyHv683HKLh8b+tFONeZlTIz1ygeEnbzYr2uthF8GPId1ydNFrZ3/tV/8XlN
+hb2AwThuDXvMHpw1GvTqbMF80kdJ5SIUIC8l6M05NyInU9kTF3mKjol2UXIpBono
+rcSeozCn6k6NLYAopmwJMGtMc4vf1SrCn59+Y2QUBrUsw8iA1yAwc2IYdteYbC1k
+4lrLPIMvGzKC8wAb2I46B3EWkR4soQ8VYfdkZKjJqdcYUSJSL9Jw2XgRuQxKrSoB
+6tc52eR0V/rPCJCTkqFB5eLIYwn4A1oD3PZllSaBlG6UxWlxZiyoAf6EzgIbORTg
+ark5NA6r9p+is7U684ysLBa1be/RX4nq7OIxdOomARErVbepDsJa8TkTyGs+CTau
+AWIi9UbYn9DZ08VMe0PPqT+7s2kcjuddIP2WX6ScYt0c3HAAr9Ky3FwOGvkMlWo2
+DETmQwt1RNhEZC0kSvJSveqHmMaaY4DcgcfPYqUpYbLKDiIzxoKkAOYeyt7X85qJ
+xpUC5nbA1G9Vp1o+zaqzOqAOLTLsI370kVj+57XoFiwIo9Fl0VlSxLoCzXJ7Rjkt
+iCu2Wwxd22W4f8OtRVXJQsL1ikAa5hsE0zkpHHypa8Wa41NqVM2H1nPjOLmcRQky
+xLkCDQRbIHcQARAA5C1Z49Fs0PsKQNjEFyMxvJGCAdGhQmH4VGqZxJQ5GRmScucG
+gPkeW2L3n3v6oguASihB8KgF+WgLgM/8Tn5u7BacY+7hcqgSDgUglVAqcGGrsNTh
+PTdO92HFcIdKfAoacF8IPRDwhCBABVUQyfhVS4yzTHbk1JbX80mcGOkaZMU1tnrF
+etfxcxEQI+1DP/JZ70A6EK+FlyGgmtjrjPRV38DJWDYV4NxHEXSO0d669dIroHBE
+LzcbarB4k9tkSnO2hmbulX2622vKKMuxDmIbKgid69AKVWD0qFaCNSYrdOcexbQr
++LNW6gS30sHhauUIvAu1pGWJefwJjvSme39uBbhG9dpP9tjeUH62RejXvqu0lb4o
+t8tRbWcOZXEZhQQy/SVJXwpMevGAKO8euOuIY9m3uNzf0wMKD7y/023odDhk4qb7
+OIjJwq1NJqkIRJLaY5MBeyAd7xwkfi0/EhrSkaoISJwbrrfHaBCgEJVBWXF+pRCB
+Tp6FotqL45AH/4j+/cw/k94kqc433u7P1ORCWWwgVxQN69X+8hZBo5BSVnW+TSvY
+MATFM/7uMvj2yk0DHPbK5jZ2sqpe5ejQTmougya3mj8kEr/vyNExM8ebo+KdSEuZ
+hnbBjRhCsQ/SgwIpY53Lbp4gZAPjN66VX0kbuJ0Nocvp2dMl+9QkVq+9im8AEQEA
+AYkCHwQYAQoACQUCWyB3EAIbDAAKCRDeT7zNgebHavfsEACp7nEkv7fYU9QgnGpP
+/14yICPtnfBJzJ5Z35WkjSStBxd+Xdbnec1UlLqzGPPuXtLOcouRvmN6WXGPynB9
+ERd2Pga1yiexDZhVVR/vucwNq5pQv6LTlufMI1Nan/p6swwlsDWpVzPfGRaOMOt1
+1OaWeTA0q2ByDWQQV1qay09fs9dJsW6YBzFCKwNNaoXhM5zrmlQj80Ces7gMBJd+
+iivRVceRC+b2vAMvceX2FReS21o4S+mTKlTCvUNDbJe46a+rUEeHzBYN5qvbLedO
+PQEXXUJLrjbj8/uA5gfW+Ib2azfpo7ezPnxYBNz4UC9CGF3JWFnYX7UHpwniB2pt
+J9RrNiKL9Ohdpg1qNgJCAXzbrE31bqpaGawTFYMtRlEYHKDTs6jglX9cIVqDiz+y
+smiMjVBM4G8PVwCvC0OfuCOqQrbieiSNfdBlcWJG5MbZqyXwQNiuRl5tdJPtB4Kx
+NfSm3khyy1/xze0gOi4OwjznHXwW7606Uu6eAKKhWcQ7mHovEJSp5ziWb+/1PXe7
+f5XpIqrYxuMpghlA6Q72sON9zcmFEKW6y+3mRKtBzOsJXLKS0AWXi19zoFlq3Kzs
+UiZuJIMHLPJK4sbOj5nMZE2163zUXz+gOVMeLqVxV/zlW6xnYu84taF4PbYIzmts
+qG1EJoF+tteqemi1ZwYipD06wA==
+=cA5d
+-END PGP PUBLIC KEY BLOCK-



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



spark git commit: [SPARK-24446][YARN] Properly quote library path for YARN.

2018-06-27 Thread vanzin
Repository: spark
Updated Branches:
  refs/heads/master 6a0b77a55 -> 78ecb6d45


[SPARK-24446][YARN] Properly quote library path for YARN.

Because the way YARN executes commands via bash -c, everything needs
to be quoted so that the whole command is fully contained inside a
bash string and is interpreted correctly when the string is read by
bash. This is a bit different than the quoting done when executing
things as if typing in a bash shell.

Tweaked unit tests to exercise the bad behavior, which would cause
existing tests to time out without the fix. Also tested on a real
cluster, verifying the shell script created by YARN to run the
container.

Author: Marcelo Vanzin 

Closes #21476 from vanzin/SPARK-24446.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/78ecb6d4
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/78ecb6d4
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/78ecb6d4

Branch: refs/heads/master
Commit: 78ecb6d457970b136a2e0e0e27d170c84ea28eac
Parents: 6a0b77a
Author: Marcelo Vanzin 
Authored: Wed Jun 27 10:57:29 2018 -0700
Committer: Marcelo Vanzin 
Committed: Wed Jun 27 10:57:29 2018 -0700

--
 .../org/apache/spark/deploy/yarn/Client.scala   | 22 ++--
 .../spark/deploy/yarn/ExecutorRunnable.scala| 11 +-
 .../deploy/yarn/BaseYarnClusterSuite.scala  |  9 
 3 files changed, 34 insertions(+), 8 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/78ecb6d4/resource-managers/yarn/src/main/scala/org/apache/spark/deploy/yarn/Client.scala
--
diff --git 
a/resource-managers/yarn/src/main/scala/org/apache/spark/deploy/yarn/Client.scala
 
b/resource-managers/yarn/src/main/scala/org/apache/spark/deploy/yarn/Client.scala
index 7225ff0..793d012 100644
--- 
a/resource-managers/yarn/src/main/scala/org/apache/spark/deploy/yarn/Client.scala
+++ 
b/resource-managers/yarn/src/main/scala/org/apache/spark/deploy/yarn/Client.scala
@@ -899,7 +899,8 @@ private[spark] class Client(
   val libraryPaths = Seq(sparkConf.get(DRIVER_LIBRARY_PATH),
 sys.props.get("spark.driver.libraryPath")).flatten
   if (libraryPaths.nonEmpty) {
-prefixEnv = Some(getClusterPath(sparkConf, 
Utils.libraryPathEnvPrefix(libraryPaths)))
+prefixEnv = 
Some(createLibraryPathPrefix(libraryPaths.mkString(File.pathSeparator),
+  sparkConf))
   }
   if (sparkConf.get(AM_JAVA_OPTIONS).isDefined) {
 logWarning(s"${AM_JAVA_OPTIONS.key} will not take effect in cluster 
mode")
@@ -921,7 +922,7 @@ private[spark] class Client(
   .map(YarnSparkHadoopUtil.escapeForShell)
   }
   sparkConf.get(AM_LIBRARY_PATH).foreach { paths =>
-prefixEnv = Some(getClusterPath(sparkConf, 
Utils.libraryPathEnvPrefix(Seq(paths
+prefixEnv = Some(createLibraryPathPrefix(paths, sparkConf))
   }
 }
 
@@ -1485,6 +1486,23 @@ private object Client extends Logging {
 YarnAppReport(report.getYarnApplicationState(), 
report.getFinalApplicationStatus(), diagsOpt)
   }
 
+  /**
+   * Create a properly quoted and escaped library path string to be added as a 
prefix to the command
+   * executed by YARN. This is different from normal quoting / escaping due to 
YARN executing the
+   * command through "bash -c".
+   */
+  def createLibraryPathPrefix(libpath: String, conf: SparkConf): String = {
+val cmdPrefix = if (Utils.isWindows) {
+  Utils.libraryPathEnvPrefix(Seq(libpath))
+} else {
+  val envName = Utils.libraryPathEnvName
+  // For quotes, escape both the quote and the escape character when 
encoding in the command
+  // string.
+  val quoted = libpath.replace("\"", "\\\"")
+  envName + "=\\\"" + quoted + File.pathSeparator + "$" + envName + "\\\""
+}
+getClusterPath(conf, cmdPrefix)
+  }
 }
 
 private[spark] class YarnClusterApplication extends SparkApplication {

http://git-wip-us.apache.org/repos/asf/spark/blob/78ecb6d4/resource-managers/yarn/src/main/scala/org/apache/spark/deploy/yarn/ExecutorRunnable.scala
--
diff --git 
a/resource-managers/yarn/src/main/scala/org/apache/spark/deploy/yarn/ExecutorRunnable.scala
 
b/resource-managers/yarn/src/main/scala/org/apache/spark/deploy/yarn/ExecutorRunnable.scala
index a2a18cd..49a0b93 100644
--- 
a/resource-managers/yarn/src/main/scala/org/apache/spark/deploy/yarn/ExecutorRunnable.scala
+++ 
b/resource-managers/yarn/src/main/scala/org/apache/spark/deploy/yarn/ExecutorRunnable.scala
@@ -131,10 +131,6 @@ private[yarn] class ExecutorRunnable(
 // Extra options for the JVM
 val javaOpts = ListBuffer[String]()
 
-// Set the environment variable through a command prefix
-// to 

svn commit: r27778 - in /dev/spark/2.3.2-SNAPSHOT-2018_06_27_14_01-6e1f5e0-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-06-27 Thread pwendell
Author: pwendell
Date: Wed Jun 27 21:15:23 2018
New Revision: 27778

Log:
Apache Spark 2.3.2-SNAPSHOT-2018_06_27_14_01-6e1f5e0 docs


[This commit notification would consist of 1443 parts, 
which exceeds the limit of 50 ones, so it was shortened to the summary.]

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



spark git commit: [SPARK-24613][SQL] Cache with UDF could not be matched with subsequent dependent caches

2018-06-27 Thread lixiao
Repository: spark
Updated Branches:
  refs/heads/branch-2.3 db538b25a -> 6e1f5e018


[SPARK-24613][SQL] Cache with UDF could not be matched with subsequent 
dependent caches

Wrap the logical plan with a `AnalysisBarrier` for execution plan compilation 
in CacheManager, in order to avoid the plan being analyzed again.

Add one test in `DatasetCacheSuite`

Author: Maryann Xue 

Closes #21602 from maryannxue/cache-mismatch.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/6e1f5e01
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/6e1f5e01
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/6e1f5e01

Branch: refs/heads/branch-2.3
Commit: 6e1f5e0182e1d111f7252c24c576674d1d2c7b91
Parents: db538b2
Author: Maryann Xue 
Authored: Thu Jun 21 11:45:30 2018 -0700
Committer: Xiao Li 
Committed: Wed Jun 27 13:20:57 2018 -0700

--
 .../apache/spark/sql/execution/CacheManager.scala   |  6 +++---
 .../org/apache/spark/sql/DatasetCacheSuite.scala| 16 
 2 files changed, 19 insertions(+), 3 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/6e1f5e01/sql/core/src/main/scala/org/apache/spark/sql/execution/CacheManager.scala
--
diff --git 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/CacheManager.scala 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/CacheManager.scala
index d68aeb2..dbdda27 100644
--- a/sql/core/src/main/scala/org/apache/spark/sql/execution/CacheManager.scala
+++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/CacheManager.scala
@@ -26,7 +26,7 @@ import org.apache.hadoop.fs.{FileSystem, Path}
 import org.apache.spark.internal.Logging
 import org.apache.spark.sql.{Dataset, SparkSession}
 import org.apache.spark.sql.catalyst.expressions.SubqueryExpression
-import org.apache.spark.sql.catalyst.plans.logical.{LogicalPlan, ResolvedHint}
+import org.apache.spark.sql.catalyst.plans.logical.{AnalysisBarrier, 
LogicalPlan, ResolvedHint}
 import org.apache.spark.sql.execution.columnar.InMemoryRelation
 import org.apache.spark.sql.execution.datasources.{HadoopFsRelation, 
LogicalRelation}
 import org.apache.spark.storage.StorageLevel
@@ -97,7 +97,7 @@ class CacheManager extends Logging {
   val inMemoryRelation = InMemoryRelation(
 sparkSession.sessionState.conf.useCompression,
 sparkSession.sessionState.conf.columnBatchSize, storageLevel,
-sparkSession.sessionState.executePlan(planToCache).executedPlan,
+
sparkSession.sessionState.executePlan(AnalysisBarrier(planToCache)).executedPlan,
 tableName,
 planToCache.stats)
   cachedData.add(CachedData(planToCache, inMemoryRelation))
@@ -146,7 +146,7 @@ class CacheManager extends Logging {
   useCompression = cd.cachedRepresentation.useCompression,
   batchSize = cd.cachedRepresentation.batchSize,
   storageLevel = cd.cachedRepresentation.storageLevel,
-  child = spark.sessionState.executePlan(cd.plan).executedPlan,
+  child = 
spark.sessionState.executePlan(AnalysisBarrier(cd.plan)).executedPlan,
   tableName = cd.cachedRepresentation.tableName,
   statsOfPlanToCache = cd.plan.stats)
 needToRecache += cd.copy(cachedRepresentation = newCache)

http://git-wip-us.apache.org/repos/asf/spark/blob/6e1f5e01/sql/core/src/test/scala/org/apache/spark/sql/DatasetCacheSuite.scala
--
diff --git 
a/sql/core/src/test/scala/org/apache/spark/sql/DatasetCacheSuite.scala 
b/sql/core/src/test/scala/org/apache/spark/sql/DatasetCacheSuite.scala
index e0561ee..f6c760e 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/DatasetCacheSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/DatasetCacheSuite.scala
@@ -17,6 +17,7 @@
 
 package org.apache.spark.sql
 
+import org.apache.spark.sql.execution.columnar.{InMemoryRelation, 
InMemoryTableScanExec}
 import org.apache.spark.sql.functions._
 import org.apache.spark.sql.test.SharedSQLContext
 import org.apache.spark.storage.StorageLevel
@@ -96,4 +97,19 @@ class DatasetCacheSuite extends QueryTest with 
SharedSQLContext {
 agged.unpersist()
 assert(agged.storageLevel == StorageLevel.NONE, "The Dataset agged should 
not be cached.")
   }
+
+  test("SPARK-24613 Cache with UDF could not be matched with subsequent 
dependent caches") {
+val udf1 = udf({x: Int => x + 1})
+val df = spark.range(0, 10).toDF("a").withColumn("b", udf1($"a"))
+val df2 = df.agg(sum(df("b")))
+
+df.cache()
+df.count()
+df2.cache()
+
+val plan = df2.queryExecution.withCachedData
+assert(plan.isInstanceOf[InMemoryRelation])
+val internalPlan = plan.asInstanceOf[InMemoryRelation].child
+

svn commit: r27774 - /dev/spark/KEYS

2018-06-27 Thread tgraves
Author: tgraves
Date: Wed Jun 27 20:22:34 2018
New Revision: 27774

Log:
Update KEYS

Modified:
dev/spark/KEYS

Modified: dev/spark/KEYS
==
--- dev/spark/KEYS (original)
+++ dev/spark/KEYS Wed Jun 27 20:22:34 2018
@@ -589,3 +589,60 @@ pp4VlK8tOgCZ3o6bWPR95GvVBvALnpoOJjS+cGwO
 7fva6arUhIPxFUb+7G0SqJkpib8s4IZ+os3wh4Rv5ZfDAavB4ftC298=
 =DFmT
 -END PGP PUBLIC KEY BLOCK-
+
+pub   4096R/81E6C76A 2018-06-13
+uid  Thomas Graves (CODE SIGNING KEY) 
+sub   4096R/FC8689C4 2018-06-13
+
+-BEGIN PGP PUBLIC KEY BLOCK-
+Version: GnuPG v2.0.22 (GNU/Linux)
+
+mQINBFsgdxABEACwfMYBgIW9v73TIIQJnpHWE7SgGHfAs60FR128uLzzcAJcAly6
+/lq4nRU6inNqiATM3f/HXKhKsv1koFtqPXz3yT6SYTgbMLP0zCU41nPi00r0AOY0
+KrRmM76bbx1hYVjw3L0gTz2froeQxcYoTUMPNqGFjGfEYHfWo04O2Ev5nf+P5kEU
+btk+QhzetcmIKRHVg1DSivPSknGYFRjVREWJTsfz8dZ/2MkN3gqPDcrIIgHJrGpM
+cAESFz1E7RLX3i+cJGH90BCDXv2DgcRtVAHJ/9b9IbfZ24pWqwk7nt7tpcGMwDPy
+9caQ2H5q4H+ECd2Rj2Tzh+ptRz6csLWUAONlS/+rP208wVTNqNCONlw5euYh8wJe
+DVrahHUlsI/SB5edBKQY6/nW+54wphHdbF1iE7GV0XFyG4t0RARkvJlaO+5SMngw
+3nxMd/CoL8r0j+0oncUkb8ZAPzJaVs0YHtEKOf/k7PpaRZW5LYCxiZzZWcjyTxtT
+kqrPZ11MwEqLVFiEMnEg4JVCX14WqP9AdWnrw1pOf9iAgs7LwyOEWxaR9CN70CLK
+Qao56vILC5/B9FN4jyee5y5U8yujqIXjU1ptKnMmYDEZn/fvns0kOix//lWSdHg5
+yt6/95Q/QXxOo4tenlKvLfORVkY4RNmWUOErZ+hbu0cysS8YToOif4+F1QARAQAB
+tDVUaG9tYXMgR3JhdmVzIChDT0RFIFNJR05JTkcgS0VZKSA8dGdyYXZlc0BhcGFj
+aGUub3JnPokCNwQTAQoAIQUCWyB3EAIbAwULCQgHAwUVCgkICwUWAgMBAAIeAQIX
+gAAKCRDeT7zNgebHam5pEACHgEA+xEsL9TYKj7lk5pwa7tVSFRIjiZqu7fausssS
+sAnCDyHv683HKLh8b+tFONeZlTIz1ygeEnbzYr2uthF8GPId1ydNFrZ3/tV/8XlN
+hb2AwThuDXvMHpw1GvTqbMF80kdJ5SIUIC8l6M05NyInU9kTF3mKjol2UXIpBono
+rcSeozCn6k6NLYAopmwJMGtMc4vf1SrCn59+Y2QUBrUsw8iA1yAwc2IYdteYbC1k
+4lrLPIMvGzKC8wAb2I46B3EWkR4soQ8VYfdkZKjJqdcYUSJSL9Jw2XgRuQxKrSoB
+6tc52eR0V/rPCJCTkqFB5eLIYwn4A1oD3PZllSaBlG6UxWlxZiyoAf6EzgIbORTg
+ark5NA6r9p+is7U684ysLBa1be/RX4nq7OIxdOomARErVbepDsJa8TkTyGs+CTau
+AWIi9UbYn9DZ08VMe0PPqT+7s2kcjuddIP2WX6ScYt0c3HAAr9Ky3FwOGvkMlWo2
+DETmQwt1RNhEZC0kSvJSveqHmMaaY4DcgcfPYqUpYbLKDiIzxoKkAOYeyt7X85qJ
+xpUC5nbA1G9Vp1o+zaqzOqAOLTLsI370kVj+57XoFiwIo9Fl0VlSxLoCzXJ7Rjkt
+iCu2Wwxd22W4f8OtRVXJQsL1ikAa5hsE0zkpHHypa8Wa41NqVM2H1nPjOLmcRQky
+xLkCDQRbIHcQARAA5C1Z49Fs0PsKQNjEFyMxvJGCAdGhQmH4VGqZxJQ5GRmScucG
+gPkeW2L3n3v6oguASihB8KgF+WgLgM/8Tn5u7BacY+7hcqgSDgUglVAqcGGrsNTh
+PTdO92HFcIdKfAoacF8IPRDwhCBABVUQyfhVS4yzTHbk1JbX80mcGOkaZMU1tnrF
+etfxcxEQI+1DP/JZ70A6EK+FlyGgmtjrjPRV38DJWDYV4NxHEXSO0d669dIroHBE
+LzcbarB4k9tkSnO2hmbulX2622vKKMuxDmIbKgid69AKVWD0qFaCNSYrdOcexbQr
++LNW6gS30sHhauUIvAu1pGWJefwJjvSme39uBbhG9dpP9tjeUH62RejXvqu0lb4o
+t8tRbWcOZXEZhQQy/SVJXwpMevGAKO8euOuIY9m3uNzf0wMKD7y/023odDhk4qb7
+OIjJwq1NJqkIRJLaY5MBeyAd7xwkfi0/EhrSkaoISJwbrrfHaBCgEJVBWXF+pRCB
+Tp6FotqL45AH/4j+/cw/k94kqc433u7P1ORCWWwgVxQN69X+8hZBo5BSVnW+TSvY
+MATFM/7uMvj2yk0DHPbK5jZ2sqpe5ejQTmougya3mj8kEr/vyNExM8ebo+KdSEuZ
+hnbBjRhCsQ/SgwIpY53Lbp4gZAPjN66VX0kbuJ0Nocvp2dMl+9QkVq+9im8AEQEA
+AYkCHwQYAQoACQUCWyB3EAIbDAAKCRDeT7zNgebHavfsEACp7nEkv7fYU9QgnGpP
+/14yICPtnfBJzJ5Z35WkjSStBxd+Xdbnec1UlLqzGPPuXtLOcouRvmN6WXGPynB9
+ERd2Pga1yiexDZhVVR/vucwNq5pQv6LTlufMI1Nan/p6swwlsDWpVzPfGRaOMOt1
+1OaWeTA0q2ByDWQQV1qay09fs9dJsW6YBzFCKwNNaoXhM5zrmlQj80Ces7gMBJd+
+iivRVceRC+b2vAMvceX2FReS21o4S+mTKlTCvUNDbJe46a+rUEeHzBYN5qvbLedO
+PQEXXUJLrjbj8/uA5gfW+Ib2azfpo7ezPnxYBNz4UC9CGF3JWFnYX7UHpwniB2pt
+J9RrNiKL9Ohdpg1qNgJCAXzbrE31bqpaGawTFYMtRlEYHKDTs6jglX9cIVqDiz+y
+smiMjVBM4G8PVwCvC0OfuCOqQrbieiSNfdBlcWJG5MbZqyXwQNiuRl5tdJPtB4Kx
+NfSm3khyy1/xze0gOi4OwjznHXwW7606Uu6eAKKhWcQ7mHovEJSp5ziWb+/1PXe7
+f5XpIqrYxuMpghlA6Q72sON9zcmFEKW6y+3mRKtBzOsJXLKS0AWXi19zoFlq3Kzs
+UiZuJIMHLPJK4sbOj5nMZE2163zUXz+gOVMeLqVxV/zlW6xnYu84taF4PbYIzmts
+qG1EJoF+tteqemi1ZwYipD06wA==
+=cA5d
+-END PGP PUBLIC KEY BLOCK-



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



spark git commit: [SPARK-24660][SHS] Show correct error pages when downloading logs

2018-06-27 Thread vanzin
Repository: spark
Updated Branches:
  refs/heads/master c04cb2d1b -> 776befbfd


[SPARK-24660][SHS] Show correct error pages when downloading logs

## What changes were proposed in this pull request?

SHS is showing bad errors when trying to download logs is not successful. This 
may happen because the requested application doesn't exist or the user doesn't 
have permissions for it, for instance.

The PR fixes the response when errors occur, so that they are displayed 
properly.

## How was this patch tested?

manual tests

**Before the patch:**
 1. Unauthorized user
![screen shot 2018-06-26 at 3 53 33 
pm](https://user-images.githubusercontent.com/8821783/41918118-f8b37e70-795b-11e8-91e8-d0250239f09d.png)

 2. Non-existing application
![screen shot 2018-06-26 at 3 25 19 
pm](https://user-images.githubusercontent.com/8821783/41918082-e3034c72-795b-11e8-970e-cee4a1eae77f.png)

**After the patch**
 1. Unauthorized user
![screen shot 2018-06-26 at 3 41 29 
pm](https://user-images.githubusercontent.com/8821783/41918155-0d950476-795c-11e8-8d26-7b7ce73e6fe1.png)

 2. Non-existing application
![screen shot 2018-06-26 at 3 40 37 
pm](https://user-images.githubusercontent.com/8821783/41918175-1a14bb88-795c-11e8-91ab-eadf29190a02.png)

Author: Marco Gaido 

Closes #21644 from mgaido91/SPARK-24660.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/776befbf
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/776befbf
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/776befbf

Branch: refs/heads/master
Commit: 776befbfd5b3c317a713d4fa3882cda6264db9ba
Parents: c04cb2d
Author: Marco Gaido 
Authored: Wed Jun 27 14:26:08 2018 -0700
Committer: Marcelo Vanzin 
Committed: Wed Jun 27 14:26:08 2018 -0700

--
 .../spark/status/api/v1/ApiRootResource.scala   | 30 
 .../status/api/v1/JacksonMessageWriter.scala|  5 +---
 .../status/api/v1/OneApplicationResource.scala  |  7 ++---
 .../scala/org/apache/spark/ui/UIUtils.scala |  5 
 4 files changed, 13 insertions(+), 34 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/776befbf/core/src/main/scala/org/apache/spark/status/api/v1/ApiRootResource.scala
--
diff --git 
a/core/src/main/scala/org/apache/spark/status/api/v1/ApiRootResource.scala 
b/core/src/main/scala/org/apache/spark/status/api/v1/ApiRootResource.scala
index d121068..84c2ad4 100644
--- a/core/src/main/scala/org/apache/spark/status/api/v1/ApiRootResource.scala
+++ b/core/src/main/scala/org/apache/spark/status/api/v1/ApiRootResource.scala
@@ -28,7 +28,7 @@ import org.glassfish.jersey.server.ServerProperties
 import org.glassfish.jersey.servlet.ServletContainer
 
 import org.apache.spark.SecurityManager
-import org.apache.spark.ui.SparkUI
+import org.apache.spark.ui.{SparkUI, UIUtils}
 
 /**
  * Main entry point for serving spark application metrics as json, using 
JAX-RS.
@@ -148,38 +148,18 @@ private[v1] trait BaseAppResource extends 
ApiRequestContext {
 }
 
 private[v1] class ForbiddenException(msg: String) extends 
WebApplicationException(
-  Response.status(Response.Status.FORBIDDEN).entity(msg).build())
+UIUtils.buildErrorResponse(Response.Status.FORBIDDEN, msg))
 
 private[v1] class NotFoundException(msg: String) extends 
WebApplicationException(
-  new NoSuchElementException(msg),
-Response
-  .status(Response.Status.NOT_FOUND)
-  .entity(ErrorWrapper(msg))
-  .build()
-)
+UIUtils.buildErrorResponse(Response.Status.NOT_FOUND, msg))
 
 private[v1] class ServiceUnavailable(msg: String) extends 
WebApplicationException(
-  new ServiceUnavailableException(msg),
-  Response
-.status(Response.Status.SERVICE_UNAVAILABLE)
-.entity(ErrorWrapper(msg))
-.build()
-)
+UIUtils.buildErrorResponse(Response.Status.SERVICE_UNAVAILABLE, msg))
 
 private[v1] class BadParameterException(msg: String) extends 
WebApplicationException(
-  new IllegalArgumentException(msg),
-  Response
-.status(Response.Status.BAD_REQUEST)
-.entity(ErrorWrapper(msg))
-.build()
-) {
+UIUtils.buildErrorResponse(Response.Status.BAD_REQUEST, msg)) {
   def this(param: String, exp: String, actual: String) = {
 this(raw"""Bad value for parameter "$param".  Expected a $exp, got 
"$actual)
   }
 }
 
-/**
- * Signal to JacksonMessageWriter to not convert the message into json (which 
would result in an
- * extra set of quotes).
- */
-private[v1] case class ErrorWrapper(s: String)

http://git-wip-us.apache.org/repos/asf/spark/blob/776befbf/core/src/main/scala/org/apache/spark/status/api/v1/JacksonMessageWriter.scala
--
diff --git 
a/core/src/main/scala/org/apache/spark/status/api/v1/JacksonMessageWriter.scala 

spark git commit: [SPARK-24533] Typesafe rebranded to lightbend. Changing the build downloads path

2018-06-27 Thread vanzin
Repository: spark
Updated Branches:
  refs/heads/master 776befbfd -> 221d03acc


[SPARK-24533] Typesafe rebranded to lightbend. Changing the build downloads path

Typesafe has rebranded to lightbend. Just changing the downloads path to avoid 
redirection

Tested by running build/mvn -DskipTests package

Author: Sanket Chintapalli 

Closes #21636 from redsanket/SPARK-24533.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/221d03ac
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/221d03ac
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/221d03ac

Branch: refs/heads/master
Commit: 221d03acca19bdf7a2624a29c180c99f098205d8
Parents: 776befb
Author: Sanket Chintapalli 
Authored: Wed Jun 27 14:37:19 2018 -0700
Committer: Marcelo Vanzin 
Committed: Wed Jun 27 14:37:24 2018 -0700

--
 build/mvn | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/221d03ac/build/mvn
--
diff --git a/build/mvn b/build/mvn
index 1405983..ae4276d 100755
--- a/build/mvn
+++ b/build/mvn
@@ -93,7 +93,7 @@ install_mvn() {
 install_zinc() {
   local zinc_path="zinc-0.3.15/bin/zinc"
   [ ! -f "${_DIR}/${zinc_path}" ] && ZINC_INSTALL_FLAG=1
-  local TYPESAFE_MIRROR=${TYPESAFE_MIRROR:-https://downloads.typesafe.com}
+  local TYPESAFE_MIRROR=${TYPESAFE_MIRROR:-https://downloads.lightbend.com}
 
   install_app \
 "${TYPESAFE_MIRROR}/zinc/0.3.15" \
@@ -109,7 +109,7 @@ install_scala() {
   # determine the Scala version used in Spark
   local scala_version=`grep "scala.version" "${_DIR}/../pom.xml" | head -n1 | 
awk -F '[<>]' '{print $3}'`
   local scala_bin="${_DIR}/scala-${scala_version}/bin/scala"
-  local TYPESAFE_MIRROR=${TYPESAFE_MIRROR:-https://downloads.typesafe.com}
+  local TYPESAFE_MIRROR=${TYPESAFE_MIRROR:-https://downloads.lightbend.com}
 
   install_app \
 "${TYPESAFE_MIRROR}/scala/${scala_version}" \


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



spark git commit: [SPARK-24645][SQL] Skip parsing when csvColumnPruning enabled and partitions scanned only

2018-06-27 Thread gurwls223
Repository: spark
Updated Branches:
  refs/heads/master c5aa54d54 -> bd32b509a


[SPARK-24645][SQL] Skip parsing when csvColumnPruning enabled and partitions 
scanned only

## What changes were proposed in this pull request?
In the master, when `csvColumnPruning`(implemented in [this 
commit](https://github.com/apache/spark/commit/64fad0b519cf35b8c0a0dec18dd3df9488a5ed25#diff-d19881aceddcaa5c60620fdcda99b4c4))
 enabled and partitions scanned only, it throws an exception below;

```
scala> val dir = "/tmp/spark-csv/csv"
scala> spark.range(10).selectExpr("id % 2 AS p", 
"id").write.mode("overwrite").partitionBy("p").csv(dir)
scala> spark.read.csv(dir).selectExpr("sum(p)").collect()
18/06/25 13:12:51 ERROR Executor: Exception in task 0.0 in stage 2.0 (TID 5)
java.lang.NullPointerException
at 
org.apache.spark.sql.execution.datasources.csv.UnivocityParser.org$apache$spark$sql$execution$datasources$csv$UnivocityParser$$convert(UnivocityParser.scala:197)
at 
org.apache.spark.sql.execution.datasources.csv.UnivocityParser.parse(UnivocityParser.scala:190)
at 
org.apache.spark.sql.execution.datasources.csv.UnivocityParser$$anonfun$5.apply(UnivocityParser.scala:309)
at 
org.apache.spark.sql.execution.datasources.csv.UnivocityParser$$anonfun$5.apply(UnivocityParser.scala:309)
at 
org.apache.spark.sql.execution.datasources.FailureSafeParser.parse(FailureSafeParser.scala:61)
...
```
This pr modified code to skip CSV parsing in the case.

## How was this patch tested?
Added tests in `CSVSuite`.

Author: Takeshi Yamamuro 

Closes #21631 from maropu/SPARK-24645.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/bd32b509
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/bd32b509
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/bd32b509

Branch: refs/heads/master
Commit: bd32b509a1728366494cba13f8f6612b7bd46ec0
Parents: c5aa54d
Author: Takeshi Yamamuro 
Authored: Thu Jun 28 09:19:25 2018 +0800
Committer: hyukjinkwon 
Committed: Thu Jun 28 09:19:25 2018 +0800

--
 .../sql/execution/datasources/csv/UnivocityParser.scala   | 10 +-
 .../spark/sql/execution/datasources/csv/CSVSuite.scala| 10 ++
 2 files changed, 19 insertions(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/bd32b509/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/csv/UnivocityParser.scala
--
diff --git 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/csv/UnivocityParser.scala
 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/csv/UnivocityParser.scala
index 5f7d569..aa545e1 100644
--- 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/csv/UnivocityParser.scala
+++ 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/csv/UnivocityParser.scala
@@ -183,11 +183,19 @@ class UnivocityParser(
 }
   }
 
+  private val doParse = if (schema.nonEmpty) {
+(input: String) => convert(tokenizer.parseLine(input))
+  } else {
+// If `columnPruning` enabled and partition attributes scanned only,
+// `schema` gets empty.
+(_: String) => InternalRow.empty
+  }
+
   /**
* Parses a single CSV string and turns it into either one resulting row or 
no row (if the
* the record is malformed).
*/
-  def parse(input: String): InternalRow = convert(tokenizer.parseLine(input))
+  def parse(input: String): InternalRow = doParse(input)
 
   private def convert(tokens: Array[String]): InternalRow = {
 if (tokens.length != schema.length) {

http://git-wip-us.apache.org/repos/asf/spark/blob/bd32b509/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/csv/CSVSuite.scala
--
diff --git 
a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/csv/CSVSuite.scala
 
b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/csv/CSVSuite.scala
index 365239d..84b91f6 100644
--- 
a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/csv/CSVSuite.scala
+++ 
b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/csv/CSVSuite.scala
@@ -1569,4 +1569,14 @@ class CSVSuite extends QueryTest with SharedSQLContext 
with SQLTestUtils with Te
 assert(testAppender2.events.asScala
   .exists(msg => msg.getRenderedMessage.contains("CSV header does not 
conform to the schema")))
   }
+
+  test("SPARK-24645 skip parsing when columnPruning enabled and partitions 
scanned only") {
+withSQLConf(SQLConf.CSV_PARSER_COLUMN_PRUNING.key -> "true") {
+  withTempPath { path =>
+val dir = path.getAbsolutePath
+spark.range(10).selectExpr("id % 2 AS p", 

spark git commit: [SPARK-24206][SQL][FOLLOW-UP] Update DataSourceReadBenchmark benchmark results

2018-06-27 Thread gurwls223
Repository: spark
Updated Branches:
  refs/heads/master bd32b509a -> 1c9acc243


[SPARK-24206][SQL][FOLLOW-UP] Update DataSourceReadBenchmark benchmark results

## What changes were proposed in this pull request?
This pr corrected the default configuration (`spark.master=local[1]`) for 
benchmarks. Also, this updated performance results on the AWS `r3.xlarge`.

## How was this patch tested?
N/A

Author: Takeshi Yamamuro 

Closes #21625 from maropu/FixDataSourceReadBenchmark.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/1c9acc24
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/1c9acc24
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/1c9acc24

Branch: refs/heads/master
Commit: 1c9acc2438f9a97134ae5213a12112b2361fbb78
Parents: bd32b50
Author: Takeshi Yamamuro 
Authored: Thu Jun 28 09:21:10 2018 +0800
Committer: hyukjinkwon 
Committed: Thu Jun 28 09:21:10 2018 +0800

--
 .../benchmark/DataSourceReadBenchmark.scala | 296 ++-
 1 file changed, 152 insertions(+), 144 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/1c9acc24/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/DataSourceReadBenchmark.scala
--
diff --git 
a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/DataSourceReadBenchmark.scala
 
b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/DataSourceReadBenchmark.scala
index fc6d8ab..8711f5a 100644
--- 
a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/DataSourceReadBenchmark.scala
+++ 
b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/DataSourceReadBenchmark.scala
@@ -39,9 +39,11 @@ import org.apache.spark.util.{Benchmark, Utils}
 object DataSourceReadBenchmark {
   val conf = new SparkConf()
 .setAppName("DataSourceReadBenchmark")
-.setIfMissing("spark.master", "local[1]")
+// Since `spark.master` always exists, overrides this value
+.set("spark.master", "local[1]")
 .setIfMissing("spark.driver.memory", "3g")
 .setIfMissing("spark.executor.memory", "3g")
+.setIfMissing("spark.ui.enabled", "false")
 
   val spark = SparkSession.builder.config(conf).getOrCreate()
 
@@ -154,73 +156,73 @@ object DataSourceReadBenchmark {
   }
 }
 
-
 /*
-Intel(R) Xeon(R) CPU E5-2686 v4 @ 2.30GHz
+OpenJDK 64-Bit Server VM 1.8.0_171-b10 on Linux 
4.14.33-51.37.amzn1.x86_64
+Intel(R) Xeon(R) CPU E5-2670 v2 @ 2.50GHz
 SQL Single TINYINT Column Scan:  Best/Avg Time(ms)Rate(M/s)   
Per Row(ns)   Relative
 

-SQL CSV 15231 / 15267  1.0 
968.3   1.0X
-SQL Json  8476 / 8498  1.9 
538.9   1.8X
-SQL Parquet Vectorized 121 /  127130.0 
  7.7 125.9X
-SQL Parquet MR1515 / 1543 10.4 
 96.3  10.1X
-SQL ORC Vectorized 164 /  171 95.9 
 10.4  92.9X
-SQL ORC Vectorized with copy   228 /  234 69.0 
 14.5  66.8X
-SQL ORC MR1297 / 1309 12.1 
 82.5  11.7X
+SQL CSV 22964 / 23096  0.7 
   1460.0   1.0X
+SQL Json  8469 / 8593  1.9 
538.4   2.7X
+SQL Parquet Vectorized 164 /  177 95.8 
 10.4 139.9X
+SQL Parquet MR1687 / 1706  9.3 
107.2  13.6X
+SQL ORC Vectorized 191 /  197 82.3 
 12.2 120.2X
+SQL ORC Vectorized with copy   215 /  219 73.2 
 13.7 106.9X
+SQL ORC MR1392 / 1412 11.3 
 88.5  16.5X
 
 
 SQL Single SMALLINT Column Scan: Best/Avg Time(ms)Rate(M/s)   
Per Row(ns)   Relative
 

-SQL CSV 16344 / 16374  1.0 
   1039.1   1.0X
-SQL Json  8634 / 8648  1.8 
548.9   1.9X
-SQL Parquet Vectorized 172 /  177 91.5 
 10.9  95.1X
-SQL Parquet MR1744 / 1746  9.0 
110.9   9.4X
-SQL ORC 

spark git commit: [SPARK-24603][SQL] Fix findTightestCommonType reference in comments

2018-06-27 Thread gurwls223
Repository: spark
Updated Branches:
  refs/heads/master 1c9acc243 -> 6a97e8eb3


[SPARK-24603][SQL] Fix findTightestCommonType reference in comments

findTightestCommonTypeOfTwo has been renamed to findTightestCommonType

## What changes were proposed in this pull request?

(Please fill in changes proposed in this fix)

## How was this patch tested?

(Please explain how this patch was tested. E.g. unit tests, integration tests, 
manual tests)
(If this patch involves UI changes, please attach a screenshot; otherwise, 
remove this)

Please review http://spark.apache.org/contributing.html before opening a pull 
request.

Author: Fokko Driesprong 

Closes #21597 from Fokko/fd-typo.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/6a97e8eb
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/6a97e8eb
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/6a97e8eb

Branch: refs/heads/master
Commit: 6a97e8eb31da76fe5af912a6304c07b63735062f
Parents: 1c9acc2
Author: Fokko Driesprong 
Authored: Thu Jun 28 09:59:00 2018 +0800
Committer: hyukjinkwon 
Committed: Thu Jun 28 09:59:00 2018 +0800

--
 .../spark/sql/execution/datasources/json/JsonInferSchema.scala   | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/6a97e8eb/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/json/JsonInferSchema.scala
--
diff --git 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/json/JsonInferSchema.scala
 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/json/JsonInferSchema.scala
index f6edc7b..8e1b430 100644
--- 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/json/JsonInferSchema.scala
+++ 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/json/JsonInferSchema.scala
@@ -334,8 +334,8 @@ private[sql] object JsonInferSchema {
   ArrayType(compatibleType(elementType1, elementType2), containsNull1 
|| containsNull2)
 
 // The case that given `DecimalType` is capable of given 
`IntegralType` is handled in
-// `findTightestCommonTypeOfTwo`. Both cases below will be executed 
only when
-// the given `DecimalType` is not capable of the given `IntegralType`.
+// `findTightestCommonType`. Both cases below will be executed only 
when the given
+// `DecimalType` is not capable of the given `IntegralType`.
 case (t1: IntegralType, t2: DecimalType) =>
   compatibleType(DecimalType.forType(t1), t2)
 case (t1: DecimalType, t2: IntegralType) =>


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



spark git commit: [SPARK-24603][SQL] Fix findTightestCommonType reference in comments

2018-06-27 Thread gurwls223
Repository: spark
Updated Branches:
  refs/heads/branch-2.2 47958270f -> a8537a5ab


[SPARK-24603][SQL] Fix findTightestCommonType reference in comments

findTightestCommonTypeOfTwo has been renamed to findTightestCommonType

## What changes were proposed in this pull request?

(Please fill in changes proposed in this fix)

## How was this patch tested?

(Please explain how this patch was tested. E.g. unit tests, integration tests, 
manual tests)
(If this patch involves UI changes, please attach a screenshot; otherwise, 
remove this)

Please review http://spark.apache.org/contributing.html before opening a pull 
request.

Author: Fokko Driesprong 

Closes #21597 from Fokko/fd-typo.

(cherry picked from commit 6a97e8eb31da76fe5af912a6304c07b63735062f)
Signed-off-by: hyukjinkwon 


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/a8537a5a
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/a8537a5a
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/a8537a5a

Branch: refs/heads/branch-2.2
Commit: a8537a5ab876e5f25c491bf3f15f339a5a7c2b79
Parents: 4795827
Author: Fokko Driesprong 
Authored: Thu Jun 28 09:59:00 2018 +0800
Committer: hyukjinkwon 
Committed: Thu Jun 28 10:00:59 2018 +0800

--
 .../spark/sql/execution/datasources/json/JsonInferSchema.scala   | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/a8537a5a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/json/JsonInferSchema.scala
--
diff --git 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/json/JsonInferSchema.scala
 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/json/JsonInferSchema.scala
index fb632cf..0987969 100644
--- 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/json/JsonInferSchema.scala
+++ 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/json/JsonInferSchema.scala
@@ -326,8 +326,8 @@ private[sql] object JsonInferSchema {
   ArrayType(compatibleType(elementType1, elementType2), containsNull1 
|| containsNull2)
 
 // The case that given `DecimalType` is capable of given 
`IntegralType` is handled in
-// `findTightestCommonTypeOfTwo`. Both cases below will be executed 
only when
-// the given `DecimalType` is not capable of the given `IntegralType`.
+// `findTightestCommonType`. Both cases below will be executed only 
when the given
+// `DecimalType` is not capable of the given `IntegralType`.
 case (t1: IntegralType, t2: DecimalType) =>
   compatibleType(DecimalType.forType(t1), t2)
 case (t1: DecimalType, t2: IntegralType) =>


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



spark git commit: [SPARK-24603][SQL] Fix findTightestCommonType reference in comments

2018-06-27 Thread gurwls223
Repository: spark
Updated Branches:
  refs/heads/branch-2.3 6e1f5e018 -> 0f534d3da


[SPARK-24603][SQL] Fix findTightestCommonType reference in comments

findTightestCommonTypeOfTwo has been renamed to findTightestCommonType

## What changes were proposed in this pull request?

(Please fill in changes proposed in this fix)

## How was this patch tested?

(Please explain how this patch was tested. E.g. unit tests, integration tests, 
manual tests)
(If this patch involves UI changes, please attach a screenshot; otherwise, 
remove this)

Please review http://spark.apache.org/contributing.html before opening a pull 
request.

Author: Fokko Driesprong 

Closes #21597 from Fokko/fd-typo.

(cherry picked from commit 6a97e8eb31da76fe5af912a6304c07b63735062f)
Signed-off-by: hyukjinkwon 


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/0f534d3d
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/0f534d3d
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/0f534d3d

Branch: refs/heads/branch-2.3
Commit: 0f534d3da284947f52b62154daaddfa77b42873c
Parents: 6e1f5e0
Author: Fokko Driesprong 
Authored: Thu Jun 28 09:59:00 2018 +0800
Committer: hyukjinkwon 
Committed: Thu Jun 28 10:00:38 2018 +0800

--
 .../spark/sql/execution/datasources/json/JsonInferSchema.scala   | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/0f534d3d/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/json/JsonInferSchema.scala
--
diff --git 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/json/JsonInferSchema.scala
 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/json/JsonInferSchema.scala
index a270a64..1702fb6 100644
--- 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/json/JsonInferSchema.scala
+++ 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/json/JsonInferSchema.scala
@@ -329,8 +329,8 @@ private[sql] object JsonInferSchema {
   ArrayType(compatibleType(elementType1, elementType2), containsNull1 
|| containsNull2)
 
 // The case that given `DecimalType` is capable of given 
`IntegralType` is handled in
-// `findTightestCommonTypeOfTwo`. Both cases below will be executed 
only when
-// the given `DecimalType` is not capable of the given `IntegralType`.
+// `findTightestCommonType`. Both cases below will be executed only 
when the given
+// `DecimalType` is not capable of the given `IntegralType`.
 case (t1: IntegralType, t2: DecimalType) =>
   compatibleType(DecimalType.forType(t1), t2)
 case (t1: DecimalType, t2: IntegralType) =>


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



spark git commit: [SPARK-24204][SQL] Verify a schema in Json/Orc/ParquetFileFormat

2018-06-27 Thread lixiao
Repository: spark
Updated Branches:
  refs/heads/master 221d03acc -> 893ea224c


[SPARK-24204][SQL] Verify a schema in Json/Orc/ParquetFileFormat

## What changes were proposed in this pull request?
This pr added code to verify a schema in Json/Orc/ParquetFileFormat along with 
CSVFileFormat.

## How was this patch tested?
Added verification tests in `FileBasedDataSourceSuite` and  
`HiveOrcSourceSuite`.

Author: Takeshi Yamamuro 

Closes #21389 from maropu/SPARK-24204.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/893ea224
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/893ea224
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/893ea224

Branch: refs/heads/master
Commit: 893ea224cc738766be207c87f4b913fe8fea4c94
Parents: 221d03a
Author: Takeshi Yamamuro 
Authored: Wed Jun 27 15:25:51 2018 -0700
Committer: Xiao Li 
Committed: Wed Jun 27 15:25:51 2018 -0700

--
 .../execution/datasources/DataSourceUtils.scala | 106 +
 .../datasources/csv/CSVFileFormat.scala |   4 +-
 .../execution/datasources/csv/CSVUtils.scala|  19 --
 .../datasources/json/JsonFileFormat.scala   |   4 +
 .../datasources/orc/OrcFileFormat.scala |   4 +
 .../datasources/parquet/ParquetFileFormat.scala |   3 +
 .../spark/sql/FileBasedDataSourceSuite.scala| 213 ++-
 .../execution/datasources/csv/CSVSuite.scala|  33 ---
 .../spark/sql/hive/orc/OrcFileFormat.scala  |   4 +
 .../spark/sql/hive/orc/HiveOrcSourceSuite.scala |  49 -
 10 files changed, 383 insertions(+), 56 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/893ea224/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSourceUtils.scala
--
diff --git 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSourceUtils.scala
 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSourceUtils.scala
new file mode 100644
index 000..c534721
--- /dev/null
+++ 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSourceUtils.scala
@@ -0,0 +1,106 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.spark.sql.execution.datasources
+
+import org.apache.spark.sql.execution.datasources.csv.CSVFileFormat
+import org.apache.spark.sql.execution.datasources.json.JsonFileFormat
+import org.apache.spark.sql.execution.datasources.orc.OrcFileFormat
+import org.apache.spark.sql.execution.datasources.parquet.ParquetFileFormat
+import org.apache.spark.sql.types._
+
+
+object DataSourceUtils {
+
+  /**
+   * Verify if the schema is supported in datasource in write path.
+   */
+  def verifyWriteSchema(format: FileFormat, schema: StructType): Unit = {
+verifySchema(format, schema, isReadPath = false)
+  }
+
+  /**
+   * Verify if the schema is supported in datasource in read path.
+   */
+  def verifyReadSchema(format: FileFormat, schema: StructType): Unit = {
+verifySchema(format, schema, isReadPath = true)
+  }
+
+  /**
+   * Verify if the schema is supported in datasource. This verification should 
be done
+   * in a driver side, e.g., `prepareWrite`, `buildReader`, and 
`buildReaderWithPartitionValues`
+   * in `FileFormat`.
+   *
+   * Unsupported data types of csv, json, orc, and parquet are as follows;
+   *  csv -> R/W: Interval, Null, Array, Map, Struct
+   *  json -> W: Interval
+   *  orc -> W: Interval, Null
+   *  parquet -> R/W: Interval, Null
+   */
+  private def verifySchema(format: FileFormat, schema: StructType, isReadPath: 
Boolean): Unit = {
+def throwUnsupportedException(dataType: DataType): Unit = {
+  throw new UnsupportedOperationException(
+s"$format data source does not support ${dataType.simpleString} data 
type.")
+}
+
+def verifyType(dataType: DataType): Unit = dataType match {
+  case BooleanType | ByteType | ShortType | IntegerType | LongType | 
FloatType | DoubleType |
+   StringType | BinaryType | DateType | 

spark git commit: [SPARK-24553][WEB-UI] http 302 fixes for href redirect

2018-06-27 Thread lixiao
Repository: spark
Updated Branches:
  refs/heads/master 893ea224c -> c5aa54d54


[SPARK-24553][WEB-UI] http 302 fixes for href redirect

## What changes were proposed in this pull request?

Updated URL/href links to include a '/' before '?id' to make links consistent 
and avoid http 302 redirect errors within UI port 4040 tabs.

## How was this patch tested?

Built a runnable distribution and executed jobs. Validated that http 302 
redirects are no longer encountered when clicking on links within UI port 4040 
tabs.

Author: Steven Kallman 

Author: Kallman, Steven 

Closes #21600 from SJKallman/{Spark-24553}{WEB-UI}-redirect-href-fixes.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/c5aa54d5
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/c5aa54d5
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/c5aa54d5

Branch: refs/heads/master
Commit: c5aa54d54b301555bad1ff0653df11293f0033ed
Parents: 893ea22
Author: Kallman, Steven 
Authored: Wed Jun 27 15:36:59 2018 -0700
Committer: Xiao Li 
Committed: Wed Jun 27 15:36:59 2018 -0700

--
 core/src/main/scala/org/apache/spark/ui/jobs/AllJobsPage.scala   | 2 +-
 core/src/main/scala/org/apache/spark/ui/jobs/StagePage.scala | 2 +-
 core/src/main/scala/org/apache/spark/ui/jobs/StageTable.scala| 2 +-
 .../org/apache/spark/sql/execution/ui/AllExecutionsPage.scala| 4 ++--
 .../scala/org/apache/spark/sql/execution/ui/ExecutionPage.scala  | 2 +-
 .../apache/spark/sql/hive/thriftserver/ui/ThriftServerPage.scala | 4 ++--
 .../spark/sql/hive/thriftserver/ui/ThriftServerSessionPage.scala | 2 +-
 .../src/main/scala/org/apache/spark/streaming/ui/BatchPage.scala | 2 +-
 8 files changed, 10 insertions(+), 10 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/c5aa54d5/core/src/main/scala/org/apache/spark/ui/jobs/AllJobsPage.scala
--
diff --git a/core/src/main/scala/org/apache/spark/ui/jobs/AllJobsPage.scala 
b/core/src/main/scala/org/apache/spark/ui/jobs/AllJobsPage.scala
index 178d2c8..90e9a7a 100644
--- a/core/src/main/scala/org/apache/spark/ui/jobs/AllJobsPage.scala
+++ b/core/src/main/scala/org/apache/spark/ui/jobs/AllJobsPage.scala
@@ -464,7 +464,7 @@ private[ui] class JobDataSource(
 
 val jobDescription = UIUtils.makeDescription(lastStageDescription, 
basePath, plainText = false)
 
-val detailUrl = "%s/jobs/job?id=%s".format(basePath, jobData.jobId)
+val detailUrl = "%s/jobs/job/?id=%s".format(basePath, jobData.jobId)
 
 new JobTableRowData(
   jobData,

http://git-wip-us.apache.org/repos/asf/spark/blob/c5aa54d5/core/src/main/scala/org/apache/spark/ui/jobs/StagePage.scala
--
diff --git a/core/src/main/scala/org/apache/spark/ui/jobs/StagePage.scala 
b/core/src/main/scala/org/apache/spark/ui/jobs/StagePage.scala
index d4e6a7b..55eb989 100644
--- a/core/src/main/scala/org/apache/spark/ui/jobs/StagePage.scala
+++ b/core/src/main/scala/org/apache/spark/ui/jobs/StagePage.scala
@@ -282,7 +282,7 @@ private[ui] class StagePage(parent: StagesTab, store: 
AppStatusStore) extends We
   val _taskTable = new TaskPagedTable(
 stageData,
 UIUtils.prependBaseUri(request, parent.basePath) +
-  s"/stages/stage?id=${stageId}=${stageAttemptId}",
+  s"/stages/stage/?id=${stageId}=${stageAttemptId}",
 currentTime,
 pageSize = taskPageSize,
 sortColumn = taskSortColumn,

http://git-wip-us.apache.org/repos/asf/spark/blob/c5aa54d5/core/src/main/scala/org/apache/spark/ui/jobs/StageTable.scala
--
diff --git a/core/src/main/scala/org/apache/spark/ui/jobs/StageTable.scala 
b/core/src/main/scala/org/apache/spark/ui/jobs/StageTable.scala
index 56e4d68..d01acda 100644
--- a/core/src/main/scala/org/apache/spark/ui/jobs/StageTable.scala
+++ b/core/src/main/scala/org/apache/spark/ui/jobs/StageTable.scala
@@ -370,7 +370,7 @@ private[ui] class StagePagedTable(
   Seq.empty
 }
 
-val nameLinkUri = 
s"$basePathUri/stages/stage?id=${s.stageId}=${s.attemptId}"
+val nameLinkUri = 
s"$basePathUri/stages/stage/?id=${s.stageId}=${s.attemptId}"
 val nameLink = {s.name}
 
 val cachedRddInfos = store.rddList().filter { rdd => 
s.rddIds.contains(rdd.id) }

http://git-wip-us.apache.org/repos/asf/spark/blob/c5aa54d5/sql/core/src/main/scala/org/apache/spark/sql/execution/ui/AllExecutionsPage.scala
--
diff --git 
a/sql/core/src/main/scala/org/apache/spark/sql/execution/ui/AllExecutionsPage.scala
 
b/sql/core/src/main/scala/org/apache/spark/sql/execution/ui/AllExecutionsPage.scala
index bf46bc4..a7a24ac 100644
--- 

svn commit: r27779 - in /dev/spark/2.4.0-SNAPSHOT-2018_06_27_16_01-c5aa54d-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-06-27 Thread pwendell
Author: pwendell
Date: Wed Jun 27 23:15:46 2018
New Revision: 27779

Log:
Apache Spark 2.4.0-SNAPSHOT-2018_06_27_16_01-c5aa54d docs


[This commit notification would consist of 1468 parts, 
which exceeds the limit of 50 ones, so it was shortened to the summary.]

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



svn commit: r27756 - in /dev/spark/2.4.0-SNAPSHOT-2018_06_27_00_01-9a76f23-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-06-27 Thread pwendell
Author: pwendell
Date: Wed Jun 27 07:17:22 2018
New Revision: 27756

Log:
Apache Spark 2.4.0-SNAPSHOT-2018_06_27_00_01-9a76f23 docs


[This commit notification would consist of 1468 parts, 
which exceeds the limit of 50 ones, so it was shortened to the summary.]

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



spark git commit: [SPARK-21335][DOC] doc changes for disallowed un-aliased subquery use case

2018-06-27 Thread wenchen
Repository: spark
Updated Branches:
  refs/heads/master 9a76f23c6 -> a1a64e358


[SPARK-21335][DOC] doc changes for disallowed un-aliased subquery use case

## What changes were proposed in this pull request?
Document a change for un-aliased subquery use case, to address the last 
question in PR #18559:
https://github.com/apache/spark/pull/18559#issuecomment-316884858

(Please fill in changes proposed in this fix)

## How was this patch tested?
 it does not affect tests.

Please review http://spark.apache.org/contributing.html before opening a pull 
request.

Author: Yuexin Zhang 

Closes #21647 from cnZach/doc_change_for_SPARK-20690_SPARK-21335.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/a1a64e35
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/a1a64e35
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/a1a64e35

Branch: refs/heads/master
Commit: a1a64e3583cfa451b4d0d2361c1da2972a5e
Parents: 9a76f23
Author: Yuexin Zhang 
Authored: Wed Jun 27 16:05:36 2018 +0800
Committer: Wenchen Fan 
Committed: Wed Jun 27 16:05:36 2018 +0800

--
 docs/sql-programming-guide.md | 1 +
 1 file changed, 1 insertion(+)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/a1a64e35/docs/sql-programming-guide.md
--
diff --git a/docs/sql-programming-guide.md b/docs/sql-programming-guide.md
index 7c4ef41..cd7329b 100644
--- a/docs/sql-programming-guide.md
+++ b/docs/sql-programming-guide.md
@@ -2017,6 +2017,7 @@ working with timestamps in `pandas_udf`s to get the best 
performance, see
 - Literal values used in SQL operations are converted to DECIMAL with the 
exact precision and scale needed by them.
 - The configuration `spark.sql.decimalOperations.allowPrecisionLoss` has 
been introduced. It defaults to `true`, which means the new behavior described 
here; if set to `false`, Spark uses previous rules, ie. it doesn't adjust the 
needed scale to represent the values and it returns NULL if an exact 
representation of the value is not possible.
   - In PySpark, `df.replace` does not allow to omit `value` when `to_replace` 
is not a dictionary. Previously, `value` could be omitted in the other cases 
and had `None` by default, which is counterintuitive and error-prone.
+  - Un-aliased subquery's semantic has not been well defined with confusing 
behaviors. Since Spark 2.3, we invalidate such confusing cases, for example: 
`SELECT v.i from (SELECT i FROM v)`, Spark will throw an analysis exception in 
this case because users should not be able to use the qualifier inside a 
subquery. See [SPARK-20690](https://issues.apache.org/jira/browse/SPARK-20690) 
and [SPARK-21335](https://issues.apache.org/jira/browse/SPARK-21335) for more 
details.
 
 ## Upgrading From Spark SQL 2.1 to 2.2
 


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



svn commit: r27763 - in /dev/spark/2.4.0-SNAPSHOT-2018_06_27_04_02-a1a64e3-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-06-27 Thread pwendell
Author: pwendell
Date: Wed Jun 27 11:21:26 2018
New Revision: 27763

Log:
Apache Spark 2.4.0-SNAPSHOT-2018_06_27_04_02-a1a64e3 docs


[This commit notification would consist of 1468 parts, 
which exceeds the limit of 50 ones, so it was shortened to the summary.]

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



svn commit: r27784 - in /dev/spark/2.3.2-SNAPSHOT-2018_06_27_22_01-0f534d3-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-06-27 Thread pwendell
Author: pwendell
Date: Thu Jun 28 05:16:02 2018
New Revision: 27784

Log:
Apache Spark 2.3.2-SNAPSHOT-2018_06_27_22_01-0f534d3 docs


[This commit notification would consist of 1443 parts, 
which exceeds the limit of 50 ones, so it was shortened to the summary.]

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



svn commit: r27783 - in /dev/spark/2.4.0-SNAPSHOT-2018_06_27_20_01-6a97e8e-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-06-27 Thread pwendell
Author: pwendell
Date: Thu Jun 28 03:15:54 2018
New Revision: 27783

Log:
Apache Spark 2.4.0-SNAPSHOT-2018_06_27_20_01-6a97e8e docs


[This commit notification would consist of 1468 parts, 
which exceeds the limit of 50 ones, so it was shortened to the summary.]

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org