[spark] branch master updated (abc8ccc -> 6fe3bf6)

2020-06-18 Thread wenchen
This is an automated email from the ASF dual-hosted git repository.

wenchen pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from abc8ccc  [SPARK-31926][SQL][TESTS][FOLLOWUP][TEST-HIVE1.2][TEST-MAVEN] 
Fix concurrency issue for ThriftCLIService to getPortNumber
 add 6fe3bf6  [SPARK-31993][SQL] Build arrays for passing variables 
generated from children for 'concat_ws' with columns having at least one of 
array type

No new revisions were added by this update.

Summary of changes:
 .../catalyst/expressions/stringExpressions.scala   | 63 --
 .../apache/spark/sql/StringFunctionsSuite.scala| 60 +
 2 files changed, 106 insertions(+), 17 deletions(-)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (abc8ccc -> 6fe3bf6)

2020-06-18 Thread wenchen
This is an automated email from the ASF dual-hosted git repository.

wenchen pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from abc8ccc  [SPARK-31926][SQL][TESTS][FOLLOWUP][TEST-HIVE1.2][TEST-MAVEN] 
Fix concurrency issue for ThriftCLIService to getPortNumber
 add 6fe3bf6  [SPARK-31993][SQL] Build arrays for passing variables 
generated from children for 'concat_ws' with columns having at least one of 
array type

No new revisions were added by this update.

Summary of changes:
 .../catalyst/expressions/stringExpressions.scala   | 63 --
 .../apache/spark/sql/StringFunctionsSuite.scala| 60 +
 2 files changed, 106 insertions(+), 17 deletions(-)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (86b54f3 -> abc8ccc)

2020-06-18 Thread wenchen
This is an automated email from the ASF dual-hosted git repository.

wenchen pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from 86b54f3  [SPARK-31894][SS] Introduce UnsafeRow format validation for 
streaming state store
 add abc8ccc  [SPARK-31926][SQL][TESTS][FOLLOWUP][TEST-HIVE1.2][TEST-MAVEN] 
Fix concurrency issue for ThriftCLIService to getPortNumber

No new revisions were added by this update.

Summary of changes:
 project/SparkBuild.scala   |  1 -
 .../sql/hive/thriftserver/SparkSQLCLIService.scala | 50 --
 .../src/test/resources/log4j.properties|  2 +-
 .../sql/hive/thriftserver/SharedThriftServer.scala | 60 --
 .../thriftserver/ThriftServerQueryTestSuite.scala  |  3 ++
 .../ThriftServerWithSparkContextSuite.scala| 11 +++-
 .../service/cli/thrift/ThriftBinaryCLIService.java | 11 +++-
 .../hive/service/cli/thrift/ThriftCLIService.java  |  3 ++
 .../service/cli/thrift/ThriftHttpCLIService.java   | 21 +---
 .../service/cli/thrift/ThriftBinaryCLIService.java | 11 +++-
 .../hive/service/cli/thrift/ThriftCLIService.java  |  3 ++
 .../service/cli/thrift/ThriftHttpCLIService.java   | 21 +---
 12 files changed, 160 insertions(+), 37 deletions(-)
 copy sql/{hive => hive-thriftserver}/src/test/resources/log4j.properties (96%)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (86b54f3 -> abc8ccc)

2020-06-18 Thread wenchen
This is an automated email from the ASF dual-hosted git repository.

wenchen pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from 86b54f3  [SPARK-31894][SS] Introduce UnsafeRow format validation for 
streaming state store
 add abc8ccc  [SPARK-31926][SQL][TESTS][FOLLOWUP][TEST-HIVE1.2][TEST-MAVEN] 
Fix concurrency issue for ThriftCLIService to getPortNumber

No new revisions were added by this update.

Summary of changes:
 project/SparkBuild.scala   |  1 -
 .../sql/hive/thriftserver/SparkSQLCLIService.scala | 50 --
 .../src/test/resources/log4j.properties|  2 +-
 .../sql/hive/thriftserver/SharedThriftServer.scala | 60 --
 .../thriftserver/ThriftServerQueryTestSuite.scala  |  3 ++
 .../ThriftServerWithSparkContextSuite.scala| 11 +++-
 .../service/cli/thrift/ThriftBinaryCLIService.java | 11 +++-
 .../hive/service/cli/thrift/ThriftCLIService.java  |  3 ++
 .../service/cli/thrift/ThriftHttpCLIService.java   | 21 +---
 .../service/cli/thrift/ThriftBinaryCLIService.java | 11 +++-
 .../hive/service/cli/thrift/ThriftCLIService.java  |  3 ++
 .../service/cli/thrift/ThriftHttpCLIService.java   | 21 +---
 12 files changed, 160 insertions(+), 37 deletions(-)
 copy sql/{hive => hive-thriftserver}/src/test/resources/log4j.properties (96%)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (86b54f3 -> abc8ccc)

2020-06-18 Thread wenchen
This is an automated email from the ASF dual-hosted git repository.

wenchen pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from 86b54f3  [SPARK-31894][SS] Introduce UnsafeRow format validation for 
streaming state store
 add abc8ccc  [SPARK-31926][SQL][TESTS][FOLLOWUP][TEST-HIVE1.2][TEST-MAVEN] 
Fix concurrency issue for ThriftCLIService to getPortNumber

No new revisions were added by this update.

Summary of changes:
 project/SparkBuild.scala   |  1 -
 .../sql/hive/thriftserver/SparkSQLCLIService.scala | 50 --
 .../src/test/resources/log4j.properties|  2 +-
 .../sql/hive/thriftserver/SharedThriftServer.scala | 60 --
 .../thriftserver/ThriftServerQueryTestSuite.scala  |  3 ++
 .../ThriftServerWithSparkContextSuite.scala| 11 +++-
 .../service/cli/thrift/ThriftBinaryCLIService.java | 11 +++-
 .../hive/service/cli/thrift/ThriftCLIService.java  |  3 ++
 .../service/cli/thrift/ThriftHttpCLIService.java   | 21 +---
 .../service/cli/thrift/ThriftBinaryCLIService.java | 11 +++-
 .../hive/service/cli/thrift/ThriftCLIService.java  |  3 ++
 .../service/cli/thrift/ThriftHttpCLIService.java   | 21 +---
 12 files changed, 160 insertions(+), 37 deletions(-)
 copy sql/{hive => hive-thriftserver}/src/test/resources/log4j.properties (96%)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (17a5007 -> 86b54f3)

2020-06-18 Thread wenchen
This is an automated email from the ASF dual-hosted git repository.

wenchen pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from 17a5007  [SPARK-30865][SQL][SS] Refactor DateTimeUtils
 add 86b54f3  [SPARK-31894][SS] Introduce UnsafeRow format validation for 
streaming state store

No new revisions were added by this update.

Summary of changes:
 .../spark/sql/catalyst/util/UnsafeRowUtils.scala   | 86 ++
 .../org/apache/spark/sql/internal/SQLConf.scala| 16 +++-
 .../sql/catalyst/util/UnsafeRowUtilsSuite.scala| 55 ++
 .../state/HDFSBackedStateStoreProvider.scala   | 13 
 .../sql/execution/streaming/state/StateStore.scala | 33 -
 .../execution/streaming/state/StateStoreConf.scala | 19 -
 .../execution/streaming/state/StateStoreRDD.scala  |  5 +-
 .../sql/execution/streaming/state/package.scala|  6 +-
 .../execution/streaming/statefulOperators.scala|  6 +-
 9 files changed, 228 insertions(+), 11 deletions(-)
 create mode 100644 
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/util/UnsafeRowUtils.scala
 create mode 100644 
sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/util/UnsafeRowUtilsSuite.scala


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (17a5007 -> 86b54f3)

2020-06-18 Thread wenchen
This is an automated email from the ASF dual-hosted git repository.

wenchen pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from 17a5007  [SPARK-30865][SQL][SS] Refactor DateTimeUtils
 add 86b54f3  [SPARK-31894][SS] Introduce UnsafeRow format validation for 
streaming state store

No new revisions were added by this update.

Summary of changes:
 .../spark/sql/catalyst/util/UnsafeRowUtils.scala   | 86 ++
 .../org/apache/spark/sql/internal/SQLConf.scala| 16 +++-
 .../sql/catalyst/util/UnsafeRowUtilsSuite.scala| 55 ++
 .../state/HDFSBackedStateStoreProvider.scala   | 13 
 .../sql/execution/streaming/state/StateStore.scala | 33 -
 .../execution/streaming/state/StateStoreConf.scala | 19 -
 .../execution/streaming/state/StateStoreRDD.scala  |  5 +-
 .../sql/execution/streaming/state/package.scala|  6 +-
 .../execution/streaming/statefulOperators.scala|  6 +-
 9 files changed, 228 insertions(+), 11 deletions(-)
 create mode 100644 
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/util/UnsafeRowUtils.scala
 create mode 100644 
sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/util/UnsafeRowUtilsSuite.scala


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (8750363 -> 17a5007)

2020-06-18 Thread wenchen
This is an automated email from the ASF dual-hosted git repository.

wenchen pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from 8750363  [MINOR][DOCS] Emphasize the Streaming tab is for DStream API
 add 17a5007  [SPARK-30865][SQL][SS] Refactor DateTimeUtils

No new revisions were added by this update.

Summary of changes:
 .../spark/sql/catalyst/expressions/Cast.scala  |  13 +-
 .../catalyst/expressions/datetimeExpressions.scala |  12 +-
 .../spark/sql/catalyst/util/DateTimeUtils.scala| 393 ++---
 .../sql/catalyst/util/TimestampFormatter.scala |   6 +-
 .../spark/sql/catalyst/expressions/CastSuite.scala |  13 +-
 .../catalyst/expressions/CsvExpressionsSuite.scala |   4 +-
 .../expressions/DateExpressionsSuite.scala |   5 +-
 .../expressions/ObjectExpressionsSuite.scala   |   4 +-
 .../sql/catalyst/json/JacksonGeneratorSuite.scala  |   8 +-
 .../sql/catalyst/util/DateTimeUtilsSuite.scala |  10 +-
 .../sql/catalyst/util/DatetimeFormatterSuite.scala |   2 +-
 .../sql/catalyst/util/RebaseDateTimeSuite.scala|   2 +-
 .../apache/spark/sql/util/ArrowUtilsSuite.scala|   5 +-
 .../execution/datasources/jdbc/JDBCRelation.scala  |   2 +-
 .../datasources/parquet/ParquetFilters.scala   |   3 +-
 .../datasources/parquet/ParquetRowConverter.scala  |   3 +-
 .../apache/spark/sql/DataFrameFunctionsSuite.scala |   6 +-
 .../spark/sql/StatisticsCollectionSuite.scala  |   7 +-
 .../sql/execution/datasources/json/JsonSuite.scala |   4 +-
 .../parquet/ParquetPartitionDiscoverySuite.scala   |   5 +-
 .../sql/sources/ResolvedDataSourceSuite.scala  |   4 +-
 .../apache/spark/sql/streaming/StreamSuite.scala   |   4 +-
 22 files changed, 241 insertions(+), 274 deletions(-)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (8750363 -> 17a5007)

2020-06-18 Thread wenchen
This is an automated email from the ASF dual-hosted git repository.

wenchen pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from 8750363  [MINOR][DOCS] Emphasize the Streaming tab is for DStream API
 add 17a5007  [SPARK-30865][SQL][SS] Refactor DateTimeUtils

No new revisions were added by this update.

Summary of changes:
 .../spark/sql/catalyst/expressions/Cast.scala  |  13 +-
 .../catalyst/expressions/datetimeExpressions.scala |  12 +-
 .../spark/sql/catalyst/util/DateTimeUtils.scala| 393 ++---
 .../sql/catalyst/util/TimestampFormatter.scala |   6 +-
 .../spark/sql/catalyst/expressions/CastSuite.scala |  13 +-
 .../catalyst/expressions/CsvExpressionsSuite.scala |   4 +-
 .../expressions/DateExpressionsSuite.scala |   5 +-
 .../expressions/ObjectExpressionsSuite.scala   |   4 +-
 .../sql/catalyst/json/JacksonGeneratorSuite.scala  |   8 +-
 .../sql/catalyst/util/DateTimeUtilsSuite.scala |  10 +-
 .../sql/catalyst/util/DatetimeFormatterSuite.scala |   2 +-
 .../sql/catalyst/util/RebaseDateTimeSuite.scala|   2 +-
 .../apache/spark/sql/util/ArrowUtilsSuite.scala|   5 +-
 .../execution/datasources/jdbc/JDBCRelation.scala  |   2 +-
 .../datasources/parquet/ParquetFilters.scala   |   3 +-
 .../datasources/parquet/ParquetRowConverter.scala  |   3 +-
 .../apache/spark/sql/DataFrameFunctionsSuite.scala |   6 +-
 .../spark/sql/StatisticsCollectionSuite.scala  |   7 +-
 .../sql/execution/datasources/json/JsonSuite.scala |   4 +-
 .../parquet/ParquetPartitionDiscoverySuite.scala   |   5 +-
 .../sql/sources/ResolvedDataSourceSuite.scala  |   4 +-
 .../apache/spark/sql/streaming/StreamSuite.scala   |   4 +-
 22 files changed, 241 insertions(+), 274 deletions(-)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (8750363 -> 17a5007)

2020-06-18 Thread wenchen
This is an automated email from the ASF dual-hosted git repository.

wenchen pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from 8750363  [MINOR][DOCS] Emphasize the Streaming tab is for DStream API
 add 17a5007  [SPARK-30865][SQL][SS] Refactor DateTimeUtils

No new revisions were added by this update.

Summary of changes:
 .../spark/sql/catalyst/expressions/Cast.scala  |  13 +-
 .../catalyst/expressions/datetimeExpressions.scala |  12 +-
 .../spark/sql/catalyst/util/DateTimeUtils.scala| 393 ++---
 .../sql/catalyst/util/TimestampFormatter.scala |   6 +-
 .../spark/sql/catalyst/expressions/CastSuite.scala |  13 +-
 .../catalyst/expressions/CsvExpressionsSuite.scala |   4 +-
 .../expressions/DateExpressionsSuite.scala |   5 +-
 .../expressions/ObjectExpressionsSuite.scala   |   4 +-
 .../sql/catalyst/json/JacksonGeneratorSuite.scala  |   8 +-
 .../sql/catalyst/util/DateTimeUtilsSuite.scala |  10 +-
 .../sql/catalyst/util/DatetimeFormatterSuite.scala |   2 +-
 .../sql/catalyst/util/RebaseDateTimeSuite.scala|   2 +-
 .../apache/spark/sql/util/ArrowUtilsSuite.scala|   5 +-
 .../execution/datasources/jdbc/JDBCRelation.scala  |   2 +-
 .../datasources/parquet/ParquetFilters.scala   |   3 +-
 .../datasources/parquet/ParquetRowConverter.scala  |   3 +-
 .../apache/spark/sql/DataFrameFunctionsSuite.scala |   6 +-
 .../spark/sql/StatisticsCollectionSuite.scala  |   7 +-
 .../sql/execution/datasources/json/JsonSuite.scala |   4 +-
 .../parquet/ParquetPartitionDiscoverySuite.scala   |   5 +-
 .../sql/sources/ResolvedDataSourceSuite.scala  |   4 +-
 .../apache/spark/sql/streaming/StreamSuite.scala   |   4 +-
 22 files changed, 241 insertions(+), 274 deletions(-)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (ac98a9a -> 8750363)

2020-06-18 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from ac98a9a  [MINOR][DOCS] Update running-on-kubernetes.md
 add 8750363  [MINOR][DOCS] Emphasize the Streaming tab is for DStream API

No new revisions were added by this update.

Summary of changes:
 docs/web-ui.md | 8 
 1 file changed, 4 insertions(+), 4 deletions(-)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch branch-3.0 updated: [MINOR][DOCS] Emphasize the Streaming tab is for DStream API

2020-06-18 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a commit to branch branch-3.0
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-3.0 by this push:
 new 73af23c  [MINOR][DOCS] Emphasize the Streaming tab is for DStream API
73af23c is described below

commit 73af23cabee6aac471fbdc9f4a35d118517fad65
Author: Yuanjian Li 
AuthorDate: Fri Jun 19 12:17:40 2020 +0900

[MINOR][DOCS] Emphasize the Streaming tab is for DStream API

### What changes were proposed in this pull request?
Emphasize the Streaming tab is for DStream API.

### Why are the changes needed?
Some users reported that it's a little confusing of the streaming tab and 
structured streaming tab.

### Does this PR introduce _any_ user-facing change?
Document change.

### How was this patch tested?
N/A

Closes #28854 from xuanyuanking/minor-doc.

Authored-by: Yuanjian Li 
Signed-off-by: HyukjinKwon 
(cherry picked from commit 8750363c8ddaa649d52e6d9001b9871d379fc4e1)
Signed-off-by: HyukjinKwon 
---
 docs/web-ui.md | 8 
 1 file changed, 4 insertions(+), 4 deletions(-)

diff --git a/docs/web-ui.md b/docs/web-ui.md
index 6653b00..134a8c8 100644
--- a/docs/web-ui.md
+++ b/docs/web-ui.md
@@ -434,10 +434,10 @@ The tracked operations are listed as follows.
 As an early-release version, the statistics page is still under development 
and will be improved in
 future releases.
 
-## Streaming Tab
-The web UI includes a Streaming tab if the application uses Spark streaming. 
This tab displays
-scheduling delay and processing time for each micro-batch in the data stream, 
which can be useful
-for troubleshooting the streaming application.
+## Streaming (DStreams) Tab
+The web UI includes a Streaming tab if the application uses Spark Streaming 
with DStream API.
+This tab displays scheduling delay and processing time for each micro-batch in 
the data stream,
+which can be useful for troubleshooting the streaming application.
 
 ## JDBC/ODBC Server Tab
 We can see this tab when Spark is running as a [distributed SQL 
engine](sql-distributed-sql-engine.html). It shows information about sessions 
and submitted SQL operations.


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (ac98a9a -> 8750363)

2020-06-18 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from ac98a9a  [MINOR][DOCS] Update running-on-kubernetes.md
 add 8750363  [MINOR][DOCS] Emphasize the Streaming tab is for DStream API

No new revisions were added by this update.

Summary of changes:
 docs/web-ui.md | 8 
 1 file changed, 4 insertions(+), 4 deletions(-)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch branch-3.0 updated: [MINOR][DOCS] Emphasize the Streaming tab is for DStream API

2020-06-18 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a commit to branch branch-3.0
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-3.0 by this push:
 new 73af23c  [MINOR][DOCS] Emphasize the Streaming tab is for DStream API
73af23c is described below

commit 73af23cabee6aac471fbdc9f4a35d118517fad65
Author: Yuanjian Li 
AuthorDate: Fri Jun 19 12:17:40 2020 +0900

[MINOR][DOCS] Emphasize the Streaming tab is for DStream API

### What changes were proposed in this pull request?
Emphasize the Streaming tab is for DStream API.

### Why are the changes needed?
Some users reported that it's a little confusing of the streaming tab and 
structured streaming tab.

### Does this PR introduce _any_ user-facing change?
Document change.

### How was this patch tested?
N/A

Closes #28854 from xuanyuanking/minor-doc.

Authored-by: Yuanjian Li 
Signed-off-by: HyukjinKwon 
(cherry picked from commit 8750363c8ddaa649d52e6d9001b9871d379fc4e1)
Signed-off-by: HyukjinKwon 
---
 docs/web-ui.md | 8 
 1 file changed, 4 insertions(+), 4 deletions(-)

diff --git a/docs/web-ui.md b/docs/web-ui.md
index 6653b00..134a8c8 100644
--- a/docs/web-ui.md
+++ b/docs/web-ui.md
@@ -434,10 +434,10 @@ The tracked operations are listed as follows.
 As an early-release version, the statistics page is still under development 
and will be improved in
 future releases.
 
-## Streaming Tab
-The web UI includes a Streaming tab if the application uses Spark streaming. 
This tab displays
-scheduling delay and processing time for each micro-batch in the data stream, 
which can be useful
-for troubleshooting the streaming application.
+## Streaming (DStreams) Tab
+The web UI includes a Streaming tab if the application uses Spark Streaming 
with DStream API.
+This tab displays scheduling delay and processing time for each micro-batch in 
the data stream,
+which can be useful for troubleshooting the streaming application.
 
 ## JDBC/ODBC Server Tab
 We can see this tab when Spark is running as a [distributed SQL 
engine](sql-distributed-sql-engine.html). It shows information about sessions 
and submitted SQL operations.


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (ac98a9a -> 8750363)

2020-06-18 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from ac98a9a  [MINOR][DOCS] Update running-on-kubernetes.md
 add 8750363  [MINOR][DOCS] Emphasize the Streaming tab is for DStream API

No new revisions were added by this update.

Summary of changes:
 docs/web-ui.md | 8 
 1 file changed, 4 insertions(+), 4 deletions(-)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch branch-3.0 updated: [MINOR][DOCS] Emphasize the Streaming tab is for DStream API

2020-06-18 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a commit to branch branch-3.0
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-3.0 by this push:
 new 73af23c  [MINOR][DOCS] Emphasize the Streaming tab is for DStream API
73af23c is described below

commit 73af23cabee6aac471fbdc9f4a35d118517fad65
Author: Yuanjian Li 
AuthorDate: Fri Jun 19 12:17:40 2020 +0900

[MINOR][DOCS] Emphasize the Streaming tab is for DStream API

### What changes were proposed in this pull request?
Emphasize the Streaming tab is for DStream API.

### Why are the changes needed?
Some users reported that it's a little confusing of the streaming tab and 
structured streaming tab.

### Does this PR introduce _any_ user-facing change?
Document change.

### How was this patch tested?
N/A

Closes #28854 from xuanyuanking/minor-doc.

Authored-by: Yuanjian Li 
Signed-off-by: HyukjinKwon 
(cherry picked from commit 8750363c8ddaa649d52e6d9001b9871d379fc4e1)
Signed-off-by: HyukjinKwon 
---
 docs/web-ui.md | 8 
 1 file changed, 4 insertions(+), 4 deletions(-)

diff --git a/docs/web-ui.md b/docs/web-ui.md
index 6653b00..134a8c8 100644
--- a/docs/web-ui.md
+++ b/docs/web-ui.md
@@ -434,10 +434,10 @@ The tracked operations are listed as follows.
 As an early-release version, the statistics page is still under development 
and will be improved in
 future releases.
 
-## Streaming Tab
-The web UI includes a Streaming tab if the application uses Spark streaming. 
This tab displays
-scheduling delay and processing time for each micro-batch in the data stream, 
which can be useful
-for troubleshooting the streaming application.
+## Streaming (DStreams) Tab
+The web UI includes a Streaming tab if the application uses Spark Streaming 
with DStream API.
+This tab displays scheduling delay and processing time for each micro-batch in 
the data stream,
+which can be useful for troubleshooting the streaming application.
 
 ## JDBC/ODBC Server Tab
 We can see this tab when Spark is running as a [distributed SQL 
engine](sql-distributed-sql-engine.html). It shows information about sessions 
and submitted SQL operations.


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (ac98a9a -> 8750363)

2020-06-18 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from ac98a9a  [MINOR][DOCS] Update running-on-kubernetes.md
 add 8750363  [MINOR][DOCS] Emphasize the Streaming tab is for DStream API

No new revisions were added by this update.

Summary of changes:
 docs/web-ui.md | 8 
 1 file changed, 4 insertions(+), 4 deletions(-)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch branch-3.0 updated: [MINOR][DOCS] Emphasize the Streaming tab is for DStream API

2020-06-18 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a commit to branch branch-3.0
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-3.0 by this push:
 new 73af23c  [MINOR][DOCS] Emphasize the Streaming tab is for DStream API
73af23c is described below

commit 73af23cabee6aac471fbdc9f4a35d118517fad65
Author: Yuanjian Li 
AuthorDate: Fri Jun 19 12:17:40 2020 +0900

[MINOR][DOCS] Emphasize the Streaming tab is for DStream API

### What changes were proposed in this pull request?
Emphasize the Streaming tab is for DStream API.

### Why are the changes needed?
Some users reported that it's a little confusing of the streaming tab and 
structured streaming tab.

### Does this PR introduce _any_ user-facing change?
Document change.

### How was this patch tested?
N/A

Closes #28854 from xuanyuanking/minor-doc.

Authored-by: Yuanjian Li 
Signed-off-by: HyukjinKwon 
(cherry picked from commit 8750363c8ddaa649d52e6d9001b9871d379fc4e1)
Signed-off-by: HyukjinKwon 
---
 docs/web-ui.md | 8 
 1 file changed, 4 insertions(+), 4 deletions(-)

diff --git a/docs/web-ui.md b/docs/web-ui.md
index 6653b00..134a8c8 100644
--- a/docs/web-ui.md
+++ b/docs/web-ui.md
@@ -434,10 +434,10 @@ The tracked operations are listed as follows.
 As an early-release version, the statistics page is still under development 
and will be improved in
 future releases.
 
-## Streaming Tab
-The web UI includes a Streaming tab if the application uses Spark streaming. 
This tab displays
-scheduling delay and processing time for each micro-batch in the data stream, 
which can be useful
-for troubleshooting the streaming application.
+## Streaming (DStreams) Tab
+The web UI includes a Streaming tab if the application uses Spark Streaming 
with DStream API.
+This tab displays scheduling delay and processing time for each micro-batch in 
the data stream,
+which can be useful for troubleshooting the streaming application.
 
 ## JDBC/ODBC Server Tab
 We can see this tab when Spark is running as a [distributed SQL 
engine](sql-distributed-sql-engine.html). It shows information about sessions 
and submitted SQL operations.


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (ac98a9a -> 8750363)

2020-06-18 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from ac98a9a  [MINOR][DOCS] Update running-on-kubernetes.md
 add 8750363  [MINOR][DOCS] Emphasize the Streaming tab is for DStream API

No new revisions were added by this update.

Summary of changes:
 docs/web-ui.md | 8 
 1 file changed, 4 insertions(+), 4 deletions(-)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch branch-3.0 updated: [MINOR][DOCS] Emphasize the Streaming tab is for DStream API

2020-06-18 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a commit to branch branch-3.0
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-3.0 by this push:
 new 73af23c  [MINOR][DOCS] Emphasize the Streaming tab is for DStream API
73af23c is described below

commit 73af23cabee6aac471fbdc9f4a35d118517fad65
Author: Yuanjian Li 
AuthorDate: Fri Jun 19 12:17:40 2020 +0900

[MINOR][DOCS] Emphasize the Streaming tab is for DStream API

### What changes were proposed in this pull request?
Emphasize the Streaming tab is for DStream API.

### Why are the changes needed?
Some users reported that it's a little confusing of the streaming tab and 
structured streaming tab.

### Does this PR introduce _any_ user-facing change?
Document change.

### How was this patch tested?
N/A

Closes #28854 from xuanyuanking/minor-doc.

Authored-by: Yuanjian Li 
Signed-off-by: HyukjinKwon 
(cherry picked from commit 8750363c8ddaa649d52e6d9001b9871d379fc4e1)
Signed-off-by: HyukjinKwon 
---
 docs/web-ui.md | 8 
 1 file changed, 4 insertions(+), 4 deletions(-)

diff --git a/docs/web-ui.md b/docs/web-ui.md
index 6653b00..134a8c8 100644
--- a/docs/web-ui.md
+++ b/docs/web-ui.md
@@ -434,10 +434,10 @@ The tracked operations are listed as follows.
 As an early-release version, the statistics page is still under development 
and will be improved in
 future releases.
 
-## Streaming Tab
-The web UI includes a Streaming tab if the application uses Spark streaming. 
This tab displays
-scheduling delay and processing time for each micro-batch in the data stream, 
which can be useful
-for troubleshooting the streaming application.
+## Streaming (DStreams) Tab
+The web UI includes a Streaming tab if the application uses Spark Streaming 
with DStream API.
+This tab displays scheduling delay and processing time for each micro-batch in 
the data stream,
+which can be useful for troubleshooting the streaming application.
 
 ## JDBC/ODBC Server Tab
 We can see this tab when Spark is running as a [distributed SQL 
engine](sql-distributed-sql-engine.html). It shows information about sessions 
and submitted SQL operations.


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] dongjoon-hyun edited a comment on pull request #273: Update Release Window from 3.0 to 3.1

2020-06-18 Thread GitBox


dongjoon-hyun edited a comment on pull request #273:
URL: https://github.com/apache/spark-website/pull/273#issuecomment-646372759


   cc @dbtsai , @holdenk , @gatorsmile , @rxin , @cloud-fan , @srowen , 
@HyukjinKwon , @maropu , @rdblue 



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] dongjoon-hyun commented on pull request #273: Update Release Window from 3.0 to 3.1

2020-06-18 Thread GitBox


dongjoon-hyun commented on pull request #273:
URL: https://github.com/apache/spark-website/pull/273#issuecomment-646372759


   cc @dbtsai , @holdenk , @gatorsmile , @rxin , @cloud-fan , @srowen , 
@HyukjinKwon , @maropu 



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] dongjoon-hyun opened a new pull request #273: Update Release Window from 3.0 to 3.1

2020-06-18 Thread GitBox


dongjoon-hyun opened a new pull request #273:
URL: https://github.com/apache/spark-website/pull/273


   According to our release cadence, this PR aims to update release window from 
3.0 to 3.1.
   > In general, feature minor releases occur about every 6 months.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch branch-3.0 updated: [MINOR][DOCS] Update running-on-kubernetes.md

2020-06-18 Thread dongjoon
This is an automated email from the ASF dual-hosted git repository.

dongjoon pushed a commit to branch branch-3.0
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-3.0 by this push:
 new cfeb7fa  [MINOR][DOCS] Update running-on-kubernetes.md
cfeb7fa is described below

commit cfeb7faf215b006463c5e676ebf7a3317f7730f1
Author: James Yu 
AuthorDate: Thu Jun 18 14:36:20 2020 -0700

[MINOR][DOCS] Update running-on-kubernetes.md

### What changes were proposed in this pull request?

Fix executor container name typo. `executor` should be 
`spark-kubernetes-executor`.

### Why are the changes needed?

The Executor pod container name the users actually get from their 
Kubernetes clusters is different from that described in the documentation.

For example, below is what a user get from an executor pod.
```
Containers:
  spark-kubernetes-executor:
Container ID:  docker://
Image: 
Image ID:  
docker-pullable://.dkr.ecr.us-east-0.amazonaws.com/spark
Port:  7079/TCP
Host Port: 0/TCP
Args:
  executor
State:  Running
  Started:  Thu, 28 May 2020 05:54:04 -0700
Ready:  True
Restart Count:  0
Limits:
  memory:  16Gi
```

### Does this PR introduce _any_ user-facing change?

Document change.

### How was this patch tested?

N/A

Closes #28862 from yuj/patch-1.

Authored-by: James Yu 
Signed-off-by: Dongjoon Hyun 
(cherry picked from commit ac98a9a07fe3e76c16bd5bb03aca553e80629e01)
Signed-off-by: Dongjoon Hyun 
---
 docs/running-on-kubernetes.md | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/docs/running-on-kubernetes.md b/docs/running-on-kubernetes.md
index 3abb891..e6a000d 100644
--- a/docs/running-on-kubernetes.md
+++ b/docs/running-on-kubernetes.md
@@ -1333,7 +1333,7 @@ The following affect the driver and executor containers. 
All other containers in
   See description
   
 The container name will be assigned by spark ("spark-kubernetes-driver" 
for the driver container, and
-"executor" for each executor container) if not defined by the pod 
template. If the container is defined by the
+"spark-kubernetes-executor" for each executor container) if not defined by 
the pod template. If the container is defined by the
 template, the template's name will be used.
   
 


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch branch-3.0 updated: [MINOR][DOCS] Update running-on-kubernetes.md

2020-06-18 Thread dongjoon
This is an automated email from the ASF dual-hosted git repository.

dongjoon pushed a commit to branch branch-3.0
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-3.0 by this push:
 new cfeb7fa  [MINOR][DOCS] Update running-on-kubernetes.md
cfeb7fa is described below

commit cfeb7faf215b006463c5e676ebf7a3317f7730f1
Author: James Yu 
AuthorDate: Thu Jun 18 14:36:20 2020 -0700

[MINOR][DOCS] Update running-on-kubernetes.md

### What changes were proposed in this pull request?

Fix executor container name typo. `executor` should be 
`spark-kubernetes-executor`.

### Why are the changes needed?

The Executor pod container name the users actually get from their 
Kubernetes clusters is different from that described in the documentation.

For example, below is what a user get from an executor pod.
```
Containers:
  spark-kubernetes-executor:
Container ID:  docker://
Image: 
Image ID:  
docker-pullable://.dkr.ecr.us-east-0.amazonaws.com/spark
Port:  7079/TCP
Host Port: 0/TCP
Args:
  executor
State:  Running
  Started:  Thu, 28 May 2020 05:54:04 -0700
Ready:  True
Restart Count:  0
Limits:
  memory:  16Gi
```

### Does this PR introduce _any_ user-facing change?

Document change.

### How was this patch tested?

N/A

Closes #28862 from yuj/patch-1.

Authored-by: James Yu 
Signed-off-by: Dongjoon Hyun 
(cherry picked from commit ac98a9a07fe3e76c16bd5bb03aca553e80629e01)
Signed-off-by: Dongjoon Hyun 
---
 docs/running-on-kubernetes.md | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/docs/running-on-kubernetes.md b/docs/running-on-kubernetes.md
index 3abb891..e6a000d 100644
--- a/docs/running-on-kubernetes.md
+++ b/docs/running-on-kubernetes.md
@@ -1333,7 +1333,7 @@ The following affect the driver and executor containers. 
All other containers in
   See description
   
 The container name will be assigned by spark ("spark-kubernetes-driver" 
for the driver container, and
-"executor" for each executor container) if not defined by the pod 
template. If the container is defined by the
+"spark-kubernetes-executor" for each executor container) if not defined by 
the pod template. If the container is defined by the
 template, the template's name will be used.
   
 


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (8a9ae01 -> ac98a9a)

2020-06-18 Thread dongjoon
This is an automated email from the ASF dual-hosted git repository.

dongjoon pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from 8a9ae01  [MINOR] update dev/create-release/known_translations
 add ac98a9a  [MINOR][DOCS] Update running-on-kubernetes.md

No new revisions were added by this update.

Summary of changes:
 docs/running-on-kubernetes.md | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch branch-3.0 updated: [MINOR][DOCS] Update running-on-kubernetes.md

2020-06-18 Thread dongjoon
This is an automated email from the ASF dual-hosted git repository.

dongjoon pushed a commit to branch branch-3.0
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-3.0 by this push:
 new cfeb7fa  [MINOR][DOCS] Update running-on-kubernetes.md
cfeb7fa is described below

commit cfeb7faf215b006463c5e676ebf7a3317f7730f1
Author: James Yu 
AuthorDate: Thu Jun 18 14:36:20 2020 -0700

[MINOR][DOCS] Update running-on-kubernetes.md

### What changes were proposed in this pull request?

Fix executor container name typo. `executor` should be 
`spark-kubernetes-executor`.

### Why are the changes needed?

The Executor pod container name the users actually get from their 
Kubernetes clusters is different from that described in the documentation.

For example, below is what a user get from an executor pod.
```
Containers:
  spark-kubernetes-executor:
Container ID:  docker://
Image: 
Image ID:  
docker-pullable://.dkr.ecr.us-east-0.amazonaws.com/spark
Port:  7079/TCP
Host Port: 0/TCP
Args:
  executor
State:  Running
  Started:  Thu, 28 May 2020 05:54:04 -0700
Ready:  True
Restart Count:  0
Limits:
  memory:  16Gi
```

### Does this PR introduce _any_ user-facing change?

Document change.

### How was this patch tested?

N/A

Closes #28862 from yuj/patch-1.

Authored-by: James Yu 
Signed-off-by: Dongjoon Hyun 
(cherry picked from commit ac98a9a07fe3e76c16bd5bb03aca553e80629e01)
Signed-off-by: Dongjoon Hyun 
---
 docs/running-on-kubernetes.md | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/docs/running-on-kubernetes.md b/docs/running-on-kubernetes.md
index 3abb891..e6a000d 100644
--- a/docs/running-on-kubernetes.md
+++ b/docs/running-on-kubernetes.md
@@ -1333,7 +1333,7 @@ The following affect the driver and executor containers. 
All other containers in
   See description
   
 The container name will be assigned by spark ("spark-kubernetes-driver" 
for the driver container, and
-"executor" for each executor container) if not defined by the pod 
template. If the container is defined by the
+"spark-kubernetes-executor" for each executor container) if not defined by 
the pod template. If the container is defined by the
 template, the template's name will be used.
   
 


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (8a9ae01 -> ac98a9a)

2020-06-18 Thread dongjoon
This is an automated email from the ASF dual-hosted git repository.

dongjoon pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from 8a9ae01  [MINOR] update dev/create-release/known_translations
 add ac98a9a  [MINOR][DOCS] Update running-on-kubernetes.md

No new revisions were added by this update.

Summary of changes:
 docs/running-on-kubernetes.md | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch branch-3.0 updated: [MINOR][DOCS] Update running-on-kubernetes.md

2020-06-18 Thread dongjoon
This is an automated email from the ASF dual-hosted git repository.

dongjoon pushed a commit to branch branch-3.0
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-3.0 by this push:
 new cfeb7fa  [MINOR][DOCS] Update running-on-kubernetes.md
cfeb7fa is described below

commit cfeb7faf215b006463c5e676ebf7a3317f7730f1
Author: James Yu 
AuthorDate: Thu Jun 18 14:36:20 2020 -0700

[MINOR][DOCS] Update running-on-kubernetes.md

### What changes were proposed in this pull request?

Fix executor container name typo. `executor` should be 
`spark-kubernetes-executor`.

### Why are the changes needed?

The Executor pod container name the users actually get from their 
Kubernetes clusters is different from that described in the documentation.

For example, below is what a user get from an executor pod.
```
Containers:
  spark-kubernetes-executor:
Container ID:  docker://
Image: 
Image ID:  
docker-pullable://.dkr.ecr.us-east-0.amazonaws.com/spark
Port:  7079/TCP
Host Port: 0/TCP
Args:
  executor
State:  Running
  Started:  Thu, 28 May 2020 05:54:04 -0700
Ready:  True
Restart Count:  0
Limits:
  memory:  16Gi
```

### Does this PR introduce _any_ user-facing change?

Document change.

### How was this patch tested?

N/A

Closes #28862 from yuj/patch-1.

Authored-by: James Yu 
Signed-off-by: Dongjoon Hyun 
(cherry picked from commit ac98a9a07fe3e76c16bd5bb03aca553e80629e01)
Signed-off-by: Dongjoon Hyun 
---
 docs/running-on-kubernetes.md | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/docs/running-on-kubernetes.md b/docs/running-on-kubernetes.md
index 3abb891..e6a000d 100644
--- a/docs/running-on-kubernetes.md
+++ b/docs/running-on-kubernetes.md
@@ -1333,7 +1333,7 @@ The following affect the driver and executor containers. 
All other containers in
   See description
   
 The container name will be assigned by spark ("spark-kubernetes-driver" 
for the driver container, and
-"executor" for each executor container) if not defined by the pod 
template. If the container is defined by the
+"spark-kubernetes-executor" for each executor container) if not defined by 
the pod template. If the container is defined by the
 template, the template's name will be used.
   
 


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (8a9ae01 -> ac98a9a)

2020-06-18 Thread dongjoon
This is an automated email from the ASF dual-hosted git repository.

dongjoon pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from 8a9ae01  [MINOR] update dev/create-release/known_translations
 add ac98a9a  [MINOR][DOCS] Update running-on-kubernetes.md

No new revisions were added by this update.

Summary of changes:
 docs/running-on-kubernetes.md | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (8a9ae01 -> ac98a9a)

2020-06-18 Thread dongjoon
This is an automated email from the ASF dual-hosted git repository.

dongjoon pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from 8a9ae01  [MINOR] update dev/create-release/known_translations
 add ac98a9a  [MINOR][DOCS] Update running-on-kubernetes.md

No new revisions were added by this update.

Summary of changes:
 docs/running-on-kubernetes.md | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] dongjoon-hyun commented on pull request #272: Remove preview download direction.

2020-06-18 Thread GitBox


dongjoon-hyun commented on pull request #272:
URL: https://github.com/apache/spark-website/pull/272#issuecomment-646240304


   Thank you, @srowen .



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] dongjoon-hyun merged pull request #272: Remove preview download direction.

2020-06-18 Thread GitBox


dongjoon-hyun merged pull request #272:
URL: https://github.com/apache/spark-website/pull/272


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark-website] branch asf-site updated: Remove preview download direction.

2020-06-18 Thread dongjoon
This is an automated email from the ASF dual-hosted git repository.

dongjoon pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/spark-website.git


The following commit(s) were added to refs/heads/asf-site by this push:
 new 38eaa66  Remove preview download direction.
38eaa66 is described below

commit 38eaa66253bce92bbe9a46fd45d83dc46978fabf
Author: Dongjoon Hyun 
AuthorDate: Thu Jun 18 11:42:30 2020 -0700

Remove preview download direction.
---
 downloads.md| 1 -
 site/downloads.html | 3 +--
 2 files changed, 1 insertion(+), 3 deletions(-)

diff --git a/downloads.md b/downloads.md
index f385cfc..2ed9870 100644
--- a/downloads.md
+++ b/downloads.md
@@ -35,7 +35,6 @@ to satisfy the legal requirements of Apache Software 
Foundation's release policy
 Preview releases are not meant to be functional, i.e. they can and highly 
likely will contain
 critical bugs or documentation errors.
 The latest preview release is Spark 3.0.0-preview2, published on Dec 23, 2019.
-You can select and download it above.
 
 ### Link with Spark
 Spark artifacts are [hosted in Maven 
Central](https://search.maven.org/search?q=g:org.apache.spark). You can add a 
Maven dependency with the following coordinates:
diff --git a/site/downloads.html b/site/downloads.html
index e8f7da6..e3b060f 100644
--- a/site/downloads.html
+++ b/site/downloads.html
@@ -234,8 +234,7 @@ Unlike nightly packages, preview releases have been audited 
by the project’
 to satisfy the legal requirements of Apache Software Foundation’s 
release policy.
 Preview releases are not meant to be functional, i.e. they can and highly 
likely will contain
 critical bugs or documentation errors.
-The latest preview release is Spark 3.0.0-preview2, published on Dec 23, 2019.
-You can select and download it above.
+The latest preview release is Spark 3.0.0-preview2, published on Dec 23, 
2019.
 
 Link with Spark
 Spark artifacts are https://search.maven.org/search?q=g:org.apache.spark";>hosted in Maven 
Central. You can add a Maven dependency with the following coordinates:


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] dongjoon-hyun commented on pull request #272: Remove preview download direction.

2020-06-18 Thread GitBox


dongjoon-hyun commented on pull request #272:
URL: https://github.com/apache/spark-website/pull/272#issuecomment-646212348


   We may want to delete the whole section, `Latest Preview Release`, but for 
the the definition and warning looks worthy to have if we have a next preview 
release someday. So, I only remove the invalid part.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] dongjoon-hyun commented on pull request #272: Remove preview download direction.

2020-06-18 Thread GitBox


dongjoon-hyun commented on pull request #272:
URL: https://github.com/apache/spark-website/pull/272#issuecomment-646211844


   cc @cloud-fan , @gatorsmile , @srowen .



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] dongjoon-hyun opened a new pull request #272: Remove preview download direction.

2020-06-18 Thread GitBox


dongjoon-hyun opened a new pull request #272:
URL: https://github.com/apache/spark-website/pull/272


   This update the download direction because preview is removed from the 
dropbox.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] gatorsmile merged pull request #271: the date of the post should be the date you create it

2020-06-18 Thread GitBox


gatorsmile merged pull request #271:
URL: https://github.com/apache/spark-website/pull/271


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] cloud-fan commented on a change in pull request #271: the date of the post should be the date you create it

2020-06-18 Thread GitBox


cloud-fan commented on a change in pull request #271:
URL: https://github.com/apache/spark-website/pull/271#discussion_r442391584



##
File path: releases/_posts/2020-06-18-spark-release-3-0-0.md
##
@@ -11,7 +11,7 @@ meta:
   _wpas_done_all: '1'
 ---
 
-Apache Spark 3.0.0 is the first release of the 3.x line. Apache Spark 3.0 
builds on many of the innovations from Spark 2.x, bringing new ideas as well as 
continuing long-term projects that have been in development. With the help of 
tremendous contributions from the open-source community, this release resolved 
more than 3400 tickets as the result of contributions from over 440 
contributors.
+Apache Spark 3.0.0 is the first release of the 3.x line. The vote passed on 
the 10th of June, 2020. This release is based on git tag `v3.0.0` which 
includes all commits up to June 10. Apache Spark 3.0 builds on many of the 
innovations from Spark 2.x, bringing new ideas as well as continuing long-term 
projects that have been in development. With the help of tremendous 
contributions from the open-source community, this release resolved more than 
3400 tickets as the result of contributions from over 440 contributors.

Review comment:
   This change is necessary as we need to mention the vote pass date 
somewhere.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] dongjoon-hyun commented on pull request #271: the date of the post should be the date you create it

2020-06-18 Thread GitBox


dongjoon-hyun commented on pull request #271:
URL: https://github.com/apache/spark-website/pull/271#issuecomment-646203798


   If you think so, okay. @srowen .



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] srowen commented on pull request #271: the date of the post should be the date you create it

2020-06-18 Thread GitBox


srowen commented on pull request #271:
URL: https://github.com/apache/spark-website/pull/271#issuecomment-646202686


   Meh, I'm sure we have roughly the same review practices here. I'd also 
generally separate concerns. In practice, it's also OK to have a 'misc minor 
edits' PR. I know we want to make sure it's fixed up ASAP now that 3.0 is 
rolling out the door, announced. So probably no big deal to just make this one 
omnibus last minute tweak change.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] dongjoon-hyun edited a comment on pull request #271: the date of the post should be the date you create it

2020-06-18 Thread GitBox


dongjoon-hyun edited a comment on pull request #271:
URL: https://github.com/apache/spark-website/pull/271#issuecomment-646194278


   Hi, @cloud-fan , @rxin , @gatorsmile .
   This PR should change news date only~
   - 2.4.4 
(https://github.com/apache/spark-website/pull/217/files#diff-2e7c596ac11621f7b55e3299dd80f1b3R25)
   - 2.4.5 
(https://github.com/apache/spark-website/pull/261#discussion_r376745966)
   - 2.3.4 
(https://github.com/apache/spark-website/pull/221/files#diff-2e7c596ac11621f7b55e3299dd80f1b3R26)



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] dongjoon-hyun commented on pull request #271: the date of the post should the date you create it

2020-06-18 Thread GitBox


dongjoon-hyun commented on pull request #271:
URL: https://github.com/apache/spark-website/pull/271#issuecomment-646197761


   Please see the 2.3.4/2.4.5 and old practices.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] dongjoon-hyun edited a comment on pull request #271: the date of the post should the date you create it

2020-06-18 Thread GitBox


dongjoon-hyun edited a comment on pull request #271:
URL: https://github.com/apache/spark-website/pull/271#issuecomment-646194278


   Hi, @cloud-fan , @rxin , @gatorsmile .
   This PR should change news date only~
   - 2.4.5 
(https://github.com/apache/spark-website/pull/261#discussion_r376745966)
   - 2.3.4 
(https://github.com/apache/spark-website/pull/221/files#diff-2e7c596ac11621f7b55e3299dd80f1b3R26)



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] cloud-fan commented on pull request #271: the date of the post should the date you create it

2020-06-18 Thread GitBox


cloud-fan commented on pull request #271:
URL: https://github.com/apache/spark-website/pull/271#issuecomment-646197363


   It's not a Spark PR, and I don't think the Spark PR policies apply here... 
We just do what's the most convenient.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] dongjoon-hyun commented on pull request #271: the date of the post should the date you create it

2020-06-18 Thread GitBox


dongjoon-hyun commented on pull request #271:
URL: https://github.com/apache/spark-website/pull/271#issuecomment-646194278


   Hi, @cloud-fan , @rxin , @gatorsmile .
   This PR should change news date only~



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] dongjoon-hyun commented on pull request #270: Release Spark 3.0.0

2020-06-18 Thread GitBox


dongjoon-hyun commented on pull request #270:
URL: https://github.com/apache/spark-website/pull/270#issuecomment-646192994


   Yes. News date should be today as I mentioned here 
(https://github.com/apache/spark-website/pull/270#discussion_r442350319)
   > The date of this part is different from news release date. 



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] gatorsmile commented on a change in pull request #271: the date of the post should the date you create it

2020-06-18 Thread GitBox


gatorsmile commented on a change in pull request #271:
URL: https://github.com/apache/spark-website/pull/271#discussion_r442380758



##
File path: releases/_posts/2020-06-18-spark-release-3-0-0.md
##
@@ -315,18 +315,18 @@ Please read the [migration 
guide](https://spark.apache.org/docs/3.0.0/sparkr-mig
 
 ### Known Issues
 
-  - Streaming queries with `dropDuplicates` operator may not be able to 
restart with the checkpoint written by Spark 2.x. 
([SPARK-31990](https://issues.apache.org/jira/browse/SPARK-31990))
-  - In Web UI, the job list page may hang for more than 40 seconds. 
([SPARK-31967](https://issues.apache.org/jira/browse/SPARK-31967))
+  - Streaming queries with `dropDuplicates` operator may not be able to 
restart with the checkpoint written by Spark 2.x. This is fixed in the next 
release 3.0.1. 
([SPARK-31990](https://issues.apache.org/jira/browse/SPARK-31990))
+  - In Web UI, the job list page may hang for more than 40 seconds. This is 
fixed in the next release 3.0.1. 
([SPARK-31967](https://issues.apache.org/jira/browse/SPARK-31967))
   - Set `io.netty.tryReflectionSetAccessible` for Arrow on JDK9+ 
([SPARK-29923](https://issues.apache.org/jira/browse/SPARK-29923))
   - With AWS SDK upgrade to 1.11.655, we strongly encourage the users that use 
S3N file system (open-source NativeS3FileSystem that is based on jets3t 
library) on Hadoop 2.7.3 to upgrade to use AWS Signature V4 and set the bucket 
endpoint or migrate to S3A (“s3a://” prefix) - jets3t library uses AWS v2 by 
default and s3.amazonaws.com as an endpoint. Otherwise, the 403 Forbidden error 
may be thrown in the following cases:
 - If a user accesses an S3 path that contains “+” characters and uses the 
legacy S3N file system, e.g. s3n://bucket/path/+file.
 - If a user has configured AWS V2 signature to sign requests to S3 with 
S3N file system.
 
 Note that if you use S3AFileSystem, e.g. (“s3a://bucket/path”) to access 
S3 in S3Select or SQS connectors, then everything will work as expected. 
([SPARK-30968](https://issues.apache.org/jira/browse/SPARK-30968))
 
-  - Parsing day of year using pattern letter 'D' returns the wrong result if 
the year field is missing. This can happen in SQL functions like `to_timestamp` 
which parses datetime string to datetime values using a pattern string. 
([SPARK-31939](https://issues.apache.org/jira/browse/SPARK-31939))
-  - Join/Window/Aggregate inside subqueries may lead to wrong results if the 
keys have values -0.0 and 0.0. 
([SPARK-31958](https://issues.apache.org/jira/browse/SPARK-31958))
-  - A window query may fail with ambiguous self-join error unexpectedly. 
([SPARK-31956](https://issues.apache.org/jira/browse/SPARK-31956))
+  - Parsing day of year using pattern letter 'D' returns the wrong result if 
the year field is missing. This can happen in SQL functions like `to_timestamp` 
which parses datetime string to datetime values using a pattern string. This is 
fixed in the next release 3.0.1. 
([SPARK-31939](https://issues.apache.org/jira/browse/SPARK-31939))

Review comment:
   => This will be fixed in the next release 3.0.1.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] srowen commented on pull request #271: the date of the post should the date you create it

2020-06-18 Thread GitBox


srowen commented on pull request #271:
URL: https://github.com/apache/spark-website/pull/271#issuecomment-646186207


   I agree with that take. The VOTE passing is just an internal procedural 
step. The release date is when it's fully available to end users, which isn't 
the same thing.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] cloud-fan commented on a change in pull request #271: the date of the post should the date you create it

2020-06-18 Thread GitBox


cloud-fan commented on a change in pull request #271:
URL: https://github.com/apache/spark-website/pull/271#discussion_r442373302



##
File path: release-process.md
##
@@ -259,7 +259,8 @@ Then, create the release notes. Go to the
 https://issues.apache.org/jira/projects/SPARK?selectedItem=com.atlassian.jira.jira-projects-plugin:release-page";>release
 page in JIRA,
 pick the release version from the list, then click on "Release Notes". Copy 
this URL and then make a short URL on
 https://s.apache.org/";>s.apache.org, sign in to your Apache 
account, and pick the ID as something like
-`spark-2.1.2`. Create a new release post under `releases/_posts` to include 
this short URL.
+`spark-2.1.2`. Create a new release post under `releases/_posts` to include 
this short URL. The date of the post should
+be the date you create it.

Review comment:
   I put it here so future releases can just follow it.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] cloud-fan commented on pull request #271: the date of the post should the date you create it

2020-06-18 Thread GitBox


cloud-fan commented on pull request #271:
URL: https://github.com/apache/spark-website/pull/271#issuecomment-646175051


   cc @dongjoon-hyun @rxin @srowen 



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] cloud-fan opened a new pull request #271: the date of the post should the date you create it

2020-06-18 Thread GitBox


cloud-fan opened a new pull request #271:
URL: https://github.com/apache/spark-website/pull/271


   After talking with @rxin offline, we think it's confusing to use the RC vote 
pass date as the date of the post. People may be wondering why they didn't see 
the post before, as the date is usually many days before.
   
   This PR tries to make it the policy clear: The date of the post should be 
the date you create it. This is the most intuitive IMO.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] gatorsmile commented on pull request #270: Release Spark 3.0.0

2020-06-18 Thread GitBox


gatorsmile commented on pull request #270:
URL: https://github.com/apache/spark-website/pull/270#issuecomment-646163523


   ![image 
(10)](https://user-images.githubusercontent.com/11567269/85048927-db1b9100-b148-11ea-9202-239a11fbcd63.png)
   Yes. Just checked the website, it shows a strange order. 



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] rxin commented on pull request #270: Release Spark 3.0.0

2020-06-18 Thread GitBox


rxin commented on pull request #270:
URL: https://github.com/apache/spark-website/pull/270#issuecomment-646162161


   I find it quite strange that the news date is not today. I think we should 
use today’s date and then say the vote passed on June 6. 



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] rxin edited a comment on pull request #270: Release Spark 3.0.0

2020-06-18 Thread GitBox


rxin edited a comment on pull request #270:
URL: https://github.com/apache/spark-website/pull/270#issuecomment-646162161


   I find it quite strange that the news date is not today. I think we should 
use today’s date and then say the vote passed on June 10 and this is based on 
tag v3.0.0 including all commits up to June 10. 



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] cloud-fan merged pull request #270: Release Spark 3.0.0

2020-06-18 Thread GitBox


cloud-fan merged pull request #270:
URL: https://github.com/apache/spark-website/pull/270


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



svn commit: r40088 - in /dev/spark: v3.0.0-rc1-bin/ v3.0.0-rc1-docs/ v3.0.0-rc2-bin/ v3.0.0-rc2-docs/ v3.0.0-rc3-docs/

2020-06-18 Thread rxin
Author: rxin
Date: Thu Jun 18 16:41:27 2020
New Revision: 40088

Log:
Removing RC artifacts.

Removed:
dev/spark/v3.0.0-rc1-bin/
dev/spark/v3.0.0-rc1-docs/
dev/spark/v3.0.0-rc2-bin/
dev/spark/v3.0.0-rc2-docs/
dev/spark/v3.0.0-rc3-docs/


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] cloud-fan commented on a change in pull request #270: Release Spark 3.0.0

2020-06-18 Thread GitBox


cloud-fan commented on a change in pull request #270:
URL: https://github.com/apache/spark-website/pull/270#discussion_r442358456



##
File path: js/downloads.js
##
@@ -23,7 +23,7 @@ var packagesV9 = [hadoop2p7, hadoop2p6, hadoopFree, 
scala2p12_hadoopFree, source
 // 3.0.0+
 var packagesV10 = [hadoop2p7, hadoop3p2, hadoopFree, sources];
 
-addRelease("3.0.0-preview2", new Date("12/23/2019"), packagesV10, true);
+addRelease("3.0.0", new Date("06/18/2020"), packagesV10, true);

Review comment:
   I recalled that we had a similar discussion at the time of 2.4.0 and we 
picked the RC vote pass date.
   
   I've updated to use June 10.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] srowen commented on a change in pull request #270: Release Spark 3.0.0

2020-06-18 Thread GitBox


srowen commented on a change in pull request #270:
URL: https://github.com/apache/spark-website/pull/270#discussion_r442355349



##
File path: js/downloads.js
##
@@ -23,7 +23,7 @@ var packagesV9 = [hadoop2p7, hadoop2p6, hadoopFree, 
scala2p12_hadoopFree, source
 // 3.0.0+
 var packagesV10 = [hadoop2p7, hadoop3p2, hadoopFree, sources];
 
-addRelease("3.0.0-preview2", new Date("12/23/2019"), packagesV10, true);
+addRelease("3.0.0", new Date("06/18/2020"), packagesV10, true);

Review comment:
   I don't think it should be the vote date - it wasn't available to users 
at that time. What you want to call the release date is a little subjective 
(when artifacts were uploaded? announced?) but anything like June 18 seems 
reasonable. It doesn't matter much.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] dongjoon-hyun commented on a change in pull request #270: Release Spark 3.0.0

2020-06-18 Thread GitBox


dongjoon-hyun commented on a change in pull request #270:
URL: https://github.com/apache/spark-website/pull/270#discussion_r442350319



##
File path: js/downloads.js
##
@@ -23,7 +23,7 @@ var packagesV9 = [hadoop2p7, hadoop2p6, hadoopFree, 
scala2p12_hadoopFree, source
 // 3.0.0+
 var packagesV10 = [hadoop2p7, hadoop3p2, hadoopFree, sources];
 
-addRelease("3.0.0-preview2", new Date("12/23/2019"), packagesV10, true);
+addRelease("3.0.0", new Date("06/18/2020"), packagesV10, true);

Review comment:
   The vote passed on `06/10/2020`, @cloud-fan . The date of this part is 
different from news release date. (cc @gatorsmile )
   - 
https://lists.apache.org/thread.html/r1f6109df463166b826b7b2d180f74885cd95ecfaa524d722b1e665d8%40%3Cdev.spark.apache.org%3E





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] dongjoon-hyun commented on a change in pull request #270: Release Spark 3.0.0

2020-06-18 Thread GitBox


dongjoon-hyun commented on a change in pull request #270:
URL: https://github.com/apache/spark-website/pull/270#discussion_r442350319



##
File path: js/downloads.js
##
@@ -23,7 +23,7 @@ var packagesV9 = [hadoop2p7, hadoop2p6, hadoopFree, 
scala2p12_hadoopFree, source
 // 3.0.0+
 var packagesV10 = [hadoop2p7, hadoop3p2, hadoopFree, sources];
 
-addRelease("3.0.0-preview2", new Date("12/23/2019"), packagesV10, true);
+addRelease("3.0.0", new Date("06/18/2020"), packagesV10, true);

Review comment:
   The vote passed on `06/10/2020`, @cloud-fan . This date is different 
from news release date. (cc @gatorsmile )
   - 
https://lists.apache.org/thread.html/r1f6109df463166b826b7b2d180f74885cd95ecfaa524d722b1e665d8%40%3Cdev.spark.apache.org%3E





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (e4f5036 -> 8a9ae01)

2020-06-18 Thread wenchen
This is an automated email from the ASF dual-hosted git repository.

wenchen pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from e4f5036  [SPARK-32020][SQL] Better error message when SPARK_HOME or 
spark.test.home is not set
 add 8a9ae01  [MINOR] update dev/create-release/known_translations

No new revisions were added by this update.

Summary of changes:
 dev/create-release/known_translations | 148 ++
 1 file changed, 148 insertions(+)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (e4f5036 -> 8a9ae01)

2020-06-18 Thread wenchen
This is an automated email from the ASF dual-hosted git repository.

wenchen pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from e4f5036  [SPARK-32020][SQL] Better error message when SPARK_HOME or 
spark.test.home is not set
 add 8a9ae01  [MINOR] update dev/create-release/known_translations

No new revisions were added by this update.

Summary of changes:
 dev/create-release/known_translations | 148 ++
 1 file changed, 148 insertions(+)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (e4f5036 -> 8a9ae01)

2020-06-18 Thread wenchen
This is an automated email from the ASF dual-hosted git repository.

wenchen pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from e4f5036  [SPARK-32020][SQL] Better error message when SPARK_HOME or 
spark.test.home is not set
 add 8a9ae01  [MINOR] update dev/create-release/known_translations

No new revisions were added by this update.

Summary of changes:
 dev/create-release/known_translations | 148 ++
 1 file changed, 148 insertions(+)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] cloud-fan commented on pull request #270: Release Spark 3.0.0

2020-06-18 Thread GitBox


cloud-fan commented on pull request #270:
URL: https://github.com/apache/spark-website/pull/270#issuecomment-646067724


   cc @rxin @gatorsmile @dongjoon-hyun @srowen 



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[GitHub] [spark-website] cloud-fan opened a new pull request #270: Release Spark 3.0.0

2020-06-18 Thread GitBox


cloud-fan opened a new pull request #270:
URL: https://github.com/apache/spark-website/pull/270


   Add the 3.0.0 [release 
notes](https://docs.google.com/document/d/1NrTqxf2f39AXDF8VTIch6kwD8VKPaIlLW1QvuqEcwR4/edit)
 and update the website.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch branch-3.0 updated: [SPARK-32020][SQL] Better error message when SPARK_HOME or spark.test.home is not set

2020-06-18 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a commit to branch branch-3.0
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-3.0 by this push:
 new 70f3ce6  [SPARK-32020][SQL] Better error message when SPARK_HOME or 
spark.test.home is not set
70f3ce6 is described below

commit 70f3ce69ee6ae01393e853ccea16a259804b5d85
Author: Dilip Biswal 
AuthorDate: Thu Jun 18 22:45:55 2020 +0900

[SPARK-32020][SQL] Better error message when SPARK_HOME or spark.test.home 
is not set

### What changes were proposed in this pull request?
Better error message when SPARK_HOME or spark,test.home is not set.

### Why are the changes needed?
Currently the error message is not easily consumable as it prints  (see 
below) the real error after printing the current environment which is rather 
long.

**Old output**
`
 time.name" -> "Java(TM) SE Runtime Environment", "sun.boot.library.path" 
-> "/Library/Java/JavaVirtualMachines/jdk1.8.0_221.jdk/Contents/Home/jre/lib",
 "java.vm.version" -> "25.221-b11",
 . . .
 . . .
 . . .
) did not contain key "SPARK_HOME" spark.test.home or SPARK_HOME is not set.
at org.scalatest.Assertions.newAssertionFailedExceptio
`

**New output**
An exception or error caused a run to abort: spark.test.home or SPARK_HOME 
is not set.
org.scalatest.exceptions.TestFailedException: spark.test.home or SPARK_HOME 
is not set
### Does this PR introduce any user-facing change?
`
No.

### How was this patch tested?
Ran the tests in intellej  manually to see the new error.

Closes #28825 from dilipbiswal/minor-spark-31950-followup.

Authored-by: Dilip Biswal 
Signed-off-by: HyukjinKwon 
(cherry picked from commit e4f503614625f81aa86455ad5f6bec3b61d525f1)
Signed-off-by: HyukjinKwon 
---
 .../sql/catalyst/parser/TableIdentifierParserSuite.scala |  6 --
 .../org/apache/spark/sql/catalyst/plans/SQLHelper.scala  |  9 +
 .../scala/org/apache/spark/sql/ExpressionsSchemaSuite.scala  |  6 --
 .../scala/org/apache/spark/sql/IntegratedUDFTestUtils.scala  |  9 +
 .../test/scala/org/apache/spark/sql/SQLQueryTestSuite.scala  | 12 +++-
 5 files changed, 13 insertions(+), 29 deletions(-)

diff --git 
a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/parser/TableIdentifierParserSuite.scala
 
b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/parser/TableIdentifierParserSuite.scala
index 04c427d..a721e17 100644
--- 
a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/parser/TableIdentifierParserSuite.scala
+++ 
b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/parser/TableIdentifierParserSuite.scala
@@ -18,7 +18,6 @@ package org.apache.spark.sql.catalyst.parser
 
 import java.io.File
 import java.nio.file.Files
-import java.util.Locale
 
 import scala.collection.JavaConverters._
 import scala.collection.mutable
@@ -295,11 +294,6 @@ class TableIdentifierParserSuite extends SparkFunSuite 
with SQLHelper {
 
   private val sqlSyntaxDefs = {
 val sqlBasePath = {
-  val sparkHome = {
-assert(sys.props.contains("spark.test.home") ||
-  sys.env.contains("SPARK_HOME"), "spark.test.home or SPARK_HOME is 
not set.")
-sys.props.getOrElse("spark.test.home", sys.env("SPARK_HOME"))
-  }
   java.nio.file.Paths.get(sparkHome, "sql", "catalyst", "src", "main", 
"antlr4", "org",
 "apache", "spark", "sql", "catalyst", "parser", "SqlBase.g4").toFile
 }
diff --git 
a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/plans/SQLHelper.scala
 
b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/plans/SQLHelper.scala
index d213743..5deab79 100644
--- 
a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/plans/SQLHelper.scala
+++ 
b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/plans/SQLHelper.scala
@@ -21,6 +21,8 @@ import java.time.ZoneId
 
 import scala.util.control.NonFatal
 
+import org.scalatest.Assertions.fail
+
 import org.apache.spark.sql.AnalysisException
 import org.apache.spark.sql.catalyst.util.DateTimeTestUtils
 import org.apache.spark.sql.catalyst.util.DateTimeUtils.getZoneId
@@ -83,4 +85,11 @@ trait SQLHelper {
   }
 }
   }
+
+  protected lazy val sparkHome: String = {
+if (!(sys.props.contains("spark.test.home") || 
sys.env.contains("SPARK_HOME"))) {
+  fail("spark.test.home or SPARK_HOME is not set.")
+}
+sys.props.getOrElse("spark.test.home", sys.env("SPARK_HOME"))
+  }
 }
diff --git 
a/sql/core/src/test/scala/org/apache/spark/sql/ExpressionsSchemaSuite.scala 
b/sql/core/src/test/scala/org/apache/spark/sql/ExpressionsSchemaSuite.scala
index d69ecd7..81c09d1 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/ExpressionsSchemaSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/ExpressionsSc

[spark] branch branch-3.0 updated: [SPARK-32020][SQL] Better error message when SPARK_HOME or spark.test.home is not set

2020-06-18 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a commit to branch branch-3.0
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-3.0 by this push:
 new 70f3ce6  [SPARK-32020][SQL] Better error message when SPARK_HOME or 
spark.test.home is not set
70f3ce6 is described below

commit 70f3ce69ee6ae01393e853ccea16a259804b5d85
Author: Dilip Biswal 
AuthorDate: Thu Jun 18 22:45:55 2020 +0900

[SPARK-32020][SQL] Better error message when SPARK_HOME or spark.test.home 
is not set

### What changes were proposed in this pull request?
Better error message when SPARK_HOME or spark,test.home is not set.

### Why are the changes needed?
Currently the error message is not easily consumable as it prints  (see 
below) the real error after printing the current environment which is rather 
long.

**Old output**
`
 time.name" -> "Java(TM) SE Runtime Environment", "sun.boot.library.path" 
-> "/Library/Java/JavaVirtualMachines/jdk1.8.0_221.jdk/Contents/Home/jre/lib",
 "java.vm.version" -> "25.221-b11",
 . . .
 . . .
 . . .
) did not contain key "SPARK_HOME" spark.test.home or SPARK_HOME is not set.
at org.scalatest.Assertions.newAssertionFailedExceptio
`

**New output**
An exception or error caused a run to abort: spark.test.home or SPARK_HOME 
is not set.
org.scalatest.exceptions.TestFailedException: spark.test.home or SPARK_HOME 
is not set
### Does this PR introduce any user-facing change?
`
No.

### How was this patch tested?
Ran the tests in intellej  manually to see the new error.

Closes #28825 from dilipbiswal/minor-spark-31950-followup.

Authored-by: Dilip Biswal 
Signed-off-by: HyukjinKwon 
(cherry picked from commit e4f503614625f81aa86455ad5f6bec3b61d525f1)
Signed-off-by: HyukjinKwon 
---
 .../sql/catalyst/parser/TableIdentifierParserSuite.scala |  6 --
 .../org/apache/spark/sql/catalyst/plans/SQLHelper.scala  |  9 +
 .../scala/org/apache/spark/sql/ExpressionsSchemaSuite.scala  |  6 --
 .../scala/org/apache/spark/sql/IntegratedUDFTestUtils.scala  |  9 +
 .../test/scala/org/apache/spark/sql/SQLQueryTestSuite.scala  | 12 +++-
 5 files changed, 13 insertions(+), 29 deletions(-)

diff --git 
a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/parser/TableIdentifierParserSuite.scala
 
b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/parser/TableIdentifierParserSuite.scala
index 04c427d..a721e17 100644
--- 
a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/parser/TableIdentifierParserSuite.scala
+++ 
b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/parser/TableIdentifierParserSuite.scala
@@ -18,7 +18,6 @@ package org.apache.spark.sql.catalyst.parser
 
 import java.io.File
 import java.nio.file.Files
-import java.util.Locale
 
 import scala.collection.JavaConverters._
 import scala.collection.mutable
@@ -295,11 +294,6 @@ class TableIdentifierParserSuite extends SparkFunSuite 
with SQLHelper {
 
   private val sqlSyntaxDefs = {
 val sqlBasePath = {
-  val sparkHome = {
-assert(sys.props.contains("spark.test.home") ||
-  sys.env.contains("SPARK_HOME"), "spark.test.home or SPARK_HOME is 
not set.")
-sys.props.getOrElse("spark.test.home", sys.env("SPARK_HOME"))
-  }
   java.nio.file.Paths.get(sparkHome, "sql", "catalyst", "src", "main", 
"antlr4", "org",
 "apache", "spark", "sql", "catalyst", "parser", "SqlBase.g4").toFile
 }
diff --git 
a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/plans/SQLHelper.scala
 
b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/plans/SQLHelper.scala
index d213743..5deab79 100644
--- 
a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/plans/SQLHelper.scala
+++ 
b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/plans/SQLHelper.scala
@@ -21,6 +21,8 @@ import java.time.ZoneId
 
 import scala.util.control.NonFatal
 
+import org.scalatest.Assertions.fail
+
 import org.apache.spark.sql.AnalysisException
 import org.apache.spark.sql.catalyst.util.DateTimeTestUtils
 import org.apache.spark.sql.catalyst.util.DateTimeUtils.getZoneId
@@ -83,4 +85,11 @@ trait SQLHelper {
   }
 }
   }
+
+  protected lazy val sparkHome: String = {
+if (!(sys.props.contains("spark.test.home") || 
sys.env.contains("SPARK_HOME"))) {
+  fail("spark.test.home or SPARK_HOME is not set.")
+}
+sys.props.getOrElse("spark.test.home", sys.env("SPARK_HOME"))
+  }
 }
diff --git 
a/sql/core/src/test/scala/org/apache/spark/sql/ExpressionsSchemaSuite.scala 
b/sql/core/src/test/scala/org/apache/spark/sql/ExpressionsSchemaSuite.scala
index d69ecd7..81c09d1 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/ExpressionsSchemaSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/ExpressionsSc

[spark] branch master updated (9b79251 -> e4f5036)

2020-06-18 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from 9b79251  [SPARK-31960][YARN][BUILD] Only populate Hadoop classpath for 
no-hadoop build
 add e4f5036  [SPARK-32020][SQL] Better error message when SPARK_HOME or 
spark.test.home is not set

No new revisions were added by this update.

Summary of changes:
 .../sql/catalyst/parser/TableIdentifierParserSuite.scala |  6 --
 .../org/apache/spark/sql/catalyst/plans/SQLHelper.scala  |  9 +
 .../scala/org/apache/spark/sql/ExpressionsSchemaSuite.scala  |  6 --
 .../scala/org/apache/spark/sql/IntegratedUDFTestUtils.scala  |  9 +
 .../test/scala/org/apache/spark/sql/SQLQueryTestSuite.scala  | 12 +++-
 5 files changed, 13 insertions(+), 29 deletions(-)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (9b79251 -> e4f5036)

2020-06-18 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from 9b79251  [SPARK-31960][YARN][BUILD] Only populate Hadoop classpath for 
no-hadoop build
 add e4f5036  [SPARK-32020][SQL] Better error message when SPARK_HOME or 
spark.test.home is not set

No new revisions were added by this update.

Summary of changes:
 .../sql/catalyst/parser/TableIdentifierParserSuite.scala |  6 --
 .../org/apache/spark/sql/catalyst/plans/SQLHelper.scala  |  9 +
 .../scala/org/apache/spark/sql/ExpressionsSchemaSuite.scala  |  6 --
 .../scala/org/apache/spark/sql/IntegratedUDFTestUtils.scala  |  9 +
 .../test/scala/org/apache/spark/sql/SQLQueryTestSuite.scala  | 12 +++-
 5 files changed, 13 insertions(+), 29 deletions(-)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org