[spark] branch branch-3.0 updated: [SPARK-31859][SPARK-31861][FOLLOWUP] Fix typo in tests

2020-06-05 Thread wenchen
This is an automated email from the ASF dual-hosted git repository.

wenchen pushed a commit to branch branch-3.0
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-3.0 by this push:
 new 8910dae  [SPARK-31859][SPARK-31861][FOLLOWUP] Fix typo in tests
8910dae is described below

commit 8910dae0966ec7bbd9886245310fa07b81426e74
Author: Juliusz Sompolski 
AuthorDate: Fri Jun 5 10:01:00 2020 +

[SPARK-31859][SPARK-31861][FOLLOWUP] Fix typo in tests

### What changes were proposed in this pull request?

It appears I have unintentionally used nested JDBC statements in the two 
tests I added.

### Why are the changes needed?

Cleanup a typo. Please merge to master/branch-3.0

### Does this PR introduce _any_ user-facing change?

No.

### How was this patch tested?

Unit tests.

Closes #28735 from juliuszsompolski/SPARK-31859-fixup.

Authored-by: Juliusz Sompolski 
Signed-off-by: Wenchen Fan 
(cherry picked from commit ea010138e95cc9c0f62603d431b906bd5104f4e3)
Signed-off-by: Wenchen Fan 
---
 .../thriftserver/HiveThriftServer2Suites.scala | 36 ++
 1 file changed, 16 insertions(+), 20 deletions(-)

diff --git 
a/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/HiveThriftServer2Suites.scala
 
b/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/HiveThriftServer2Suites.scala
index 21256ad..4e6d4e1 100644
--- 
a/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/HiveThriftServer2Suites.scala
+++ 
b/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/HiveThriftServer2Suites.scala
@@ -813,31 +813,27 @@ class HiveThriftBinaryServerSuite extends 
HiveThriftJdbcTest {
   }
 
   test("SPARK-31859 Thriftserver works with 
spark.sql.datetime.java8API.enabled=true") {
-withJdbcStatement() { statement =>
-  withJdbcStatement() { st =>
-st.execute("set spark.sql.datetime.java8API.enabled=true")
-val rs = st.executeQuery("select date '2020-05-28', timestamp 
'2020-05-28 00:00:00'")
-rs.next()
-assert(rs.getDate(1).toString() == "2020-05-28")
-assert(rs.getTimestamp(2).toString() == "2020-05-28 00:00:00.0")
-  }
+withJdbcStatement() { st =>
+  st.execute("set spark.sql.datetime.java8API.enabled=true")
+  val rs = st.executeQuery("select date '2020-05-28', timestamp 
'2020-05-28 00:00:00'")
+  rs.next()
+  assert(rs.getDate(1).toString() == "2020-05-28")
+  assert(rs.getTimestamp(2).toString() == "2020-05-28 00:00:00.0")
 }
   }
 
   test("SPARK-31861 Thriftserver respects spark.sql.session.timeZone") {
-withJdbcStatement() { statement =>
-  withJdbcStatement() { st =>
-st.execute("set spark.sql.session.timeZone=+03:15") // different than 
Thriftserver's JVM tz
+withJdbcStatement() { st =>
+  st.execute("set spark.sql.session.timeZone=+03:15") // different than 
Thriftserver's JVM tz
   val rs = st.executeQuery("select timestamp '2020-05-28 10:00:00'")
-rs.next()
-// The timestamp as string is the same as the literal
-assert(rs.getString(1) == "2020-05-28 10:00:00.0")
-// Parsing it to java.sql.Timestamp in the client will always result 
in a timestamp
-// in client default JVM timezone. The string value of the Timestamp 
will match the literal,
-// but if the JDBC application cares about the internal timezone and 
UTC offset of the
-// Timestamp object, it should set spark.sql.session.timeZone to match 
its client JVM tz.
-assert(rs.getTimestamp(1).toString() == "2020-05-28 10:00:00.0")
-  }
+  rs.next()
+  // The timestamp as string is the same as the literal
+  assert(rs.getString(1) == "2020-05-28 10:00:00.0")
+  // Parsing it to java.sql.Timestamp in the client will always result in 
a timestamp
+  // in client default JVM timezone. The string value of the Timestamp 
will match the literal,
+  // but if the JDBC application cares about the internal timezone and UTC 
offset of the
+  // Timestamp object, it should set spark.sql.session.timeZone to match 
its client JVM tz.
+  assert(rs.getTimestamp(1).toString() == "2020-05-28 10:00:00.0")
 }
   }
 


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch branch-3.0 updated: [SPARK-31859][SPARK-31861][FOLLOWUP] Fix typo in tests

2020-06-05 Thread wenchen
This is an automated email from the ASF dual-hosted git repository.

wenchen pushed a commit to branch branch-3.0
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-3.0 by this push:
 new 8910dae  [SPARK-31859][SPARK-31861][FOLLOWUP] Fix typo in tests
8910dae is described below

commit 8910dae0966ec7bbd9886245310fa07b81426e74
Author: Juliusz Sompolski 
AuthorDate: Fri Jun 5 10:01:00 2020 +

[SPARK-31859][SPARK-31861][FOLLOWUP] Fix typo in tests

### What changes were proposed in this pull request?

It appears I have unintentionally used nested JDBC statements in the two 
tests I added.

### Why are the changes needed?

Cleanup a typo. Please merge to master/branch-3.0

### Does this PR introduce _any_ user-facing change?

No.

### How was this patch tested?

Unit tests.

Closes #28735 from juliuszsompolski/SPARK-31859-fixup.

Authored-by: Juliusz Sompolski 
Signed-off-by: Wenchen Fan 
(cherry picked from commit ea010138e95cc9c0f62603d431b906bd5104f4e3)
Signed-off-by: Wenchen Fan 
---
 .../thriftserver/HiveThriftServer2Suites.scala | 36 ++
 1 file changed, 16 insertions(+), 20 deletions(-)

diff --git 
a/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/HiveThriftServer2Suites.scala
 
b/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/HiveThriftServer2Suites.scala
index 21256ad..4e6d4e1 100644
--- 
a/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/HiveThriftServer2Suites.scala
+++ 
b/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/HiveThriftServer2Suites.scala
@@ -813,31 +813,27 @@ class HiveThriftBinaryServerSuite extends 
HiveThriftJdbcTest {
   }
 
   test("SPARK-31859 Thriftserver works with 
spark.sql.datetime.java8API.enabled=true") {
-withJdbcStatement() { statement =>
-  withJdbcStatement() { st =>
-st.execute("set spark.sql.datetime.java8API.enabled=true")
-val rs = st.executeQuery("select date '2020-05-28', timestamp 
'2020-05-28 00:00:00'")
-rs.next()
-assert(rs.getDate(1).toString() == "2020-05-28")
-assert(rs.getTimestamp(2).toString() == "2020-05-28 00:00:00.0")
-  }
+withJdbcStatement() { st =>
+  st.execute("set spark.sql.datetime.java8API.enabled=true")
+  val rs = st.executeQuery("select date '2020-05-28', timestamp 
'2020-05-28 00:00:00'")
+  rs.next()
+  assert(rs.getDate(1).toString() == "2020-05-28")
+  assert(rs.getTimestamp(2).toString() == "2020-05-28 00:00:00.0")
 }
   }
 
   test("SPARK-31861 Thriftserver respects spark.sql.session.timeZone") {
-withJdbcStatement() { statement =>
-  withJdbcStatement() { st =>
-st.execute("set spark.sql.session.timeZone=+03:15") // different than 
Thriftserver's JVM tz
+withJdbcStatement() { st =>
+  st.execute("set spark.sql.session.timeZone=+03:15") // different than 
Thriftserver's JVM tz
   val rs = st.executeQuery("select timestamp '2020-05-28 10:00:00'")
-rs.next()
-// The timestamp as string is the same as the literal
-assert(rs.getString(1) == "2020-05-28 10:00:00.0")
-// Parsing it to java.sql.Timestamp in the client will always result 
in a timestamp
-// in client default JVM timezone. The string value of the Timestamp 
will match the literal,
-// but if the JDBC application cares about the internal timezone and 
UTC offset of the
-// Timestamp object, it should set spark.sql.session.timeZone to match 
its client JVM tz.
-assert(rs.getTimestamp(1).toString() == "2020-05-28 10:00:00.0")
-  }
+  rs.next()
+  // The timestamp as string is the same as the literal
+  assert(rs.getString(1) == "2020-05-28 10:00:00.0")
+  // Parsing it to java.sql.Timestamp in the client will always result in 
a timestamp
+  // in client default JVM timezone. The string value of the Timestamp 
will match the literal,
+  // but if the JDBC application cares about the internal timezone and UTC 
offset of the
+  // Timestamp object, it should set spark.sql.session.timeZone to match 
its client JVM tz.
+  assert(rs.getTimestamp(1).toString() == "2020-05-28 10:00:00.0")
 }
   }
 


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch branch-3.0 updated: [SPARK-31859][SPARK-31861][FOLLOWUP] Fix typo in tests

2020-06-05 Thread wenchen
This is an automated email from the ASF dual-hosted git repository.

wenchen pushed a commit to branch branch-3.0
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-3.0 by this push:
 new 8910dae  [SPARK-31859][SPARK-31861][FOLLOWUP] Fix typo in tests
8910dae is described below

commit 8910dae0966ec7bbd9886245310fa07b81426e74
Author: Juliusz Sompolski 
AuthorDate: Fri Jun 5 10:01:00 2020 +

[SPARK-31859][SPARK-31861][FOLLOWUP] Fix typo in tests

### What changes were proposed in this pull request?

It appears I have unintentionally used nested JDBC statements in the two 
tests I added.

### Why are the changes needed?

Cleanup a typo. Please merge to master/branch-3.0

### Does this PR introduce _any_ user-facing change?

No.

### How was this patch tested?

Unit tests.

Closes #28735 from juliuszsompolski/SPARK-31859-fixup.

Authored-by: Juliusz Sompolski 
Signed-off-by: Wenchen Fan 
(cherry picked from commit ea010138e95cc9c0f62603d431b906bd5104f4e3)
Signed-off-by: Wenchen Fan 
---
 .../thriftserver/HiveThriftServer2Suites.scala | 36 ++
 1 file changed, 16 insertions(+), 20 deletions(-)

diff --git 
a/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/HiveThriftServer2Suites.scala
 
b/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/HiveThriftServer2Suites.scala
index 21256ad..4e6d4e1 100644
--- 
a/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/HiveThriftServer2Suites.scala
+++ 
b/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/HiveThriftServer2Suites.scala
@@ -813,31 +813,27 @@ class HiveThriftBinaryServerSuite extends 
HiveThriftJdbcTest {
   }
 
   test("SPARK-31859 Thriftserver works with 
spark.sql.datetime.java8API.enabled=true") {
-withJdbcStatement() { statement =>
-  withJdbcStatement() { st =>
-st.execute("set spark.sql.datetime.java8API.enabled=true")
-val rs = st.executeQuery("select date '2020-05-28', timestamp 
'2020-05-28 00:00:00'")
-rs.next()
-assert(rs.getDate(1).toString() == "2020-05-28")
-assert(rs.getTimestamp(2).toString() == "2020-05-28 00:00:00.0")
-  }
+withJdbcStatement() { st =>
+  st.execute("set spark.sql.datetime.java8API.enabled=true")
+  val rs = st.executeQuery("select date '2020-05-28', timestamp 
'2020-05-28 00:00:00'")
+  rs.next()
+  assert(rs.getDate(1).toString() == "2020-05-28")
+  assert(rs.getTimestamp(2).toString() == "2020-05-28 00:00:00.0")
 }
   }
 
   test("SPARK-31861 Thriftserver respects spark.sql.session.timeZone") {
-withJdbcStatement() { statement =>
-  withJdbcStatement() { st =>
-st.execute("set spark.sql.session.timeZone=+03:15") // different than 
Thriftserver's JVM tz
+withJdbcStatement() { st =>
+  st.execute("set spark.sql.session.timeZone=+03:15") // different than 
Thriftserver's JVM tz
   val rs = st.executeQuery("select timestamp '2020-05-28 10:00:00'")
-rs.next()
-// The timestamp as string is the same as the literal
-assert(rs.getString(1) == "2020-05-28 10:00:00.0")
-// Parsing it to java.sql.Timestamp in the client will always result 
in a timestamp
-// in client default JVM timezone. The string value of the Timestamp 
will match the literal,
-// but if the JDBC application cares about the internal timezone and 
UTC offset of the
-// Timestamp object, it should set spark.sql.session.timeZone to match 
its client JVM tz.
-assert(rs.getTimestamp(1).toString() == "2020-05-28 10:00:00.0")
-  }
+  rs.next()
+  // The timestamp as string is the same as the literal
+  assert(rs.getString(1) == "2020-05-28 10:00:00.0")
+  // Parsing it to java.sql.Timestamp in the client will always result in 
a timestamp
+  // in client default JVM timezone. The string value of the Timestamp 
will match the literal,
+  // but if the JDBC application cares about the internal timezone and UTC 
offset of the
+  // Timestamp object, it should set spark.sql.session.timeZone to match 
its client JVM tz.
+  assert(rs.getTimestamp(1).toString() == "2020-05-28 10:00:00.0")
 }
   }
 


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch branch-3.0 updated: [SPARK-31859][SPARK-31861][FOLLOWUP] Fix typo in tests

2020-06-05 Thread wenchen
This is an automated email from the ASF dual-hosted git repository.

wenchen pushed a commit to branch branch-3.0
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-3.0 by this push:
 new 8910dae  [SPARK-31859][SPARK-31861][FOLLOWUP] Fix typo in tests
8910dae is described below

commit 8910dae0966ec7bbd9886245310fa07b81426e74
Author: Juliusz Sompolski 
AuthorDate: Fri Jun 5 10:01:00 2020 +

[SPARK-31859][SPARK-31861][FOLLOWUP] Fix typo in tests

### What changes were proposed in this pull request?

It appears I have unintentionally used nested JDBC statements in the two 
tests I added.

### Why are the changes needed?

Cleanup a typo. Please merge to master/branch-3.0

### Does this PR introduce _any_ user-facing change?

No.

### How was this patch tested?

Unit tests.

Closes #28735 from juliuszsompolski/SPARK-31859-fixup.

Authored-by: Juliusz Sompolski 
Signed-off-by: Wenchen Fan 
(cherry picked from commit ea010138e95cc9c0f62603d431b906bd5104f4e3)
Signed-off-by: Wenchen Fan 
---
 .../thriftserver/HiveThriftServer2Suites.scala | 36 ++
 1 file changed, 16 insertions(+), 20 deletions(-)

diff --git 
a/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/HiveThriftServer2Suites.scala
 
b/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/HiveThriftServer2Suites.scala
index 21256ad..4e6d4e1 100644
--- 
a/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/HiveThriftServer2Suites.scala
+++ 
b/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/HiveThriftServer2Suites.scala
@@ -813,31 +813,27 @@ class HiveThriftBinaryServerSuite extends 
HiveThriftJdbcTest {
   }
 
   test("SPARK-31859 Thriftserver works with 
spark.sql.datetime.java8API.enabled=true") {
-withJdbcStatement() { statement =>
-  withJdbcStatement() { st =>
-st.execute("set spark.sql.datetime.java8API.enabled=true")
-val rs = st.executeQuery("select date '2020-05-28', timestamp 
'2020-05-28 00:00:00'")
-rs.next()
-assert(rs.getDate(1).toString() == "2020-05-28")
-assert(rs.getTimestamp(2).toString() == "2020-05-28 00:00:00.0")
-  }
+withJdbcStatement() { st =>
+  st.execute("set spark.sql.datetime.java8API.enabled=true")
+  val rs = st.executeQuery("select date '2020-05-28', timestamp 
'2020-05-28 00:00:00'")
+  rs.next()
+  assert(rs.getDate(1).toString() == "2020-05-28")
+  assert(rs.getTimestamp(2).toString() == "2020-05-28 00:00:00.0")
 }
   }
 
   test("SPARK-31861 Thriftserver respects spark.sql.session.timeZone") {
-withJdbcStatement() { statement =>
-  withJdbcStatement() { st =>
-st.execute("set spark.sql.session.timeZone=+03:15") // different than 
Thriftserver's JVM tz
+withJdbcStatement() { st =>
+  st.execute("set spark.sql.session.timeZone=+03:15") // different than 
Thriftserver's JVM tz
   val rs = st.executeQuery("select timestamp '2020-05-28 10:00:00'")
-rs.next()
-// The timestamp as string is the same as the literal
-assert(rs.getString(1) == "2020-05-28 10:00:00.0")
-// Parsing it to java.sql.Timestamp in the client will always result 
in a timestamp
-// in client default JVM timezone. The string value of the Timestamp 
will match the literal,
-// but if the JDBC application cares about the internal timezone and 
UTC offset of the
-// Timestamp object, it should set spark.sql.session.timeZone to match 
its client JVM tz.
-assert(rs.getTimestamp(1).toString() == "2020-05-28 10:00:00.0")
-  }
+  rs.next()
+  // The timestamp as string is the same as the literal
+  assert(rs.getString(1) == "2020-05-28 10:00:00.0")
+  // Parsing it to java.sql.Timestamp in the client will always result in 
a timestamp
+  // in client default JVM timezone. The string value of the Timestamp 
will match the literal,
+  // but if the JDBC application cares about the internal timezone and UTC 
offset of the
+  // Timestamp object, it should set spark.sql.session.timeZone to match 
its client JVM tz.
+  assert(rs.getTimestamp(1).toString() == "2020-05-28 10:00:00.0")
 }
   }
 


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch branch-3.0 updated: [SPARK-31859][SPARK-31861][FOLLOWUP] Fix typo in tests

2020-06-05 Thread wenchen
This is an automated email from the ASF dual-hosted git repository.

wenchen pushed a commit to branch branch-3.0
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-3.0 by this push:
 new 8910dae  [SPARK-31859][SPARK-31861][FOLLOWUP] Fix typo in tests
8910dae is described below

commit 8910dae0966ec7bbd9886245310fa07b81426e74
Author: Juliusz Sompolski 
AuthorDate: Fri Jun 5 10:01:00 2020 +

[SPARK-31859][SPARK-31861][FOLLOWUP] Fix typo in tests

### What changes were proposed in this pull request?

It appears I have unintentionally used nested JDBC statements in the two 
tests I added.

### Why are the changes needed?

Cleanup a typo. Please merge to master/branch-3.0

### Does this PR introduce _any_ user-facing change?

No.

### How was this patch tested?

Unit tests.

Closes #28735 from juliuszsompolski/SPARK-31859-fixup.

Authored-by: Juliusz Sompolski 
Signed-off-by: Wenchen Fan 
(cherry picked from commit ea010138e95cc9c0f62603d431b906bd5104f4e3)
Signed-off-by: Wenchen Fan 
---
 .../thriftserver/HiveThriftServer2Suites.scala | 36 ++
 1 file changed, 16 insertions(+), 20 deletions(-)

diff --git 
a/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/HiveThriftServer2Suites.scala
 
b/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/HiveThriftServer2Suites.scala
index 21256ad..4e6d4e1 100644
--- 
a/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/HiveThriftServer2Suites.scala
+++ 
b/sql/hive-thriftserver/src/test/scala/org/apache/spark/sql/hive/thriftserver/HiveThriftServer2Suites.scala
@@ -813,31 +813,27 @@ class HiveThriftBinaryServerSuite extends 
HiveThriftJdbcTest {
   }
 
   test("SPARK-31859 Thriftserver works with 
spark.sql.datetime.java8API.enabled=true") {
-withJdbcStatement() { statement =>
-  withJdbcStatement() { st =>
-st.execute("set spark.sql.datetime.java8API.enabled=true")
-val rs = st.executeQuery("select date '2020-05-28', timestamp 
'2020-05-28 00:00:00'")
-rs.next()
-assert(rs.getDate(1).toString() == "2020-05-28")
-assert(rs.getTimestamp(2).toString() == "2020-05-28 00:00:00.0")
-  }
+withJdbcStatement() { st =>
+  st.execute("set spark.sql.datetime.java8API.enabled=true")
+  val rs = st.executeQuery("select date '2020-05-28', timestamp 
'2020-05-28 00:00:00'")
+  rs.next()
+  assert(rs.getDate(1).toString() == "2020-05-28")
+  assert(rs.getTimestamp(2).toString() == "2020-05-28 00:00:00.0")
 }
   }
 
   test("SPARK-31861 Thriftserver respects spark.sql.session.timeZone") {
-withJdbcStatement() { statement =>
-  withJdbcStatement() { st =>
-st.execute("set spark.sql.session.timeZone=+03:15") // different than 
Thriftserver's JVM tz
+withJdbcStatement() { st =>
+  st.execute("set spark.sql.session.timeZone=+03:15") // different than 
Thriftserver's JVM tz
   val rs = st.executeQuery("select timestamp '2020-05-28 10:00:00'")
-rs.next()
-// The timestamp as string is the same as the literal
-assert(rs.getString(1) == "2020-05-28 10:00:00.0")
-// Parsing it to java.sql.Timestamp in the client will always result 
in a timestamp
-// in client default JVM timezone. The string value of the Timestamp 
will match the literal,
-// but if the JDBC application cares about the internal timezone and 
UTC offset of the
-// Timestamp object, it should set spark.sql.session.timeZone to match 
its client JVM tz.
-assert(rs.getTimestamp(1).toString() == "2020-05-28 10:00:00.0")
-  }
+  rs.next()
+  // The timestamp as string is the same as the literal
+  assert(rs.getString(1) == "2020-05-28 10:00:00.0")
+  // Parsing it to java.sql.Timestamp in the client will always result in 
a timestamp
+  // in client default JVM timezone. The string value of the Timestamp 
will match the literal,
+  // but if the JDBC application cares about the internal timezone and UTC 
offset of the
+  // Timestamp object, it should set spark.sql.session.timeZone to match 
its client JVM tz.
+  assert(rs.getTimestamp(1).toString() == "2020-05-28 10:00:00.0")
 }
   }
 


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org