Repository: hive Updated Branches: refs/heads/master 3a677303f -> e0eee6bef
HIVE-18596: Synchronize value of hive.spark.client.connect.timeout across unit tests (Sahil Takiar, reviewed by Peter Vary) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/e0eee6be Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/e0eee6be Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/e0eee6be Branch: refs/heads/master Commit: e0eee6bef3fa5816b8ca60b7e737a12a99dcecfa Parents: 3a67730 Author: Sahil Takiar <takiar.sa...@gmail.com> Authored: Mon Feb 5 11:43:26 2018 -0600 Committer: Sahil Takiar <stak...@cloudera.com> Committed: Mon Feb 5 11:44:03 2018 -0600 ---------------------------------------------------------------------- data/conf/spark/standalone/hive-site.xml | 5 +++++ .../org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java | 1 + .../hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java | 1 + .../test/java/org/apache/hive/spark/client/TestSparkClient.java | 4 ++++ 4 files changed, 11 insertions(+) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hive/blob/e0eee6be/data/conf/spark/standalone/hive-site.xml ---------------------------------------------------------------------- diff --git a/data/conf/spark/standalone/hive-site.xml b/data/conf/spark/standalone/hive-site.xml index 1e5bd65..7095979 100644 --- a/data/conf/spark/standalone/hive-site.xml +++ b/data/conf/spark/standalone/hive-site.xml @@ -255,6 +255,11 @@ <description>Internal marker for test. Used for masking env-dependent values</description> </property> +<property> + <name>hive.spark.client.connect.timeout</name> + <value>30000ms</value> +</property> + <!-- remove this after HIVE-18139 --> <property> <name>hive.stats.column.autogather</name> http://git-wip-us.apache.org/repos/asf/hive/blob/e0eee6be/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java ---------------------------------------------------------------------- diff --git a/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java b/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java index 0fee3c6..2a4da20 100644 --- a/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java +++ b/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java @@ -70,6 +70,7 @@ public class TestJdbcWithLocalClusterSpark { conf.set("hive.execution.engine", "spark"); conf.set("spark.serializer", "org.apache.spark.serializer.KryoSerializer"); conf.set("spark.master", "local-cluster[2,2,1024]"); + conf.set("hive.spark.client.connect.timeout", "30000ms"); // FIXME: Hadoop3 made the incompatible change for dfs.client.datanode-restart.timeout // while spark2 is still using Hadoop2. // Spark requires Hive to support Hadoop3 first then Spark can start http://git-wip-us.apache.org/repos/asf/hive/blob/e0eee6be/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java ---------------------------------------------------------------------- diff --git a/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java b/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java index 93fe776..9f72e51 100644 --- a/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java +++ b/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java @@ -79,6 +79,7 @@ public class TestMultiSessionsHS2WithLocalClusterSpark { conf.set("spark.serializer", "org.apache.spark.serializer.KryoSerializer"); conf.set("spark.master", "local-cluster[2,2,1024]"); conf.set("spark.deploy.defaultCores", "2"); + conf.set("hive.spark.client.connect.timeout", "30000ms"); // FIXME: Hadoop3 made the incompatible change for dfs.client.datanode-restart.timeout // while spark2 is still using Hadoop2. // Spark requires Hive to support Hadoop3 first then Spark can start http://git-wip-us.apache.org/repos/asf/hive/blob/e0eee6be/spark-client/src/test/java/org/apache/hive/spark/client/TestSparkClient.java ---------------------------------------------------------------------- diff --git a/spark-client/src/test/java/org/apache/hive/spark/client/TestSparkClient.java b/spark-client/src/test/java/org/apache/hive/spark/client/TestSparkClient.java index 23df792..579ca07 100644 --- a/spark-client/src/test/java/org/apache/hive/spark/client/TestSparkClient.java +++ b/spark-client/src/test/java/org/apache/hive/spark/client/TestSparkClient.java @@ -68,6 +68,10 @@ public class TestSparkClient { private static final long TIMEOUT = 20; private static final HiveConf HIVECONF = new HiveConf(); + static { + HIVECONF.set("hive.spark.client.connect.timeout", "30000ms"); + } + private Map<String, String> createConf() { Map<String, String> conf = new HashMap<String, String>();