[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-28 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14228289#comment-14228289
 ] 

Xuefu Zhang commented on HIVE-8836:
---

To summarize the above Spark related test failures:

1. custom_input_output_format.q, tracked by HIVE-8991
2. infer_bucket_sort_convert_join.q and  parquet_join.q, tracked by HIVE-8992
3. mapjoin_hook.q, tracked by HIVE-8981

Above tez and MR related failrues should be unrelated. They need to be fixed in 
trunk.

Thank everyone for making this happen.

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.13-spark.patch, HIVE-8836.14-spark.patch, 
 HIVE-8836.14-spark.patch, HIVE-8836.7-spark.patch, HIVE-8836.8-spark.patch, 
 HIVE-8836.9-spark.patch, additional-enable-spark-log.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-27 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14227411#comment-14227411
 ] 

Hive QA commented on HIVE-8836:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12684005/HIVE-8836.11-spark.patch

{color:red}ERROR:{color} -1 due to 13 failed/errored test(s), 7180 tests 
executed
*Failed tests:*
{noformat}
TestHS2ImpersonationWithRemoteMS - did not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample_islocalmode_hook
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_vector_decimal_aggregate
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_optimize_nullscan
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_custom_input_output_format
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_multi_insert_common_distinct
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_multi_single_reducer
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_infer_bucket_sort_convert_join
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapjoin_hook
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_parquet_join
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_smb_mapjoin_3
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_stats_counter
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_table_access_keys_stats
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/458/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/458/console
Test logs: 
http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/PreCommit-HIVE-SPARK-Build-458/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 13 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12684005 - PreCommit-HIVE-SPARK-Build

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.10-spark.patch, 
 HIVE-8836.10-spark.patch, HIVE-8836.11-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, 
 HIVE-8836.7-spark.patch, HIVE-8836.8-spark.patch, HIVE-8836.9-spark.patch, 
 additional-enable-spark-log.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-27 Thread Szehon Ho (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14227424#comment-14227424
 ] 

Szehon Ho commented on HIVE-8836:
-

HIVE-8924 committed and fixed some errors.  

Mapjoin_hook tracked by HIVE-8981, feel free to ignore for now.

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.10-spark.patch, 
 HIVE-8836.10-spark.patch, HIVE-8836.11-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, 
 HIVE-8836.7-spark.patch, HIVE-8836.8-spark.patch, HIVE-8836.9-spark.patch, 
 additional-enable-spark-log.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-27 Thread Szehon Ho (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14227460#comment-14227460
 ] 

Szehon Ho commented on HIVE-8836:
-

Just looking through and one more quick observation, parquet_join is failing 
because the table parquet_jointable1_bucketed_sorted is corrupted, with 3 files 
but created with 1 bucket.  It seems a problem with number of reducers in the 
insert statement that populate the table, but not sure if I will have a chance 
to look soon.

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.10-spark.patch, 
 HIVE-8836.10-spark.patch, HIVE-8836.11-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, 
 HIVE-8836.7-spark.patch, HIVE-8836.8-spark.patch, HIVE-8836.9-spark.patch, 
 additional-enable-spark-log.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-27 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14227797#comment-14227797
 ] 

Brock Noland commented on HIVE-8836:


The latest rev missed some changes related to spark.home so it ran in process. 
I will update shortly.
{noformat}
2014-11-27 08:14:40,046 INFO  [main]: client.SparkClientImpl 
(SparkClientImpl.java:startDriver(198)) - No spark.home provided, calling 
SparkSubmit directly.
2014-11-27 08:14:40,046 DEBUG [main]: client.SparkClientImpl 
(SparkClientImpl.java:startDriver(248)) - Running client driver with argv: 
/usr/java/jdk1.7.0_45-cloudera/jre/bin/java org.apache.spark.deploy.SparkSubmit 
--properties-file 
/home/hiveptest/54.177.147.244-hiveptest-0/apache-svn-spark-source/itests/qtest-spark/target/tmp/spark-submit.3800861402634267765.properties
 --class org.apache.hive.spark.client.RemoteDriver 
/home/hiveptest/54.177.147.244-hiveptest-0/maven/org/apache/hive/hive-exec/0.15.0-SNAPSHOT/hive-exec-0.15.0-SNAPSHOT.jar
 --remote 
akka.tcp://73d8f0b4-dd16-4893-92f7-18a91307f929@10.232.134.247:35405/user/SparkClient-327f65bb-a5fd-43f2-917d-a5829a58465a
2014-11-27 08:14:40,134 WARN  [Driver]: client.SparkClientImpl 
(SparkClientImpl.java:run(264)) - Child process exited with code 1.
{noformat}

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.10-spark.patch, 
 HIVE-8836.10-spark.patch, HIVE-8836.11-spark.patch, HIVE-8836.12-spark.patch, 
 HIVE-8836.2-spark.patch, HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, 
 HIVE-8836.5-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.7-spark.patch, HIVE-8836.8-spark.patch, 
 HIVE-8836.9-spark.patch, additional-enable-spark-log.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-27 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14227799#comment-14227799
 ] 

Hive QA commented on HIVE-8836:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12684069/HIVE-8836.12-spark.patch

{color:red}ERROR:{color} -1 due to 498 failed/errored test(s), 7167 tests 
executed
*Failed tests:*
{noformat}
TestSparkCliDriver-escape_distributeby1.q-auto_sortmerge_join_7.q-bucketmapjoin_negative3.q-and-12-more
 - did not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join_without_localtask
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample_islocalmode_hook
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_vector_decimal_aggregate
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_optimize_nullscan
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_add_part_multiple
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_alter_merge_orc
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_alter_merge_stats_orc
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_annotate_stats_join
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join0
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join10
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join11
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join12
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join13
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join14
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join15
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join16
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join17
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join18
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join18_multi_distinct
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join19
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join2
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join20
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join21
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join22
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join23
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join24
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join26
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join27
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join28
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join29
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join3
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join30
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join31
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join32
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join4
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join6
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join7
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join8
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join9
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join_filters
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join_nulls
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join_reordering_values
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join_without_localtask
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_smb_mapjoin_14
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_10
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_11
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_12
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_13
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_14
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_15
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_16
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_2
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_3
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_4

[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-27 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14227801#comment-14227801
 ] 

Brock Noland commented on HIVE-8836:


The latest rev also missed the changes to the query plan (number of reducers). 
I will include that in my next patch.

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.10-spark.patch, 
 HIVE-8836.10-spark.patch, HIVE-8836.11-spark.patch, HIVE-8836.12-spark.patch, 
 HIVE-8836.2-spark.patch, HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, 
 HIVE-8836.5-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.7-spark.patch, HIVE-8836.8-spark.patch, 
 HIVE-8836.9-spark.patch, additional-enable-spark-log.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-27 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14227812#comment-14227812
 ] 

Brock Noland commented on HIVE-8836:


bq. Just looking through and one more quick observation, parquet_join is 
failing because the table parquet_jointable1_bucketed_sorted is corrupted, with 
3 files but created with 1 bucket. It seems a problem with number of reducers 
in the insert statement that populate the table, but not sure if I will have a 
chance to look soon.

Ahh interesting. We'll probably need to fall back to the number of buckets for 
bucketed tables.

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.10-spark.patch, 
 HIVE-8836.10-spark.patch, HIVE-8836.11-spark.patch, HIVE-8836.12-spark.patch, 
 HIVE-8836.2-spark.patch, HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, 
 HIVE-8836.5-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.7-spark.patch, HIVE-8836.8-spark.patch, 
 HIVE-8836.9-spark.patch, additional-enable-spark-log.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-27 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14227878#comment-14227878
 ] 

Hive QA commented on HIVE-8836:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12684085/HIVE-8836.13-spark.patch

{color:red}ERROR:{color} -1 due to 11 failed/errored test(s), 7182 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join_without_localtask
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample_islocalmode_hook
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_vector_decimal_aggregate
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_custom_input_output_format
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_multi_insert_common_distinct
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_multi_single_reducer
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_infer_bucket_sort_convert_join
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapjoin_hook
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_parquet_join
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_smb_mapjoin_3
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_stats_counter
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/461/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/461/console
Test logs: 
http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/PreCommit-HIVE-SPARK-Build-461/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 11 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12684085 - PreCommit-HIVE-SPARK-Build

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.11-spark.patch, HIVE-8836.12-spark.patch, 
 HIVE-8836.13-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, 
 HIVE-8836.7-spark.patch, HIVE-8836.8-spark.patch, HIVE-8836.9-spark.patch, 
 additional-enable-spark-log.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-27 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14227883#comment-14227883
 ] 

Brock Noland commented on HIVE-8836:


Latest run looks pretty good:
# One test which needs the number of reducers updated (must have just been 
committed)
# A few failures due to instability in RSC (fixing this will be a long term 
project which we can do after commit IMO)
# A few ordering issues (fixing in trunk via HIVE-8989)
# A few other known issues like parquet_join, custom_input_output_format, and 
map_hook.

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.11-spark.patch, HIVE-8836.12-spark.patch, 
 HIVE-8836.13-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, 
 HIVE-8836.7-spark.patch, HIVE-8836.8-spark.patch, HIVE-8836.9-spark.patch, 
 additional-enable-spark-log.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-27 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14227904#comment-14227904
 ] 

Xuefu Zhang commented on HIVE-8836:
---

Nice. Once we have reduced the number of failures to a trackable few, we can 
commit this and track remaining issues in individual JIRAs. I'm sure more 
issues will be discovered with this on the way, but that's a good thing.

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.11-spark.patch, HIVE-8836.12-spark.patch, 
 HIVE-8836.13-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, 
 HIVE-8836.7-spark.patch, HIVE-8836.8-spark.patch, HIVE-8836.9-spark.patch, 
 additional-enable-spark-log.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-27 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14227949#comment-14227949
 ] 

Brock Noland commented on HIVE-8836:


Agreed this latest patch includes HIVE-8989. It might not apply due to the 
recent commits to SVN and the slow svn to git publishing. If the results look 
similar to last time with some improvement, I will commit the change.

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.13-spark.patch, HIVE-8836.14-spark.patch, 
 HIVE-8836.7-spark.patch, HIVE-8836.8-spark.patch, HIVE-8836.9-spark.patch, 
 additional-enable-spark-log.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-27 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14227982#comment-14227982
 ] 

Hive QA commented on HIVE-8836:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12684116/HIVE-8836.14-spark.patch

{color:red}ERROR:{color} -1 due to 9 failed/errored test(s), 7181 tests executed
*Failed tests:*
{noformat}
TestHS2ImpersonationWithRemoteMS - did not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join_without_localtask
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample_islocalmode_hook
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_vector_decimal_aggregate
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_custom_input_output_format
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_infer_bucket_sort_convert_join
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapjoin_hook
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_parquet_join
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_stats_counter
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/462/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/462/console
Test logs: 
http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/PreCommit-HIVE-SPARK-Build-462/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 9 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12684116 - PreCommit-HIVE-SPARK-Build

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.13-spark.patch, HIVE-8836.14-spark.patch, 
 HIVE-8836.7-spark.patch, HIVE-8836.8-spark.patch, HIVE-8836.9-spark.patch, 
 additional-enable-spark-log.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-27 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14228059#comment-14228059
 ] 

Hive QA commented on HIVE-8836:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12684127/HIVE-8836.14-spark.patch

{color:red}ERROR:{color} -1 due to 8 failed/errored test(s), 7182 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join_without_localtask
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample_islocalmode_hook
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_vector_decimal_aggregate
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_optimize_nullscan
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_custom_input_output_format
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_infer_bucket_sort_convert_join
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapjoin_hook
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_parquet_join
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/463/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/463/console
Test logs: 
http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/PreCommit-HIVE-SPARK-Build-463/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 8 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12684127 - PreCommit-HIVE-SPARK-Build

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.13-spark.patch, HIVE-8836.14-spark.patch, 
 HIVE-8836.14-spark.patch, HIVE-8836.7-spark.patch, HIVE-8836.8-spark.patch, 
 HIVE-8836.9-spark.patch, additional-enable-spark-log.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-26 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14226270#comment-14226270
 ] 

Xuefu Zhang commented on HIVE-8836:
---

[~ruili], I think the number of reducers changed because of the cluster 
changes. Previously the plan is generated with one node with 4 cores 
(local[4]). Now the cluster has 2 nodes and one core each. Memory configuration 
is also different. I guess it's hard to tweek the cluster configuration so that 
the same number of reducer results.

For now, I think we have to go thru the list and analyze failures one by one. 
It's a long list, and maybe it can be divided among people so that each only 
take a slice of it.

Briefly checking the result, it seems the failures are caused by any of the 
following reasons:
1. reducer number change, which is okay.
2. result diff. It could be a matter of ordering, but could be different result 
also.
3. test failed to run.

I noticed that we are using local-cluster[2,1,2048]. Maybe we should have a 
more general case where one node has more than one core. Also, we may need to 
adjust the memory settings. Once we have a representative of a small cluster, 
we probably will stay with it for some time. 



 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-26 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14226461#comment-14226461
 ] 

Brock Noland commented on HIVE-8836:


I will change it to two cores and then re-generate the outputs. This should 
allow us to differentiate between failed tests, changes outputs, and just 
reducer changes.

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-26 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14226730#comment-14226730
 ] 

Brock Noland commented on HIVE-8836:


Attached patch has regenerated output for queries which had a different plan 
(number of reducers). It does not update the following:

*Query Result differences*
{noformat}
auto_join_without_localtask.q.out
count.q.out
join_filters_overlap.q.out
limit_pushdown.q.out
mapreduce2.q.out
multi_insert_gby3.q.out
multi_join_union.q.out
ppd_outer_join3.q.out
ptf_decimal.q.out
ptf_general_queries.q.out
smb_mapjoin_1.q.out
smb_mapjoin_2.q.out
smb_mapjoin_4.q.out
smb_mapjoin_5.q.out
smb_mapjoin_8.q.out
stats_counter.q.out
table_access_keys_stats.q.out
uniquejoin.q.out
vector_decimal_aggregate.q.out
vectorization_13.q.out
join_reorder.q.out
outer_join_ppr.q.out

*Failed*
{noformat}
bucketmapjoin1.q.out
groupby_multi_insert_common_distinct.q.out
groupby_multi_single_reducer.q.out
infer_bucket_sort_convert_join.q.out
mapjoin_hook.q.out
smb_mapjoin9
{noformat}



 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, 
 HIVE-8836.7-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-26 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14226820#comment-14226820
 ] 

Hive QA commented on HIVE-8836:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12683900/HIVE-8836.7-spark.patch

{color:red}ERROR:{color} -1 due to 48 failed/errored test(s), 7177 tests 
executed
*Failed tests:*
{noformat}
TestHS2ImpersonationWithRemoteMS - did not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample_islocalmode_hook
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join_without_localtask
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_bucketmapjoin1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_bucketmapjoin7
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_count
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_custom_input_output_format
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_complex_types_multi_single_reducer
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_multi_insert_common_distinct
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_multi_single_reducer
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_infer_bucket_sort_convert_join
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join_filters_overlap
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join_reorder
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_limit_pushdown
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapjoin_hook
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapreduce2
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_multi_insert_gby3
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_multi_join_union
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_optimize_nullscan
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_outer_join_ppr
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_parquet_join
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_ppd_outer_join3
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_ptf_decimal
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_ptf_general_queries
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_smb_mapjoin9
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_smb_mapjoin_1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_smb_mapjoin_2
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_smb_mapjoin_3
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_smb_mapjoin_4
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_smb_mapjoin_5
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_smb_mapjoin_8
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_stats_counter
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_table_access_keys_stats
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_timestamp_1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_timestamp_2
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_timestamp_3
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_timestamp_lazy
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_uniquejoin
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_vector_between_in
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_vector_data_types
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_vector_decimal_aggregate
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_vectorization_13
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_vectorization_14
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_vectorization_15
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_vectorization_16
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_vectorization_9
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_vectorization_short_regress
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_vectorized_timestamp_funcs
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/449/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/449/console
Test logs: 
http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/PreCommit-HIVE-SPARK-Build-449/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 48 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12683900 - PreCommit-HIVE-SPARK-Build

 Enable 

[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-26 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14226883#comment-14226883
 ] 

Brock Noland commented on HIVE-8836:


I am making the following tests deterministic over in HIVE-8976.

{noformat}
auto_join_without_localtask.q
count.q
limit_pushdown.q
mapreduce2.q
multi_insert_gby3.q
multi_join_union.q
ppd_outer_join3.q
ptf_decimal.q
ptf_general_queries.q
{noformat}

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, 
 HIVE-8836.7-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-26 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14227073#comment-14227073
 ] 

Hive QA commented on HIVE-8836:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12683943/HIVE-8836.9-spark.patch

{color:red}ERROR:{color} -1 due to 33 failed/errored test(s), 7178 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_auto_join_without_localtask
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_count
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_limit_pushdown
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_mapreduce2
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_multi_insert_gby3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_multi_join_union
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ppd_outer_join3
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ptf_decimal
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ptf_general_queries
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample_islocalmode_hook
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_bucketmapjoin1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_bucketmapjoin7
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_custom_input_output_format
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_multi_insert_common_distinct
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_multi_single_reducer
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_infer_bucket_sort_convert_join
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join_reorder
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapjoin_hook
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_optimize_nullscan
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_outer_join_ppr
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_parquet_join
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_smb_mapjoin9
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_smb_mapjoin_1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_smb_mapjoin_2
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_smb_mapjoin_3
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_smb_mapjoin_4
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_smb_mapjoin_5
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_smb_mapjoin_8
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_stats_counter
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_table_access_keys_stats
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_uniquejoin
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_vector_decimal_aggregate
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_vectorization_13
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/451/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/451/console
Test logs: 
http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/PreCommit-HIVE-SPARK-Build-451/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 33 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12683943 - PreCommit-HIVE-SPARK-Build

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, 
 HIVE-8836.7-spark.patch, HIVE-8836.8-spark.patch, HIVE-8836.9-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-26 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14227075#comment-14227075
 ] 

Brock Noland commented on HIVE-8836:


The non-spark failures are my fault due to not including the changes in 
HIVE-8976 for MR in this patch.

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, 
 HIVE-8836.7-spark.patch, HIVE-8836.8-spark.patch, HIVE-8836.9-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-26 Thread Rui Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14227118#comment-14227118
 ] 

Rui Li commented on HIVE-8836:
--

Thanks [~brocknoland] and [~xuefuz] for the work. Maybe we can split this job a 
little bit? I can take a look at the tests that failed to run.

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, 
 HIVE-8836.7-spark.patch, HIVE-8836.8-spark.patch, HIVE-8836.9-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-26 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14227184#comment-14227184
 ] 

Brock Noland commented on HIVE-8836:


Thank you [~chengxiang li]! I will apply that on the current patch. I've also 
created HIVE-8983 to have PTest backup the spark.log file as part of the 
testing process.

Sync'ed with [~lirui] offline and he'll be looking at some failures.

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, 
 HIVE-8836.7-spark.patch, HIVE-8836.8-spark.patch, HIVE-8836.9-spark.patch, 
 additional-enable-spark-log.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-26 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14227201#comment-14227201
 ] 

Brock Noland commented on HIVE-8836:


FYI I am generating a new patch which incorporates Jimmy's changes in HIVE-8978.

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, 
 HIVE-8836.7-spark.patch, HIVE-8836.8-spark.patch, HIVE-8836.9-spark.patch, 
 additional-enable-spark-log.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-26 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14227316#comment-14227316
 ] 

Hive QA commented on HIVE-8836:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12683990/HIVE-8836.10-spark.patch

{color:red}ERROR:{color} -1 due to 33 failed/errored test(s), 7180 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample_islocalmode_hook
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_vector_decimal_aggregate
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_bucketmapjoin1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_bucketmapjoin7
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_custom_input_output_format
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_multi_insert_common_distinct
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_multi_single_reducer
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_infer_bucket_sort_convert_join
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_input14
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_input17
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_input18
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_insert_into1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_insert_into2
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_insert_into3
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapjoin_hook
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapreduce1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapreduce2
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_optimize_nullscan
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_parquet_join
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_ppd_transform
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_script_env_var1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_script_env_var2
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_script_pipe
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_scriptfile1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_smb_mapjoin9
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_smb_mapjoin_3
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_stats_counter
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_table_access_keys_stats
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_transform1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_transform2
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_transform_ppr1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_transform_ppr2
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_union23
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/456/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/456/console
Test logs: 
http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/PreCommit-HIVE-SPARK-Build-456/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 33 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12683990 - PreCommit-HIVE-SPARK-Build

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.10-spark.patch, 
 HIVE-8836.2-spark.patch, HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, 
 HIVE-8836.5-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.7-spark.patch, HIVE-8836.8-spark.patch, 
 HIVE-8836.9-spark.patch, additional-enable-spark-log.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client.[Spark Branch]

2014-11-25 Thread Marcelo Vanzin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225049#comment-14225049
 ] 

Marcelo Vanzin commented on HIVE-8836:
--

I talked briefly with Brock about this, but the main thing here is that, right 
now, Spark is not very friendly to applications that are trying to embed it. As 
you've noticed, the assembly jar, which contains almost everything you need to 
run Spark, is not published in maven or anywhere. And not all artifacts used to 
build the assembly are published - for example, the Yarn backend cannot be 
found anywhere in maven, so without the assembly you cannot submit jobs to Yarn.

I've suggested it in the past, but I think right now, or until Spark makes 
itself more friendly to such use cases, Hive should require a full Spark 
install to work. If desired we could use the hacks I added to the remote client 
to not need the full install for unit tests, but even those are very limited; 
it probably only works with a local master as some of you may have noticed.

 Enable automatic tests with remote spark client.[Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Attachments: HIVE-8836-brock-1.patch, HIVE-8836-brock-2.patch, 
 HIVE-8836-brock-3.patch, HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client.[Spark Branch]

2014-11-25 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225114#comment-14225114
 ] 

Brock Noland commented on HIVE-8836:


I think it makes sense to move forward with the tarball approach since it works 
today. Then we can file a follow-on jira to come up  with a cleaner approach. 
That approach might need some enhancements to Spark like publishing the 
assembly to Maven.

 Enable automatic tests with remote spark client.[Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Attachments: HIVE-8836-brock-1.patch, HIVE-8836-brock-2.patch, 
 HIVE-8836-brock-3.patch, HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client.[Spark Branch]

2014-11-25 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225336#comment-14225336
 ] 

Brock Noland commented on HIVE-8836:


Tests of this patch are in progress here: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/433/console

 Enable automatic tests with remote spark client.[Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client.[Spark Branch]

2014-11-25 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225407#comment-14225407
 ] 

Xuefu Zhang commented on HIVE-8836:
---

It seems the test is stuck.


On Tue, Nov 25, 2014 at 2:45 PM, Brock Noland (JIRA) j...@apache.org



 Enable automatic tests with remote spark client.[Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client.[Spark Branch]

2014-11-25 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225412#comment-14225412
 ] 

Brock Noland commented on HIVE-8836:


Yes it looks like two commands are taking a while. Looking into it.
{noformat}
501   7997  0.0  0.0  60236  3804 ?S17:31   0:00  \_ ssh -v -i 
/home/hiveptest/.ssh/hive-ptest-user-key -l hiveptest 50.18.64.184 bash 
/home/hiveptest//50.18.64.184-hiveptest-1/scratch/hiveptest-TestSparkCliDriver-join11.q-join18.q-groupby2.q-and-12-more.sh
501   8211  0.0  0.0  60236  3804 ?S17:36   0:00  \_ ssh -v -i 
/home/hiveptest/.ssh/hive-ptest-user-key -l hiveptest 50.18.64.184 bash 
/home/hiveptest//50.18.64.184-hiveptest-2/scratch/hiveptest-TestSparkCliDriver-union_remove_9.q-ppd_multi_insert.q-auto_sortmerge_join_16.q-and-12-more.sh
{noformat}

 Enable automatic tests with remote spark client.[Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client.[Spark Branch]

2014-11-25 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225417#comment-14225417
 ] 

Brock Noland commented on HIVE-8836:


One of the stuck processes:

{noformat}
2014-11-25 14:30:44,874 INFO  ql.Driver (SessionState.java:printInfo(828)) - 
Query ID = hiveptest_20141125143030_c56b5f72-8552-4122-930d-7df9dea96638
2014-11-25 14:30:44,874 INFO  ql.Driver (SessionState.java:printInfo(828)) - 
Total jobs = 1
2014-11-25 14:30:44,874 INFO  ql.Driver (SessionState.java:printInfo(828)) - 
Launching Job 1 out of 1
2014-11-25 14:30:44,875 INFO  ql.Driver (Driver.java:launchTask(1643)) - 
Starting task [Stage-1:MAPRED] in serial mode
2014-11-25 14:30:44,875 INFO  exec.Task (SessionState.java:printInfo(828)) - In 
order to change the average load for a reducer (in bytes):
2014-11-25 14:30:44,876 INFO  exec.Task (SessionState.java:printInfo(828)) -   
set hive.exec.reducers.bytes.per.reducer=number
2014-11-25 14:30:44,876 INFO  exec.Task (SessionState.java:printInfo(828)) - In 
order to limit the maximum number of reducers:
2014-11-25 14:30:44,876 INFO  exec.Task (SessionState.java:printInfo(828)) -   
set hive.exec.reducers.max=number
2014-11-25 14:30:44,876 INFO  exec.Task (SessionState.java:printInfo(828)) - In 
order to set a constant number of reducers:
2014-11-25 14:30:44,876 INFO  exec.Task (SessionState.java:printInfo(828)) -   
set mapreduce.job.reduces=number
2014-11-25 14:30:44,876 INFO  spark.HiveSparkClientFactory 
(HiveSparkClientFactory.java:initiateSparkConf(105)) - load spark configuration 
from hive configuration (spark.master - local-cluster[2,1,2048]).
2014-11-25 14:30:44,894 INFO  slf4j.Slf4jLogger 
(Slf4jLogger.scala:applyOrElse(80)) - Slf4jLogger started
2014-11-25 14:30:44,899 INFO  Remoting (Slf4jLogger.scala:apply$mcV$sp(74)) - 
Starting remoting
2014-11-25 14:30:44,907 INFO  Remoting (Slf4jLogger.scala:apply$mcV$sp(74)) - 
Remoting started; listening on addresses 
:[akka.tcp://d84d934e-e10e-4744-b42c-ed86a49ebbd3@10.227.4.181:56697]
2014-11-25 14:30:44,909 DEBUG client.SparkClientImpl 
(SparkClientImpl.java:startDriver(252)) - Running client driver with argv: 
/home/hiveptest/50.18.64.184-hiveptest-1/apache-svn-spark-source/itests/qtest-spark/../../itests/qtest-spark/target/spark/bin/spark-submit
 --properties-file 
/home/hiveptest/50.18.64.184-hiveptest-1/apache-svn-spark-source/itests/qtest-spark/target/tmp/spark-submit.8721943354350626566.properties
 --class org.apache.hive.spark.client.RemoteDriver 
/home/hiveptest/50.18.64.184-hiveptest-1/maven/org/apache/hive/hive-exec/0.15.0-SNAPSHOT/hive-exec-0.15.0-SNAPSHOT.jar
 --remote 
akka.tcp://d84d934e-e10e-4744-b42c-ed86a49ebbd3@10.227.4.181:56697/user/SparkClient-aa5b3525-a031-41d7-ab04-8a18b0aa3fcf
2014-11-25 14:30:48,058 INFO  client.SparkClientImpl 
(SparkClientImpl.java:onReceive(312)) - Received hello from 
akka.tcp://92e75da1-125e-4576-b63e-dc3166653dbe@10.227.4.181:35948/user/RemoteDriver
2014-11-25 14:30:48,059 DEBUG session.SparkSessionManagerImpl 
(SparkSessionManagerImpl.java:getSession(126)) - New session 
(1743e7e0-1fb3-4766-9249-cc138f88a2a7) is created.
2014-11-25 14:30:48,085 INFO  ql.Context (Context.java:getMRScratchDir(266)) - 
New scratch dir is 
file:/home/hiveptest/50.18.64.184-hiveptest-1/apache-svn-spark-source/itests/qtest-spark/target/tmp/scratchdir/hiveptest/cf1f8ff8-a03d-413c-9922-d7c5c3c25e18/hive_2014-11-25_14-30-44_827_2636220880717371319-1
2014-11-25 14:30:48,283 INFO  client.SparkClientImpl 
(SparkClientImpl.java:onReceive(329)) - Received result for 
fa96fc57-3999-4103-9bbd-a4bab346a324
2014-11-25 14:30:48,535 INFO  client.SparkClientImpl 
(SparkClientImpl.java:onReceive(329)) - Received result for 
00716575-c666-4b2a-bffe-682780684df8
2014-11-25 14:47:28,067 INFO  transport.ProtocolStateActor 
(Slf4jLogger.scala:apply$mcV$sp(74)) - No response from remote. Handshake timed 
out or transport failure detector triggered.
2014-11-25 14:47:28,071 WARN  remote.ReliableDeliverySupervisor 
(Slf4jLogger.scala:apply$mcV$sp(71)) - Association with remote system 
[akka.tcp://92e75da1-125e-4576-b63e-dc3166653dbe@10.227.4.181:35948] has 
failed, address is now gated for [5000] ms. Reason is: [Disassociated].
{noformat}

 Enable automatic tests with remote spark client.[Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch


 In real production environment, remote spark client should be used to submit 
 

[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client.[Spark Branch]

2014-11-25 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225423#comment-14225423
 ] 

Brock Noland commented on HIVE-8836:


Similar warning on the other:
{noformat}
2014-11-25 14:39:26,714 INFO  ql.Driver (SessionState.java:printInfo(828)) - 
Query ID = hiveptest_20141125143939_77f4ba53-fc27-4efe-ad3d-85e23ea54748
2014-11-25 14:39:26,714 INFO  ql.Driver (SessionState.java:printInfo(828)) - 
Total jobs = 2
2014-11-25 14:39:26,714 INFO  ql.Driver (SessionState.java:printInfo(828)) - 
Launching Job 1 out of 2
2014-11-25 14:39:26,715 INFO  ql.Driver (Driver.java:launchTask(1643)) - 
Starting task [Stage-3:MAPRED] in serial mode
2014-11-25 14:39:26,715 INFO  exec.Task (SessionState.java:printInfo(828)) - In 
order to change the average load for a reducer (in bytes):
2014-11-25 14:39:26,715 INFO  exec.Task (SessionState.java:printInfo(828)) -   
set hive.exec.reducers.bytes.per.reducer=number
2014-11-25 14:39:26,715 INFO  exec.Task (SessionState.java:printInfo(828)) - In 
order to limit the maximum number of reducers:
2014-11-25 14:39:26,715 INFO  exec.Task (SessionState.java:printInfo(828)) -   
set hive.exec.reducers.max=number
2014-11-25 14:39:26,715 INFO  exec.Task (SessionState.java:printInfo(828)) - In 
order to set a constant number of reducers:
2014-11-25 14:39:26,715 INFO  exec.Task (SessionState.java:printInfo(828)) -   
set mapreduce.job.reduces=number
2014-11-25 14:39:26,715 DEBUG session.SparkSessionManagerImpl 
(SparkSessionManagerImpl.java:getSession(107)) - Existing session 
(34e37f91-2cac-4a31-aba7-85b711d8dad3) is reused.
2014-11-25 14:39:26,728 INFO  ql.Context (Context.java:getMRScratchDir(266)) - 
New scratch dir is 
file:/home/hiveptest/50.18.64.184-hiveptest-2/apache-svn-spark-source/itests/qtest-spark/target/tmp/scratchdir/hiveptest/e9216630-f66e-4b31-bc30-58078678a976/hive_2014-11-25_14-39-26_638_5903751068675432492-1
2014-11-25 14:39:26,775 INFO  client.SparkClientImpl 
(SparkClientImpl.java:onReceive(329)) - Received result for 
e288f226-2429-469a-9c53-07fabda12db3
2014-11-25 14:55:54,771 WARN  remote.ReliableDeliverySupervisor 
(Slf4jLogger.scala:apply$mcV$sp(71)) - Association with remote system 
[akka.tcp://8fe8195c-f1be-45e3-a9d1-b94b7caafcd9@10.227.4.181:38320] has 
failed, address is now gated for [5000] ms. Reason is: [Disassociated].
{noformat}

 Enable automatic tests with remote spark client.[Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client.[Spark Branch]

2014-11-25 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225429#comment-14225429
 ] 

Brock Noland commented on HIVE-8836:


Complete log: 
http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/akka-error/hive.log.txt

 Enable automatic tests with remote spark client.[Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client.[Spark Branch]

2014-11-25 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225438#comment-14225438
 ] 

Brock Noland commented on HIVE-8836:


Bunch of debug logs: 
http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/akka-error/

 Enable automatic tests with remote spark client.[Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client.[Spark Branch]

2014-11-25 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225492#comment-14225492
 ] 

Hive QA commented on HIVE-8836:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12683650/HIVE-8836.4-spark.patch

{color:red}ERROR:{color} -1 due to 96 failed/errored test(s), 7148 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample_islocalmode_hook
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join20
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join29
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join30
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join32
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join_without_localtask
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_smb_mapjoin_14
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_10
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_6
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_9
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_bucket_map_join_tez1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_bucket_map_join_tez2
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_bucketmapjoin1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_bucketmapjoin7
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_count
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_ctas
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby10
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby8
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby9
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_complex_types_multi_single_reducer
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_cube1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_multi_insert_common_distinct
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_multi_single_reducer
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_multi_single_reducer3
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_position
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_rollup1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_sort_1_23
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_sort_skew_1_23
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_having
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_infer_bucket_sort_convert_join
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join20
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join38
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join40
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join_filters_overlap
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join_reorder
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_limit_pushdown
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapjoin_distinct
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapjoin_filter_on_outerjoin
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapjoin_hook
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapjoin_test_outer
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapreduce2
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_multi_insert
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_multi_insert_gby
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_multi_insert_gby3
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_multi_insert_lateral_view
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_multi_insert_move_tasks_share_dependencies
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_multi_join_union
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_outer_join_ppr
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_parquet_join
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_ppd_join4
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_ppd_join_filter
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_ppd_outer_join3
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_ppd_transform
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_ptf_decimal
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_ptf_general_queries
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_sample10
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_semijoin

[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-25 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225535#comment-14225535
 ] 

Hive QA commented on HIVE-8836:
---



{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12683721/HIVE-8836.5-spark.patch

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/437/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/437/console
Test logs: 
http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/PreCommit-HIVE-SPARK-Build-437/

Messages:
{noformat}
 This message was trimmed, see log for full details 
Reverted 'ql/src/test/results/clientpositive/spark/load_dyn_part10.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/stats5.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/smb_mapjoin_7.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/ppd_join3.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/skewjoin.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/mapjoin_distinct.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/join12.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/smb_mapjoin_16.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/input17.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/ppd_join_filter.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/join35.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/groupby1.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/join_cond_pushdown_4.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/bucketmapjoin2.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/mapjoin1.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/insert_into1.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/mapjoin_subquery2.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/skewjoinopt16.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/smb_mapjoin9.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/groupby11.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/column_access_stats.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/join21.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/smb_mapjoin_25.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/join_merging.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/join_map_ppr.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/stats0.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/smb_mapjoin_2.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/auto_sortmerge_join_12.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/ppd_multi_insert.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/join9.q.out'
Reverted 
'ql/src/test/results/clientpositive/spark/skewjoin_union_remove_1.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/input12.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/join30.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/auto_join6.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/mapjoin_hook.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/union_remove_19.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/subquery_multiinsert.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/skewjoinopt11.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/mapreduce2.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/index_auto_self_join.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/join_nullsafe.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/union_remove_6.q.out'
Reverted 
'ql/src/test/results/clientpositive/spark/join_merge_multi_expressions.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/vector_decimal_mapjoin.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/auto_join16.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/smb_mapjoin_20.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/bucket3.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/ppd_outer_join5.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/vector_mapjoin_reduce.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/semijoin.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/stats_partscan_1_23.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/skewjoinopt20.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/merge2.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/join4.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/bucket_map_join_tez2.q.out'
Reverted 'ql/src/test/results/clientpositive/spark/auto_join25.q.out'
Reverted 
'ql/src/test/results/clientpositive/spark/join_cond_pushdown_unqual4.q.out'
Reverted 

[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-25 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225578#comment-14225578
 ] 

Hive QA commented on HIVE-8836:
---



{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12683726/HIVE-8836.6-spark.patch

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/438/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/438/console
Test logs: 
http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/PreCommit-HIVE-SPARK-Build-438/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: SSHExecutionException: SSHResult [command=pkill -f java, 
getExitCode()=1, getException()=null, getUser()=hiveptest, 
getHost()=54.176.195.138, getInstance()=0]: 'OpenSSH_5.3p1, OpenSSL 1.0.0-fips 
29 Mar 2010
debug1: Reading configuration data /home/hiveptest/.ssh/config
debug1: Reading configuration data /etc/ssh/ssh_config
debug1: Applying options for *
debug1: Connecting to 54.176.195.138 [54.176.195.138] port 22.
debug1: fd 3 clearing O_NONBLOCK
debug1: Connection established.
debug1: identity file /home/hiveptest/.ssh/hive-ptest-user-key type -1
debug1: Remote protocol version 2.0, remote software version OpenSSH_5.3
debug1: match: OpenSSH_5.3 pat OpenSSH*
debug1: Enabling compatibility mode for protocol 2.0
debug1: Local version string SSH-2.0-OpenSSH_5.3
debug1: SSH2_MSG_KEXINIT sent
debug1: SSH2_MSG_KEXINIT received
debug1: kex: server-client aes128-ctr hmac-md5 none
debug1: kex: client-server aes128-ctr hmac-md5 none
debug1: SSH2_MSG_KEX_DH_GEX_REQUEST(102410248192) sent
debug1: expecting SSH2_MSG_KEX_DH_GEX_GROUP
debug1: SSH2_MSG_KEX_DH_GEX_INIT sent
debug1: expecting SSH2_MSG_KEX_DH_GEX_REPLY
debug1: Host '54.176.195.138' is known and matches the RSA host key.
debug1: Found key in /home/hiveptest/.ssh/known_hosts:1194
debug1: ssh_rsa_verify: signature correct
debug1: SSH2_MSG_NEWKEYS sent
debug1: expecting SSH2_MSG_NEWKEYS
debug1: SSH2_MSG_NEWKEYS received
debug1: SSH2_MSG_SERVICE_REQUEST sent
debug1: SSH2_MSG_SERVICE_ACCEPT received
debug1: Authentications that can continue: 
publickey,gssapi-keyex,gssapi-with-mic,password
debug1: Next authentication method: gssapi-keyex
debug1: No valid Key exchange context
debug1: Next authentication method: gssapi-with-mic
Address 54.176.195.138 maps to 
ec2-54-176-195-138.us-west-1.compute.amazonaws.com, but this does not map back 
to the address - POSSIBLE BREAK-IN ATTEMPT!
debug1: Unspecified GSS failure.  Minor code may provide more information
Credentials cache file '/tmp/krb5cc_501' not found

debug1: Unspecified GSS failure.  Minor code may provide more information
Credentials cache file '/tmp/krb5cc_501' not found

debug1: Unspecified GSS failure.  Minor code may provide more information


debug1: Unspecified GSS failure.  Minor code may provide more information
Credentials cache file '/tmp/krb5cc_501' not found

debug1: Next authentication method: publickey
debug1: Trying private key: /home/hiveptest/.ssh/hive-ptest-user-key
debug1: read PEM private key done: type RSA
debug1: Authentication succeeded (publickey).
debug1: channel 0: new [client-session]
debug1: Requesting no-more-sessi...@openssh.com
debug1: Entering interactive session.
debug1: Sending environment.
debug1: Sending env LANG = en_US.UTF-8
debug1: Sending command: pkill -f java
debug1: client_input_channel_req: channel 0 rtype exit-status reply 0
debug1: client_input_channel_req: channel 0 rtype e...@openssh.com reply 0
debug1: channel 0: free: client-session, nchannels 1
debug1: fd 0 clearing O_NONBLOCK
debug1: fd 1 clearing O_NONBLOCK
Transferred: sent 1952, received 1976 bytes, in 0.0 seconds
Bytes per second: sent 59126.3, received 59853.3
debug1: Exit status 1
'
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12683726 - PreCommit-HIVE-SPARK-Build

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, 
 HIVE-8836.6-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--

[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-25 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225716#comment-14225716
 ] 

Brock Noland commented on HIVE-8836:


Again with the latest patch two tests are stuck. I copied all relavent info to:

http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/akka-error/2/

if you look at the end of the hive.log:

http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/akka-error/2/qtest-spark/target/tmp/log/

it seems to get into a state where {{SparkJobMonitor}} is only getting an 
{{UKNOWN}} state from {{SparkJobStatus}}

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-25 Thread Rui Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225721#comment-14225721
 ] 

Rui Li commented on HIVE-8836:
--

Hi [~brocknoland], this may be related to HIVE-8956. Do you know which two 
tests are stuck?

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-25 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225730#comment-14225730
 ] 

Brock Noland commented on HIVE-8836:


Looks like it's:

{noformat}
Begin query: optimize_nullscan.q
Begin query: custom_input_output_format.q
{noformat}

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-25 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225732#comment-14225732
 ] 

Brock Noland commented on HIVE-8836:


[~lirui] that does look like it might be related! Nice find.

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-25 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225742#comment-14225742
 ] 

Hive QA commented on HIVE-8836:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12683735/HIVE-8836.6-spark.patch

{color:red}ERROR:{color} -1 due to 105 failed/errored test(s), 7148 tests 
executed
*Failed tests:*
{noformat}
TestAuthorizationApiAuthorizer - did not produce a TEST-*.xml file
TestGenericUDFOPNumeric - did not produce a TEST-*.xml file
TestHBaseKeyFactory - did not produce a TEST-*.xml file
TestHBaseKeyFactory2 - did not produce a TEST-*.xml file
TestHBaseKeyFactory3 - did not produce a TEST-*.xml file
TestHBasePredicateDecomposer - did not produce a TEST-*.xml file
TestSparkCliDriver-join11.q-join18.q-groupby2.q-and-12-more - did not produce a 
TEST-*.xml file
TestSparkCliDriver-union_remove_9.q-ppd_multi_insert.q-auto_sortmerge_join_16.q-and-12-more
 - did not produce a TEST-*.xml file
TestTezSessionState - did not produce a TEST-*.xml file
TestURLHook - did not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample_islocalmode_hook
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join20
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join29
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join30
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join32
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join_without_localtask
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_smb_mapjoin_14
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_10
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_6
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_9
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_bucket_map_join_tez1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_bucket_map_join_tez2
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_bucketmapjoin1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_bucketmapjoin7
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_count
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_ctas
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby10
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby8
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby9
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_complex_types_multi_single_reducer
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_cube1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_multi_insert_common_distinct
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_multi_single_reducer
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_multi_single_reducer3
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_position
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_rollup1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_sort_1_23
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_sort_skew_1_23
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_having
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_infer_bucket_sort_convert_join
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join20
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join38
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join40
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join_filters_overlap
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join_reorder
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_limit_pushdown
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapjoin_distinct
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapjoin_filter_on_outerjoin
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapjoin_hook
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapjoin_test_outer
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapreduce2
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_multi_insert
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_multi_insert_gby
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_multi_insert_gby3
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_multi_insert_lateral_view
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_multi_insert_move_tasks_share_dependencies
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_multi_join_union
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_outer_join_ppr

[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-25 Thread Rui Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225751#comment-14225751
 ] 

Rui Li commented on HIVE-8836:
--

OK I'll try out the failed tests.

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-25 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225754#comment-14225754
 ] 

Brock Noland commented on HIVE-8836:


{noformat}
TestSparkCliDriver-join11.q-join18.q-groupby2.q-and-12-more - did not produce a 
TEST-*.xml file
TestSparkCliDriver-union_remove_9.q-ppd_multi_insert.q-auto_sortmerge_join_16.q-and-12-more
 - did not produce a TEST-*.xml file
{noformat}

The two tests above were the hung tests. Other did not produce a ... items 
will be resolved by HIVE-8971. We'll commit HIVE-8956 to fix the hangs.

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-25 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225779#comment-14225779
 ] 

Xuefu Zhang commented on HIVE-8836:
---

Many test failures above are shown as plan diff, somehting like: 
{code}
211c211
 Reducer 2 - Map 1 (GROUP, 3)
---
 Reducer 2 - Map 1 (GROUP, 1)
{code}
This can be explained by the fact that the current algorithm determining the 
number of reducers has included the factors such as number of executors in the 
cluster. Because we switched to local-cluster with two nodes, the number of 
reducers is different from previous value. This is not good, because every time 
we change the testing cluster, we might have a different plan.

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-25 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225795#comment-14225795
 ] 

Brock Noland commented on HIVE-8836:


bq. This is not good, because every time we change the testing cluster, we 
might have a different plan.

How often do you think we'll be changing the local cluster size? My guess is 
not that often?

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-25 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225803#comment-14225803
 ] 

Xuefu Zhang commented on HIVE-8836:
---

Yeah. but it also means that when dev generates the .out file, same cluster 
configuration is required. I guess we can live with that for now.

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-25 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225849#comment-14225849
 ] 

Hive QA commented on HIVE-8836:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12683770/HIVE-8836.6-spark.patch

{color:red}ERROR:{color} -1 due to 108 failed/errored test(s), 7178 tests 
executed
*Failed tests:*
{noformat}
TestAuthorizationApiAuthorizer - did not produce a TEST-*.xml file
TestGenericUDFOPNumeric - did not produce a TEST-*.xml file
TestHBaseKeyFactory - did not produce a TEST-*.xml file
TestHBaseKeyFactory2 - did not produce a TEST-*.xml file
TestHBaseKeyFactory3 - did not produce a TEST-*.xml file
TestHBasePredicateDecomposer - did not produce a TEST-*.xml file
TestTezSessionState - did not produce a TEST-*.xml file
TestURLHook - did not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample_islocalmode_hook
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_transform_acid
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join20
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join29
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join30
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join32
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join_without_localtask
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_smb_mapjoin_14
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_10
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_6
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_9
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_bucket_map_join_tez1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_bucket_map_join_tez2
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_bucketmapjoin1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_bucketmapjoin7
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_count
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_ctas
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_custom_input_output_format
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby10
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby8
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby9
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_complex_types_multi_single_reducer
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_cube1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_multi_insert_common_distinct
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_multi_single_reducer
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_multi_single_reducer3
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_position
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_rollup1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_sort_1_23
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_groupby_sort_skew_1_23
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_having
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_infer_bucket_sort_convert_join
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join20
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join38
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join40
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join_filters_overlap
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join_reorder
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_limit_pushdown
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapjoin_distinct
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapjoin_filter_on_outerjoin
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapjoin_hook
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapjoin_test_outer
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapreduce2
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_multi_insert
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_multi_insert_gby
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_multi_insert_gby3
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_multi_insert_lateral_view
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_multi_insert_move_tasks_share_dependencies
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_multi_join_union
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_optimize_nullscan
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_outer_join_ppr

[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client [Spark Branch]

2014-11-25 Thread Rui Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14225860#comment-14225860
 ] 

Rui Li commented on HIVE-8836:
--

Shall we use same #reducers as before for now? So that we can tell which tests 
are actually failing and which are just affected by the #reducers change.

 Enable automatic tests with remote spark client [Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Fix For: spark-branch

 Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, 
 HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch, HIVE-8836.5-spark.patch, 
 HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch, HIVE-8836.6-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client.[Spark Branch]

2014-11-24 Thread Chengxiang Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14224065#comment-14224065
 ] 

Chengxiang Li commented on HIVE-8836:
-

The join0.q output diff should because of HIVE-8855 which jimmy is working on.

 Enable automatic tests with remote spark client.[Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Attachments: HIVE-8836-brock-1.patch, HIVE-8836-brock-2.patch, 
 HIVE-8836-brock-3.patch, HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client.[Spark Branch]

2014-11-24 Thread Szehon Ho (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14224098#comment-14224098
 ] 

Szehon Ho commented on HIVE-8836:
-

bq. Spark assembly dependency can be fetched from public maven reposiotry

Hi Chengxiang, I dont understand this part, how are you downloading this from 
public repo?  When I try your patch I get the error:
{noformat}
Could not resolve dependencies for project 
org.apache.hive:hive-it-qfile-spark:jar:0.15.0-SNAPSHOT: Could not find 
artifact org.apache.spark:spark-assembly_2.10:jar:1.2.0-SNAPSHOT in 
spark-snapshot 
(http://ec2-50-18-79-139.us-west-1.compute.amazonaws.com/data/spark_2.10-1.2-SNAPSHOT/
{noformat}

And we cant find it in any public repo.  Thats why we assumed we had to build 
it and upload to hosted location.

Also another question, as we were trying to set spark.home, which looks for 
bin/spark-submit, which then pulled in scripts like compute-classpath.sh, 
load-spark-env.sh, spark-class, and finally spark-assembly itself.  I see you 
are using another way (spark.test.home, spark.testing), how does that avoid 
looking for these artifacts to start the spark process?

 Enable automatic tests with remote spark client.[Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Attachments: HIVE-8836-brock-1.patch, HIVE-8836-brock-2.patch, 
 HIVE-8836-brock-3.patch, HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client.[Spark Branch]

2014-11-24 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14224100#comment-14224100
 ] 

Brock Noland commented on HIVE-8836:


bq. Spark assembly dependency can be fetched from public maven reposiotry, it 
should be more maintainable than download a whole spark installation. what do 
you think?

I don't think spark assembly is published to maven. The Sentry project has used 
the download a tarball (of Hive and Hadoop) for over a year now and it works 
well.

However, if there is a lightweight method of making this work, I am all for 
it.

 Enable automatic tests with remote spark client.[Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Attachments: HIVE-8836-brock-1.patch, HIVE-8836-brock-2.patch, 
 HIVE-8836-brock-3.patch, HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client.[Spark Branch]

2014-11-24 Thread Chengxiang Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14224124#comment-14224124
 ] 

Chengxiang Li commented on HIVE-8836:
-

Hi, [~szehon] and [~brocknoland]
I'm not 100% percent sure spark assembly jar would be published to public maven 
repository, but I find spark assembly at 
[here|http://mvnrepository.com/artifact/org.apache.spark/spark-assembly_2.10/1.1.0],
 maybe [~vanzin] know more about this. There is no 
org.apache.spark:spark-assembly_2.10:jar:1.2.0-SNAPSHOT in any public maven 
repository yet as it's still in SNAPSHOT status, but we can publish it to 
http://ec2-50-18-79-139.us-west-1.compute.amazonaws.com/data as what we have 
done for spark core. I build spark and public to local maven repository in my 
local test.
{quote}
Also another question, as we were trying to set spark.home, which looks for 
bin/spark-submit, which then pulled in scripts like compute-classpath.sh, 
load-spark-env.sh, spark-class, and finally spark-assembly itself. I see you 
are using another way (spark.test.home, spark.testing), how does that avoid 
looking for these artifacts to start the spark process?
{quote}

First, bin/spark-submit is optional for Remote Spark Context. 
Then, local-cluster spark only need compute-classpath.sh for launch executor, 
which is used to add spark related jars into classpath(Hive unit test should 
only need spark-assembly). spark.test.home and spark.testing are used to set 
spark home to dummy spark installation, you can check 
org.apache.spark.deploy.worker.Worker::line101 for why. I create dummy spark 
installation with empty compute-classpath.sh as compute-classpath.sh is 
required, and add spark assembly to spark executor classpath through 
spark.executor.extraClassPath.


 Enable automatic tests with remote spark client.[Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Attachments: HIVE-8836-brock-1.patch, HIVE-8836-brock-2.patch, 
 HIVE-8836-brock-3.patch, HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client.[Spark Branch]

2014-11-24 Thread Szehon Ho (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14224170#comment-14224170
 ] 

Szehon Ho commented on HIVE-8836:
-

Thanks for the details, Chengxiang.

 Enable automatic tests with remote spark client.[Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
Assignee: Rui Li
  Labels: Spark-M3
 Attachments: HIVE-8836-brock-1.patch, HIVE-8836-brock-2.patch, 
 HIVE-8836-brock-3.patch, HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch


 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client.[Spark Branch]

2014-11-12 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14208088#comment-14208088
 ] 

Xuefu Zhang commented on HIVE-8836:
---

Hi [~chengxiang li],  With HIVE-8795, this should get enabled, right? I 
understand that we need to verify this. Thanks.

 Enable automatic tests with remote spark client.[Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
  Labels: Spark-M3

 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8836) Enable automatic tests with remote spark client.[Spark Branch]

2014-11-12 Thread Chengxiang Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14209035#comment-14209035
 ] 

Chengxiang Li commented on HIVE-8836:
-

Yes, [~xuefuz], this would be enabled with HIVE-8795 and HIVE-8833, I suspect 
several tests would failed, this JIRA is listed here to notify us to track the 
failed tests.

 Enable automatic tests with remote spark client.[Spark Branch]
 --

 Key: HIVE-8836
 URL: https://issues.apache.org/jira/browse/HIVE-8836
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
  Labels: Spark-M3

 In real production environment, remote spark client should be used to submit 
 spark job for Hive mostly, we should enable automatic test with remote spark 
 client to make sure the Hive feature workable with it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)