[jira] [Commented] (HIVE-9059) Remove wrappers for SparkJobInfo and SparkStageInfo

2014-12-13 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9059?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245249#comment-14245249
 ] 

Hive QA commented on HIVE-9059:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12686183/HIVE-9059.1-spark.patch

{color:red}ERROR:{color} -1 due to 408 failed/errored test(s), 6706 tests 
executed
*Failed tests:*
{noformat}
TestHWISessionManager - did not produce a TEST-*.xml file
TestSparkCliDriver-auto_join30.q-sort_merge_join_desc_3.q-join2.q-and-12-more - 
did not produce a TEST-*.xml file
TestSparkCliDriver-avro_joins.q-join36.q-join4.q-and-12-more - did not produce 
a TEST-*.xml file
TestSparkCliDriver-bucketmapjoin_negative.q-parallel_join0.q-insert_into1.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-bucketsortoptimize_insert_7.q-skewjoin_noskew.q-mapreduce1.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-groupby10.q-union5.q-load_dyn_part15.q-and-12-more - did not 
produce a TEST-*.xml file
TestSparkCliDriver-groupby3_map.q-skewjoinopt8.q-union_remove_1.q-and-12-more - 
did not produce a TEST-*.xml file
TestSparkCliDriver-groupby4.q-tez_joins_explain.q-load_dyn_part3.q-and-12-more 
- did not produce a TEST-*.xml file
TestSparkCliDriver-groupby_complex_types.q-auto_join9.q-groupby_map_ppr.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-groupby_map_ppr_multi_distinct.q-vectorization_16.q-multi_insert_mixed.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-input1_limit.q-groupby8_map.q-varchar_join1.q-and-12-more - 
did not produce a TEST-*.xml file
TestSparkCliDriver-join11.q-join18.q-groupby2.q-and-12-more - did not produce a 
TEST-*.xml file
TestSparkCliDriver-join13.q-sample2.q-join_reorder3.q-and-12-more - did not 
produce a TEST-*.xml file
TestSparkCliDriver-join39.q-bucketmapjoin10.q-bucket_map_join_tez1.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-join9.q-ppd_outer_join2.q-join_casesensitive.q-and-12-more - 
did not produce a TEST-*.xml file
TestSparkCliDriver-join_cond_pushdown_3.q-groupby7.q-auto_join17.q-and-12-more 
- did not produce a TEST-*.xml file
TestSparkCliDriver-join_cond_pushdown_unqual4.q-bucketmapjoin12.q-avro_decimal_native.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-join_filters.q-bucketmapjoin6.q-smb_mapjoin_4.q-and-12-more 
- did not produce a TEST-*.xml file
TestSparkCliDriver-list_bucket_dml_2.q-multi_insert.q-join5.q-and-12-more - did 
not produce a TEST-*.xml file
TestSparkCliDriver-load_dyn_part5.q-load_dyn_part2.q-auto_join10.q-and-12-more 
- did not produce a TEST-*.xml file
TestSparkCliDriver-parallel_join1.q-escape_distributeby1.q-timestamp_null.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-ppd_join4.q-skewjoinopt3.q-auto_join1.q-and-12-more - did 
not produce a TEST-*.xml file
TestSparkCliDriver-ppd_transform.q-auto_sortmerge_join_7.q-date_udf.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-ptf_general_queries.q-bucketmapjoin3.q-enforce_order.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-ptf_seqfile.q-auto_join18.q-vectorization_10.q-and-12-more - 
did not produce a TEST-*.xml file
TestSparkCliDriver-skewjoin_union_remove_2.q-auto_join14.q-auto_join20.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-skewjoinopt15.q-union_remove_7.q-avro_joins_native.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-smb_mapjoin_15.q-auto_join18_multi_distinct.q-mapreduce2.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-sort_merge_join_desc_1.q-order.q-union2.q-and-12-more - did 
not produce a TEST-*.xml file
TestSparkCliDriver-stats12.q-auto_join24.q-vector_decimal_aggregate.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-stats13.q-stats2.q-ppd_gby_join.q-and-12-more - did not 
produce a TEST-*.xml file
TestSparkCliDriver-table_access_keys_stats.q-bucketsortoptimize_insert_4.q-union_remove_9.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-timestamp_lazy.q-auto_join_reordering_values.q-union29.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-transform_ppr2.q-join20.q-decimal_join.q-and-11-more - did 
not produce a TEST-*.xml file
TestSparkCliDriver-vector_distinct_2.q-join15.q-union19.q-and-12-more - did not 
produce a TEST-*.xml file
TestSparkCliDriver-vectorization_13.q-auto_sortmerge_join_13.q-bucket3.q-and-12-more
 - did not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample_islocalmode_hook
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_vector_cast_constant
org.apache.hadoop.hive.ql.TestMTQueries.testMTQueries1
org.apache.hadoop.hive.ql.exec.TestExecDriver.initializationError

[jira] [Commented] (HIVE-9091) Add additional unit tests for HiveSQLException

2014-12-13 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9091?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245255#comment-14245255
 ] 

Hive QA commented on HIVE-9091:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12686938/HIVE-9091.1.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 6708 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_optimize_nullscan
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2069/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2069/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2069/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12686938 - PreCommit-HIVE-TRUNK-Build

 Add additional unit tests for HiveSQLException
 --

 Key: HIVE-9091
 URL: https://issues.apache.org/jira/browse/HIVE-9091
 Project: Hive
  Issue Type: Test
  Components: JDBC
Reporter: Aihua Xu
Assignee: Aihua Xu
Priority: Minor
  Labels: easyfix, test
 Fix For: 0.15.0

 Attachments: HIVE-9091.1.patch

   Original Estimate: 12h
  Remaining Estimate: 12h

 Change the function scope for the functions in HiveSQLException to be 
 private, including 
   toString(Throwable cause, StackTraceElement[] parent) , 
   enroll(Throwable ex, StackTraceElement[] trace, int max),
   toStackTrace(ListString details, StackTraceElement[] parent, int index) 
 since these serve as helper function internally. 
 Add comments for the public functions in HiveSQLException.
 To add additional unit test cases for HiveSQLException such as exception with 
 unknown source, test TStatus conversion, etc. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8911) Enable mapjoin hints [Spark Branch]

2014-12-13 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8911?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245259#comment-14245259
 ] 

Hive QA commented on HIVE-8911:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12686982/HIVE-8911.6-spark.patch

{color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 7233 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample_islocalmode_hook
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_vector_cast_constant
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_bucketmapjoin10
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_ppd_join4
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_smb_mapjoin_25
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/534/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/534/console
Test logs: 
http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/PreCommit-HIVE-SPARK-Build-534/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 5 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12686982 - PreCommit-HIVE-SPARK-Build

 Enable mapjoin hints [Spark Branch]
 ---

 Key: HIVE-8911
 URL: https://issues.apache.org/jira/browse/HIVE-8911
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Szehon Ho
Assignee: Chao
 Attachments: HIVE-8911.1-spark.patch, HIVE-8911.2-spark.patch, 
 HIVE-8911.3-spark.patch, HIVE-8911.4-spark.patch, HIVE-8911.5-spark.patch, 
 HIVE-8911.6-spark.patch


 Currently the big table selection in a mapjoin is based on stats.
 We should also enable the big-table selection based on hints.  See class 
 MapJoinProcessor.  This is a logical-optimizer class, so we should be able to 
 re-use this without too many changes to hook up with SparkMapJoinResolver.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8866) Vectorization on partitioned table throws ArrayIndexOutOfBoundsException when partitions are not of same #of columns

2014-12-13 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8866?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245265#comment-14245265
 ] 

Hive QA commented on HIVE-8866:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12686946/HIVE-8866.08.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 6704 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_udaf_covar_pop
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2070/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2070/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2070/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12686946 - PreCommit-HIVE-TRUNK-Build

 Vectorization on partitioned table throws ArrayIndexOutOfBoundsException when 
 partitions are not of same #of columns
 

 Key: HIVE-8866
 URL: https://issues.apache.org/jira/browse/HIVE-8866
 Project: Hive
  Issue Type: Bug
  Components: Vectorization
Affects Versions: 0.14.1
Reporter: Matt McCline
Assignee: Matt McCline
Priority: Critical
 Fix For: 0.14.1

 Attachments: HIVE-8866.01.patch, HIVE-8866.02.patch, 
 HIVE-8866.03.patch, HIVE-8866.05.patch, HIVE-8866.06.patch, 
 HIVE-8866.07.patch, HIVE-8866.08.patch


 Vectorization assumes partitions are of same number of columns, and takes 
 upon # of columns on first read. consequent addPartitionColsToBatch throws 
 ArrayIndexOutOfboundsException if the # columns is bigger.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9084) Investigate IOContext object initialization problem [Spark Branch]

2014-12-13 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9084?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245281#comment-14245281
 ] 

Hive QA commented on HIVE-9084:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12687021/HIVE-9084.2-spark.patch

{color:red}ERROR:{color} -1 due to 7 failed/errored test(s), 7261 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_nonmr_fetch
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample_islocalmode_hook
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_virtual_column
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_vector_cast_constant
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_ppd_join4
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_smb_mapjoin_25
org.apache.hadoop.hive.ql.exec.TestOperators.testMapOperator
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/535/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/535/console
Test logs: 
http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/PreCommit-HIVE-SPARK-Build-535/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 7 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12687021 - PreCommit-HIVE-SPARK-Build

 Investigate IOContext object initialization problem [Spark Branch]
 --

 Key: HIVE-9084
 URL: https://issues.apache.org/jira/browse/HIVE-9084
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Xuefu Zhang
Assignee: Xuefu Zhang
 Attachments: HIVE-9084.1-spark.patch, HIVE-9084.2-spark.patch


 In recent ptest run (Test results: 
 http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/511/testReport),
  test groupby_multi_single_reducer.q failed w/ the following stacktrace:
 {code}
 java.lang.RuntimeException: Map operator initialization failed
   at 
 org.apache.hadoop.hive.ql.exec.spark.SparkMapRecordHandler.init(SparkMapRecordHandler.java:136)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveMapFunction.call(HiveMapFunction.java:54)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveMapFunction.call(HiveMapFunction.java:29)
   at 
 org.apache.spark.api.java.JavaRDDLike$$anonfun$fn$7$1.apply(JavaRDDLike.scala:167)
   at 
 org.apache.spark.api.java.JavaRDDLike$$anonfun$fn$7$1.apply(JavaRDDLike.scala:167)
   at org.apache.spark.rdd.RDD$$anonfun$13.apply(RDD.scala:601)
   at org.apache.spark.rdd.RDD$$anonfun$13.apply(RDD.scala:601)
   at 
 org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)
   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:263)
   at org.apache.spark.rdd.RDD.iterator(RDD.scala:230)
   at 
 org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
   at 
 org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
   at org.apache.spark.scheduler.Task.run(Task.scala:56)
   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:196)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:744)
 Caused by: java.lang.NullPointerException
   at org.apache.hadoop.hive.ql.io.IOContext.copy(IOContext.java:119)
   at 
 org.apache.hadoop.hive.ql.exec.spark.SparkMapRecordHandler.init(SparkMapRecordHandler.java:97)
   ... 16 more
 {code}
 This failure is again about IOContext object, which needs further 
 investigation.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8937) fix description of hive.security.authorization.sqlstd.confwhitelist.* params

2014-12-13 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8937?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245284#comment-14245284
 ] 

Lefty Leverenz commented on HIVE-8937:
--

Should I document *hive.security.authorization.sqlstd.confwhitelist.append* 
with its original description (HIVE-8534) or wait for this revision?

 fix description of hive.security.authorization.sqlstd.confwhitelist.* params
 

 Key: HIVE-8937
 URL: https://issues.apache.org/jira/browse/HIVE-8937
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.14.0
Reporter: Thejas M Nair
Assignee: Thejas M Nair

 hive.security.authorization.sqlstd.confwhitelist.* param description in 
 HiveConf is incorrect. The expected value is a regex, not comma separated 
 regexes.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8534) sql std auth : update configuration whitelist for 0.14

2014-12-13 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8534?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245285#comment-14245285
 ] 

Lefty Leverenz commented on HIVE-8534:
--

Partially documented:

* [Configuration Properties -- hive.security.authorization.sqlstd.confwhitelist 
| 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-hive.security.authorization.sqlstd.confwhitelist]

 sql std auth : update configuration whitelist for 0.14
 --

 Key: HIVE-8534
 URL: https://issues.apache.org/jira/browse/HIVE-8534
 Project: Hive
  Issue Type: Bug
  Components: Authorization, SQLStandardAuthorization
Reporter: Thejas M Nair
Assignee: Thejas M Nair
Priority: Blocker
  Labels: TODOC14
 Fix For: 0.14.0

 Attachments: HIVE-8534.1.patch, HIVE-8534.2.patch, HIVE-8534.3.patch, 
 HIVE-8534.4.patch, HIVE-8534.5.patch


 New config parameters have been introduced in hive 0.14. SQL standard 
 authorization needs to be updated to allow some new parameters to be set, 
 when the authorization mode is enabled.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8950) Add support in ParquetHiveSerde to create table schema from a parquet file

2014-12-13 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8950?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245286#comment-14245286
 ] 

Hive QA commented on HIVE-8950:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12686977/HIVE-8950.5.patch

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 6715 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_optimize_nullscan
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2071/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2071/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2071/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12686977 - PreCommit-HIVE-TRUNK-Build

 Add support in ParquetHiveSerde to create table schema from a parquet file
 --

 Key: HIVE-8950
 URL: https://issues.apache.org/jira/browse/HIVE-8950
 Project: Hive
  Issue Type: Improvement
Reporter: Ashish Kumar Singh
Assignee: Ashish Kumar Singh
 Attachments: HIVE-8950.1.patch, HIVE-8950.2.patch, HIVE-8950.3.patch, 
 HIVE-8950.4.patch, HIVE-8950.5.patch, HIVE-8950.patch


 PARQUET-76 and PARQUET-47 ask for creating parquet backed tables without 
 having to specify the column names and types. As, parquet files store schema 
 in their footer, it is possible to generate hive schema from parquet file's 
 metadata. This will improve usability of parquet backed tables.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8982) IndexOutOfBounds exception in mapjoin [Spark Branch]

2014-12-13 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8982?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245305#comment-14245305
 ] 

Hive QA commented on HIVE-8982:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12687022/HIVE-8982.2-spark.patch

{color:red}ERROR:{color} -1 due to 4 failed/errored test(s), 7261 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample_islocalmode_hook
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_optimize_nullscan
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_vector_cast_constant
org.apache.hive.hcatalog.streaming.TestStreaming.testTransactionBatchEmptyCommit
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/536/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/536/console
Test logs: 
http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/PreCommit-HIVE-SPARK-Build-536/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 4 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12687022 - PreCommit-HIVE-SPARK-Build

 IndexOutOfBounds exception in mapjoin [Spark Branch]
 

 Key: HIVE-8982
 URL: https://issues.apache.org/jira/browse/HIVE-8982
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Szehon Ho
Assignee: Chao
 Attachments: HIVE-8982.1-spark.patch, HIVE-8982.2-spark.patch


 There are sometimes random failures in spark mapjoin during unit tests like:
 {noformat}
 org.apache.hadoop.hive.ql.metadata.HiveException: 
 java.lang.IndexOutOfBoundsException: Index: 1, Size: 1
   at 
 org.apache.hadoop.hive.ql.exec.SparkHashTableSinkOperator.closeOp(SparkHashTableSinkOperator.java:83)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:598)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:610)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:610)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:610)
   at 
 org.apache.hadoop.hive.ql.exec.spark.SparkMapRecordHandler.close(SparkMapRecordHandler.java:185)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveMapFunctionResultList.closeRecordProcessor(HiveMapFunctionResultList.java:57)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveBaseFunctionResultList$ResultIterator.hasNext(HiveBaseFunctionResultList.java:108)
   at 
 scala.collection.convert.Wrappers$JIteratorWrapper.hasNext(Wrappers.scala:41)
   at scala.collection.Iterator$class.foreach(Iterator.scala:727)
   at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
   at 
 org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$2.apply(AsyncRDDActions.scala:115)
   at 
 org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$2.apply(AsyncRDDActions.scala:115)
   at 
 org.apache.spark.SparkContext$$anonfun$30.apply(SparkContext.scala:1365)
   at 
 org.apache.spark.SparkContext$$anonfun$30.apply(SparkContext.scala:1365)
   at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:61)
   at org.apache.spark.scheduler.Task.run(Task.scala:56)
   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:196)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:744)
 Caused by: java.lang.IndexOutOfBoundsException: Index: 1, Size: 1
   at java.util.ArrayList.rangeCheck(ArrayList.java:635)
   at java.util.ArrayList.get(ArrayList.java:411)
   at 
 org.apache.hadoop.hive.ql.exec.persistence.MapJoinEagerRowContainer.first(MapJoinEagerRowContainer.java:70)
   at 
 org.apache.hadoop.hive.ql.exec.persistence.MapJoinEagerRowContainer.write(MapJoinEagerRowContainer.java:150)
   at 
 org.apache.hadoop.hive.ql.exec.persistence.MapJoinTableContainerSerDe.persist(MapJoinTableContainerSerDe.java:167)
   at 
 org.apache.hadoop.hive.ql.exec.SparkHashTableSinkOperator.flushToFile(SparkHashTableSinkOperator.java:128)
   at 
 org.apache.hadoop.hive.ql.exec.SparkHashTableSinkOperator.closeOp(SparkHashTableSinkOperator.java:77)
   ... 20 more
 org.apache.hadoop.hive.ql.metadata.HiveException: 
 java.lang.IndexOutOfBoundsException: Index: 1, Size: 1
  

[jira] [Commented] (HIVE-9053) select constant in union all followed by group by gives wrong result

2014-12-13 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9053?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245318#comment-14245318
 ] 

Hive QA commented on HIVE-9053:
---



{color:green}Overall{color}: +1 all checks pass

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12686979/HIVE-9053.03.patch

{color:green}SUCCESS:{color} +1 6703 tests passed

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2072/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2072/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2072/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12686979 - PreCommit-HIVE-TRUNK-Build

 select constant in union all followed by group by gives wrong result
 

 Key: HIVE-9053
 URL: https://issues.apache.org/jira/browse/HIVE-9053
 Project: Hive
  Issue Type: Bug
Reporter: Pengcheng Xiong
Assignee: Pengcheng Xiong
 Attachments: HIVE-9053.01.patch, HIVE-9053.02.patch, 
 HIVE-9053.03.patch


 Here is the the way to reproduce with q test:
 select key from (select '1' as key from src union all select key from src)tab 
 group by key;
 will give
 OK
 NULL
 1
 This is not correct as src contains many other keys.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8395) CBO: enable by default

2014-12-13 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8395?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245367#comment-14245367
 ] 

Hive QA commented on HIVE-8395:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12686988/HIVE-8395.32.patch

{color:red}ERROR:{color} -1 due to 27 failed/errored test(s), 6702 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_constantPropagateForSubQuery
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_filter_join_breaktask
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_sort_1_23
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_groupby_sort_skew_1_23
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_input23
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join17
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join32
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join32_lessSize
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join33
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join34
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join35
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_join9
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_louter_join_ppr
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_mapjoin_mapjoin
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_metadataonly1
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_optimize_nullscan
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_outer_join_ppr
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_pcr
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ppd_join_filter
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ppd_vc
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_router_join_ppr
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_udf_case_column_pruning
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_optimize_nullscan
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_constprog_partitioner
org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_join1
org.apache.hadoop.hive.ql.TestMTQueries.testMTQueries1
org.apache.hive.hcatalog.streaming.TestStreaming.testEndpointConnection
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2073/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2073/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2073/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 27 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12686988 - PreCommit-HIVE-TRUNK-Build

 CBO: enable by default
 --

 Key: HIVE-8395
 URL: https://issues.apache.org/jira/browse/HIVE-8395
 Project: Hive
  Issue Type: Improvement
  Components: CBO
Reporter: Sergey Shelukhin
Assignee: Sergey Shelukhin
 Fix For: 0.15.0

 Attachments: HIVE-8395-27-28-delta.patch, 
 HIVE-8395-28-29-delta.patch, HIVE-8395.01.patch, HIVE-8395.02.patch, 
 HIVE-8395.03.patch, HIVE-8395.04.patch, HIVE-8395.05.patch, 
 HIVE-8395.06.patch, HIVE-8395.07.patch, HIVE-8395.08.patch, 
 HIVE-8395.09.patch, HIVE-8395.10.patch, HIVE-8395.11.patch, 
 HIVE-8395.12.patch, HIVE-8395.12.patch, HIVE-8395.13.patch, 
 HIVE-8395.13.patch, HIVE-8395.14.patch, HIVE-8395.15.patch, 
 HIVE-8395.16.patch, HIVE-8395.17.patch, HIVE-8395.18.patch, 
 HIVE-8395.18.patch, HIVE-8395.19.patch, HIVE-8395.20.patch, 
 HIVE-8395.21.patch, HIVE-8395.22.patch, HIVE-8395.23.patch, 
 HIVE-8395.23.withon.patch, HIVE-8395.24.patch, HIVE-8395.25.patch, 
 HIVE-8395.25.patch, HIVE-8395.26.patch, HIVE-8395.27.patch, 
 HIVE-8395.28.patch, HIVE-8395.29.patch, HIVE-8395.30.patch, 
 HIVE-8395.31.patch, HIVE-8395.32.patch, HIVE-8395.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8982) IndexOutOfBounds exception in mapjoin [Spark Branch]

2014-12-13 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8982?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-8982:
--
   Resolution: Fixed
Fix Version/s: spark-branch
   Status: Resolved  (was: Patch Available)

Committed to Spark branch. Thanks to Chao for this nice finding and fix.

 IndexOutOfBounds exception in mapjoin [Spark Branch]
 

 Key: HIVE-8982
 URL: https://issues.apache.org/jira/browse/HIVE-8982
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Szehon Ho
Assignee: Chao
 Fix For: spark-branch

 Attachments: HIVE-8982.1-spark.patch, HIVE-8982.2-spark.patch


 There are sometimes random failures in spark mapjoin during unit tests like:
 {noformat}
 org.apache.hadoop.hive.ql.metadata.HiveException: 
 java.lang.IndexOutOfBoundsException: Index: 1, Size: 1
   at 
 org.apache.hadoop.hive.ql.exec.SparkHashTableSinkOperator.closeOp(SparkHashTableSinkOperator.java:83)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:598)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:610)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:610)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:610)
   at 
 org.apache.hadoop.hive.ql.exec.spark.SparkMapRecordHandler.close(SparkMapRecordHandler.java:185)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveMapFunctionResultList.closeRecordProcessor(HiveMapFunctionResultList.java:57)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveBaseFunctionResultList$ResultIterator.hasNext(HiveBaseFunctionResultList.java:108)
   at 
 scala.collection.convert.Wrappers$JIteratorWrapper.hasNext(Wrappers.scala:41)
   at scala.collection.Iterator$class.foreach(Iterator.scala:727)
   at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
   at 
 org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$2.apply(AsyncRDDActions.scala:115)
   at 
 org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$2.apply(AsyncRDDActions.scala:115)
   at 
 org.apache.spark.SparkContext$$anonfun$30.apply(SparkContext.scala:1365)
   at 
 org.apache.spark.SparkContext$$anonfun$30.apply(SparkContext.scala:1365)
   at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:61)
   at org.apache.spark.scheduler.Task.run(Task.scala:56)
   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:196)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:744)
 Caused by: java.lang.IndexOutOfBoundsException: Index: 1, Size: 1
   at java.util.ArrayList.rangeCheck(ArrayList.java:635)
   at java.util.ArrayList.get(ArrayList.java:411)
   at 
 org.apache.hadoop.hive.ql.exec.persistence.MapJoinEagerRowContainer.first(MapJoinEagerRowContainer.java:70)
   at 
 org.apache.hadoop.hive.ql.exec.persistence.MapJoinEagerRowContainer.write(MapJoinEagerRowContainer.java:150)
   at 
 org.apache.hadoop.hive.ql.exec.persistence.MapJoinTableContainerSerDe.persist(MapJoinTableContainerSerDe.java:167)
   at 
 org.apache.hadoop.hive.ql.exec.SparkHashTableSinkOperator.flushToFile(SparkHashTableSinkOperator.java:128)
   at 
 org.apache.hadoop.hive.ql.exec.SparkHashTableSinkOperator.closeOp(SparkHashTableSinkOperator.java:77)
   ... 20 more
 org.apache.hadoop.hive.ql.metadata.HiveException: 
 java.lang.IndexOutOfBoundsException: Index: 1, Size: 1
   at 
 org.apache.hadoop.hive.ql.exec.SparkHashTableSinkOperator.closeOp(SparkHashTableSinkOperator.java:83)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:598)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:610)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:610)
   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:610)
   at 
 org.apache.hadoop.hive.ql.exec.spark.SparkMapRecordHandler.close(SparkMapRecordHandler.java:185)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveMapFunctionResultList.closeRecordProcessor(HiveMapFunctionResultList.java:57)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveBaseFunctionResultList$ResultIterator.hasNext(HiveBaseFunctionResultList.java:108)
   at 
 scala.collection.convert.Wrappers$JIteratorWrapper.hasNext(Wrappers.scala:41)
   at scala.collection.Iterator$class.foreach(Iterator.scala:727)
   at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
   at 
 org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$2.apply(AsyncRDDActions.scala:115)
   at 
 

[jira] [Commented] (HIVE-8911) Enable mapjoin hints [Spark Branch]

2014-12-13 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8911?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245386#comment-14245386
 ] 

Xuefu Zhang commented on HIVE-8911:
---

ppd_join4 and smb_mapjoin_25 are not related, but bucketmapjoin10 might be.

 Enable mapjoin hints [Spark Branch]
 ---

 Key: HIVE-8911
 URL: https://issues.apache.org/jira/browse/HIVE-8911
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Szehon Ho
Assignee: Chao
 Attachments: HIVE-8911.1-spark.patch, HIVE-8911.2-spark.patch, 
 HIVE-8911.3-spark.patch, HIVE-8911.4-spark.patch, HIVE-8911.5-spark.patch, 
 HIVE-8911.6-spark.patch


 Currently the big table selection in a mapjoin is based on stats.
 We should also enable the big-table selection based on hints.  See class 
 MapJoinProcessor.  This is a logical-optimizer class, so we should be able to 
 re-use this without too many changes to hook up with SparkMapJoinResolver.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9059) Remove wrappers for SparkJobInfo and SparkStageInfo

2014-12-13 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9059?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-9059:
--
Attachment: HIVE-9059.1-spark.patch

 Remove wrappers for SparkJobInfo and SparkStageInfo
 ---

 Key: HIVE-9059
 URL: https://issues.apache.org/jira/browse/HIVE-9059
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Xuefu Zhang
Assignee: Chengxiang Li
 Attachments: HIVE-9059.1-spark.patch, HIVE-9059.1-spark.patch


 SPARK-4567 is resolved. We can remove the wrappers we added to solve the 
 serailization issues.
 CLEAR LIBRARY CACHE



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9059) Remove wrappers for SparkJobInfo and SparkStageInfo

2014-12-13 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9059?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-9059:
--
Description: SPARK-4567 is resolved. We can remove the wrappers we added to 
solve the serailization issues.  (was: SPARK-4567 is resolved. We can remove 
the wrappers we added to solve the serailization issues.

CLEAR LIBRARY CACHE)

 Remove wrappers for SparkJobInfo and SparkStageInfo
 ---

 Key: HIVE-9059
 URL: https://issues.apache.org/jira/browse/HIVE-9059
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Xuefu Zhang
Assignee: Chengxiang Li
 Attachments: HIVE-9059.1-spark.patch, HIVE-9059.1-spark.patch


 SPARK-4567 is resolved. We can remove the wrappers we added to solve the 
 serailization issues.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 28941: HIVE-8988

2014-12-13 Thread Jesús Camacho Rodríguez

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/28941/
---

(Updated Dec. 13, 2014, 4:06 p.m.)


Review request for hive, John Pullokkaran and Julian Hyde.


Changes
---

Created HiveGroupingID operator.


Bugs: HIVE-8988
https://issues.apache.org/jira/browse/HIVE-8988


Repository: hive-git


Description
---

HIVE-8988


Diffs (updated)
-

  
ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/reloperators/HiveGroupingID.java
 PRE-CREATION 
  
ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ASTBuilder.java
 e6e6fe3a1152df82c0176bc9617ddad0ed404f0b 
  
ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/ASTConverter.java
 c02a65e2041e4742a56cf4a935da0a7c04d18fdb 
  
ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/translator/RexNodeConverter.java
 56cb4e81e16a077c5e4fc11b9d651ef8c3c241af 
  ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java 
6c8624f59d9cab53fe096e76d23852c922bbbc30 
  ql/src/test/queries/clientpositive/groupby_cube1.q 
c12720b27059075050fc92d9f31420c081303699 
  ql/src/test/results/clientpositive/groupby_cube1.q.out 
7b5d70ae8ffce47a4b351ed9dfedcd15ab1e139c 

Diff: https://reviews.apache.org/r/28941/diff/


Testing
---


Thanks,

Jesús Camacho Rodríguez



[jira] [Updated] (HIVE-8988) Support advanced aggregation in Hive to Calcite path

2014-12-13 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8988?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-8988:
--
Attachment: HIVE-8988.02.patch

[~julianhyde], thanks for the feedback. I uploaded a new patch.

 Support advanced aggregation in Hive to Calcite path 
 -

 Key: HIVE-8988
 URL: https://issues.apache.org/jira/browse/HIVE-8988
 Project: Hive
  Issue Type: Improvement
Affects Versions: 0.15.0
Reporter: Jesus Camacho Rodriguez
Assignee: Jesus Camacho Rodriguez
  Labels: grouping, logical, optiq
 Fix For: 0.15.0

 Attachments: HIVE-8988.01.patch, HIVE-8988.02.patch, HIVE-8988.patch


 CLEAR LIBRARY CACHE
 To close the gap between Hive and Calcite, we need to support the translation 
 of GroupingSets into Calcite; currently this is not implemented.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9084) Investigate IOContext object initialization problem [Spark Branch]

2014-12-13 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9084?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-9084:
--
Attachment: HIVE-9084.2-spark.patch

 Investigate IOContext object initialization problem [Spark Branch]
 --

 Key: HIVE-9084
 URL: https://issues.apache.org/jira/browse/HIVE-9084
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Xuefu Zhang
Assignee: Xuefu Zhang
 Attachments: HIVE-9084.1-spark.patch, HIVE-9084.2-spark.patch, 
 HIVE-9084.2-spark.patch


 In recent ptest run (Test results: 
 http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/511/testReport),
  test groupby_multi_single_reducer.q failed w/ the following stacktrace:
 {code}
 java.lang.RuntimeException: Map operator initialization failed
   at 
 org.apache.hadoop.hive.ql.exec.spark.SparkMapRecordHandler.init(SparkMapRecordHandler.java:136)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveMapFunction.call(HiveMapFunction.java:54)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveMapFunction.call(HiveMapFunction.java:29)
   at 
 org.apache.spark.api.java.JavaRDDLike$$anonfun$fn$7$1.apply(JavaRDDLike.scala:167)
   at 
 org.apache.spark.api.java.JavaRDDLike$$anonfun$fn$7$1.apply(JavaRDDLike.scala:167)
   at org.apache.spark.rdd.RDD$$anonfun$13.apply(RDD.scala:601)
   at org.apache.spark.rdd.RDD$$anonfun$13.apply(RDD.scala:601)
   at 
 org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)
   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:263)
   at org.apache.spark.rdd.RDD.iterator(RDD.scala:230)
   at 
 org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
   at 
 org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
   at org.apache.spark.scheduler.Task.run(Task.scala:56)
   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:196)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:744)
 Caused by: java.lang.NullPointerException
   at org.apache.hadoop.hive.ql.io.IOContext.copy(IOContext.java:119)
   at 
 org.apache.hadoop.hive.ql.exec.spark.SparkMapRecordHandler.init(SparkMapRecordHandler.java:97)
   ... 16 more
 {code}
 This failure is again about IOContext object, which needs further 
 investigation.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8911) Enable mapjoin hints [Spark Branch]

2014-12-13 Thread Chao (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8911?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245401#comment-14245401
 ] 

Chao commented on HIVE-8911:


bucketmapjoin10 failed because of the IndexOutOfBoundException:

{noformat}
java.lang.RuntimeException: Hive Runtime Error while closing operators
at 
org.apache.hadoop.hive.ql.exec.spark.SparkMapRecordHandler.close(SparkMapRecordHandler.java:207)
at 
org.apache.hadoop.hive.ql.exec.spark.HiveMapFunctionResultList.closeRecordProcessor(HiveMapFunctionResultList.java:57)
at 
org.apache.hadoop.hive.ql.exec.spark.HiveBaseFunctionResultList$ResultIterator.hasNext(HiveBaseFunctionResultList.java:108)
at 
scala.collection.convert.Wrappers$JIteratorWrapper.hasNext(Wrappers.scala:41)
at scala.collection.Iterator$class.foreach(Iterator.scala:727)
at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
at 
org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$2.apply(AsyncRDDActions.scala:115)
at 
org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$2.apply(AsyncRDDActions.scala:115)
at 
org.apache.spark.SparkContext$$anonfun$30.apply(SparkContext.scala:1390)
at 
org.apache.spark.SparkContext$$anonfun$30.apply(SparkContext.scala:1390)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:61)
at org.apache.spark.scheduler.Task.run(Task.scala:56)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:196)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:744)
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
java.lang.IndexOutOfBoundsException: Index: 1, Size: 1
at 
org.apache.hadoop.hive.ql.exec.SparkHashTableSinkOperator.closeOp(SparkHashTableSinkOperator.java:87)
at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:598)
at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:610)
at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:610)
at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:610)
at 
org.apache.hadoop.hive.ql.exec.spark.SparkMapRecordHandler.close(SparkMapRecordHandler.java:185)
... 15 more
Caused by: java.lang.IndexOutOfBoundsException: Index: 1, Size: 1
at java.util.ArrayList.rangeCheck(ArrayList.java:635)
at java.util.ArrayList.get(ArrayList.java:411)
at 
org.apache.hadoop.hive.ql.exec.persistence.MapJoinEagerRowContainer.first(MapJoinEagerRowContainer.java:70)
at 
org.apache.hadoop.hive.ql.exec.persistence.MapJoinEagerRowContainer.write(MapJoinEagerRowContainer.java:149)
at 
org.apache.hadoop.hive.ql.exec.persistence.MapJoinTableContainerSerDe.persist(MapJoinTableContainerSerDe.java:170)
at 
org.apache.hadoop.hive.ql.exec.SparkHashTableSinkOperator.flushToFile(SparkHashTableSinkOperator.java:142)
at 
org.apache.hadoop.hive.ql.exec.SparkHashTableSinkOperator.closeOp(SparkHashTableSinkOperator.java:81)
... 20 more
{noformat}

This run may have not used the patch from HIVE-8982 - it is committed at 
07:33am, while the run ended at 07:49am.

 Enable mapjoin hints [Spark Branch]
 ---

 Key: HIVE-8911
 URL: https://issues.apache.org/jira/browse/HIVE-8911
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Szehon Ho
Assignee: Chao
 Attachments: HIVE-8911.1-spark.patch, HIVE-8911.2-spark.patch, 
 HIVE-8911.3-spark.patch, HIVE-8911.4-spark.patch, HIVE-8911.5-spark.patch, 
 HIVE-8911.6-spark.patch


 Currently the big table selection in a mapjoin is based on stats.
 We should also enable the big-table selection based on hints.  See class 
 MapJoinProcessor.  This is a logical-optimizer class, so we should be able to 
 re-use this without too many changes to hook up with SparkMapJoinResolver.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8988) Support advanced aggregation in Hive to Calcite path

2014-12-13 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8988?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245408#comment-14245408
 ] 

Hive QA commented on HIVE-8988:
---



{color:red}Overall{color}: -1 no tests executed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12687044/HIVE-8988.02.patch

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2074/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2074/console
Test logs: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2074/

Messages:
{noformat}
 This message was trimmed, see log for full details 
warning(200): IdentifiersParser.g:68:4: 
Decision can match input such as LPAREN LPAREN KW_FALSE using multiple 
alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:68:4: 
Decision can match input such as LPAREN KW_NOT SmallintLiteral using multiple 
alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:115:5: 
Decision can match input such as KW_CLUSTER KW_BY LPAREN using multiple 
alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:127:5: 
Decision can match input such as KW_PARTITION KW_BY LPAREN using multiple 
alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:138:5: 
Decision can match input such as KW_DISTRIBUTE KW_BY LPAREN using multiple 
alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:149:5: 
Decision can match input such as KW_SORT KW_BY LPAREN using multiple 
alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:166:7: 
Decision can match input such as STAR using multiple alternatives: 1, 2

As a result, alternative(s) 2 were disabled for that input
warning(200): IdentifiersParser.g:179:5: 
Decision can match input such as KW_ARRAY using multiple alternatives: 2, 6

As a result, alternative(s) 6 were disabled for that input
warning(200): IdentifiersParser.g:179:5: 
Decision can match input such as KW_STRUCT using multiple alternatives: 4, 6

As a result, alternative(s) 6 were disabled for that input
warning(200): IdentifiersParser.g:179:5: 
Decision can match input such as KW_UNIONTYPE using multiple alternatives: 5, 
6

As a result, alternative(s) 6 were disabled for that input
warning(200): IdentifiersParser.g:270:5: 
Decision can match input such as KW_TRUE using multiple alternatives: 2, 7

As a result, alternative(s) 7 were disabled for that input
warning(200): IdentifiersParser.g:270:5: 
Decision can match input such as KW_FALSE using multiple alternatives: 2, 7

As a result, alternative(s) 7 were disabled for that input
warning(200): IdentifiersParser.g:270:5: 
Decision can match input such as KW_NULL using multiple alternatives: 1, 7

As a result, alternative(s) 7 were disabled for that input
warning(200): IdentifiersParser.g:401:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_SORT KW_BY 
using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:401:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_MAP LPAREN 
using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:401:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_GROUP 
KW_BY using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:401:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_CLUSTER 
KW_BY using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:401:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_DISTRIBUTE 
KW_BY using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:401:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_INSERT 
KW_OVERWRITE using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:401:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_UNION 
KW_ALL using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that input
warning(200): IdentifiersParser.g:401:5: 
Decision can match input such as {KW_LIKE, KW_REGEXP, KW_RLIKE} KW_INSERT 
KW_INTO using multiple alternatives: 2, 9

As a result, alternative(s) 9 were disabled for that 

Hive-0.14 - Build # 776 - Fixed

2014-12-13 Thread Apache Jenkins Server
Changes for Build #775

Changes for Build #776



No tests ran.

The Apache Jenkins build system has built Hive-0.14 (build #776)

Status: Fixed

Check console output at https://builds.apache.org/job/Hive-0.14/776/ to view 
the results.

[jira] [Updated] (HIVE-8911) Enable mapjoin hints [Spark Branch]

2014-12-13 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8911?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-8911:
--
   Resolution: Fixed
Fix Version/s: spark-branch
   Status: Resolved  (was: Patch Available)

Awesome! Patch committed to Spark branch. Thanks, Chao.

 Enable mapjoin hints [Spark Branch]
 ---

 Key: HIVE-8911
 URL: https://issues.apache.org/jira/browse/HIVE-8911
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Szehon Ho
Assignee: Chao
 Fix For: spark-branch

 Attachments: HIVE-8911.1-spark.patch, HIVE-8911.2-spark.patch, 
 HIVE-8911.3-spark.patch, HIVE-8911.4-spark.patch, HIVE-8911.5-spark.patch, 
 HIVE-8911.6-spark.patch


 Currently the big table selection in a mapjoin is based on stats.
 We should also enable the big-table selection based on hints.  See class 
 MapJoinProcessor.  This is a logical-optimizer class, so we should be able to 
 re-use this without too many changes to hook up with SparkMapJoinResolver.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9053) select constant in union all followed by group by gives wrong result

2014-12-13 Thread Pengcheng Xiong (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9053?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245445#comment-14245445
 ] 

Pengcheng Xiong commented on HIVE-9053:
---

Good to hear that all tests passed. I still would like to investigate another 
small corner issue before checked in... Thanks!

 select constant in union all followed by group by gives wrong result
 

 Key: HIVE-9053
 URL: https://issues.apache.org/jira/browse/HIVE-9053
 Project: Hive
  Issue Type: Bug
Reporter: Pengcheng Xiong
Assignee: Pengcheng Xiong
 Attachments: HIVE-9053.01.patch, HIVE-9053.02.patch, 
 HIVE-9053.03.patch


 Here is the the way to reproduce with q test:
 select key from (select '1' as key from src union all select key from src)tab 
 group by key;
 will give
 OK
 NULL
 1
 This is not correct as src contains many other keys.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9059) Remove wrappers for SparkJobInfo and SparkStageInfo

2014-12-13 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9059?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245495#comment-14245495
 ] 

Xuefu Zhang commented on HIVE-9059:
---

[~chengxiang li], the patch seems making the ptest hang for some reason.

 Remove wrappers for SparkJobInfo and SparkStageInfo
 ---

 Key: HIVE-9059
 URL: https://issues.apache.org/jira/browse/HIVE-9059
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Xuefu Zhang
Assignee: Chengxiang Li
 Attachments: HIVE-9059.1-spark.patch, HIVE-9059.1-spark.patch


 SPARK-4567 is resolved. We can remove the wrappers we added to solve the 
 serailization issues.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8950) Add support in ParquetHiveSerde to create table schema from a parquet file

2014-12-13 Thread Ashish Kumar Singh (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8950?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245509#comment-14245509
 ] 

Ashish Kumar Singh commented on HIVE-8950:
--

The test failure above does not look related to this patch.

 Add support in ParquetHiveSerde to create table schema from a parquet file
 --

 Key: HIVE-8950
 URL: https://issues.apache.org/jira/browse/HIVE-8950
 Project: Hive
  Issue Type: Improvement
Reporter: Ashish Kumar Singh
Assignee: Ashish Kumar Singh
 Attachments: HIVE-8950.1.patch, HIVE-8950.2.patch, HIVE-8950.3.patch, 
 HIVE-8950.4.patch, HIVE-8950.5.patch, HIVE-8950.patch


 PARQUET-76 and PARQUET-47 ask for creating parquet backed tables without 
 having to specify the column names and types. As, parquet files store schema 
 in their footer, it is possible to generate hive schema from parquet file's 
 metadata. This will improve usability of parquet backed tables.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9059) Remove wrappers for SparkJobInfo and SparkStageInfo

2014-12-13 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9059?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245529#comment-14245529
 ] 

Hive QA commented on HIVE-9059:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12687043/HIVE-9059.1-spark.patch

{color:red}ERROR:{color} -1 due to 39 failed/errored test(s), 6731 tests 
executed
*Failed tests:*
{noformat}
TestJdbcWithLocalClusterSpark - did not produce a TEST-*.xml file
TestSparkCliDriver-auto_join30.q-sort_merge_join_desc_3.q-join2.q-and-12-more - 
did not produce a TEST-*.xml file
TestSparkCliDriver-avro_joins.q-join36.q-join4.q-and-12-more - did not produce 
a TEST-*.xml file
TestSparkCliDriver-bucketmapjoin_negative.q-parallel_join0.q-insert_into1.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-bucketsortoptimize_insert_7.q-skewjoin_noskew.q-mapreduce1.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-groupby10.q-union5.q-load_dyn_part15.q-and-12-more - did not 
produce a TEST-*.xml file
TestSparkCliDriver-groupby3_map.q-skewjoinopt8.q-union_remove_1.q-and-12-more - 
did not produce a TEST-*.xml file
TestSparkCliDriver-groupby4.q-tez_joins_explain.q-load_dyn_part3.q-and-12-more 
- did not produce a TEST-*.xml file
TestSparkCliDriver-groupby_complex_types.q-auto_join9.q-groupby_map_ppr.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-groupby_map_ppr_multi_distinct.q-vectorization_16.q-multi_insert_mixed.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-input1_limit.q-groupby8_map.q-varchar_join1.q-and-12-more - 
did not produce a TEST-*.xml file
TestSparkCliDriver-join11.q-join18.q-groupby2.q-and-12-more - did not produce a 
TEST-*.xml file
TestSparkCliDriver-join13.q-sample2.q-join_reorder3.q-and-12-more - did not 
produce a TEST-*.xml file
TestSparkCliDriver-join39.q-bucketmapjoin10.q-bucket_map_join_tez1.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-join9.q-ppd_outer_join2.q-join_casesensitive.q-and-12-more - 
did not produce a TEST-*.xml file
TestSparkCliDriver-join_cond_pushdown_3.q-groupby7.q-auto_join17.q-and-12-more 
- did not produce a TEST-*.xml file
TestSparkCliDriver-join_cond_pushdown_unqual4.q-bucketmapjoin12.q-avro_decimal_native.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-join_filters.q-bucketmapjoin6.q-smb_mapjoin_4.q-and-12-more 
- did not produce a TEST-*.xml file
TestSparkCliDriver-list_bucket_dml_2.q-multi_insert.q-join5.q-and-12-more - did 
not produce a TEST-*.xml file
TestSparkCliDriver-load_dyn_part5.q-load_dyn_part2.q-auto_join10.q-and-12-more 
- did not produce a TEST-*.xml file
TestSparkCliDriver-parallel_join1.q-escape_distributeby1.q-timestamp_null.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-ppd_join4.q-skewjoinopt3.q-auto_join1.q-and-12-more - did 
not produce a TEST-*.xml file
TestSparkCliDriver-ppd_transform.q-auto_sortmerge_join_7.q-date_udf.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-ptf_general_queries.q-bucketmapjoin3.q-enforce_order.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-ptf_seqfile.q-auto_join18.q-vectorization_10.q-and-12-more - 
did not produce a TEST-*.xml file
TestSparkCliDriver-skewjoin_union_remove_2.q-auto_join14.q-auto_join20.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-skewjoinopt15.q-union_remove_7.q-avro_joins_native.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-smb_mapjoin_15.q-auto_join18_multi_distinct.q-mapreduce2.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-sort_merge_join_desc_1.q-order.q-union2.q-and-12-more - did 
not produce a TEST-*.xml file
TestSparkCliDriver-stats12.q-auto_join24.q-vector_decimal_aggregate.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-stats13.q-stats2.q-ppd_gby_join.q-and-12-more - did not 
produce a TEST-*.xml file
TestSparkCliDriver-table_access_keys_stats.q-bucketsortoptimize_insert_4.q-union_remove_9.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-timestamp_lazy.q-auto_join_reordering_values.q-union29.q-and-12-more
 - did not produce a TEST-*.xml file
TestSparkCliDriver-transform_ppr2.q-join20.q-decimal_join.q-and-11-more - did 
not produce a TEST-*.xml file
TestSparkCliDriver-vector_distinct_2.q-join15.q-union19.q-and-12-more - did not 
produce a TEST-*.xml file
TestSparkCliDriver-vectorization_13.q-auto_sortmerge_join_13.q-bucket3.q-and-12-more
 - did not produce a TEST-*.xml file
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample_islocalmode_hook
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_optimize_nullscan
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_vector_cast_constant
{noformat}

Test results: 

[jira] [Commented] (HIVE-9084) Investigate IOContext object initialization problem [Spark Branch]

2014-12-13 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9084?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245582#comment-14245582
 ] 

Hive QA commented on HIVE-9084:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12687045/HIVE-9084.2-spark.patch

{color:red}ERROR:{color} -1 due to 7 failed/errored test(s), 7233 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_nonmr_fetch
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample_islocalmode_hook
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_virtual_column
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_optimize_nullscan
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_vector_cast_constant
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_outer_join_ppr
org.apache.hadoop.hive.ql.exec.TestOperators.testMapOperator
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/538/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/538/console
Test logs: 
http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/PreCommit-HIVE-SPARK-Build-538/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 7 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12687045 - PreCommit-HIVE-SPARK-Build

 Investigate IOContext object initialization problem [Spark Branch]
 --

 Key: HIVE-9084
 URL: https://issues.apache.org/jira/browse/HIVE-9084
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Xuefu Zhang
Assignee: Xuefu Zhang
 Attachments: HIVE-9084.1-spark.patch, HIVE-9084.2-spark.patch, 
 HIVE-9084.2-spark.patch


 In recent ptest run (Test results: 
 http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/511/testReport),
  test groupby_multi_single_reducer.q failed w/ the following stacktrace:
 {code}
 java.lang.RuntimeException: Map operator initialization failed
   at 
 org.apache.hadoop.hive.ql.exec.spark.SparkMapRecordHandler.init(SparkMapRecordHandler.java:136)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveMapFunction.call(HiveMapFunction.java:54)
   at 
 org.apache.hadoop.hive.ql.exec.spark.HiveMapFunction.call(HiveMapFunction.java:29)
   at 
 org.apache.spark.api.java.JavaRDDLike$$anonfun$fn$7$1.apply(JavaRDDLike.scala:167)
   at 
 org.apache.spark.api.java.JavaRDDLike$$anonfun$fn$7$1.apply(JavaRDDLike.scala:167)
   at org.apache.spark.rdd.RDD$$anonfun$13.apply(RDD.scala:601)
   at org.apache.spark.rdd.RDD$$anonfun$13.apply(RDD.scala:601)
   at 
 org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)
   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:263)
   at org.apache.spark.rdd.RDD.iterator(RDD.scala:230)
   at 
 org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
   at 
 org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
   at org.apache.spark.scheduler.Task.run(Task.scala:56)
   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:196)
   at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   at java.lang.Thread.run(Thread.java:744)
 Caused by: java.lang.NullPointerException
   at org.apache.hadoop.hive.ql.io.IOContext.copy(IOContext.java:119)
   at 
 org.apache.hadoop.hive.ql.exec.spark.SparkMapRecordHandler.init(SparkMapRecordHandler.java:97)
   ... 16 more
 {code}
 This failure is again about IOContext object, which needs further 
 investigation.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8911) Enable mapjoin hints [Spark Branch]

2014-12-13 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8911?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245668#comment-14245668
 ] 

Lefty Leverenz commented on HIVE-8911:
--

Does this need documentation?  If so, please add the TODOC-SPARK label.

 Enable mapjoin hints [Spark Branch]
 ---

 Key: HIVE-8911
 URL: https://issues.apache.org/jira/browse/HIVE-8911
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Szehon Ho
Assignee: Chao
 Fix For: spark-branch

 Attachments: HIVE-8911.1-spark.patch, HIVE-8911.2-spark.patch, 
 HIVE-8911.3-spark.patch, HIVE-8911.4-spark.patch, HIVE-8911.5-spark.patch, 
 HIVE-8911.6-spark.patch


 Currently the big table selection in a mapjoin is based on stats.
 We should also enable the big-table selection based on hints.  See class 
 MapJoinProcessor.  This is a logical-optimizer class, so we should be able to 
 re-use this without too many changes to hook up with SparkMapJoinResolver.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9079) Hive hangs while failed to get executorCount[Spark Branch]

2014-12-13 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9079?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-9079:
--
Attachment: HIVE-9079.1-spark.patch

 Hive hangs while failed to get executorCount[Spark Branch]
 --

 Key: HIVE-9079
 URL: https://issues.apache.org/jira/browse/HIVE-9079
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
Assignee: Chengxiang Li
  Labels: Spark-M4
 Fix For: spark-branch

 Attachments: HIVE-9079.1-spark.patch, HIVE-9079.1-spark.patch


 Hive on Spark get executorCount from RSC to dynamically set reduce number, it 
 use future.get() to wait result, which may hangs forever if remote side 
 failed with no notification. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HIVE-9096) GenericUDF may be left unclosed in PartitionPrune#visitCall()

2014-12-13 Thread Ted Yu (JIRA)
Ted Yu created HIVE-9096:


 Summary: GenericUDF may be left unclosed in 
PartitionPrune#visitCall()
 Key: HIVE-9096
 URL: https://issues.apache.org/jira/browse/HIVE-9096
 Project: Hive
  Issue Type: Bug
Reporter: Ted Yu
Priority: Minor


In 
ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/PartitionPrune.java
 :
{code}
  GenericUDF hiveUDF = SqlFunctionConverter.getHiveUDF(call.getOperator(),
  call.getType(), call.operands.size());
  if (hiveUDF != null 
  !FunctionRegistry.isDeterministic(hiveUDF)) {
return null;
  }
{code}
hiveUDF should be closed upon return.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-8911) Enable mapjoin hints [Spark Branch]

2014-12-13 Thread Chao (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-8911?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245672#comment-14245672
 ] 

Chao commented on HIVE-8911:


Yes, I think this need doc. Thanks [~leftylev] for reminding me!

 Enable mapjoin hints [Spark Branch]
 ---

 Key: HIVE-8911
 URL: https://issues.apache.org/jira/browse/HIVE-8911
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Szehon Ho
Assignee: Chao
  Labels: TODOC-SPARK
 Fix For: spark-branch

 Attachments: HIVE-8911.1-spark.patch, HIVE-8911.2-spark.patch, 
 HIVE-8911.3-spark.patch, HIVE-8911.4-spark.patch, HIVE-8911.5-spark.patch, 
 HIVE-8911.6-spark.patch


 Currently the big table selection in a mapjoin is based on stats.
 We should also enable the big-table selection based on hints.  See class 
 MapJoinProcessor.  This is a logical-optimizer class, so we should be able to 
 re-use this without too many changes to hook up with SparkMapJoinResolver.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-8911) Enable mapjoin hints [Spark Branch]

2014-12-13 Thread Chao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-8911?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chao updated HIVE-8911:
---
Labels: TODOC-SPARK  (was: )

 Enable mapjoin hints [Spark Branch]
 ---

 Key: HIVE-8911
 URL: https://issues.apache.org/jira/browse/HIVE-8911
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Affects Versions: spark-branch
Reporter: Szehon Ho
Assignee: Chao
  Labels: TODOC-SPARK
 Fix For: spark-branch

 Attachments: HIVE-8911.1-spark.patch, HIVE-8911.2-spark.patch, 
 HIVE-8911.3-spark.patch, HIVE-8911.4-spark.patch, HIVE-8911.5-spark.patch, 
 HIVE-8911.6-spark.patch


 Currently the big table selection in a mapjoin is based on stats.
 We should also enable the big-table selection based on hints.  See class 
 MapJoinProcessor.  This is a logical-optimizer class, so we should be able to 
 re-use this without too many changes to hook up with SparkMapJoinResolver.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9053) select constant in union all followed by group by gives wrong result

2014-12-13 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9053?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-9053:
--
Attachment: HIVE-9053.04.patch

 select constant in union all followed by group by gives wrong result
 

 Key: HIVE-9053
 URL: https://issues.apache.org/jira/browse/HIVE-9053
 Project: Hive
  Issue Type: Bug
Reporter: Pengcheng Xiong
Assignee: Pengcheng Xiong
 Attachments: HIVE-9053.01.patch, HIVE-9053.02.patch, 
 HIVE-9053.03.patch, HIVE-9053.04.patch


 Here is the the way to reproduce with q test:
 select key from (select '1' as key from src union all select key from src)tab 
 group by key;
 will give
 OK
 NULL
 1
 This is not correct as src contains many other keys.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9053) select constant in union all followed by group by gives wrong result

2014-12-13 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9053?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-9053:
--
Attachment: HIVE-9053.04.patch

add a new test case which will still fail on previous patch.

 select constant in union all followed by group by gives wrong result
 

 Key: HIVE-9053
 URL: https://issues.apache.org/jira/browse/HIVE-9053
 Project: Hive
  Issue Type: Bug
Reporter: Pengcheng Xiong
Assignee: Pengcheng Xiong
 Attachments: HIVE-9053.01.patch, HIVE-9053.02.patch, 
 HIVE-9053.03.patch, HIVE-9053.04.patch


 Here is the the way to reproduce with q test:
 select key from (select '1' as key from src union all select key from src)tab 
 group by key;
 will give
 OK
 NULL
 1
 This is not correct as src contains many other keys.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9053) select constant in union all followed by group by gives wrong result

2014-12-13 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9053?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-9053:
--
Status: Open  (was: Patch Available)

 select constant in union all followed by group by gives wrong result
 

 Key: HIVE-9053
 URL: https://issues.apache.org/jira/browse/HIVE-9053
 Project: Hive
  Issue Type: Bug
Reporter: Pengcheng Xiong
Assignee: Pengcheng Xiong
 Attachments: HIVE-9053.01.patch, HIVE-9053.02.patch, 
 HIVE-9053.03.patch, HIVE-9053.04.patch


 Here is the the way to reproduce with q test:
 select key from (select '1' as key from src union all select key from src)tab 
 group by key;
 will give
 OK
 NULL
 1
 This is not correct as src contains many other keys.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9053) select constant in union all followed by group by gives wrong result

2014-12-13 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9053?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-9053:
--
Attachment: (was: HIVE-9053.04.patch)

 select constant in union all followed by group by gives wrong result
 

 Key: HIVE-9053
 URL: https://issues.apache.org/jira/browse/HIVE-9053
 Project: Hive
  Issue Type: Bug
Reporter: Pengcheng Xiong
Assignee: Pengcheng Xiong
 Attachments: HIVE-9053.01.patch, HIVE-9053.02.patch, 
 HIVE-9053.03.patch, HIVE-9053.04.patch


 Here is the the way to reproduce with q test:
 select key from (select '1' as key from src union all select key from src)tab 
 group by key;
 will give
 OK
 NULL
 1
 This is not correct as src contains many other keys.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9053) select constant in union all followed by group by gives wrong result

2014-12-13 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9053?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-9053:
--
Status: Patch Available  (was: Open)

 select constant in union all followed by group by gives wrong result
 

 Key: HIVE-9053
 URL: https://issues.apache.org/jira/browse/HIVE-9053
 Project: Hive
  Issue Type: Bug
Reporter: Pengcheng Xiong
Assignee: Pengcheng Xiong
 Attachments: HIVE-9053.01.patch, HIVE-9053.02.patch, 
 HIVE-9053.03.patch, HIVE-9053.04.patch


 Here is the the way to reproduce with q test:
 select key from (select '1' as key from src union all select key from src)tab 
 group by key;
 will give
 OK
 NULL
 1
 This is not correct as src contains many other keys.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 28973: select constant in union all followed by group by gives wrong result

2014-12-13 Thread pengcheng xiong

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/28973/
---

(Updated Dec. 13, 2014, 11:52 p.m.)


Review request for hive, Ashutosh Chauhan and John Pullokkaran.


Changes
---

add one more test case


Repository: hive-git


Description
---

Here is the the way to reproduce with q test:
select key from (select '1' as key from src union all select key from src)tab 
group by key;
will give
OK
NULL
1
This is not correct as src contains many other keys.


Diffs (updated)
-

  
ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/MetadataOnlyOptimizer.java
 8551a5b 
  
ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/NullScanTaskDispatcher.java
 957c327 
  ql/src/test/queries/clientpositive/metadataOnlyOptimizer.q PRE-CREATION 
  ql/src/test/results/clientpositive/metadataOnlyOptimizer.q.out PRE-CREATION 
  ql/src/test/results/clientpositive/optimize_nullscan.q.out 02636ba 

Diff: https://reviews.apache.org/r/28973/diff/


Testing
---


Thanks,

pengcheng xiong



[jira] [Commented] (HIVE-9053) select constant in union all followed by group by gives wrong result

2014-12-13 Thread Pengcheng Xiong (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9053?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245772#comment-14245772
 ] 

Pengcheng Xiong commented on HIVE-9053:
---

[~ashutoshc], I think the test failure is unrelated and the patch is safe to 
check in. Could you please take a look? Thanks!

 select constant in union all followed by group by gives wrong result
 

 Key: HIVE-9053
 URL: https://issues.apache.org/jira/browse/HIVE-9053
 Project: Hive
  Issue Type: Bug
Reporter: Pengcheng Xiong
Assignee: Pengcheng Xiong
 Attachments: HIVE-9053.01.patch, HIVE-9053.02.patch, 
 HIVE-9053.03.patch, HIVE-9053.04.patch


 Here is the the way to reproduce with q test:
 select key from (select '1' as key from src union all select key from src)tab 
 group by key;
 will give
 OK
 NULL
 1
 This is not correct as src contains many other keys.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9055) Tez: union all followed by group by followed by another union all gives error

2014-12-13 Thread Pengcheng Xiong (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9055?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245773#comment-14245773
 ] 

Pengcheng Xiong commented on HIVE-9055:
---

[~jpullokkaran], could you please help check in the patch so that I can test 
union distinct on Tez? This one blocks any union distinct to run on Tez. Thanks!

 Tez: union all followed by group by followed by another union all gives error
 -

 Key: HIVE-9055
 URL: https://issues.apache.org/jira/browse/HIVE-9055
 Project: Hive
  Issue Type: Bug
Reporter: Pengcheng Xiong
Assignee: Vikram Dixit K
 Attachments: HIVE-9055.1.patch, HIVE-9055.WIP.patch


 Here is the way to produce it:
 in Hive q test setting (with src table)
 set hive.execution.engine=tez;
 select key from 
 (
 select key from src
 union all 
 select key from src
 ) tab group by key
 union all
 select key from src;
 will give you
 ERROR
 2014-12-09 11:38:48,316 ERROR ql.Driver (SessionState.java:printError(834)) - 
 FAILED: IndexOutOfBoundsException Index: -1, Size: 1
 java.lang.IndexOutOfBoundsException: Index: -1, Size: 1
 at java.util.LinkedList.checkElementIndex(LinkedList.java:553)
 at java.util.LinkedList.get(LinkedList.java:474)
 at 
 org.apache.hadoop.hive.ql.parse.GenTezWork.process(GenTezWork.java:354)
 at 
 org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:90)
 at 
 org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatchAndReturn(DefaultGraphWalker.java:94)
 at 
 org.apache.hadoop.hive.ql.parse.GenTezWorkWalker.walk(GenTezWorkWalker.java:87)
 at 
 org.apache.hadoop.hive.ql.parse.GenTezWorkWalker.walk(GenTezWorkWalker.java:103)
 at 
 org.apache.hadoop.hive.ql.parse.GenTezWorkWalker.walk(GenTezWorkWalker.java:103)
 at 
 org.apache.hadoop.hive.ql.parse.GenTezWorkWalker.walk(GenTezWorkWalker.java:103)
 at 
 org.apache.hadoop.hive.ql.parse.GenTezWorkWalker.walk(GenTezWorkWalker.java:103)
 at 
 org.apache.hadoop.hive.ql.parse.GenTezWorkWalker.startWalking(GenTezWorkWalker.java:69)
 at 
 org.apache.hadoop.hive.ql.parse.TezCompiler.generateTaskTree(TezCompiler.java:368)
 at 
 org.apache.hadoop.hive.ql.parse.TaskCompiler.compile(TaskCompiler.java:202)
 at 
 org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:10202)
 at 
 org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:221)
 at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:420)
 at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:306)
 at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1108)
 at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1170)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1045)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1035)
 at 
 org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:199)
 at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:151)
 at 
 org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:362)
 at 
 org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:297)
 at 
 org.apache.hadoop.hive.ql.QTestUtil.executeClient(QTestUtil.java:834)
 at 
 org.apache.hadoop.hive.cli.TestMiniTezCliDriver.runTest(TestMiniTezCliDriver.java:136)
 at 
 org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_uniontez(TestMiniTezCliDriver.java:120)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 btw: there is not problem when it is run with MR



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-7816) Enable map-join tests which Tez executes [Spark Branch]

2014-12-13 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7816?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-7816:
--
Attachment: HIVE-7816.1-spark.patch

 Enable map-join tests which Tez executes [Spark Branch]
 ---

 Key: HIVE-7816
 URL: https://issues.apache.org/jira/browse/HIVE-7816
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Brock Noland
Assignee: Rui Li
 Attachments: HIVE-7816.1-spark.patch, HIVE-7816.1-spark.patch


  
 {noformat}
   auto_join0.q,\
   auto_join1.q,\
   cross_join.q,\
   cross_product_check_1.q,\
   cross_product_check_2.q,\
 {noformat}
 {noformat}
 filter_join_breaktask.q,\
 filter_join_breaktask2.q
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-7816) Enable map-join tests which Tez executes [Spark Branch]

2014-12-13 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7816?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-7816:
--
Attachment: (was: HIVE-7816.1-spark.patch)

 Enable map-join tests which Tez executes [Spark Branch]
 ---

 Key: HIVE-7816
 URL: https://issues.apache.org/jira/browse/HIVE-7816
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Brock Noland
Assignee: Rui Li
 Attachments: HIVE-7816.1-spark.patch, HIVE-7816.1-spark.patch


  
 {noformat}
   auto_join0.q,\
   auto_join1.q,\
   cross_join.q,\
   cross_product_check_1.q,\
   cross_product_check_2.q,\
 {noformat}
 {noformat}
 filter_join_breaktask.q,\
 filter_join_breaktask2.q
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-7816) Enable map-join tests which Tez executes [Spark Branch]

2014-12-13 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7816?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-7816:
--
Attachment: (was: HIVE-7816.1-spark.patch)

 Enable map-join tests which Tez executes [Spark Branch]
 ---

 Key: HIVE-7816
 URL: https://issues.apache.org/jira/browse/HIVE-7816
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Brock Noland
Assignee: Rui Li
 Attachments: HIVE-7816.1-spark.patch


  
 {noformat}
   auto_join0.q,\
   auto_join1.q,\
   cross_join.q,\
   cross_product_check_1.q,\
   cross_product_check_2.q,\
 {noformat}
 {noformat}
 filter_join_breaktask.q,\
 filter_join_breaktask2.q
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9078) Hive should not submit second SparkTask while previous one has failed.[Spark Branch]

2014-12-13 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9078?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-9078:
--
Attachment: HIVE-9078.2-spark.patch

 Hive should not submit second SparkTask while previous one has failed.[Spark 
 Branch]
 

 Key: HIVE-9078
 URL: https://issues.apache.org/jira/browse/HIVE-9078
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
Assignee: Chengxiang Li
  Labels: Spark-M4
 Attachments: HIVE-9078.1-spark.patch, HIVE-9078.2-spark.patch


 {noformat}
 hive select n_name, c_name from nation, customer where nation.n_nationkey = 
 customer.c_nationkey limit 10;
 Query ID = root_20141211135050_51e5ae15-49a3-4a46-826f-e27ee314ccb2
 Total jobs = 2
 Launching Job 1 out of 2
 In order to change the average load for a reducer (in bytes):
   set hive.exec.reducers.bytes.per.reducer=number
 In order to limit the maximum number of reducers:
   set hive.exec.reducers.max=number
 In order to set a constant number of reducers:
   set mapreduce.job.reduces=number
 Status: Failed
 Launching Job 2 out of 2
 In order to change the average load for a reducer (in bytes):
   set hive.exec.reducers.bytes.per.reducer=number
 In order to limit the maximum number of reducers:
   set hive.exec.reducers.max=number
 In order to set a constant number of reducers:
   set mapreduce.job.reduces=number
 Status: Failed
 OK
 Time taken: 68.53 seconds
 {noformat}
 2 issue in the above CLI output.
 # For a query which would be translated into multi SparkTask, is previous 
 SparkTask failed, Hive should failed right away, the following SparkTask 
 should not be submitted any more.
 # Print failed info in Hive console while query failed.
 The correct CLI output while query failed:
 {noformat}
 hive select n_name, c_name from nation, customer where nation.n_nationkey = 
 customer.c_nationkey limit 10;
 Query ID = root_20141211142929_ddb7f205-8422-44b4-96bd-96a1c9291895
 Total jobs = 2
 Launching Job 1 out of 2
 In order to change the average load for a reducer (in bytes):
   set hive.exec.reducers.bytes.per.reducer=number
 In order to limit the maximum number of reducers:
   set hive.exec.reducers.max=number
 In order to set a constant number of reducers:
   set mapreduce.job.reduces=number
 Status: Failed
 FAILED: Execution Error, return code 2 from 
 org.apache.hadoop.hive.ql.exec.spark.SparkTask
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9078) Hive should not submit second SparkTask while previous one has failed.[Spark Branch]

2014-12-13 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9078?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-9078:
--
Attachment: (was: HIVE-9078.2-spark.patch)

 Hive should not submit second SparkTask while previous one has failed.[Spark 
 Branch]
 

 Key: HIVE-9078
 URL: https://issues.apache.org/jira/browse/HIVE-9078
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
Assignee: Chengxiang Li
  Labels: Spark-M4
 Attachments: HIVE-9078.1-spark.patch, HIVE-9078.2-spark.patch


 {noformat}
 hive select n_name, c_name from nation, customer where nation.n_nationkey = 
 customer.c_nationkey limit 10;
 Query ID = root_20141211135050_51e5ae15-49a3-4a46-826f-e27ee314ccb2
 Total jobs = 2
 Launching Job 1 out of 2
 In order to change the average load for a reducer (in bytes):
   set hive.exec.reducers.bytes.per.reducer=number
 In order to limit the maximum number of reducers:
   set hive.exec.reducers.max=number
 In order to set a constant number of reducers:
   set mapreduce.job.reduces=number
 Status: Failed
 Launching Job 2 out of 2
 In order to change the average load for a reducer (in bytes):
   set hive.exec.reducers.bytes.per.reducer=number
 In order to limit the maximum number of reducers:
   set hive.exec.reducers.max=number
 In order to set a constant number of reducers:
   set mapreduce.job.reduces=number
 Status: Failed
 OK
 Time taken: 68.53 seconds
 {noformat}
 2 issue in the above CLI output.
 # For a query which would be translated into multi SparkTask, is previous 
 SparkTask failed, Hive should failed right away, the following SparkTask 
 should not be submitted any more.
 # Print failed info in Hive console while query failed.
 The correct CLI output while query failed:
 {noformat}
 hive select n_name, c_name from nation, customer where nation.n_nationkey = 
 customer.c_nationkey limit 10;
 Query ID = root_20141211142929_ddb7f205-8422-44b4-96bd-96a1c9291895
 Total jobs = 2
 Launching Job 1 out of 2
 In order to change the average load for a reducer (in bytes):
   set hive.exec.reducers.bytes.per.reducer=number
 In order to limit the maximum number of reducers:
   set hive.exec.reducers.max=number
 In order to set a constant number of reducers:
   set mapreduce.job.reduces=number
 Status: Failed
 FAILED: Execution Error, return code 2 from 
 org.apache.hadoop.hive.ql.exec.spark.SparkTask
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9078) Hive should not submit second SparkTask while previous one has failed.[Spark Branch]

2014-12-13 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9078?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-9078:
--
Attachment: (was: HIVE-9078.2-spark.patch)

 Hive should not submit second SparkTask while previous one has failed.[Spark 
 Branch]
 

 Key: HIVE-9078
 URL: https://issues.apache.org/jira/browse/HIVE-9078
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
Assignee: Chengxiang Li
  Labels: Spark-M4
 Attachments: HIVE-9078.1-spark.patch, HIVE-9078.2-spark.patch


 {noformat}
 hive select n_name, c_name from nation, customer where nation.n_nationkey = 
 customer.c_nationkey limit 10;
 Query ID = root_20141211135050_51e5ae15-49a3-4a46-826f-e27ee314ccb2
 Total jobs = 2
 Launching Job 1 out of 2
 In order to change the average load for a reducer (in bytes):
   set hive.exec.reducers.bytes.per.reducer=number
 In order to limit the maximum number of reducers:
   set hive.exec.reducers.max=number
 In order to set a constant number of reducers:
   set mapreduce.job.reduces=number
 Status: Failed
 Launching Job 2 out of 2
 In order to change the average load for a reducer (in bytes):
   set hive.exec.reducers.bytes.per.reducer=number
 In order to limit the maximum number of reducers:
   set hive.exec.reducers.max=number
 In order to set a constant number of reducers:
   set mapreduce.job.reduces=number
 Status: Failed
 OK
 Time taken: 68.53 seconds
 {noformat}
 2 issue in the above CLI output.
 # For a query which would be translated into multi SparkTask, is previous 
 SparkTask failed, Hive should failed right away, the following SparkTask 
 should not be submitted any more.
 # Print failed info in Hive console while query failed.
 The correct CLI output while query failed:
 {noformat}
 hive select n_name, c_name from nation, customer where nation.n_nationkey = 
 customer.c_nationkey limit 10;
 Query ID = root_20141211142929_ddb7f205-8422-44b4-96bd-96a1c9291895
 Total jobs = 2
 Launching Job 1 out of 2
 In order to change the average load for a reducer (in bytes):
   set hive.exec.reducers.bytes.per.reducer=number
 In order to limit the maximum number of reducers:
   set hive.exec.reducers.max=number
 In order to set a constant number of reducers:
   set mapreduce.job.reduces=number
 Status: Failed
 FAILED: Execution Error, return code 2 from 
 org.apache.hadoop.hive.ql.exec.spark.SparkTask
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9078) Hive should not submit second SparkTask while previous one has failed.[Spark Branch]

2014-12-13 Thread Xuefu Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9078?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuefu Zhang updated HIVE-9078:
--
Attachment: (was: HIVE-9078.1-spark.patch)

 Hive should not submit second SparkTask while previous one has failed.[Spark 
 Branch]
 

 Key: HIVE-9078
 URL: https://issues.apache.org/jira/browse/HIVE-9078
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
Assignee: Chengxiang Li
  Labels: Spark-M4
 Attachments: HIVE-9078.1-spark.patch, HIVE-9078.2-spark.patch


 {noformat}
 hive select n_name, c_name from nation, customer where nation.n_nationkey = 
 customer.c_nationkey limit 10;
 Query ID = root_20141211135050_51e5ae15-49a3-4a46-826f-e27ee314ccb2
 Total jobs = 2
 Launching Job 1 out of 2
 In order to change the average load for a reducer (in bytes):
   set hive.exec.reducers.bytes.per.reducer=number
 In order to limit the maximum number of reducers:
   set hive.exec.reducers.max=number
 In order to set a constant number of reducers:
   set mapreduce.job.reduces=number
 Status: Failed
 Launching Job 2 out of 2
 In order to change the average load for a reducer (in bytes):
   set hive.exec.reducers.bytes.per.reducer=number
 In order to limit the maximum number of reducers:
   set hive.exec.reducers.max=number
 In order to set a constant number of reducers:
   set mapreduce.job.reduces=number
 Status: Failed
 OK
 Time taken: 68.53 seconds
 {noformat}
 2 issue in the above CLI output.
 # For a query which would be translated into multi SparkTask, is previous 
 SparkTask failed, Hive should failed right away, the following SparkTask 
 should not be submitted any more.
 # Print failed info in Hive console while query failed.
 The correct CLI output while query failed:
 {noformat}
 hive select n_name, c_name from nation, customer where nation.n_nationkey = 
 customer.c_nationkey limit 10;
 Query ID = root_20141211142929_ddb7f205-8422-44b4-96bd-96a1c9291895
 Total jobs = 2
 Launching Job 1 out of 2
 In order to change the average load for a reducer (in bytes):
   set hive.exec.reducers.bytes.per.reducer=number
 In order to limit the maximum number of reducers:
   set hive.exec.reducers.max=number
 In order to set a constant number of reducers:
   set mapreduce.job.reduces=number
 Status: Failed
 FAILED: Execution Error, return code 2 from 
 org.apache.hadoop.hive.ql.exec.spark.SparkTask
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-9090) Rename Tez File Merge Work to smaller name

2014-12-13 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-9090?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran updated HIVE-9090:

   Resolution: Fixed
Fix Version/s: 0.14.1
   0.15.0
   Status: Resolved  (was: Patch Available)

Committed to trunk and branch-0.14.1

 Rename Tez File Merge Work to smaller name
 

 Key: HIVE-9090
 URL: https://issues.apache.org/jira/browse/HIVE-9090
 Project: Hive
  Issue Type: Bug
Affects Versions: 0.14.0, 0.15.0, 0.14.1
Reporter: Prasanth Jayachandran
Assignee: Prasanth Jayachandran
Priority: Trivial
  Labels: tez-ui
 Fix For: 0.15.0, 0.14.1

 Attachments: HIVE-9090.1.patch


 This is just a cosmetic change. The Tez File Merge Work vertex name is long 
 that in-place update UI showing vertex name is out of place.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9078) Hive should not submit second SparkTask while previous one has failed.[Spark Branch]

2014-12-13 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9078?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245837#comment-14245837
 ] 

Hive QA commented on HIVE-9078:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12687095/HIVE-9078.2-spark.patch

{color:red}ERROR:{color} -1 due to 6 failed/errored test(s), 7233 tests executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample_islocalmode_hook
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_optimize_nullscan
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_vector_cast_constant
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_bucket_map_join_spark4
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_infer_bucket_sort_convert_join
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_mapjoin_hook
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/539/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/539/console
Test logs: 
http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/PreCommit-HIVE-SPARK-Build-539/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 6 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12687095 - PreCommit-HIVE-SPARK-Build

 Hive should not submit second SparkTask while previous one has failed.[Spark 
 Branch]
 

 Key: HIVE-9078
 URL: https://issues.apache.org/jira/browse/HIVE-9078
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Chengxiang Li
Assignee: Chengxiang Li
  Labels: Spark-M4
 Attachments: HIVE-9078.1-spark.patch, HIVE-9078.2-spark.patch


 {noformat}
 hive select n_name, c_name from nation, customer where nation.n_nationkey = 
 customer.c_nationkey limit 10;
 Query ID = root_20141211135050_51e5ae15-49a3-4a46-826f-e27ee314ccb2
 Total jobs = 2
 Launching Job 1 out of 2
 In order to change the average load for a reducer (in bytes):
   set hive.exec.reducers.bytes.per.reducer=number
 In order to limit the maximum number of reducers:
   set hive.exec.reducers.max=number
 In order to set a constant number of reducers:
   set mapreduce.job.reduces=number
 Status: Failed
 Launching Job 2 out of 2
 In order to change the average load for a reducer (in bytes):
   set hive.exec.reducers.bytes.per.reducer=number
 In order to limit the maximum number of reducers:
   set hive.exec.reducers.max=number
 In order to set a constant number of reducers:
   set mapreduce.job.reduces=number
 Status: Failed
 OK
 Time taken: 68.53 seconds
 {noformat}
 2 issue in the above CLI output.
 # For a query which would be translated into multi SparkTask, is previous 
 SparkTask failed, Hive should failed right away, the following SparkTask 
 should not be submitted any more.
 # Print failed info in Hive console while query failed.
 The correct CLI output while query failed:
 {noformat}
 hive select n_name, c_name from nation, customer where nation.n_nationkey = 
 customer.c_nationkey limit 10;
 Query ID = root_20141211142929_ddb7f205-8422-44b4-96bd-96a1c9291895
 Total jobs = 2
 Launching Job 1 out of 2
 In order to change the average load for a reducer (in bytes):
   set hive.exec.reducers.bytes.per.reducer=number
 In order to limit the maximum number of reducers:
   set hive.exec.reducers.max=number
 In order to set a constant number of reducers:
   set mapreduce.job.reduces=number
 Status: Failed
 FAILED: Execution Error, return code 2 from 
 org.apache.hadoop.hive.ql.exec.spark.SparkTask
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-7816) Enable map-join tests which Tez executes [Spark Branch]

2014-12-13 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-7816?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14245858#comment-14245858
 ] 

Hive QA commented on HIVE-7816:
---



{color:red}Overall{color}: -1 at least one tests failed

Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12687094/HIVE-7816.1-spark.patch

{color:red}ERROR:{color} -1 due to 22 failed/errored test(s), 7237 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_sample_islocalmode_hook
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_optimize_nullscan
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_vector_cast_constant
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join0
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join23
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join_filters
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_join_nulls
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_auto_sortmerge_join_12
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_cross_join
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_ctas
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join0
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join23
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join_alt_syntax
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join_cond_pushdown_1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join_cond_pushdown_3
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join_cond_pushdown_unqual1
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_join_cond_pushdown_unqual3
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_multi_insert_lateral_view
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_optimize_nullscan
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_parallel_join0
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_ppd_join5
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_subquery_multiinsert
{noformat}

Test results: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/540/testReport
Console output: 
http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-SPARK-Build/540/console
Test logs: 
http://ec2-50-18-27-0.us-west-1.compute.amazonaws.com/logs/PreCommit-HIVE-SPARK-Build-540/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 22 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12687094 - PreCommit-HIVE-SPARK-Build

 Enable map-join tests which Tez executes [Spark Branch]
 ---

 Key: HIVE-7816
 URL: https://issues.apache.org/jira/browse/HIVE-7816
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Brock Noland
Assignee: Rui Li
 Attachments: HIVE-7816.1-spark.patch


  
 {noformat}
   auto_join0.q,\
   auto_join1.q,\
   cross_join.q,\
   cross_product_check_1.q,\
   cross_product_check_2.q,\
 {noformat}
 {noformat}
 filter_join_breaktask.q,\
 filter_join_breaktask2.q
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-7816) Enable map-join tests which Tez executes [Spark Branch]

2014-12-13 Thread Chao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7816?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chao updated HIVE-7816:
---
Attachment: HIVE-7816.2-spark.patch

Regenerated golden files.

 Enable map-join tests which Tez executes [Spark Branch]
 ---

 Key: HIVE-7816
 URL: https://issues.apache.org/jira/browse/HIVE-7816
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Brock Noland
Assignee: Rui Li
 Attachments: HIVE-7816.1-spark.patch, HIVE-7816.2-spark.patch


  
 {noformat}
   auto_join0.q,\
   auto_join1.q,\
   cross_join.q,\
   cross_product_check_1.q,\
   cross_product_check_2.q,\
 {noformat}
 {noformat}
 filter_join_breaktask.q,\
 filter_join_breaktask2.q
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-7816) Enable map-join tests which Tez executes [Spark Branch]

2014-12-13 Thread Chao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7816?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chao updated HIVE-7816:
---
Attachment: HIVE-7816.2-spark.patch

 Enable map-join tests which Tez executes [Spark Branch]
 ---

 Key: HIVE-7816
 URL: https://issues.apache.org/jira/browse/HIVE-7816
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Brock Noland
Assignee: Rui Li
 Attachments: HIVE-7816.1-spark.patch, HIVE-7816.2-spark.patch


  
 {noformat}
   auto_join0.q,\
   auto_join1.q,\
   cross_join.q,\
   cross_product_check_1.q,\
   cross_product_check_2.q,\
 {noformat}
 {noformat}
 filter_join_breaktask.q,\
 filter_join_breaktask2.q
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-7816) Enable map-join tests which Tez executes [Spark Branch]

2014-12-13 Thread Chao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-7816?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chao updated HIVE-7816:
---
Attachment: (was: HIVE-7816.2-spark.patch)

 Enable map-join tests which Tez executes [Spark Branch]
 ---

 Key: HIVE-7816
 URL: https://issues.apache.org/jira/browse/HIVE-7816
 Project: Hive
  Issue Type: Sub-task
  Components: Spark
Reporter: Brock Noland
Assignee: Rui Li
 Attachments: HIVE-7816.1-spark.patch, HIVE-7816.2-spark.patch


  
 {noformat}
   auto_join0.q,\
   auto_join1.q,\
   cross_join.q,\
   cross_product_check_1.q,\
   cross_product_check_2.q,\
 {noformat}
 {noformat}
 filter_join_breaktask.q,\
 filter_join_breaktask2.q
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Review Request 28933: HIVE-8131:Support timestamp in Avro

2014-12-13 Thread Mohit Sabharwal

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/28933/#review65039
---



ql/src/test/results/clientpositive/avro_timestamp.q.out
https://reviews.apache.org/r/28933/#comment107977

Looks like we are truncating the timestamp from millis to nanos because 
Avro spec expects millis.

Any idea why the precision for the partition timestamp column is printed in 
nanos whenever timestamp is in where clause ? 

Also maybe add another query in the test with millis in the where clause ?



serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerDe.java
https://reviews.apache.org/r/28933/#comment107976

should be timestamp-milllis per the spec...



serde/src/test/org/apache/hadoop/hive/serde2/avro/TestTypeInfoToSchema.java
https://reviews.apache.org/r/28933/#comment107975

Shouldn't the logical type be timestamp-millis according to Avro spec ?
 
See: 
https://issues.apache.org/jira/secure/attachment/12663245/AVRO-739-update-spec.diff



serde/src/test/org/apache/hadoop/hive/serde2/avro/TestTypeInfoToSchema.java
https://reviews.apache.org/r/28933/#comment107974

Test for timestamp instead of Test for date


- Mohit Sabharwal


On Dec. 11, 2014, 5:57 a.m., cheng xu wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/28933/
 ---
 
 (Updated Dec. 11, 2014, 5:57 a.m.)
 
 
 Review request for hive.
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 The patch includes:
 1.add timestamp support for AvroSerde 
 2.add related test cases
 
 
 Diffs
 -
 
   data/files/avro_timestamp.txt PRE-CREATION 
   ql/src/test/queries/clientpositive/avro_timestamp.q PRE-CREATION 
   ql/src/test/results/clientpositive/avro_timestamp.q.out PRE-CREATION 
   serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroDeserializer.java 
 07c5ecf 
   serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerDe.java 7639a2b 
   serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerializer.java 
 c8eac89 
   serde/src/java/org/apache/hadoop/hive/serde2/avro/SchemaToTypeInfo.java 
 c84b1a0 
   serde/src/java/org/apache/hadoop/hive/serde2/avro/TypeInfoToSchema.java 
 8cb2dc3 
   serde/src/test/org/apache/hadoop/hive/serde2/avro/TestTypeInfoToSchema.java 
 cd5a0fa 
 
 Diff: https://reviews.apache.org/r/28933/diff/
 
 
 Testing
 ---
 
 Test passed for added cases
 
 
 Thanks,
 
 cheng xu
 




Re: Review Request 28933: HIVE-8131:Support timestamp in Avro

2014-12-13 Thread Mohit Sabharwal


 On Dec. 14, 2014, 7:50 a.m., Mohit Sabharwal wrote:
  ql/src/test/results/clientpositive/avro_timestamp.q.out, line 101
  https://reviews.apache.org/r/28933/diff/1/?file=789140#file789140line101
 
  Looks like we are truncating the timestamp from millis to nanos because 
  Avro spec expects millis.
  
  Any idea why the precision for the partition timestamp column is 
  printed in nanos whenever timestamp is in where clause ? 
  
  Also maybe add another query in the test with millis in the where 
  clause ?

Typo... I meant truncating the timestamp from nanos to millis 


- Mohit


---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/28933/#review65039
---


On Dec. 11, 2014, 5:57 a.m., cheng xu wrote:
 
 ---
 This is an automatically generated e-mail. To reply, visit:
 https://reviews.apache.org/r/28933/
 ---
 
 (Updated Dec. 11, 2014, 5:57 a.m.)
 
 
 Review request for hive.
 
 
 Repository: hive-git
 
 
 Description
 ---
 
 The patch includes:
 1.add timestamp support for AvroSerde 
 2.add related test cases
 
 
 Diffs
 -
 
   data/files/avro_timestamp.txt PRE-CREATION 
   ql/src/test/queries/clientpositive/avro_timestamp.q PRE-CREATION 
   ql/src/test/results/clientpositive/avro_timestamp.q.out PRE-CREATION 
   serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroDeserializer.java 
 07c5ecf 
   serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerDe.java 7639a2b 
   serde/src/java/org/apache/hadoop/hive/serde2/avro/AvroSerializer.java 
 c8eac89 
   serde/src/java/org/apache/hadoop/hive/serde2/avro/SchemaToTypeInfo.java 
 c84b1a0 
   serde/src/java/org/apache/hadoop/hive/serde2/avro/TypeInfoToSchema.java 
 8cb2dc3 
   serde/src/test/org/apache/hadoop/hive/serde2/avro/TestTypeInfoToSchema.java 
 cd5a0fa 
 
 Diff: https://reviews.apache.org/r/28933/diff/
 
 
 Testing
 ---
 
 Test passed for added cases
 
 
 Thanks,
 
 cheng xu