[jira] [Commented] (HIVE-15202) Concurrent compactions for the same partition may generate malformed folder structure

2016-11-16 Thread Rui Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15202?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15672903#comment-15672903
 ] 

Rui Li commented on HIVE-15202:
---

Thank you [~ekoifman]. That's great to know.

> Concurrent compactions for the same partition may generate malformed folder 
> structure
> -
>
> Key: HIVE-15202
> URL: https://issues.apache.org/jira/browse/HIVE-15202
> Project: Hive
>  Issue Type: Bug
>Reporter: Rui Li
>
> If two compactions run concurrently on a single partition, it may generate 
> folder structure like this: (nested base dir)
> {noformat}
> drwxr-xr-x   - root supergroup  0 2016-11-14 22:23 
> /user/hive/warehouse/test/z=1/base_007/base_007
> -rw-r--r--   3 root supergroup201 2016-11-14 21:46 
> /user/hive/warehouse/test/z=1/base_007/bucket_0
> -rw-r--r--   3 root supergroup611 2016-11-14 21:46 
> /user/hive/warehouse/test/z=1/base_007/bucket_1
> -rw-r--r--   3 root supergroup614 2016-11-14 21:46 
> /user/hive/warehouse/test/z=1/base_007/bucket_2
> -rw-r--r--   3 root supergroup621 2016-11-14 21:46 
> /user/hive/warehouse/test/z=1/base_007/bucket_3
> -rw-r--r--   3 root supergroup621 2016-11-14 21:46 
> /user/hive/warehouse/test/z=1/base_007/bucket_4
> -rw-r--r--   3 root supergroup201 2016-11-14 21:46 
> /user/hive/warehouse/test/z=1/base_007/bucket_5
> -rw-r--r--   3 root supergroup201 2016-11-14 21:46 
> /user/hive/warehouse/test/z=1/base_007/bucket_6
> -rw-r--r--   3 root supergroup201 2016-11-14 21:46 
> /user/hive/warehouse/test/z=1/base_007/bucket_7
> -rw-r--r--   3 root supergroup201 2016-11-14 21:46 
> /user/hive/warehouse/test/z=1/base_007/bucket_8
> -rw-r--r--   3 root supergroup201 2016-11-14 21:46 
> /user/hive/warehouse/test/z=1/base_007/bucket_9
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15227) Optimize join + gby into semijoin

2016-11-16 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15227?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15672817#comment-15672817
 ] 

Ashutosh Chauhan commented on HIVE-15227:
-

yes.. there is. You may note (from the plan in attached patch) that Gby is only 
in mapper, not in reducer. This is possible since LSJ operator understands the 
semantics, such that GBy in reducer is obviated. Net perf win here is to not 
compute GBy on reducer side.

> Optimize join + gby into semijoin
> -
>
> Key: HIVE-15227
> URL: https://issues.apache.org/jira/browse/HIVE-15227
> Project: Hive
>  Issue Type: Bug
>  Components: Logical Optimizer
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-15227.patch
>
>
> Calcite has a rule which can do this transformation. Lets take advantage of 
> this since Hive has native Left semi join operator.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15199) INSERT INTO data on S3 is replacing the old rows with the new ones

2016-11-16 Thread Sahil Takiar (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15199?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15672593#comment-15672593
 ] 

Sahil Takiar commented on HIVE-15199:
-

* Is the goal to trigger mvFile when the destination file is a blobstore? I 
don't think thats the right approach because a {{FileUtils.copy}} will do a 
client-side copy when running on S3, data will be downloaded from HDFS to HS2 
and then uploaded to S3; the target should be to do a server-side copy (happens 
internally on S3). A server side copy can only be triggered by called 
{{FileSystem.rename}}.
* The listing optimization can be applied to HDFS too, right? It should 
increase perf when running on HDFS too.
* A bit orthogonal to this JIRA, but {{mvFile}} should probably be called 
copyFile because it always copies data.

> INSERT INTO data on S3 is replacing the old rows with the new ones
> --
>
> Key: HIVE-15199
> URL: https://issues.apache.org/jira/browse/HIVE-15199
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Sergio Peña
>Assignee: Sergio Peña
>Priority: Critical
> Attachments: HIVE-15199.1.patch, HIVE-15199.2.patch, 
> HIVE-15199.3.patch
>
>
> Any INSERT INTO statement run on S3 tables and when the scratch directory is 
> saved on S3 is deleting old rows of the table.
> {noformat}
> hive> set hive.blobstore.use.blobstore.as.scratchdir=true;
> hive> create table t1 (id int, name string) location 's3a://spena-bucket/t1';
> hive> insert into table t1 values (1,'name1');
> hive> select * from t1;
> 1   name1
> hive> insert into table t1 values (2,'name2');
> hive> select * from t1;
> 2   name2
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HIVE-14990) run all tests for MM tables and fix the issues that are found

2016-11-16 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-14990?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15668996#comment-15668996
 ] 

Sergey Shelukhin edited comment on HIVE-14990 at 11/17/16 3:02 AM:
---

Updated test list to fix/declare irrelevant before closing this. Only updated 
the CliDriver list actually, haven't made my way thru it yet
{panel}
TestCliDriver:
stats_list_bucket
show_tablestatus
-vector_udf2-
list_bucket_dml_14
autoColumnStats_9
stats_noscan_2
symlink_text_input_format
temp_table_precedence
offset_limit_global_optimizer
rand_partitionpruner2
materialized_view_authorization_sqlstd,materialized_*
merge_dynamic_partition, merge_dynamic_partition*
orc_vectorization_ppd
parquet_join2
repl_3_exim_metadata
sample6
sample_islocalmode_hook
smb_mapjoin_2,smb_mapjoin_3,smb_mapjoin_7
orc_createas1
exim_16_part_external,exim_17_part_managed,


TestEncryptedHDFSCliDriver:
encryption_ctas
encryption_drop_partition 
encryption_insert_values
encryption_join_unencrypted_tbl
encryption_load_data_to_encrypted_tables

MiniLlapLocal:
exchgpartition2lel
cbo_rp_lineage2
create_merge_compressed
deleteAnalyze
delete_where_no_match
delete_where_non_partitioned
dynpart_sort_optimization
escape2
insert1
lineage2
lineage3
orc_llap
schema_evol_orc_nonvec_part
schema_evol_orc_vec_part
schema_evol_text_nonvec_part
schema_evol_text_vec_part
schema_evol_text_vecrow_part
smb_mapjoin_6
tez_dml
union_fast_stats
update_all_types
update_tmp_table
update_where_no_match
update_where_non_partitioned
vector_outer_join1
vector_outer_join4

MiniLlap:
load_fs2
orc_ppd_basic
external_table_with_space_in_location_path
file_with_header_footer
import_exported_table
schemeAuthority,schemeAuthority2
table_nonprintable

Minimr:
infer_bucket_sort_map_operators
infer_bucket_sort_merge
infer_bucket_sort_reducers_power_two
root_dir_external_table
scriptfile1

TestSymlinkTextInputFormat#testCombine 
TestJdbcWithLocalClusterSpark, etc.
{panel}


was (Author: sershe):
Updated test list to fix/declare irrelevant before closing this. Only updated 
the CliDriver list actually, haven't made my way thru it yet
{noformat}
TestCliDriver:
stats_list_bucket
show_tablestatus
vector_udf2
list_bucket_dml_14
autoColumnStats_9
stats_noscan_2
symlink_text_input_format
temp_table_precedence
offset_limit_global_optimizer
rand_partitionpruner2
materialized_view_authorization_sqlstd,materialized_*
merge_dynamic_partition, merge_dynamic_partition*
orc_vectorization_ppd
parquet_join2
repl_3_exim_metadata
sample6
sample_islocalmode_hook
smb_mapjoin_2,smb_mapjoin_3,smb_mapjoin_7
orc_createas1
exim_16_part_external,exim_17_part_managed,


TestEncryptedHDFSCliDriver:
encryption_ctas
encryption_drop_partition 
encryption_insert_values
encryption_join_unencrypted_tbl
encryption_load_data_to_encrypted_tables

MiniLlapLocal:
exchgpartition2lel
cbo_rp_lineage2
create_merge_compressed
deleteAnalyze
delete_where_no_match
delete_where_non_partitioned
dynpart_sort_optimization
escape2
insert1
lineage2
lineage3
orc_llap
schema_evol_orc_nonvec_part
schema_evol_orc_vec_part
schema_evol_text_nonvec_part
schema_evol_text_vec_part
schema_evol_text_vecrow_part
smb_mapjoin_6
tez_dml
union_fast_stats
update_all_types
update_tmp_table
update_where_no_match
update_where_non_partitioned
vector_outer_join1
vector_outer_join4

MiniLlap:
load_fs2
orc_ppd_basic
external_table_with_space_in_location_path
file_with_header_footer
import_exported_table
schemeAuthority,schemeAuthority2
table_nonprintable

Minimr:
infer_bucket_sort_map_operators
infer_bucket_sort_merge
infer_bucket_sort_reducers_power_two
root_dir_external_table
scriptfile1

TestSymlinkTextInputFormat#testCombine 
TestJdbcWithLocalClusterSpark, etc.
{noformat}

> run all tests for MM tables and fix the issues that are found
> -
>
> Key: HIVE-14990
> URL: https://issues.apache.org/jira/browse/HIVE-14990
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-14990.01.patch, HIVE-14990.02.patch, 
> HIVE-14990.03.patch, HIVE-14990.04.patch, HIVE-14990.04.patch, 
> HIVE-14990.05.patch, HIVE-14990.05.patch, HIVE-14990.06.patch, 
> HIVE-14990.06.patch, HIVE-14990.07.patch, HIVE-14990.08.patch, 
> HIVE-14990.09.patch, HIVE-14990.10.patch, HIVE-14990.10.patch, 
> HIVE-14990.10.patch, HIVE-14990.patch
>
>
> Expected failures 
> 1) All HCat tests (cannot write MM tables via the HCat writer)
> 2) Almost all merge tests (alter .. concat is not supported).
> 3) Tests that run dfs commands with specific paths (path changes).
> 4) Truncate column (not supported).
> 5) Describe formatted will have the new table fields in the output (before 
> merging MM with ACID).
> 6) Many tests w/explain extended - diff in partition "base file name" (path 
> changes).
> 7) 

[jira] [Commented] (HIVE-14189) backport HIVE-13945 to branch-1

2016-11-16 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-14189?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15672491#comment-15672491
 ] 

Sergey Shelukhin commented on HIVE-14189:
-

[~cartershanklin] [~jcamachorodriguez] [~ashutoshc] do you have access to the 
text of a later SQL standard than 99? That one is pretty vague.
Seems like there's implicit cast to string in these cases on master:
{noformat}
POSTHOOK: query: select 'NaN' < 2.0 from src limit 1
false
POSTHOOK: query: select 1.0 < 'NaN' from src limit 1
true
{noformat}

branch-1 with this patch returns null by happenstance (because something else 
inserts a cast to decimal from string), so the operation returns null.
I wonder if SQL standard has anything to say on this... 99 doesn't seem to.

> backport HIVE-13945 to branch-1
> ---
>
> Key: HIVE-14189
> URL: https://issues.apache.org/jira/browse/HIVE-14189
> Project: Hive
>  Issue Type: Bug
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
>  Labels: TODOC1.3
> Attachments: HIVE-14189-branch-1.patch, HIVE-14189.01-branch-1.patch, 
> HIVE-14189.02-branch-1.patch, HIVE-14189.03-branch-1.patch, 
> HIVE-14189.04-branch-1.patch, HIVE-14189.05-branch-1.patch, 
> HIVE-14189.06-branch-1.patch, HIVE-14189.07-branch-1.patch, 
> HIVE-14189.08-branch-1.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-14189) backport HIVE-13945 to branch-1

2016-11-16 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-14189?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15672468#comment-15672468
 ] 

Sergey Shelukhin commented on HIVE-14189:
-

Hmm, different types notwithstanding, this seems like the correct output. 
Master seems to be comparing strings. Let me see what SQL standard sayeth.

> backport HIVE-13945 to branch-1
> ---
>
> Key: HIVE-14189
> URL: https://issues.apache.org/jira/browse/HIVE-14189
> Project: Hive
>  Issue Type: Bug
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
>  Labels: TODOC1.3
> Attachments: HIVE-14189-branch-1.patch, HIVE-14189.01-branch-1.patch, 
> HIVE-14189.02-branch-1.patch, HIVE-14189.03-branch-1.patch, 
> HIVE-14189.04-branch-1.patch, HIVE-14189.05-branch-1.patch, 
> HIVE-14189.06-branch-1.patch, HIVE-14189.07-branch-1.patch, 
> HIVE-14189.08-branch-1.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-14189) backport HIVE-13945 to branch-1

2016-11-16 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-14189?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15672440#comment-15672440
 ] 

Sergey Shelukhin commented on HIVE-14189:
-

Difference (with CBO disabled) on master vs branch-1
{noformat}
2016-11-16 17:37:28,788 DEBUG [main]: optimizer.ConstantPropagateProcFactory 
(ConstantPropagateProcFactory.java:foldExprFull(310)) - 
Folding expression:GenericUDFOPLessThan(GenericUDFToDecimal(Const string NaN), 
Const int 2) -> Const boolean null

2016-11-16T18:06:46,788 DEBUG [5682f00b-91b1-4106-a205-b55e98f03ad0 main] 
optimizer.ConstantPropagateProcFactory:
Folding expression:GenericUDFOPLessThan(Const string NaN, Const decimal(1,0) 2) 
-> Const boolean false
{noformat}

> backport HIVE-13945 to branch-1
> ---
>
> Key: HIVE-14189
> URL: https://issues.apache.org/jira/browse/HIVE-14189
> Project: Hive
>  Issue Type: Bug
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
>  Labels: TODOC1.3
> Attachments: HIVE-14189-branch-1.patch, HIVE-14189.01-branch-1.patch, 
> HIVE-14189.02-branch-1.patch, HIVE-14189.03-branch-1.patch, 
> HIVE-14189.04-branch-1.patch, HIVE-14189.05-branch-1.patch, 
> HIVE-14189.06-branch-1.patch, HIVE-14189.07-branch-1.patch, 
> HIVE-14189.08-branch-1.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15225) QueryPlan.getJSONValue should code against empty string values

2016-11-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15225?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15672413#comment-15672413
 ] 

Hive QA commented on HIVE-15225:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12839260/HIVE-15225.1.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 4 failed/errored test(s), 10694 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[transform_ppr2] 
(batchId=133)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[join_acid_non_acid]
 (batchId=150)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[union_fast_stats]
 (batchId=145)
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver[explainanalyze_1] 
(batchId=90)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/2161/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/2161/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-2161/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 4 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12839260 - PreCommit-HIVE-Build

> QueryPlan.getJSONValue should code against empty string values
> --
>
> Key: HIVE-15225
> URL: https://issues.apache.org/jira/browse/HIVE-15225
> Project: Hive
>  Issue Type: Bug
>Reporter: Yibing Shi
>Assignee: Yibing Shi
> Attachments: HIVE-15225.1.patch
>
>
> The current {{QueryPlan.getJSONValue}} implementation is as below:
> {code}
>   public String getJSONValue(Object value) {
> String v = "null";
> if (value != null) {
>   v = value.toString();
>   if (v.charAt(0) != '[' && v.charAt(0) != '{') {
> v = "\"" + v + "\"";
>   }
> }
> return v;
>   }
> {code}
> When {{value.toString()}} returns an empty string, a 
> StringIndexOutOfRangeException would be thrown out when "v.charAt(0)" is 
> evaluated.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-14189) backport HIVE-13945 to branch-1

2016-11-16 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-14189?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-14189:

Attachment: HIVE-14189.08-branch-1.patch

Most of the out file changes are valid; updated them.
Some are spurious (and so are timeouts, I suspect) - rerunning.
ops_comparison seems to be a real problem that doesn't happen on master version 
of the same patch, need to take a look

> backport HIVE-13945 to branch-1
> ---
>
> Key: HIVE-14189
> URL: https://issues.apache.org/jira/browse/HIVE-14189
> Project: Hive
>  Issue Type: Bug
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
>  Labels: TODOC1.3
> Attachments: HIVE-14189-branch-1.patch, HIVE-14189.01-branch-1.patch, 
> HIVE-14189.02-branch-1.patch, HIVE-14189.03-branch-1.patch, 
> HIVE-14189.04-branch-1.patch, HIVE-14189.05-branch-1.patch, 
> HIVE-14189.06-branch-1.patch, HIVE-14189.07-branch-1.patch, 
> HIVE-14189.08-branch-1.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-1478) Non-boolean expression in WHERE should be rejected

2016-11-16 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1478?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15672332#comment-15672332
 ] 

Ashutosh Chauhan commented on HIVE-1478:


This should be closed as won't fix since as discussed on HIVE-13659 this is 
already supported in Hive and there is no good reason to  withdra support  for 
this. 
By the way, on master these queries should not throw any exception. MR or 
otherwise.

> Non-boolean expression in WHERE should be rejected
> --
>
> Key: HIVE-1478
> URL: https://issues.apache.org/jira/browse/HIVE-1478
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.7.0
>Reporter: Paul Yang
>Assignee: Zoltan Haindrich
>Priority: Minor
> Attachments: HIVE-1478.1.patch, HIVE-1478.2.patch
>
>
> Automatically casting strings or other types into boolean may confuse even 
> the user - and somehow it doesn't always work (HIVE-15089)
> sql2011 states that "where expression" should accept a boolean expression.
> Original reported problem:
> If the expression in the where clause does not evaluate to a boolean, the job 
> will fail with the following exception in the task logs:
> Query:
> SELECT key FROM src WHERE 1;
> Exception in mapper:
> 2010-07-21 17:00:31,460 FATAL ExecMapper: 
> org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while 
> processing row {"key":"238","value":"val_238"}
>   at 
> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:417)
>   at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:180)
>   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
>   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:358)
>   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:307)
>   at org.apache.hadoop.mapred.Child.main(Child.java:159)
> Caused by: java.lang.ClassCastException: java.lang.Integer cannot be cast to 
> java.lang.Boolean
>   at 
> org.apache.hadoop.hive.ql.exec.FilterOperator.processOp(FilterOperator.java:84)
>   at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
>   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
>   at 
> org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:45)
>   at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
>   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
>   at 
> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:400)
>   ... 5 more



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15227) Optimize join + gby into semijoin

2016-11-16 Thread Vineet Garg (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15227?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15672331#comment-15672331
 ] 

Vineet Garg commented on HIVE-15227:


[~ashutoshc] I discussed this with [~jcamachorodriguez] today and it looks like 
Hive does group by before doing left semi join anyway. e.g. {noformat} select a 
from t1 left semi join t2 on t1.a=t2.b {noformat} will generate GBY on t2.b 
before doing Left Semi Join. 
Is there any improvement which physical Left Semi Join operator has in this 
case ?

> Optimize join + gby into semijoin
> -
>
> Key: HIVE-15227
> URL: https://issues.apache.org/jira/browse/HIVE-15227
> Project: Hive
>  Issue Type: Bug
>  Components: Logical Optimizer
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-15227.patch
>
>
> Calcite has a rule which can do this transformation. Lets take advantage of 
> this since Hive has native Left semi join operator.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-1478) Non-boolean expression in WHERE should be rejected

2016-11-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1478?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15672326#comment-15672326
 ] 

Hive QA commented on HIVE-1478:
---



Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12839245/HIVE-1478.2.patch

{color:green}SUCCESS:{color} +1 due to 2 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 8 failed/errored test(s), 10673 tests 
executed
*Failed tests:*
{noformat}
TestSparkCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=95)

[ptf_general_queries.q,auto_join_reordering_values.q,sample2.q,join1.q,decimal_join.q,mapjoin_subquery2.q,join32_lessSize.q,mapjoin1.q,order2.q,skewjoinopt18.q,union_remove_18.q,join25.q,groupby9.q,bucketsortoptimize_insert_6.q,ctas.q]
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[orc_ppd_schema_evol_3a]
 (batchId=133)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[transform_ppr2] 
(batchId=133)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[join_acid_non_acid]
 (batchId=150)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[union_fast_stats]
 (batchId=145)
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver[explainanalyze_2] 
(batchId=91)
org.apache.hadoop.hive.ql.security.TestStorageBasedMetastoreAuthorizationDrops.testDropPartition
 (batchId=204)
org.apache.hive.hcatalog.api.repl.commands.TestCommands.org.apache.hive.hcatalog.api.repl.commands.TestCommands
 (batchId=166)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/2159/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/2159/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-2159/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 8 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12839245 - PreCommit-HIVE-Build

> Non-boolean expression in WHERE should be rejected
> --
>
> Key: HIVE-1478
> URL: https://issues.apache.org/jira/browse/HIVE-1478
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.7.0
>Reporter: Paul Yang
>Assignee: Zoltan Haindrich
>Priority: Minor
> Attachments: HIVE-1478.1.patch, HIVE-1478.2.patch
>
>
> Automatically casting strings or other types into boolean may confuse even 
> the user - and somehow it doesn't always work (HIVE-15089)
> sql2011 states that "where expression" should accept a boolean expression.
> Original reported problem:
> If the expression in the where clause does not evaluate to a boolean, the job 
> will fail with the following exception in the task logs:
> Query:
> SELECT key FROM src WHERE 1;
> Exception in mapper:
> 2010-07-21 17:00:31,460 FATAL ExecMapper: 
> org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while 
> processing row {"key":"238","value":"val_238"}
>   at 
> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:417)
>   at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:180)
>   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
>   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:358)
>   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:307)
>   at org.apache.hadoop.mapred.Child.main(Child.java:159)
> Caused by: java.lang.ClassCastException: java.lang.Integer cannot be cast to 
> java.lang.Boolean
>   at 
> org.apache.hadoop.hive.ql.exec.FilterOperator.processOp(FilterOperator.java:84)
>   at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
>   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
>   at 
> org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:45)
>   at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
>   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
>   at 
> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:400)
>   ... 5 more



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10901) Optimize multi column distinct queries

2016-11-16 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10901?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15672321#comment-15672321
 ] 

Ashutosh Chauhan commented on HIVE-10901:
-

+1

> Optimize  multi column distinct queries 
> 
>
> Key: HIVE-10901
> URL: https://issues.apache.org/jira/browse/HIVE-10901
> Project: Hive
>  Issue Type: New Feature
>  Components: CBO, Logical Optimizer
>Affects Versions: 1.2.0
>Reporter: Mostafa Mokhtar
>Assignee: Pengcheng Xiong
> Attachments: HIVE-10901.02.patch, HIVE-10901.03.patch, 
> HIVE-10901.04.patch, HIVE-10901.patch
>
>
> HIVE-10568 is useful only when there is a distinct on one column. It can be 
> expanded for multiple column cases too.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15227) Optimize join + gby into semijoin

2016-11-16 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15227?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-15227:

Status: Patch Available  (was: Open)

> Optimize join + gby into semijoin
> -
>
> Key: HIVE-15227
> URL: https://issues.apache.org/jira/browse/HIVE-15227
> Project: Hive
>  Issue Type: Bug
>  Components: Logical Optimizer
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-15227.patch
>
>
> Calcite has a rule which can do this transformation. Lets take advantage of 
> this since Hive has native Left semi join operator.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15227) Optimize join + gby into semijoin

2016-11-16 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15227?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-15227:

Attachment: HIVE-15227.patch

> Optimize join + gby into semijoin
> -
>
> Key: HIVE-15227
> URL: https://issues.apache.org/jira/browse/HIVE-15227
> Project: Hive
>  Issue Type: Bug
>  Components: Logical Optimizer
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-15227.patch
>
>
> Calcite has a rule which can do this transformation. Lets take advantage of 
> this since Hive has native Left semi join operator.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10901) Optimize multi column distinct queries

2016-11-16 Thread Pengcheng Xiong (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10901?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15672237#comment-15672237
 ] 

Pengcheng Xiong commented on HIVE-10901:


[~ashutoshc], could u take a look? Thanks.

> Optimize  multi column distinct queries 
> 
>
> Key: HIVE-10901
> URL: https://issues.apache.org/jira/browse/HIVE-10901
> Project: Hive
>  Issue Type: New Feature
>  Components: CBO, Logical Optimizer
>Affects Versions: 1.2.0
>Reporter: Mostafa Mokhtar
>Assignee: Pengcheng Xiong
> Attachments: HIVE-10901.02.patch, HIVE-10901.03.patch, 
> HIVE-10901.04.patch, HIVE-10901.patch
>
>
> HIVE-10568 is useful only when there is a distinct on one column. It can be 
> expanded for multiple column cases too.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15219) LLAP: Allow additional slider global parameters to be set while creating the LLAP package

2016-11-16 Thread Siddharth Seth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15219?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Siddharth Seth updated HIVE-15219:
--
Attachment: HIVE-15219.02.patch

Updated patch.
Works with --slider-appconfig-global propertName=value

[~gopalv] - ready for review when you get a chance. Tested locally.

> LLAP: Allow additional slider global parameters to be set while creating the 
> LLAP package
> -
>
> Key: HIVE-15219
> URL: https://issues.apache.org/jira/browse/HIVE-15219
> Project: Hive
>  Issue Type: Task
>  Components: llap
>Reporter: Siddharth Seth
>Assignee: Siddharth Seth
> Attachments: HIVE-15219.02.patch, HIVE-15219.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15226) Add a different masking comment to qtests blobstore output

2016-11-16 Thread Sahil Takiar (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15226?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15672205#comment-15672205
 ] 

Sahil Takiar commented on HIVE-15226:
-

Overall looks good. Only comment is that {{patternsWithMaskComments}} should 
probably return a {{Pair}} instead of a {{String[][]}}. 
{{Pattern.compile(...)}} is an expensive call, so you really only want to do it 
once per Hive QA run.

> Add a different masking comment to qtests blobstore output
> --
>
> Key: HIVE-15226
> URL: https://issues.apache.org/jira/browse/HIVE-15226
> Project: Hive
>  Issue Type: Task
>  Components: Hive
>Reporter: Sergio Peña
>Assignee: Sergio Peña
>Priority: Minor
> Attachments: HIVE-15226.1.patch
>
>
> The output of TestBlobstoreCliDriver is masking all s3a patch so that we can 
> use the tests with any other blobstore scheme.
> It should be good to have a specific masking comment for those paths instead 
> of the generic " A masked pattern was here " so that we can verify 
> that certain tests are indeed using the blobstore path.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-10901) Optimize multi column distinct queries

2016-11-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-10901?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15672177#comment-15672177
 ] 

Hive QA commented on HIVE-10901:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12839247/HIVE-10901.04.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 10666 tests 
executed
*Failed tests:*
{noformat}
TestSparkCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=96)

[groupby_map_ppr.q,nullgroup4_multi_distinct.q,join_rc.q,union14.q,smb_mapjoin_12.q,vector_cast_constant.q,union_remove_4.q,auto_join11.q,load_dyn_part7.q,udaf_collect_set.q,vectorization_12.q,groupby_sort_skew_1.q,groupby_sort_skew_1_23.q,smb_mapjoin_25.q,skewjoinopt12.q]
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[transform_ppr2] 
(batchId=133)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[join_acid_non_acid]
 (batchId=150)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[union_fast_stats]
 (batchId=145)
org.apache.hadoop.hive.cli.TestSparkCliDriver.org.apache.hadoop.hive.cli.TestSparkCliDriver
 (batchId=118)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/2158/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/2158/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-2158/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 5 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12839247 - PreCommit-HIVE-Build

> Optimize  multi column distinct queries 
> 
>
> Key: HIVE-10901
> URL: https://issues.apache.org/jira/browse/HIVE-10901
> Project: Hive
>  Issue Type: New Feature
>  Components: CBO, Logical Optimizer
>Affects Versions: 1.2.0
>Reporter: Mostafa Mokhtar
>Assignee: Pengcheng Xiong
> Attachments: HIVE-10901.02.patch, HIVE-10901.03.patch, 
> HIVE-10901.04.patch, HIVE-10901.patch
>
>
> HIVE-10568 is useful only when there is a distinct on one column. It can be 
> expanded for multiple column cases too.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15226) Add a different masking comment to qtests blobstore output

2016-11-16 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/HIVE-15226?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergio Peña updated HIVE-15226:
---
Status: Patch Available  (was: Open)

[~stakiar] What do you think about this approach? It uses a specific mask 
comment, such as "### BLOBSTORE_PATH ###" when a s3a, swift or any other 
blobstore scheme is found on the output.

> Add a different masking comment to qtests blobstore output
> --
>
> Key: HIVE-15226
> URL: https://issues.apache.org/jira/browse/HIVE-15226
> Project: Hive
>  Issue Type: Task
>  Components: Hive
>Reporter: Sergio Peña
>Assignee: Sergio Peña
>Priority: Minor
> Attachments: HIVE-15226.1.patch
>
>
> The output of TestBlobstoreCliDriver is masking all s3a patch so that we can 
> use the tests with any other blobstore scheme.
> It should be good to have a specific masking comment for those paths instead 
> of the generic " A masked pattern was here " so that we can verify 
> that certain tests are indeed using the blobstore path.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15226) Add a different masking comment to qtests blobstore output

2016-11-16 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/HIVE-15226?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergio Peña updated HIVE-15226:
---
Attachment: HIVE-15226.1.patch

> Add a different masking comment to qtests blobstore output
> --
>
> Key: HIVE-15226
> URL: https://issues.apache.org/jira/browse/HIVE-15226
> Project: Hive
>  Issue Type: Task
>  Components: Hive
>Reporter: Sergio Peña
>Assignee: Sergio Peña
>Priority: Minor
> Attachments: HIVE-15226.1.patch
>
>
> The output of TestBlobstoreCliDriver is masking all s3a patch so that we can 
> use the tests with any other blobstore scheme.
> It should be good to have a specific masking comment for those paths instead 
> of the generic " A masked pattern was here " so that we can verify 
> that certain tests are indeed using the blobstore path.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15192) Use Calcite to de-correlate and plan subqueries

2016-11-16 Thread Vineet Garg (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15192?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-15192:
---
Status: Patch Available  (was: Open)

> Use Calcite to de-correlate and plan subqueries
> ---
>
> Key: HIVE-15192
> URL: https://issues.apache.org/jira/browse/HIVE-15192
> Project: Hive
>  Issue Type: Task
>  Components: Logical Optimizer
>Reporter: Vineet Garg
>Assignee: Vineet Garg
> Attachments: HIVE-15192.patch
>
>
> Currently support of subqueries is limited [Link to original spec | 
> https://issues.apache.org/jira/secure/attachment/12614003/SubQuerySpec.pdf].
> Using Calcite to plan and de-correlate subqueries will help Hive get rid of 
> these limitations.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15173) Allow dec as an alias for decimal

2016-11-16 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15173?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-15173:

Attachment: HIVE-15173.1.patch

Updated test case.

> Allow dec as an alias for decimal
> -
>
> Key: HIVE-15173
> URL: https://issues.apache.org/jira/browse/HIVE-15173
> Project: Hive
>  Issue Type: Sub-task
>  Components: Parser
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-15173.1.patch, HIVE-15173.patch
>
>
> Standard allows dec as an alias for decimal



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15173) Allow dec as an alias for decimal

2016-11-16 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15173?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-15173:

Status: Patch Available  (was: Open)

> Allow dec as an alias for decimal
> -
>
> Key: HIVE-15173
> URL: https://issues.apache.org/jira/browse/HIVE-15173
> Project: Hive
>  Issue Type: Sub-task
>  Components: Parser
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-15173.1.patch, HIVE-15173.patch
>
>
> Standard allows dec as an alias for decimal



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15173) Allow dec as an alias for decimal

2016-11-16 Thread Ashutosh Chauhan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15173?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-15173:

Status: Open  (was: Patch Available)

> Allow dec as an alias for decimal
> -
>
> Key: HIVE-15173
> URL: https://issues.apache.org/jira/browse/HIVE-15173
> Project: Hive
>  Issue Type: Sub-task
>  Components: Parser
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
> Attachments: HIVE-15173.patch
>
>
> Standard allows dec as an alias for decimal



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-14189) backport HIVE-13945 to branch-1

2016-11-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-14189?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15672020#comment-15672020
 ] 

Hive QA commented on HIVE-14189:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12839226/HIVE-14189.07-branch-1.patch

{color:green}SUCCESS:{color} +1 due to 8 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 201 failed/errored test(s), 8083 tests 
executed
*Failed tests:*
{noformat}
TestAcidOnTez - did not produce a TEST-*.xml file (likely timed out) 
(batchId=377)
TestAdminUser - did not produce a TEST-*.xml file (likely timed out) 
(batchId=359)
TestAuthorizationPreEventListener - did not produce a TEST-*.xml file (likely 
timed out) (batchId=392)
TestAuthzApiEmbedAuthorizerInEmbed - did not produce a TEST-*.xml file (likely 
timed out) (batchId=369)
TestAuthzApiEmbedAuthorizerInRemote - did not produce a TEST-*.xml file (likely 
timed out) (batchId=375)
TestBeeLineWithArgs - did not produce a TEST-*.xml file (likely timed out) 
(batchId=399)
TestCLIAuthzSessionContext - did not produce a TEST-*.xml file (likely timed 
out) (batchId=417)
TestClearDanglingScratchDir - did not produce a TEST-*.xml file (likely timed 
out) (batchId=384)
TestClientSideAuthorizationProvider - did not produce a TEST-*.xml file (likely 
timed out) (batchId=391)
TestCompactor - did not produce a TEST-*.xml file (likely timed out) 
(batchId=380)
TestCreateUdfEntities - did not produce a TEST-*.xml file (likely timed out) 
(batchId=379)
TestCustomAuthentication - did not produce a TEST-*.xml file (likely timed out) 
(batchId=400)
TestDBTokenStore - did not produce a TEST-*.xml file (likely timed out) 
(batchId=343)
TestDDLWithRemoteMetastoreSecondNamenode - did not produce a TEST-*.xml file 
(likely timed out) (batchId=378)
TestDynamicSerDe - did not produce a TEST-*.xml file (likely timed out) 
(batchId=346)
TestEmbeddedHiveMetaStore - did not produce a TEST-*.xml file (likely timed 
out) (batchId=356)
TestEmbeddedThriftBinaryCLIService - did not produce a TEST-*.xml file (likely 
timed out) (batchId=403)
TestFilterHooks - did not produce a TEST-*.xml file (likely timed out) 
(batchId=351)
TestFolderPermissions - did not produce a TEST-*.xml file (likely timed out) 
(batchId=386)
TestHS2AuthzContext - did not produce a TEST-*.xml file (likely timed out) 
(batchId=420)
TestHS2AuthzSessionContext - did not produce a TEST-*.xml file (likely timed 
out) (batchId=421)
TestHS2ClearDanglingScratchDir - did not produce a TEST-*.xml file (likely 
timed out) (batchId=407)
TestHS2ImpersonationWithRemoteMS - did not produce a TEST-*.xml file (likely 
timed out) (batchId=408)
TestHiveAuthorizerCheckInvocation - did not produce a TEST-*.xml file (likely 
timed out) (batchId=395)
TestHiveAuthorizerShowFilters - did not produce a TEST-*.xml file (likely timed 
out) (batchId=394)
TestHiveHistory - did not produce a TEST-*.xml file (likely timed out) 
(batchId=397)
TestHiveMetaStoreTxns - did not produce a TEST-*.xml file (likely timed out) 
(batchId=371)
TestHiveMetaStoreWithEnvironmentContext - did not produce a TEST-*.xml file 
(likely timed out) (batchId=361)
TestHiveMetaTool - did not produce a TEST-*.xml file (likely timed out) 
(batchId=374)
TestHiveServer2 - did not produce a TEST-*.xml file (likely timed out) 
(batchId=423)
TestHiveServer2SessionTimeout - did not produce a TEST-*.xml file (likely timed 
out) (batchId=424)
TestHiveSessionImpl - did not produce a TEST-*.xml file (likely timed out) 
(batchId=404)
TestHs2Hooks - did not produce a TEST-*.xml file (likely timed out) 
(batchId=376)
TestJdbcDriver2 - did not produce a TEST-*.xml file (likely timed out) 
(batchId=411)
TestJdbcMetadataApiAuth - did not produce a TEST-*.xml file (likely timed out) 
(batchId=422)
TestJdbcWithLocalClusterSpark - did not produce a TEST-*.xml file (likely timed 
out) (batchId=416)
TestJdbcWithMiniHS2 - did not produce a TEST-*.xml file (likely timed out) 
(batchId=413)
TestJdbcWithMiniMr - did not produce a TEST-*.xml file (likely timed out) 
(batchId=412)
TestJdbcWithSQLAuthUDFBlacklist - did not produce a TEST-*.xml file (likely 
timed out) (batchId=418)
TestJdbcWithSQLAuthorization - did not produce a TEST-*.xml file (likely timed 
out) (batchId=419)
TestLocationQueries - did not produce a TEST-*.xml file (likely timed out) 
(batchId=383)
TestMTQueries - did not produce a TEST-*.xml file (likely timed out) 
(batchId=381)
TestMarkPartition - did not produce a TEST-*.xml file (likely timed out) 
(batchId=368)
TestMarkPartitionRemote - did not produce a TEST-*.xml file (likely timed out) 
(batchId=372)
TestMetaStoreAuthorization - did not produce a TEST-*.xml file (likely timed 
out) (batchId=357)
TestMetaStoreConnectionUrlHook - did not produce a TEST-*.xml file (likely 
timed out) (batchId=355)
TestMetaStoreEndFunctionListener - did not produce a TEST-*.xml file (likely 
timed out) 

[jira] [Updated] (HIVE-15225) QueryPlan.getJSONValue should code against empty string values

2016-11-16 Thread Yibing Shi (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15225?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yibing Shi updated HIVE-15225:
--
Status: Patch Available  (was: Open)

> QueryPlan.getJSONValue should code against empty string values
> --
>
> Key: HIVE-15225
> URL: https://issues.apache.org/jira/browse/HIVE-15225
> Project: Hive
>  Issue Type: Bug
>Reporter: Yibing Shi
> Attachments: HIVE-15225.1.patch
>
>
> The current {{QueryPlan.getJSONValue}} implementation is as below:
> {code}
>   public String getJSONValue(Object value) {
> String v = "null";
> if (value != null) {
>   v = value.toString();
>   if (v.charAt(0) != '[' && v.charAt(0) != '{') {
> v = "\"" + v + "\"";
>   }
> }
> return v;
>   }
> {code}
> When {{value.toString()}} returns an empty string, a 
> StringIndexOutOfRangeException would be thrown out when "v.charAt(0)" is 
> evaluated.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (HIVE-15225) QueryPlan.getJSONValue should code against empty string values

2016-11-16 Thread Yibing Shi (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15225?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yibing Shi reassigned HIVE-15225:
-

Assignee: Yibing Shi

> QueryPlan.getJSONValue should code against empty string values
> --
>
> Key: HIVE-15225
> URL: https://issues.apache.org/jira/browse/HIVE-15225
> Project: Hive
>  Issue Type: Bug
>Reporter: Yibing Shi
>Assignee: Yibing Shi
> Attachments: HIVE-15225.1.patch
>
>
> The current {{QueryPlan.getJSONValue}} implementation is as below:
> {code}
>   public String getJSONValue(Object value) {
> String v = "null";
> if (value != null) {
>   v = value.toString();
>   if (v.charAt(0) != '[' && v.charAt(0) != '{') {
> v = "\"" + v + "\"";
>   }
> }
> return v;
>   }
> {code}
> When {{value.toString()}} returns an empty string, a 
> StringIndexOutOfRangeException would be thrown out when "v.charAt(0)" is 
> evaluated.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15225) QueryPlan.getJSONValue should code against empty string values

2016-11-16 Thread Yibing Shi (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15225?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yibing Shi updated HIVE-15225:
--
Attachment: HIVE-15225.1.patch

Attach a quick patch

> QueryPlan.getJSONValue should code against empty string values
> --
>
> Key: HIVE-15225
> URL: https://issues.apache.org/jira/browse/HIVE-15225
> Project: Hive
>  Issue Type: Bug
>Reporter: Yibing Shi
> Attachments: HIVE-15225.1.patch
>
>
> The current {{QueryPlan.getJSONValue}} implementation is as below:
> {code}
>   public String getJSONValue(Object value) {
> String v = "null";
> if (value != null) {
>   v = value.toString();
>   if (v.charAt(0) != '[' && v.charAt(0) != '{') {
> v = "\"" + v + "\"";
>   }
> }
> return v;
>   }
> {code}
> When {{value.toString()}} returns an empty string, a 
> StringIndexOutOfRangeException would be thrown out when "v.charAt(0)" is 
> evaluated.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-14982) Remove some reserved keywords in 2.2

2016-11-16 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-14982?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671972#comment-15671972
 ] 

Ashutosh Chauhan commented on HIVE-14982:
-

+1

> Remove some reserved keywords in 2.2
> 
>
> Key: HIVE-14982
> URL: https://issues.apache.org/jira/browse/HIVE-14982
> Project: Hive
>  Issue Type: Bug
>Reporter: Pengcheng Xiong
>Assignee: Pengcheng Xiong
> Attachments: HIVE-14982.01.patch
>
>
> It seems that CACHE, DAYOFWEEK, VIEWS are reserved keywords in master. This 
> conflicts with SQL2011 standard.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13282) GroupBy and select operator encounter ArrayIndexOutOfBoundsException

2016-11-16 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/HIVE-13282?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergio Peña updated HIVE-13282:
---
Target Version/s: 1.2.2, 2.2.0  (was: 1.2.2, 2.2.0, 2.1.1)

> GroupBy and select operator encounter ArrayIndexOutOfBoundsException
> 
>
> Key: HIVE-13282
> URL: https://issues.apache.org/jira/browse/HIVE-13282
> Project: Hive
>  Issue Type: Bug
>  Components: Query Processor
>Affects Versions: 1.2.1, 2.0.0, 2.1.0
>Reporter: Vikram Dixit K
>Assignee: Matt McCline
>Priority: Blocker
> Attachments: HIVE-13282.01.patch, smb_fail_issue.patch, 
> smb_groupby.q, smb_groupby.q.out
>
>
> The group by and select operators run into the ArrayIndexOutOfBoundsException 
> when they incorrectly initialize themselves with tag 0 but the incoming tag 
> id is different.
> {code}
> select count(*) from
> (select rt1.id from
> (select t1.key as id, t1.value as od from tab t1 group by key, value) rt1) vt1
> join
> (select rt2.id from
> (select t2.key as id, t2.value as od from tab_part t2 group by key, value) 
> rt2) vt2
> where vt1.id=vt2.id;
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15199) INSERT INTO data on S3 is replacing the old rows with the new ones

2016-11-16 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/HIVE-15199?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergio Peña updated HIVE-15199:
---
Attachment: HIVE-15199.3.patch

> INSERT INTO data on S3 is replacing the old rows with the new ones
> --
>
> Key: HIVE-15199
> URL: https://issues.apache.org/jira/browse/HIVE-15199
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Sergio Peña
>Assignee: Sergio Peña
>Priority: Critical
> Attachments: HIVE-15199.1.patch, HIVE-15199.2.patch, 
> HIVE-15199.3.patch
>
>
> Any INSERT INTO statement run on S3 tables and when the scratch directory is 
> saved on S3 is deleting old rows of the table.
> {noformat}
> hive> set hive.blobstore.use.blobstore.as.scratchdir=true;
> hive> create table t1 (id int, name string) location 's3a://spena-bucket/t1';
> hive> insert into table t1 values (1,'name1');
> hive> select * from t1;
> 1   name1
> hive> insert into table t1 values (2,'name2');
> hive> select * from t1;
> 2   name2
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-1478) Non-boolean expression in WHERE should be rejected

2016-11-16 Thread Zoltan Haindrich (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1478?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltan Haindrich updated HIVE-1478:
---
Attachment: HIVE-1478.2.patch

patch#2)

removed qtest cases which contained invalid where expressions.

it turned out that there was a ticket to do support this conversion for 
vectorization(HIVE-13659)...not sure what to do now:

* close this as won't fix - and open an another issue that this fails for mr
* i've undone the {{VectorizationContext}} change of HIVE-13659 in this patch - 
as it will became an unreachable codepath after this change.

I think it would be better to stick to the standard...because casting anything 
into a boolean may help a bad query to hide in the woods ;)

[~ashutoshc],[~mmccline] what's your opinion?

> Non-boolean expression in WHERE should be rejected
> --
>
> Key: HIVE-1478
> URL: https://issues.apache.org/jira/browse/HIVE-1478
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.7.0
>Reporter: Paul Yang
>Assignee: Zoltan Haindrich
>Priority: Minor
> Attachments: HIVE-1478.1.patch, HIVE-1478.2.patch
>
>
> Automatically casting strings or other types into boolean may confuse even 
> the user - and somehow it doesn't always work (HIVE-15089)
> sql2011 states that "where expression" should accept a boolean expression.
> Original reported problem:
> If the expression in the where clause does not evaluate to a boolean, the job 
> will fail with the following exception in the task logs:
> Query:
> SELECT key FROM src WHERE 1;
> Exception in mapper:
> 2010-07-21 17:00:31,460 FATAL ExecMapper: 
> org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while 
> processing row {"key":"238","value":"val_238"}
>   at 
> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:417)
>   at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:180)
>   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
>   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:358)
>   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:307)
>   at org.apache.hadoop.mapred.Child.main(Child.java:159)
> Caused by: java.lang.ClassCastException: java.lang.Integer cannot be cast to 
> java.lang.Boolean
>   at 
> org.apache.hadoop.hive.ql.exec.FilterOperator.processOp(FilterOperator.java:84)
>   at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
>   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
>   at 
> org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:45)
>   at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
>   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
>   at 
> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:400)
>   ... 5 more



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10901) Optimize mutli column distinct queries

2016-11-16 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10901?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-10901:
---
Status: Patch Available  (was: Open)

> Optimize  mutli column distinct queries 
> 
>
> Key: HIVE-10901
> URL: https://issues.apache.org/jira/browse/HIVE-10901
> Project: Hive
>  Issue Type: New Feature
>  Components: CBO, Logical Optimizer
>Affects Versions: 1.2.0
>Reporter: Mostafa Mokhtar
>Assignee: Pengcheng Xiong
> Attachments: HIVE-10901.02.patch, HIVE-10901.03.patch, 
> HIVE-10901.04.patch, HIVE-10901.patch
>
>
> HIVE-10568 is useful only when there is a distinct on one column. It can be 
> expanded for multiple column cases too.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10901) Optimize multi column distinct queries

2016-11-16 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10901?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-10901:
---
Summary: Optimize  multi column distinct queries   (was: Optimize  mutli 
column distinct queries )

> Optimize  multi column distinct queries 
> 
>
> Key: HIVE-10901
> URL: https://issues.apache.org/jira/browse/HIVE-10901
> Project: Hive
>  Issue Type: New Feature
>  Components: CBO, Logical Optimizer
>Affects Versions: 1.2.0
>Reporter: Mostafa Mokhtar
>Assignee: Pengcheng Xiong
> Attachments: HIVE-10901.02.patch, HIVE-10901.03.patch, 
> HIVE-10901.04.patch, HIVE-10901.patch
>
>
> HIVE-10568 is useful only when there is a distinct on one column. It can be 
> expanded for multiple column cases too.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10901) Optimize mutli column distinct queries

2016-11-16 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10901?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-10901:
---
Status: Open  (was: Patch Available)

> Optimize  mutli column distinct queries 
> 
>
> Key: HIVE-10901
> URL: https://issues.apache.org/jira/browse/HIVE-10901
> Project: Hive
>  Issue Type: New Feature
>  Components: CBO, Logical Optimizer
>Affects Versions: 1.2.0
>Reporter: Mostafa Mokhtar
>Assignee: Pengcheng Xiong
> Attachments: HIVE-10901.02.patch, HIVE-10901.03.patch, 
> HIVE-10901.04.patch, HIVE-10901.patch
>
>
> HIVE-10568 is useful only when there is a distinct on one column. It can be 
> expanded for multiple column cases too.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-10901) Optimize mutli column distinct queries

2016-11-16 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-10901?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-10901:
---
Attachment: HIVE-10901.04.patch

> Optimize  mutli column distinct queries 
> 
>
> Key: HIVE-10901
> URL: https://issues.apache.org/jira/browse/HIVE-10901
> Project: Hive
>  Issue Type: New Feature
>  Components: CBO, Logical Optimizer
>Affects Versions: 1.2.0
>Reporter: Mostafa Mokhtar
>Assignee: Pengcheng Xiong
> Attachments: HIVE-10901.02.patch, HIVE-10901.03.patch, 
> HIVE-10901.04.patch, HIVE-10901.patch
>
>
> HIVE-10568 is useful only when there is a distinct on one column. It can be 
> expanded for multiple column cases too.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15199) INSERT INTO data on S3 is replacing the old rows with the new ones

2016-11-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15199?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671747#comment-15671747
 ] 

Hive QA commented on HIVE-15199:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12839221/HIVE-15199.2.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 209 failed/errored test(s), 10664 tests 
executed
*Failed tests:*
{noformat}
TestSparkCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=109)

[enforce_order.q,ppd_join2.q,smb_mapjoin_21.q,load_dyn_part15.q,udf_min.q,groupby_resolution.q,mapjoin_memcheck.q,subquery_exists.q,join27.q,alter_merge_stats_orc.q,union_remove_2.q,vector_orderby_5.q,groupby6_map_skew.q,join12.q,union9.q]
TestSparkCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=128)

[union_remove_15.q,bucket_map_join_tez1.q,groupby7_noskew.q,bucketmapjoin1.q,subquery_multiinsert.q,auto_join8.q,auto_join6.q,groupby2_map_skew.q,lateral_view_explode2.q,join28.q,load_dyn_part1.q,skewjoinopt17.q,skewjoin_union_remove_1.q,union_remove_20.q,bucketmapjoin5.q]
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[add_part_multiple] 
(batchId=62)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[alter_merge_2_orc] 
(batchId=67)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[alter_partition_coltype] 
(batchId=23)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[alter_table_add_partition]
 (batchId=16)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[alter_table_update_status]
 (batchId=71)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[array_map_access_nonconstant]
 (batchId=20)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[autoColumnStats_3] 
(batchId=50)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[autoColumnStats_5] 
(batchId=37)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[avro_add_column2] 
(batchId=80)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[avrocountemptytbl] 
(batchId=73)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[ba_table_udfs] 
(batchId=22)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[cbo_rp_insert] 
(batchId=10)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[columnStatsUpdateForStatsOptimizer_2]
 (batchId=27)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[column_names_with_leading_and_trailing_spaces]
 (batchId=21)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[complex_alias] 
(batchId=15)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[concat_op] (batchId=66)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[constantPropWhen] 
(batchId=31)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[constantfolding] 
(batchId=67)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[constprog_when_case] 
(batchId=52)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[create_merge_compressed] 
(batchId=38)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[cte_5] (batchId=30)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[cte_7] (batchId=23)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[cte_mat_5] (batchId=2)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[dbtxnmgr_query1] 
(batchId=65)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[decimal_precision] 
(batchId=47)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[distinct_stats] 
(batchId=64)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[except_all] (batchId=41)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[exim_01_nonpart] 
(batchId=48)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[exim_03_nonpart_over_compat]
 (batchId=5)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[exim_08_nonpart_rename] 
(batchId=55)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[exim_10_external_managed]
 (batchId=62)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[exim_12_external_location]
 (batchId=49)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[exim_13_managed_location]
 (batchId=35)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[exim_14_managed_location_over_existing]
 (batchId=48)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[exim_22_import_exist_authsuccess]
 (batchId=17)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[exim_24_import_nonexist_authsuccess]
 (batchId=17)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[folder_predicate] 
(batchId=4)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[groupby_distinct_samekey]
 (batchId=51)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[groupby_grouping_window] 
(batchId=29)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[groupby_nullvalues] 
(batchId=72)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[implicit_decimal] 
(batchId=61)

[jira] [Commented] (HIVE-15219) LLAP: Allow additional slider global parameters to be set while creating the LLAP package

2016-11-16 Thread Siddharth Seth (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15219?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671739#comment-15671739
 ] 

Siddharth Seth commented on HIVE-15219:
---

The intent was for the user to provide free-form text,
site.global.prop1=val1,site.global.prop2=val2.

Think I'm going to simplify this to allow specific properties to be set, using 
the same mechanism used for hiveconf. i.e. --slider_appconfig_global 
site.global.prop1=val1 --slider_appconfig_global site.global.prop2=val2

> LLAP: Allow additional slider global parameters to be set while creating the 
> LLAP package
> -
>
> Key: HIVE-15219
> URL: https://issues.apache.org/jira/browse/HIVE-15219
> Project: Hive
>  Issue Type: Task
>  Components: llap
>Reporter: Siddharth Seth
>Assignee: Siddharth Seth
> Attachments: HIVE-15219.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13282) GroupBy and select operator encounter ArrayIndexOutOfBoundsException

2016-11-16 Thread Matt McCline (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13282?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671710#comment-15671710
 ] 

Matt McCline commented on HIVE-13282:
-

No, this patch is not needed for 2.1.1

> GroupBy and select operator encounter ArrayIndexOutOfBoundsException
> 
>
> Key: HIVE-13282
> URL: https://issues.apache.org/jira/browse/HIVE-13282
> Project: Hive
>  Issue Type: Bug
>  Components: Query Processor
>Affects Versions: 1.2.1, 2.0.0, 2.1.0
>Reporter: Vikram Dixit K
>Assignee: Matt McCline
>Priority: Blocker
> Attachments: HIVE-13282.01.patch, smb_fail_issue.patch, 
> smb_groupby.q, smb_groupby.q.out
>
>
> The group by and select operators run into the ArrayIndexOutOfBoundsException 
> when they incorrectly initialize themselves with tag 0 but the incoming tag 
> id is different.
> {code}
> select count(*) from
> (select rt1.id from
> (select t1.key as id, t1.value as od from tab t1 group by key, value) rt1) vt1
> join
> (select rt2.id from
> (select t2.key as id, t2.value as od from tab_part t2 group by key, value) 
> rt2) vt2
> where vt1.id=vt2.id;
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13282) GroupBy and select operator encounter ArrayIndexOutOfBoundsException

2016-11-16 Thread JIRA

[ 
https://issues.apache.org/jira/browse/HIVE-13282?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671668#comment-15671668
 ] 

Sergio Peña commented on HIVE-13282:


[~vikram.dixit] Is this patch still needed for 2.1.1? We're looking to release 
an RC this or next week.

> GroupBy and select operator encounter ArrayIndexOutOfBoundsException
> 
>
> Key: HIVE-13282
> URL: https://issues.apache.org/jira/browse/HIVE-13282
> Project: Hive
>  Issue Type: Bug
>  Components: Query Processor
>Affects Versions: 1.2.1, 2.0.0, 2.1.0
>Reporter: Vikram Dixit K
>Assignee: Matt McCline
>Priority: Blocker
> Attachments: HIVE-13282.01.patch, smb_fail_issue.patch, 
> smb_groupby.q, smb_groupby.q.out
>
>
> The group by and select operators run into the ArrayIndexOutOfBoundsException 
> when they incorrectly initialize themselves with tag 0 but the incoming tag 
> id is different.
> {code}
> select count(*) from
> (select rt1.id from
> (select t1.key as id, t1.value as od from tab t1 group by key, value) rt1) vt1
> join
> (select rt2.id from
> (select t2.key as id, t2.value as od from tab_part t2 group by key, value) 
> rt2) vt2
> where vt1.id=vt2.id;
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-1478) Non-boolean expression in WHERE should be rejected

2016-11-16 Thread Zoltan Haindrich (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1478?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltan Haindrich updated HIVE-1478:
---
Description: 
Automatically casting strings or other types into boolean may confuse even the 
user - and somehow it doesn't always work (HIVE-15089)

sql2011 states that "where expression" should accept a boolean expression.


Original reported problem:

If the expression in the where clause does not evaluate to a boolean, the job 
will fail with the following exception in the task logs:

Query:

SELECT key FROM src WHERE 1;

Exception in mapper:

2010-07-21 17:00:31,460 FATAL ExecMapper: 
org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while 
processing row {"key":"238","value":"val_238"}
at 
org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:417)
at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:180)
at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:358)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:307)
at org.apache.hadoop.mapred.Child.main(Child.java:159)
Caused by: java.lang.ClassCastException: java.lang.Integer cannot be cast to 
java.lang.Boolean
at 
org.apache.hadoop.hive.ql.exec.FilterOperator.processOp(FilterOperator.java:84)
at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
at 
org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:45)
at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
at 
org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:400)
... 5 more


  was:
This problem is still present (

Currently:

If the expression in the where clause does not evaluate to a boolean, the job 
will fail with the following exception in the task logs:

Query:

SELECT key FROM src WHERE 1;

Exception in mapper:

2010-07-21 17:00:31,460 FATAL ExecMapper: 
org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while 
processing row {"key":"238","value":"val_238"}
at 
org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:417)
at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:180)
at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:358)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:307)
at org.apache.hadoop.mapred.Child.main(Child.java:159)
Caused by: java.lang.ClassCastException: java.lang.Integer cannot be cast to 
java.lang.Boolean
at 
org.apache.hadoop.hive.ql.exec.FilterOperator.processOp(FilterOperator.java:84)
at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
at 
org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:45)
at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
at 
org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:400)
... 5 more



> Non-boolean expression in WHERE should be rejected
> --
>
> Key: HIVE-1478
> URL: https://issues.apache.org/jira/browse/HIVE-1478
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.7.0
>Reporter: Paul Yang
>Assignee: Zoltan Haindrich
>Priority: Minor
> Attachments: HIVE-1478.1.patch
>
>
> Automatically casting strings or other types into boolean may confuse even 
> the user - and somehow it doesn't always work (HIVE-15089)
> sql2011 states that "where expression" should accept a boolean expression.
> Original reported problem:
> If the expression in the where clause does not evaluate to a boolean, the job 
> will fail with the following exception in the task logs:
> Query:
> SELECT key FROM src WHERE 1;
> Exception in mapper:
> 2010-07-21 17:00:31,460 FATAL ExecMapper: 
> org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while 
> processing row {"key":"238","value":"val_238"}
>   at 
> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:417)
>   at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:180)
>   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
>   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:358)
>   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:307)
>   at 

[jira] [Updated] (HIVE-1478) Non-boolean expression in WHERE should be rejected

2016-11-16 Thread Zoltan Haindrich (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1478?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltan Haindrich updated HIVE-1478:
---
Summary: Non-boolean expression in WHERE should be rejected  (was: 
Non-boolean expression in WHERE clause throws exception)

> Non-boolean expression in WHERE should be rejected
> --
>
> Key: HIVE-1478
> URL: https://issues.apache.org/jira/browse/HIVE-1478
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.7.0
>Reporter: Paul Yang
>Assignee: Zoltan Haindrich
>Priority: Minor
> Attachments: HIVE-1478.1.patch
>
>
> If the expression in the where clause does not evaluate to a boolean, the job 
> will fail with the following exception in the task logs:
> Query:
> SELECT key FROM src WHERE 1;
> Exception in mapper:
> 2010-07-21 17:00:31,460 FATAL ExecMapper: 
> org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while 
> processing row {"key":"238","value":"val_238"}
>   at 
> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:417)
>   at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:180)
>   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
>   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:358)
>   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:307)
>   at org.apache.hadoop.mapred.Child.main(Child.java:159)
> Caused by: java.lang.ClassCastException: java.lang.Integer cannot be cast to 
> java.lang.Boolean
>   at 
> org.apache.hadoop.hive.ql.exec.FilterOperator.processOp(FilterOperator.java:84)
>   at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
>   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
>   at 
> org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:45)
>   at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
>   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
>   at 
> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:400)
>   ... 5 more



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-1478) Non-boolean expression in WHERE should be rejected

2016-11-16 Thread Zoltan Haindrich (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-1478?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltan Haindrich updated HIVE-1478:
---
Description: 
This problem is still present (

Currently:

If the expression in the where clause does not evaluate to a boolean, the job 
will fail with the following exception in the task logs:

Query:

SELECT key FROM src WHERE 1;

Exception in mapper:

2010-07-21 17:00:31,460 FATAL ExecMapper: 
org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while 
processing row {"key":"238","value":"val_238"}
at 
org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:417)
at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:180)
at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:358)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:307)
at org.apache.hadoop.mapred.Child.main(Child.java:159)
Caused by: java.lang.ClassCastException: java.lang.Integer cannot be cast to 
java.lang.Boolean
at 
org.apache.hadoop.hive.ql.exec.FilterOperator.processOp(FilterOperator.java:84)
at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
at 
org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:45)
at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
at 
org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:400)
... 5 more


  was:
If the expression in the where clause does not evaluate to a boolean, the job 
will fail with the following exception in the task logs:

Query:

SELECT key FROM src WHERE 1;

Exception in mapper:

2010-07-21 17:00:31,460 FATAL ExecMapper: 
org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while 
processing row {"key":"238","value":"val_238"}
at 
org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:417)
at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:180)
at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:358)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:307)
at org.apache.hadoop.mapred.Child.main(Child.java:159)
Caused by: java.lang.ClassCastException: java.lang.Integer cannot be cast to 
java.lang.Boolean
at 
org.apache.hadoop.hive.ql.exec.FilterOperator.processOp(FilterOperator.java:84)
at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
at 
org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:45)
at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
at 
org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:400)
... 5 more



> Non-boolean expression in WHERE should be rejected
> --
>
> Key: HIVE-1478
> URL: https://issues.apache.org/jira/browse/HIVE-1478
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.7.0
>Reporter: Paul Yang
>Assignee: Zoltan Haindrich
>Priority: Minor
> Attachments: HIVE-1478.1.patch
>
>
> This problem is still present (
> Currently:
> If the expression in the where clause does not evaluate to a boolean, the job 
> will fail with the following exception in the task logs:
> Query:
> SELECT key FROM src WHERE 1;
> Exception in mapper:
> 2010-07-21 17:00:31,460 FATAL ExecMapper: 
> org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while 
> processing row {"key":"238","value":"val_238"}
>   at 
> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:417)
>   at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:180)
>   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
>   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:358)
>   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:307)
>   at org.apache.hadoop.mapred.Child.main(Child.java:159)
> Caused by: java.lang.ClassCastException: java.lang.Integer cannot be cast to 
> java.lang.Boolean
>   at 
> org.apache.hadoop.hive.ql.exec.FilterOperator.processOp(FilterOperator.java:84)
>   at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
>   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
>   at 
> 

[jira] [Commented] (HIVE-15215) Files on S3 are deleted one by one in INSERT OVERWRITE queries

2016-11-16 Thread JIRA

[ 
https://issues.apache.org/jira/browse/HIVE-15215?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671633#comment-15671633
 ] 

Sergio Peña commented on HIVE-15215:


It might be good to use approach #1 where it sets the staging directory under 
the scratch directory only for blobstorage so Hive can do the final move at the 
end. This move may be in parallel as renames are not allowed on S3, so it may 
be faster.

> Files on S3 are deleted one by one in INSERT OVERWRITE queries
> --
>
> Key: HIVE-15215
> URL: https://issues.apache.org/jira/browse/HIVE-15215
> Project: Hive
>  Issue Type: Sub-task
>  Components: Hive
>Reporter: Sahil Takiar
>
> When running {{INSERT OVERWRITE}} queries the files to overwrite are deleted 
> one by one. The reason is that, by default, hive.exec.stagingdir is inside 
> the target table directory.
> Ideally Hive would just delete the entire table directory, but it can't do 
> that since the staging data is also inside the directory. Instead it deletes 
> each file one-by-one, which is very slow.
> There are a few ways to fix this:
> 1: Move the staging directory outside the table location. This can be done by 
>  setting hive.exec.stagingdir to a different location when running on S3. It 
> would be nice if users didn't have to explicitly set this when running on S3 
> and things just worked out-of-the-box. My understanding is that 
> hive.exec.stagingdir was only added to support HDFS encryption zones. Since 
> S3 doesn't have encryption zones, there should be no problem with using the 
> value of hive.exec.scratchdir to store all intermediate data instead.
> 2: Multi-thread the delete operations
> 3: See if the {{S3AFileSystem}} can expose some type of bulk delete op



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15180) Extend JSONMessageFactory to store additional information about metadata objects on different table events

2016-11-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15180?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671568#comment-15671568
 ] 

Hive QA commented on HIVE-15180:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12839218/HIVE-15180.3.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 7 failed/errored test(s), 10694 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[transform_ppr2] 
(batchId=133)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[join_acid_non_acid]
 (batchId=150)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[union_fast_stats]
 (batchId=145)
org.apache.hive.hcatalog.api.TestHCatClientNotification.addPartition 
(batchId=217)
org.apache.hive.hcatalog.api.TestHCatClientNotification.createTable 
(batchId=217)
org.apache.hive.hcatalog.api.TestHCatClientNotification.dropPartition 
(batchId=217)
org.apache.hive.hcatalog.api.TestHCatClientNotification.dropTable (batchId=217)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/2155/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/2155/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-2155/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 7 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12839218 - PreCommit-HIVE-Build

> Extend JSONMessageFactory to store additional information about metadata 
> objects on different table events
> --
>
> Key: HIVE-15180
> URL: https://issues.apache.org/jira/browse/HIVE-15180
> Project: Hive
>  Issue Type: Sub-task
>  Components: repl
>Reporter: Vaibhav Gumashta
>Assignee: Vaibhav Gumashta
> Attachments: HIVE-15180.1.patch, HIVE-15180.2.patch, 
> HIVE-15180.3.patch, HIVE-15180.3.patch
>
>
> We want the {{NOTIFICATION_LOG}} table to capture additional information 
> about the metadata objects when {{DbNotificationListener}} captures different 
> events for a table (create/drop/alter).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-6536) Reduce dependencies of org.apache.hive:hive-jdbc maven module

2016-11-16 Thread Tero Paananen (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-6536?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671508#comment-15671508
 ] 

Tero Paananen commented on HIVE-6536:
-

This issue has become much, much worse in the 2.1.0 version of the JDBC driver.

There's something like 80MB of dependencies in the 2.1.0 version. I'm not sure 
how much bigger it is compared to 1.2.1 I was using before, but it's at least 
2x - 4x bigger.

> Reduce dependencies of org.apache.hive:hive-jdbc maven module
> -
>
> Key: HIVE-6536
> URL: https://issues.apache.org/jira/browse/HIVE-6536
> Project: Hive
>  Issue Type: Improvement
>  Components: JDBC
>Affects Versions: 0.12.0
> Environment: org.apache.hive:hive-jdbc:jar:0.12.0
>Reporter: Kevin Minder
> Attachments: hive-jdbc-maven-dependencies-0-13.log, 
> hive-jdbc-maven-dependencies.log
>
>
> The Hive JDBC driver maven module requires a significant number of 
> dependencies that are likely unnecessary and will result in bloating of 
> consumers.  Most of this is a result of the dependency on 
> org.apache.hive:hive-cli.  I have attached a portion of the output from mvn 
> depedency:tree output for a client that depends on the 
> org.apache.hive:hive-jdbc module.  Note the extra 2.0.6.1-102 in the output 
> is the result of our local build and publish to a local nexus repo.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-14189) backport HIVE-13945 to branch-1

2016-11-16 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-14189?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-14189:

Attachment: HIVE-14189.07-branch-1.patch

Trying again...

> backport HIVE-13945 to branch-1
> ---
>
> Key: HIVE-14189
> URL: https://issues.apache.org/jira/browse/HIVE-14189
> Project: Hive
>  Issue Type: Bug
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
>  Labels: TODOC1.3
> Attachments: HIVE-14189-branch-1.patch, HIVE-14189.01-branch-1.patch, 
> HIVE-14189.02-branch-1.patch, HIVE-14189.03-branch-1.patch, 
> HIVE-14189.04-branch-1.patch, HIVE-14189.05-branch-1.patch, 
> HIVE-14189.06-branch-1.patch, HIVE-14189.07-branch-1.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-14902) HiveServer2: Cleanup code which checks for ThriftJDBCSerde usage

2016-11-16 Thread Ziyang Zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-14902?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671439#comment-15671439
 ] 

Ziyang Zhao commented on HIVE-14902:


Failed tests passed in my local. Seems unrelated.

> HiveServer2: Cleanup code which checks for ThriftJDBCSerde usage
> 
>
> Key: HIVE-14902
> URL: https://issues.apache.org/jira/browse/HIVE-14902
> Project: Hive
>  Issue Type: Sub-task
>  Components: HiveServer2, JDBC, ODBC
>Affects Versions: 2.1.0
>Reporter: Vaibhav Gumashta
>Assignee: Ziyang Zhao
> Attachments: HIVE-14902.2.patch
>
>
> As part of HIVE-14039, we simplified the condition checking by setting a 
> boolean in SessionState (from HiveSessionImpl) and then using that in other 
> layers to decide whether we should use ThriftJDBCSerde. However, there are 
> some parts of the code (e.g. FileSinkOperator.closeOp, Driver.getResults, 
> SQLOperation.getNextRowSet) where the condition checking is still complicated 
> - we should refactor that.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15199) INSERT INTO data on S3 is replacing the old rows with the new ones

2016-11-16 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/HIVE-15199?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergio Peña updated HIVE-15199:
---
Attachment: HIVE-15199.2.patch

[~stakiar] Thanks. I updated the patch with:

- Using an hybrid solution that checks if a file exists on the list status or 
if exists on the FS.
- Change the renameNonLocal to renameIsAllowed

For the Hive lock, yes, Hive should have a lock to avoid another client inserts 
data on the same table.

> INSERT INTO data on S3 is replacing the old rows with the new ones
> --
>
> Key: HIVE-15199
> URL: https://issues.apache.org/jira/browse/HIVE-15199
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Sergio Peña
>Assignee: Sergio Peña
>Priority: Critical
> Attachments: HIVE-15199.1.patch, HIVE-15199.2.patch
>
>
> Any INSERT INTO statement run on S3 tables and when the scratch directory is 
> saved on S3 is deleting old rows of the table.
> {noformat}
> hive> set hive.blobstore.use.blobstore.as.scratchdir=true;
> hive> create table t1 (id int, name string) location 's3a://spena-bucket/t1';
> hive> insert into table t1 values (1,'name1');
> hive> select * from t1;
> 1   name1
> hive> insert into table t1 values (2,'name2');
> hive> select * from t1;
> 2   name2
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15200) Support setOp in subQuery with parentheses

2016-11-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15200?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671428#comment-15671428
 ] 

Hive QA commented on HIVE-15200:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12839211/HIVE-15200.01.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 80 failed/errored test(s), 10665 tests 
executed
*Failed tests:*
{noformat}
TestSparkCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=104)

[skewjoin_union_remove_2.q,avro_decimal_native.q,skewjoinopt8.q,bucketmapjoin_negative3.q,union32.q,stats6.q,groupby2_map.q,stats_only_null.q,insert_into3.q,join18_multi_distinct.q,vectorization_6.q,cross_join.q,stats9.q,timestamp_1.q,join24.q]
TestSparkCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=128)

[union_remove_15.q,bucket_map_join_tez1.q,groupby7_noskew.q,bucketmapjoin1.q,subquery_multiinsert.q,auto_join8.q,auto_join6.q,groupby2_map_skew.q,lateral_view_explode2.q,join28.q,load_dyn_part1.q,skewjoinopt17.q,skewjoin_union_remove_1.q,union_remove_20.q,bucketmapjoin5.q]
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[auto_join4] (batchId=63)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[auto_join5] (batchId=65)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[auto_join6] (batchId=77)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[auto_join7] (batchId=24)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[auto_join8] (batchId=77)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[druid_basic2] 
(batchId=10)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[groupby_bigdata] 
(batchId=31)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[groupby_position] 
(batchId=35)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[input14] (batchId=69)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[input14_limit] 
(batchId=55)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[input17] (batchId=33)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[input18] (batchId=16)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[input20] (batchId=12)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[input33] (batchId=54)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[input34] (batchId=16)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[input35] (batchId=53)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[input36] (batchId=13)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[input37] (batchId=68)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[input38] (batchId=12)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[input5] (batchId=12)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[join4] (batchId=58)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[join5] (batchId=62)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[join6] (batchId=67)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[join7] (batchId=4)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[join8] (batchId=43)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[nonreserved_keywords_input37]
 (batchId=27)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[partcols1] (batchId=57)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[ppd_transform] 
(batchId=68)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[ppd_union] (batchId=42)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[rcfile_bigdata] 
(batchId=19)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[rcfile_null_value] 
(batchId=25)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[regexp_extract] 
(batchId=48)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[setop_subq] (batchId=5)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[subq] (batchId=39)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[subquery_alias] 
(batchId=22)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[transform_ppr1] 
(batchId=13)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[transform_ppr2] 
(batchId=39)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[union] (batchId=4)
org.apache.hadoop.hive.cli.TestContribCliDriver.testCliDriver[java_mr_example] 
(batchId=218)
org.apache.hadoop.hive.cli.TestContribCliDriver.testCliDriver[serde_typedbytes2]
 (batchId=218)
org.apache.hadoop.hive.cli.TestContribCliDriver.testCliDriver[serde_typedbytes3]
 (batchId=218)
org.apache.hadoop.hive.cli.TestContribCliDriver.testCliDriver[serde_typedbytes4]
 (batchId=218)
org.apache.hadoop.hive.cli.TestContribCliDriver.testCliDriver[serde_typedbytes5]
 (batchId=218)
org.apache.hadoop.hive.cli.TestContribCliDriver.testCliDriver[serde_typedbytes] 
(batchId=218)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[transform_ppr1] 
(batchId=131)

[jira] [Commented] (HIVE-15217) Add watch mode to llap status tool

2016-11-16 Thread Prasanth Jayachandran (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15217?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671407#comment-15671407
 ] 

Prasanth Jayachandran commented on HIVE-15217:
--

i see what you mean. Makes sense. Will update the patch.

> Add watch mode to llap status tool
> --
>
> Key: HIVE-15217
> URL: https://issues.apache.org/jira/browse/HIVE-15217
> Project: Hive
>  Issue Type: Improvement
>  Components: llap
>Affects Versions: 2.2.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
>Priority: Minor
> Attachments: HIVE-15217.1.patch
>
>
> There is few seconds overhead for launching the llap status command. To avoid 
> we can add "watch" mode to llap status tool that refreshes the status after 
> configured interval. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15180) Extend JSONMessageFactory to store additional information about metadata objects on different table events

2016-11-16 Thread Vaibhav Gumashta (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15180?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vaibhav Gumashta updated HIVE-15180:

Attachment: HIVE-15180.3.patch

> Extend JSONMessageFactory to store additional information about metadata 
> objects on different table events
> --
>
> Key: HIVE-15180
> URL: https://issues.apache.org/jira/browse/HIVE-15180
> Project: Hive
>  Issue Type: Sub-task
>  Components: repl
>Reporter: Vaibhav Gumashta
>Assignee: Vaibhav Gumashta
> Attachments: HIVE-15180.1.patch, HIVE-15180.2.patch, 
> HIVE-15180.3.patch, HIVE-15180.3.patch
>
>
> We want the {{NOTIFICATION_LOG}} table to capture additional information 
> about the metadata objects when {{DbNotificationListener}} captures different 
> events for a table (create/drop/alter).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15220) WebHCat test driver not capturing end time of test accurately

2016-11-16 Thread Deepesh Khandelwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15220?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671380#comment-15671380
 ] 

Deepesh Khandelwal commented on HIVE-15220:
---

Thanks [~thejas] for the review and commit!

> WebHCat test driver not capturing end time of test accurately
> -
>
> Key: HIVE-15220
> URL: https://issues.apache.org/jira/browse/HIVE-15220
> Project: Hive
>  Issue Type: Bug
>  Components: Tests
>Reporter: Deepesh Khandelwal
>Assignee: Deepesh Khandelwal
>Priority: Trivial
> Fix For: 2.2.0
>
> Attachments: HIVE-15220.1.patch
>
>
> Webhcat e2e testsuite prints message while ending test run:
> {noformat}
> Ending test  at 1479264720
> {noformat}
> Currently it is not capturing the end time correctly.
> NO PRECOMMIT TESTS



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15217) Add watch mode to llap status tool

2016-11-16 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15217?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671355#comment-15671355
 ] 

Sergey Shelukhin commented on HIVE-15217:
-

Why print output at all in wait mode? We can print the last one if anything at 
all.

> Add watch mode to llap status tool
> --
>
> Key: HIVE-15217
> URL: https://issues.apache.org/jira/browse/HIVE-15217
> Project: Hive
>  Issue Type: Improvement
>  Components: llap
>Affects Versions: 2.2.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
>Priority: Minor
> Attachments: HIVE-15217.1.patch
>
>
> There is few seconds overhead for launching the llap status command. To avoid 
> we can add "watch" mode to llap status tool that refreshes the status after 
> configured interval. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15217) Add watch mode to llap status tool

2016-11-16 Thread Prasanth Jayachandran (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15217?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671346#comment-15671346
 ] 

Prasanth Jayachandran commented on HIVE-15217:
--

I will add another option that waits/watches until RUNNING (or any state) state.

The problem with keeping shorter refresh interval is with the way we are 
viewing the output. Currently the output is printed as json which is only 
machine readable. The output that gets printed is way too much even for 10s of 
nodes. IMO we should change the output format to something like tez job monitor 
and update in-place but that's a broader change and optionally output json. I 
think both formats will be useful. Refreshing too often and spitting json out 
is completely unreadable. Thoughts?

> Add watch mode to llap status tool
> --
>
> Key: HIVE-15217
> URL: https://issues.apache.org/jira/browse/HIVE-15217
> Project: Hive
>  Issue Type: Improvement
>  Components: llap
>Affects Versions: 2.2.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
>Priority: Minor
> Attachments: HIVE-15217.1.patch
>
>
> There is few seconds overhead for launching the llap status command. To avoid 
> we can add "watch" mode to llap status tool that refreshes the status after 
> configured interval. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15217) Add watch mode to llap status tool

2016-11-16 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15217?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671298#comment-15671298
 ] 

Sergey Shelukhin commented on HIVE-15217:
-

1) Is it possible to wait for specific status, e.g. "running"?
2) I think the wait interval can be much shorter <1sec. 

> Add watch mode to llap status tool
> --
>
> Key: HIVE-15217
> URL: https://issues.apache.org/jira/browse/HIVE-15217
> Project: Hive
>  Issue Type: Improvement
>  Components: llap
>Affects Versions: 2.2.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
>Priority: Minor
> Attachments: HIVE-15217.1.patch
>
>
> There is few seconds overhead for launching the llap status command. To avoid 
> we can add "watch" mode to llap status tool that refreshes the status after 
> configured interval. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15220) WebHCat test driver not capturing end time of test accurately

2016-11-16 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15220?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-15220:
-
   Resolution: Fixed
Fix Version/s: 2.2.0
   Status: Resolved  (was: Patch Available)

Committed to master.
Thanks for the patch [~deepesh]!

> WebHCat test driver not capturing end time of test accurately
> -
>
> Key: HIVE-15220
> URL: https://issues.apache.org/jira/browse/HIVE-15220
> Project: Hive
>  Issue Type: Bug
>  Components: Tests
>Reporter: Deepesh Khandelwal
>Assignee: Deepesh Khandelwal
>Priority: Trivial
> Fix For: 2.2.0
>
> Attachments: HIVE-15220.1.patch
>
>
> Webhcat e2e testsuite prints message while ending test run:
> {noformat}
> Ending test  at 1479264720
> {noformat}
> Currently it is not capturing the end time correctly.
> NO PRECOMMIT TESTS



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15199) INSERT INTO data on S3 is replacing the old rows with the new ones

2016-11-16 Thread Sahil Takiar (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15199?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671234#comment-15671234
 ] 

Sahil Takiar commented on HIVE-15199:
-

@spena a few comments:

* It may be better to take a hybrid of the list files approach + the exists 
approach; for blobstores like S3 listfiles is only eventually consistent; this 
means listfiles may not return all the files that are actually there. One way 
to get around this is to first do the listfiles, and then checks if the 
targetFilename exists or not. This has the advantage of the perf gains of using 
listfiles, but avoids the consistency problems
* I remember we discussed offline about concerns w.r.t multiple INSERT INTO 
queries running against the same table, but I just remembered that Hive Locking 
(https://cwiki.apache.org/confluence/display/Hive/Locking) should prevent that 
from ever happening, correct?
* It would be nice (although not necessary) if we changed the name of 
{{renameNonLocal}} to something more descriptive

> INSERT INTO data on S3 is replacing the old rows with the new ones
> --
>
> Key: HIVE-15199
> URL: https://issues.apache.org/jira/browse/HIVE-15199
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Sergio Peña
>Assignee: Sergio Peña
>Priority: Critical
> Attachments: HIVE-15199.1.patch
>
>
> Any INSERT INTO statement run on S3 tables and when the scratch directory is 
> saved on S3 is deleting old rows of the table.
> {noformat}
> hive> set hive.blobstore.use.blobstore.as.scratchdir=true;
> hive> create table t1 (id int, name string) location 's3a://spena-bucket/t1';
> hive> insert into table t1 values (1,'name1');
> hive> select * from t1;
> 1   name1
> hive> insert into table t1 values (2,'name2');
> hive> select * from t1;
> 2   name2
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HIVE-15199) INSERT INTO data on S3 is replacing the old rows with the new ones

2016-11-16 Thread Sahil Takiar (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15199?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671234#comment-15671234
 ] 

Sahil Takiar edited comment on HIVE-15199 at 11/16/16 6:34 PM:
---

[~spena] a few comments:

* It may be better to take a hybrid of the list files approach + the exists 
approach; for blobstores like S3 listfiles is only eventually consistent; this 
means listfiles may not return all the files that are actually there. One way 
to get around this is to first do the listfiles, and then checks if the 
targetFilename exists or not. This has the advantage of the perf gains of using 
listfiles, but avoids the consistency problems
* I remember we discussed offline about concerns w.r.t multiple INSERT INTO 
queries running against the same table, but I just remembered that Hive Locking 
(https://cwiki.apache.org/confluence/display/Hive/Locking) should prevent that 
from ever happening, correct?
* It would be nice (although not necessary) if we changed the name of 
{{renameNonLocal}} to something more descriptive


was (Author: stakiar):
@spena a few comments:

* It may be better to take a hybrid of the list files approach + the exists 
approach; for blobstores like S3 listfiles is only eventually consistent; this 
means listfiles may not return all the files that are actually there. One way 
to get around this is to first do the listfiles, and then checks if the 
targetFilename exists or not. This has the advantage of the perf gains of using 
listfiles, but avoids the consistency problems
* I remember we discussed offline about concerns w.r.t multiple INSERT INTO 
queries running against the same table, but I just remembered that Hive Locking 
(https://cwiki.apache.org/confluence/display/Hive/Locking) should prevent that 
from ever happening, correct?
* It would be nice (although not necessary) if we changed the name of 
{{renameNonLocal}} to something more descriptive

> INSERT INTO data on S3 is replacing the old rows with the new ones
> --
>
> Key: HIVE-15199
> URL: https://issues.apache.org/jira/browse/HIVE-15199
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Sergio Peña
>Assignee: Sergio Peña
>Priority: Critical
> Attachments: HIVE-15199.1.patch
>
>
> Any INSERT INTO statement run on S3 tables and when the scratch directory is 
> saved on S3 is deleting old rows of the table.
> {noformat}
> hive> set hive.blobstore.use.blobstore.as.scratchdir=true;
> hive> create table t1 (id int, name string) location 's3a://spena-bucket/t1';
> hive> insert into table t1 values (1,'name1');
> hive> select * from t1;
> 1   name1
> hive> insert into table t1 values (2,'name2');
> hive> select * from t1;
> 2   name2
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15200) Support setOp in subQuery with parentheses

2016-11-16 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15200?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-15200:
---
Status: Patch Available  (was: Open)

> Support setOp in subQuery with parentheses
> --
>
> Key: HIVE-15200
> URL: https://issues.apache.org/jira/browse/HIVE-15200
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Pengcheng Xiong
>Assignee: Pengcheng Xiong
> Attachments: HIVE-15200.01.patch
>
>
> {code}
> explain select key from ((select key from src) union (select key from 
> src))subq;
> {code}
> will throw
> {code}
> FAILED: ParseException line 1:47 cannot recognize input near 'union' '(' 
> 'select' in subquery source
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15200) Support setOp in subQuery with parentheses

2016-11-16 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15200?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-15200:
---
Attachment: (was: HIVE-15200.01.patch)

> Support setOp in subQuery with parentheses
> --
>
> Key: HIVE-15200
> URL: https://issues.apache.org/jira/browse/HIVE-15200
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Pengcheng Xiong
>Assignee: Pengcheng Xiong
> Attachments: HIVE-15200.01.patch
>
>
> {code}
> explain select key from ((select key from src) union (select key from 
> src))subq;
> {code}
> will throw
> {code}
> FAILED: ParseException line 1:47 cannot recognize input near 'union' '(' 
> 'select' in subquery source
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15200) Support setOp in subQuery with parentheses

2016-11-16 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15200?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-15200:
---
Attachment: HIVE-15200.01.patch

> Support setOp in subQuery with parentheses
> --
>
> Key: HIVE-15200
> URL: https://issues.apache.org/jira/browse/HIVE-15200
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Pengcheng Xiong
>Assignee: Pengcheng Xiong
> Attachments: HIVE-15200.01.patch
>
>
> {code}
> explain select key from ((select key from src) union (select key from 
> src))subq;
> {code}
> will throw
> {code}
> FAILED: ParseException line 1:47 cannot recognize input near 'union' '(' 
> 'select' in subquery source
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15200) Support setOp in subQuery with parentheses

2016-11-16 Thread Pengcheng Xiong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15200?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Pengcheng Xiong updated HIVE-15200:
---
Status: Open  (was: Patch Available)

> Support setOp in subQuery with parentheses
> --
>
> Key: HIVE-15200
> URL: https://issues.apache.org/jira/browse/HIVE-15200
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Pengcheng Xiong
>Assignee: Pengcheng Xiong
> Attachments: HIVE-15200.01.patch
>
>
> {code}
> explain select key from ((select key from src) union (select key from 
> src))subq;
> {code}
> will throw
> {code}
> FAILED: ParseException line 1:47 cannot recognize input near 'union' '(' 
> 'select' in subquery source
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-9482) Hive parquet timestamp compatibility

2016-11-16 Thread Vitalii Diravka (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9482?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671191#comment-15671191
 ] 

Vitalii Diravka commented on HIVE-9482:
---

Why this hive.parquet.timestamp.skip.conversion option is enabled by default? 
Since according [parquet 
spec|https://github.com/Parquet/parquet-format/blob/master/LogicalTypes.md#timestamp_millis],
 parquet files don't keep local timezone. And we cann't distinguish from file 
what was the value of that option while parquet file was generating.

> Hive parquet timestamp compatibility
> 
>
> Key: HIVE-9482
> URL: https://issues.apache.org/jira/browse/HIVE-9482
> Project: Hive
>  Issue Type: Bug
>  Components: File Formats
>Affects Versions: 0.15.0
>Reporter: Szehon Ho
>Assignee: Szehon Ho
> Fix For: 1.2.0
>
> Attachments: HIVE-9482.2.patch, HIVE-9482.patch, HIVE-9482.patch, 
> parquet_external_time.parq
>
>
> In current Hive implementation, timestamps are stored in UTC (converted from 
> current timezone), based on original parquet timestamp spec.
> However, we find this is not compatibility with other tools, and after some 
> investigation it is not the way of the other file formats, or even some 
> databases (Hive Timestamp is more equivalent of 'timestamp without timezone' 
> datatype).
> This is the first part of the fix, which will restore compatibility with 
> parquet-timestamp files generated by external tools by skipping conversion on 
> reading.
> Later fix will change the write path to not convert, and stop the 
> read-conversion even for files written by Hive itself.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15114) Remove extra MoveTask operators

2016-11-16 Thread JIRA

[ 
https://issues.apache.org/jira/browse/HIVE-15114?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671167#comment-15671167
 ] 

Sergio Peña commented on HIVE-15114:


[~stakiar] The conditional task created on the GenMapRedUtils class also 
creates tasks for Spark or Tez. The MoveTask should be linked to any task 
regardless the execution engine. 

The .q tests added to the TestBlobstorecliDriver can be executed with any 
engine as well, so nothing to change. However, I did not verify those tests 
were running correctly on them. I will run them and verify they work.

> Remove extra MoveTask operators
> ---
>
> Key: HIVE-15114
> URL: https://issues.apache.org/jira/browse/HIVE-15114
> Project: Hive
>  Issue Type: Sub-task
>  Components: Hive
>Affects Versions: 2.1.0
>Reporter: Sahil Takiar
>Assignee: Sergio Peña
> Attachments: HIVE-15114.WIP.1.patch, HIVE-15114.WIP.2.patch
>
>
> When running simple insert queries (e.g. {{INSERT INTO TABLE ... VALUES 
> ...}}) there an extraneous {{MoveTask}s is created.
> This is problematic when the scratch directory is on S3 since renames require 
> copying the entire dataset.
> For simple queries (like the one above), there are two MoveTasks. The first 
> one moves the output data from one file in the scratch directory to another 
> file in the scratch directory. The second MoveTask moves the data from the 
> scratch directory to its final table location.
> The first MoveTask should not be necessary. The goal of this JIRA it to 
> remove it. This should help improve performance when running on S3.
> It seems that the first Move might be caused by a dependency resolution 
> problem in the optimizer, where a dependent task doesn't get properly removed 
> when the task it depends on is filtered by a condition resolver.
> A dummy {{MoveTask}} is added in the 
> {{GenMapRedUtils.createMRWorkForMergingFiles}} method. This method creates a 
> conditional task which launches a job to merge tasks at the end of the file. 
> At the end of the conditional job there is a MoveTask.
> Even though Hive decides that the conditional merge job is no needed, it 
> seems the MoveTask is still added to the plan.
> Seems this extra {{MoveTask}} may have been added intentionally. Not sure why 
> yet. The {{ConditionalResolverMergeFiles}} says that one of three tasks will 
> be returned: move task only, merge task only, merge task followed by a move 
> task.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15199) INSERT INTO data on S3 is replacing the old rows with the new ones

2016-11-16 Thread JIRA

[ 
https://issues.apache.org/jira/browse/HIVE-15199?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671136#comment-15671136
 ] 

Sergio Peña commented on HIVE-15199:


Guys, [~ashutoshc], could you help me review the patch?

What it does is to use the alternative condition {{mvFile}} when the 
destination filesystem is a blobstore. And, because this {{mvFile}} was calling 
{{destFs.exists()}} for every file on S3, then I changed it to get a list of 
files, and check whether the {{destf}} exists or not on that list.

> INSERT INTO data on S3 is replacing the old rows with the new ones
> --
>
> Key: HIVE-15199
> URL: https://issues.apache.org/jira/browse/HIVE-15199
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Sergio Peña
>Assignee: Sergio Peña
>Priority: Critical
> Attachments: HIVE-15199.1.patch
>
>
> Any INSERT INTO statement run on S3 tables and when the scratch directory is 
> saved on S3 is deleting old rows of the table.
> {noformat}
> hive> set hive.blobstore.use.blobstore.as.scratchdir=true;
> hive> create table t1 (id int, name string) location 's3a://spena-bucket/t1';
> hive> insert into table t1 values (1,'name1');
> hive> select * from t1;
> 1   name1
> hive> insert into table t1 values (2,'name2');
> hive> select * from t1;
> 2   name2
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15202) Concurrent compactions for the same partition may generate malformed folder structure

2016-11-16 Thread Eugene Koifman (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15202?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671134#comment-15671134
 ] 

Eugene Koifman commented on HIVE-15202:
---

When compactions are triggered by the system automatically they have logic to 
avoid concurrent invocations but not when triggered by the user.  It should and 
will be fixed (almost certainly before 2.2 is released)

> Concurrent compactions for the same partition may generate malformed folder 
> structure
> -
>
> Key: HIVE-15202
> URL: https://issues.apache.org/jira/browse/HIVE-15202
> Project: Hive
>  Issue Type: Bug
>Reporter: Rui Li
>
> If two compactions run concurrently on a single partition, it may generate 
> folder structure like this: (nested base dir)
> {noformat}
> drwxr-xr-x   - root supergroup  0 2016-11-14 22:23 
> /user/hive/warehouse/test/z=1/base_007/base_007
> -rw-r--r--   3 root supergroup201 2016-11-14 21:46 
> /user/hive/warehouse/test/z=1/base_007/bucket_0
> -rw-r--r--   3 root supergroup611 2016-11-14 21:46 
> /user/hive/warehouse/test/z=1/base_007/bucket_1
> -rw-r--r--   3 root supergroup614 2016-11-14 21:46 
> /user/hive/warehouse/test/z=1/base_007/bucket_2
> -rw-r--r--   3 root supergroup621 2016-11-14 21:46 
> /user/hive/warehouse/test/z=1/base_007/bucket_3
> -rw-r--r--   3 root supergroup621 2016-11-14 21:46 
> /user/hive/warehouse/test/z=1/base_007/bucket_4
> -rw-r--r--   3 root supergroup201 2016-11-14 21:46 
> /user/hive/warehouse/test/z=1/base_007/bucket_5
> -rw-r--r--   3 root supergroup201 2016-11-14 21:46 
> /user/hive/warehouse/test/z=1/base_007/bucket_6
> -rw-r--r--   3 root supergroup201 2016-11-14 21:46 
> /user/hive/warehouse/test/z=1/base_007/bucket_7
> -rw-r--r--   3 root supergroup201 2016-11-14 21:46 
> /user/hive/warehouse/test/z=1/base_007/bucket_8
> -rw-r--r--   3 root supergroup201 2016-11-14 21:46 
> /user/hive/warehouse/test/z=1/base_007/bucket_9
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15144) JSON.org license is now CatX

2016-11-16 Thread Alan Gates (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15144?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671129#comment-15671129
 ] 

Alan Gates commented on HIVE-15144:
---

We may have a bit more time in which to get this done.  Based on a discussion 
on legal-discuss we may have a 6 month grace period.  This will allow us to 
keep doing releases while we do the work to get this switched out and tested.

> JSON.org license is now CatX
> 
>
> Key: HIVE-15144
> URL: https://issues.apache.org/jira/browse/HIVE-15144
> Project: Hive
>  Issue Type: Bug
>Reporter: Robert Kanter
>Priority: Blocker
> Fix For: 2.2.0
>
>
> per [update resolved legal|http://www.apache.org/legal/resolved.html#json]:
> {quote}
> CAN APACHE PRODUCTS INCLUDE WORKS LICENSED UNDER THE JSON LICENSE?
> No. As of 2016-11-03 this has been moved to the 'Category X' license list. 
> Prior to this, use of the JSON Java library was allowed. See Debian's page 
> for a list of alternatives.
> {quote}
> I'm not sure when this dependency was first introduced, but it looks like 
> it's currently used in a few places:
> https://github.com/apache/hive/search?p=1=%22org.json%22=%E2%9C%93



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15224) replace org.json usage in branch-1 with as minor changes as possible

2016-11-16 Thread Alan Gates (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15224?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671124#comment-15671124
 ] 

Alan Gates commented on HIVE-15224:
---

Based on a thread going on in legal discuss we may have a bit before we have to 
do this.  I'm working on getting a 6 month grace period so we can still do 
maintenance releases without waiting until we've tested and validated a 
replacement JSON library.

> replace org.json usage in branch-1 with as minor changes as possible
> 
>
> Key: HIVE-15224
> URL: https://issues.apache.org/jira/browse/HIVE-15224
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Zoltan Haindrich
>Assignee: Daniel Voros
> Fix For: 2.2.0
>
>
> branch-1 / master have diverged in many ways - StatsCollector have changed; 
> EximUtil supports new replication
> ...so backporting any changes from master would be hard.
> maybe we should use some drop-in replacement like the android one.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15199) INSERT INTO data on S3 is replacing the old rows with the new ones

2016-11-16 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/HIVE-15199?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergio Peña updated HIVE-15199:
---
Status: Patch Available  (was: Open)

> INSERT INTO data on S3 is replacing the old rows with the new ones
> --
>
> Key: HIVE-15199
> URL: https://issues.apache.org/jira/browse/HIVE-15199
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Sergio Peña
>Assignee: Sergio Peña
>Priority: Critical
> Attachments: HIVE-15199.1.patch
>
>
> Any INSERT INTO statement run on S3 tables and when the scratch directory is 
> saved on S3 is deleting old rows of the table.
> {noformat}
> hive> set hive.blobstore.use.blobstore.as.scratchdir=true;
> hive> create table t1 (id int, name string) location 's3a://spena-bucket/t1';
> hive> insert into table t1 values (1,'name1');
> hive> select * from t1;
> 1   name1
> hive> insert into table t1 values (2,'name2');
> hive> select * from t1;
> 2   name2
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15199) INSERT INTO data on S3 is replacing the old rows with the new ones

2016-11-16 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/HIVE-15199?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergio Peña updated HIVE-15199:
---
Attachment: HIVE-15199.1.patch

> INSERT INTO data on S3 is replacing the old rows with the new ones
> --
>
> Key: HIVE-15199
> URL: https://issues.apache.org/jira/browse/HIVE-15199
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Sergio Peña
>Assignee: Sergio Peña
>Priority: Critical
> Attachments: HIVE-15199.1.patch
>
>
> Any INSERT INTO statement run on S3 tables and when the scratch directory is 
> saved on S3 is deleting old rows of the table.
> {noformat}
> hive> set hive.blobstore.use.blobstore.as.scratchdir=true;
> hive> create table t1 (id int, name string) location 's3a://spena-bucket/t1';
> hive> insert into table t1 values (1,'name1');
> hive> select * from t1;
> 1   name1
> hive> insert into table t1 values (2,'name2');
> hive> select * from t1;
> 2   name2
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15220) WebHCat test driver not capturing end time of test accurately

2016-11-16 Thread Thejas M Nair (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15220?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671091#comment-15671091
 ] 

Thejas M Nair commented on HIVE-15220:
--

+1

> WebHCat test driver not capturing end time of test accurately
> -
>
> Key: HIVE-15220
> URL: https://issues.apache.org/jira/browse/HIVE-15220
> Project: Hive
>  Issue Type: Bug
>  Components: Tests
>Reporter: Deepesh Khandelwal
>Assignee: Deepesh Khandelwal
>Priority: Trivial
> Attachments: HIVE-15220.1.patch
>
>
> Webhcat e2e testsuite prints message while ending test run:
> {noformat}
> Ending test  at 1479264720
> {noformat}
> Currently it is not capturing the end time correctly.
> NO PRECOMMIT TESTS



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15057) Support other types of operators (other than SELECT)

2016-11-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15057?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15671073#comment-15671073
 ] 

Hive QA commented on HIVE-15057:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12839188/HIVE-15057.wip.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 4 failed/errored test(s), 10696 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[orc_ppd_schema_evol_3a]
 (batchId=133)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[transform_ppr2] 
(batchId=133)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[join_acid_non_acid]
 (batchId=150)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[union_fast_stats]
 (batchId=145)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/2152/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/2152/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-2152/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 4 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12839188 - PreCommit-HIVE-Build

> Support other types of operators (other than SELECT)
> 
>
> Key: HIVE-15057
> URL: https://issues.apache.org/jira/browse/HIVE-15057
> Project: Hive
>  Issue Type: Sub-task
>  Components: Logical Optimizer, Physical Optimizer
>Reporter: Chao Sun
>Assignee: Chao Sun
> Attachments: HIVE-15057.wip.patch
>
>
> Currently only SELECT operators are supported for nested column pruning. We 
> should add support for other types of operators so the optimization can work 
> for complex queries.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15208) Query string should be HTML encoded for Web UI

2016-11-16 Thread Jimmy Xiang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15208?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jimmy Xiang updated HIVE-15208:
---
   Resolution: Fixed
 Hadoop Flags: Reviewed
Fix Version/s: 2.2.0
   Status: Resolved  (was: Patch Available)

Pushed to the master branch. Thanks Xuefu for the review.

> Query string should be HTML encoded for Web UI
> --
>
> Key: HIVE-15208
> URL: https://issues.apache.org/jira/browse/HIVE-15208
> Project: Hive
>  Issue Type: Bug
>  Components: Web UI
>Reporter: Jimmy Xiang
>Assignee: Jimmy Xiang
>Priority: Minor
> Fix For: 2.2.0
>
> Attachments: HIVE-15208.1.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-14582) Add trunc(numeric) udf

2016-11-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-14582?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15670933#comment-15670933
 ] 

Hive QA commented on HIVE-14582:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12839177/HIVE-14582.2.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 10680 tests 
executed
*Failed tests:*
{noformat}
TestSparkCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=120)

[groupby3_map.q,union11.q,union26.q,mapreduce1.q,mapjoin_addjar.q,bucket_map_join_spark1.q,udf_example_add.q,multi_insert_with_join.q,sample7.q,auto_join_nulls.q,ppd_outer_join4.q,load_dyn_part8.q,sample6.q,bucket_map_join_1.q,auto_sortmerge_join_9.q]
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[transform_ppr2] 
(batchId=133)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[join_acid_non_acid]
 (batchId=150)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[union_fast_stats]
 (batchId=145)
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver[explainanalyze_2] 
(batchId=91)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/2151/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/2151/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-2151/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 5 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12839177 - PreCommit-HIVE-Build

> Add trunc(numeric) udf
> --
>
> Key: HIVE-14582
> URL: https://issues.apache.org/jira/browse/HIVE-14582
> Project: Hive
>  Issue Type: Sub-task
>  Components: SQL
>Reporter: Ashutosh Chauhan
>Assignee: Chinna Rao Lalam
> Attachments: HIVE-14582.1.patch, HIVE-14582.2.patch, HIVE-14582.patch
>
>
> https://docs.oracle.com/cd/B19306_01/server.102/b14200/functions200.htm



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15057) Support other types of operators (other than SELECT)

2016-11-16 Thread Chao Sun (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15057?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chao Sun updated HIVE-15057:

Attachment: HIVE-15057.wip.patch

> Support other types of operators (other than SELECT)
> 
>
> Key: HIVE-15057
> URL: https://issues.apache.org/jira/browse/HIVE-15057
> Project: Hive
>  Issue Type: Sub-task
>  Components: Logical Optimizer, Physical Optimizer
>Reporter: Chao Sun
>Assignee: Chao Sun
> Attachments: HIVE-15057.wip.patch
>
>
> Currently only SELECT operators are supported for nested column pruning. We 
> should add support for other types of operators so the optimization can work 
> for complex queries.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15057) Support other types of operators (other than SELECT)

2016-11-16 Thread Chao Sun (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15057?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chao Sun updated HIVE-15057:

Attachment: (was: HIVE-15057.wip.patch)

> Support other types of operators (other than SELECT)
> 
>
> Key: HIVE-15057
> URL: https://issues.apache.org/jira/browse/HIVE-15057
> Project: Hive
>  Issue Type: Sub-task
>  Components: Logical Optimizer, Physical Optimizer
>Reporter: Chao Sun
>Assignee: Chao Sun
> Attachments: HIVE-15057.wip.patch
>
>
> Currently only SELECT operators are supported for nested column pruning. We 
> should add support for other types of operators so the optimization can work 
> for complex queries.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (HIVE-15224) replace org.json usage in branch-1 with as minor changes as possible

2016-11-16 Thread Daniel Voros (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15224?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daniel Voros reassigned HIVE-15224:
---

Assignee: Daniel Voros

> replace org.json usage in branch-1 with as minor changes as possible
> 
>
> Key: HIVE-15224
> URL: https://issues.apache.org/jira/browse/HIVE-15224
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Zoltan Haindrich
>Assignee: Daniel Voros
> Fix For: 2.2.0
>
>
> branch-1 / master have diverged in many ways - StatsCollector have changed; 
> EximUtil supports new replication
> ...so backporting any changes from master would be hard.
> maybe we should use some drop-in replacement like the android one.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-14582) Add trunc(numeric) udf

2016-11-16 Thread Chinna Rao Lalam (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-14582?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chinna Rao Lalam updated HIVE-14582:

Status: Patch Available  (was: Open)

> Add trunc(numeric) udf
> --
>
> Key: HIVE-14582
> URL: https://issues.apache.org/jira/browse/HIVE-14582
> Project: Hive
>  Issue Type: Sub-task
>  Components: SQL
>Reporter: Ashutosh Chauhan
>Assignee: Chinna Rao Lalam
> Attachments: HIVE-14582.1.patch, HIVE-14582.2.patch, HIVE-14582.patch
>
>
> https://docs.oracle.com/cd/B19306_01/server.102/b14200/functions200.htm



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-14582) Add trunc(numeric) udf

2016-11-16 Thread Chinna Rao Lalam (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-14582?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chinna Rao Lalam updated HIVE-14582:

Attachment: HIVE-14582.2.patch

{quote}
udf_trunc.q, 
udf_trunc_number.q 
{quote}
tests failed because of function description change, it was not updated in test 
files. Remaining failures are not related to this patch.

Patch.2 Updated with new function description in test files.

> Add trunc(numeric) udf
> --
>
> Key: HIVE-14582
> URL: https://issues.apache.org/jira/browse/HIVE-14582
> Project: Hive
>  Issue Type: Sub-task
>  Components: SQL
>Reporter: Ashutosh Chauhan
>Assignee: Chinna Rao Lalam
> Attachments: HIVE-14582.1.patch, HIVE-14582.2.patch, HIVE-14582.patch
>
>
> https://docs.oracle.com/cd/B19306_01/server.102/b14200/functions200.htm



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-14582) Add trunc(numeric) udf

2016-11-16 Thread Chinna Rao Lalam (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-14582?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chinna Rao Lalam updated HIVE-14582:

Status: Open  (was: Patch Available)

> Add trunc(numeric) udf
> --
>
> Key: HIVE-14582
> URL: https://issues.apache.org/jira/browse/HIVE-14582
> Project: Hive
>  Issue Type: Sub-task
>  Components: SQL
>Reporter: Ashutosh Chauhan
>Assignee: Chinna Rao Lalam
> Attachments: HIVE-14582.1.patch, HIVE-14582.patch
>
>
> https://docs.oracle.com/cd/B19306_01/server.102/b14200/functions200.htm



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-14840) MSCK not adding the missing partitions to Hive Metastore when the partition names are not in lowercase

2016-11-16 Thread Sushil Kumar S (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-14840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sushil Kumar S updated HIVE-14840:
--
Assignee: (was: Sushil Kumar S)

> MSCK not adding the missing partitions to Hive Metastore when the partition 
> names are not in lowercase
> --
>
> Key: HIVE-14840
> URL: https://issues.apache.org/jira/browse/HIVE-14840
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 1.2.1
>Reporter: Sushil Kumar S
>Priority: Minor
>  Labels: hive
>
> Hi,
>   There's is a bug while running MSCK REPAIR TABLE EXTERNAL_TABLE_NAME on 
> Hive 1.2.1, all the partition that are not present in the metastore are being 
> listed but not added if the partition names are not in lowercase, in other 
> words if an external path has a camel case based name and value i.e 
> s3n://some_external_path/myPartition=01 it just gets listed as partition not 
> found in metastore but doesn’t add it.
> However, am not able to run ALTER TABLE MY_EXTERNAL_TABLE RECOVER PARTITIONS; 
> on hive 1.2 and based on the source code from hive-exec am able to see under 
> org/apache/hadoop/hive/ql/parse/HiveParser.g:1001:1 that there's no token 
> matching in the grammar for RECOVER PARTITIONS.
> Example:
> - When external path = s3n://some_external_path/myPartition=01
>hive> MSCK REPAIR TABLE my_external_table;
>Partitions not in metastore: my_external_table:mypartition=01
>Time taken: 1.729 seconds, Fetched: 2 row(s)
> hive> show partitions foster.ola_raven_raven_users_raw;
> OK
> Time taken: 0.901 seconds, Fetched: 0 row(s)
> - When external path = s3n://some_external_path/mypartition=01
> hive> MSCK REPAIR TABLE my_external_table;
> Partitions not in metastore: my_external_table:mypartition=01
> Repair: Added partition to metastore my_external_table:mypartition=01
> Time taken: 1.729 seconds, Fetched: 2 row(s)
>  hive> show partitions my_external_table;
>  OK
>  mypartition=01
>  Time taken: 1.101 seconds, Fetched: 1 row(s)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Reopened] (HIVE-13539) HiveHFileOutputFormat searching the wrong directory for HFiles

2016-11-16 Thread Chaoyu Tang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13539?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chaoyu Tang reopened HIVE-13539:

  Assignee: Tim Robertson  (was: Matt McCline)

> HiveHFileOutputFormat searching the wrong directory for HFiles
> --
>
> Key: HIVE-13539
> URL: https://issues.apache.org/jira/browse/HIVE-13539
> Project: Hive
>  Issue Type: Bug
>  Components: HBase Handler
>Affects Versions: 1.1.0
> Environment: Built into CDH 5.4.7
>Reporter: Tim Robertson
>Assignee: Tim Robertson
>Priority: Blocker
> Attachments: hive_hfile_output_format.q, 
> hive_hfile_output_format.q.out
>
>
> When creating HFiles for a bulkload in HBase I believe it is looking in the 
> wrong directory to find the HFiles, resulting in the following exception:
> {code}
> Error: java.lang.RuntimeException: Hive Runtime Error while closing 
> operators: java.io.IOException: Multiple family directories found in 
> hdfs://c1n1.gbif.org:8020/user/hive/warehouse/tim.db/coords_hbase/_temporary/2/_temporary
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.close(ExecReducer.java:295)
>   at 
> org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:453)
>   at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:392)
>   at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:163)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:415)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1671)
>   at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
> java.io.IOException: Multiple family directories found in 
> hdfs://c1n1.gbif.org:8020/user/hive/warehouse/tim.db/coords_hbase/_temporary/2/_temporary
>   at 
> org.apache.hadoop.hive.ql.exec.FileSinkOperator$FSPaths.closeWriters(FileSinkOperator.java:188)
>   at 
> org.apache.hadoop.hive.ql.exec.FileSinkOperator.closeOp(FileSinkOperator.java:958)
>   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:598)
>   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:610)
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.close(ExecReducer.java:287)
>   ... 7 more
> Caused by: java.io.IOException: Multiple family directories found in 
> hdfs://c1n1.gbif.org:8020/user/hive/warehouse/tim.db/coords_hbase/_temporary/2/_temporary
>   at 
> org.apache.hadoop.hive.hbase.HiveHFileOutputFormat$1.close(HiveHFileOutputFormat.java:158)
>   at 
> org.apache.hadoop.hive.ql.exec.FileSinkOperator$FSPaths.closeWriters(FileSinkOperator.java:185)
>   ... 11 more
> {code}
> The issue is that is looks for the HFiles in 
> {{hdfs://c1n1.gbif.org:8020/user/hive/warehouse/tim.db/coords_hbase/_temporary/2/_temporary}}
>  when I believe it should be looking in the task attempt subfolder, such as 
> {{hdfs://c1n1.gbif.org:8020/user/hive/warehouse/tim.db/coords_hbase/_temporary/2/_temporary/attempt_1461004169450_0002_r_00_1000}}.
> This can be reproduced in any HFile creation such as:
> {code:sql}
> CREATE TABLE coords_hbase(id INT, x DOUBLE, y DOUBLE)
> STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
> WITH SERDEPROPERTIES (
>   'hbase.columns.mapping' = ':key,o:x,o:y',
>   'hbase.table.default.storage.type' = 'binary');
> SET hfile.family.path=/tmp/coords_hfiles/o; 
> SET hive.hbase.generatehfiles=true;
> INSERT OVERWRITE TABLE coords_hbase 
> SELECT id, decimalLongitude, decimalLatitude
> FROM source
> CLUSTER BY id; 
> {code}
> Any advice greatly appreciated



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13539) HiveHFileOutputFormat searching the wrong directory for HFiles

2016-11-16 Thread Chaoyu Tang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13539?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15670315#comment-15670315
 ] 

Chaoyu Tang commented on HIVE-13539:


The issue could only be reproduced when multiple reducers are used. 
[~timrobertson100], would you like to provide a new qtest?

> HiveHFileOutputFormat searching the wrong directory for HFiles
> --
>
> Key: HIVE-13539
> URL: https://issues.apache.org/jira/browse/HIVE-13539
> Project: Hive
>  Issue Type: Bug
>  Components: HBase Handler
>Affects Versions: 1.1.0
> Environment: Built into CDH 5.4.7
>Reporter: Tim Robertson
>Assignee: Matt McCline
>Priority: Blocker
> Attachments: hive_hfile_output_format.q, 
> hive_hfile_output_format.q.out
>
>
> When creating HFiles for a bulkload in HBase I believe it is looking in the 
> wrong directory to find the HFiles, resulting in the following exception:
> {code}
> Error: java.lang.RuntimeException: Hive Runtime Error while closing 
> operators: java.io.IOException: Multiple family directories found in 
> hdfs://c1n1.gbif.org:8020/user/hive/warehouse/tim.db/coords_hbase/_temporary/2/_temporary
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.close(ExecReducer.java:295)
>   at 
> org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:453)
>   at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:392)
>   at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:163)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:415)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1671)
>   at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
> java.io.IOException: Multiple family directories found in 
> hdfs://c1n1.gbif.org:8020/user/hive/warehouse/tim.db/coords_hbase/_temporary/2/_temporary
>   at 
> org.apache.hadoop.hive.ql.exec.FileSinkOperator$FSPaths.closeWriters(FileSinkOperator.java:188)
>   at 
> org.apache.hadoop.hive.ql.exec.FileSinkOperator.closeOp(FileSinkOperator.java:958)
>   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:598)
>   at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:610)
>   at 
> org.apache.hadoop.hive.ql.exec.mr.ExecReducer.close(ExecReducer.java:287)
>   ... 7 more
> Caused by: java.io.IOException: Multiple family directories found in 
> hdfs://c1n1.gbif.org:8020/user/hive/warehouse/tim.db/coords_hbase/_temporary/2/_temporary
>   at 
> org.apache.hadoop.hive.hbase.HiveHFileOutputFormat$1.close(HiveHFileOutputFormat.java:158)
>   at 
> org.apache.hadoop.hive.ql.exec.FileSinkOperator$FSPaths.closeWriters(FileSinkOperator.java:185)
>   ... 11 more
> {code}
> The issue is that is looks for the HFiles in 
> {{hdfs://c1n1.gbif.org:8020/user/hive/warehouse/tim.db/coords_hbase/_temporary/2/_temporary}}
>  when I believe it should be looking in the task attempt subfolder, such as 
> {{hdfs://c1n1.gbif.org:8020/user/hive/warehouse/tim.db/coords_hbase/_temporary/2/_temporary/attempt_1461004169450_0002_r_00_1000}}.
> This can be reproduced in any HFile creation such as:
> {code:sql}
> CREATE TABLE coords_hbase(id INT, x DOUBLE, y DOUBLE)
> STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
> WITH SERDEPROPERTIES (
>   'hbase.columns.mapping' = ':key,o:x,o:y',
>   'hbase.table.default.storage.type' = 'binary');
> SET hfile.family.path=/tmp/coords_hfiles/o; 
> SET hive.hbase.generatehfiles=true;
> INSERT OVERWRITE TABLE coords_hbase 
> SELECT id, decimalLongitude, decimalLatitude
> FROM source
> CLUSTER BY id; 
> {code}
> Any advice greatly appreciated



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13557) Make interval keyword optional while specifying DAY in interval arithmetic

2016-11-16 Thread Zoltan Haindrich (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13557?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15670308#comment-15670308
 ] 

Zoltan Haindrich commented on HIVE-13557:
-

failures are unrelated; HIVE-15201 HIVE-15116 HIVE-15115

[~pxiong] can you please take a look at the new changes?

> Make interval keyword optional while specifying DAY in interval arithmetic
> --
>
> Key: HIVE-13557
> URL: https://issues.apache.org/jira/browse/HIVE-13557
> Project: Hive
>  Issue Type: Sub-task
>  Components: Types
>Reporter: Ashutosh Chauhan
>Assignee: Zoltan Haindrich
> Attachments: HIVE-13557.1.patch, HIVE-13557.1.patch, 
> HIVE-13557.1.patch, HIVE-13557.2.patch, HIVE-13557.3.patch, HIVE-13557.3.patch
>
>
> Currently we support expressions like: {code}
> WHERE SOLD_DATE BETWEEN ((DATE('2000-01-31'))  - INTERVAL '30' DAY) AND 
> DATE('2000-01-31')
> {code}
> We should support:
> {code}
> WHERE SOLD_DATE BETWEEN ((DATE('2000-01-31')) + (-30) DAY) AND 
> DATE('2000-01-31')
> {code}
>   



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13557) Make interval keyword optional while specifying DAY in interval arithmetic

2016-11-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13557?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15670270#comment-15670270
 ] 

Hive QA commented on HIVE-13557:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12839158/HIVE-13557.3.patch

{color:green}SUCCESS:{color} +1 due to 3 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 3 failed/errored test(s), 10710 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[transform_ppr2] 
(batchId=133)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[join_acid_non_acid]
 (batchId=150)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[union_fast_stats]
 (batchId=145)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/2150/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/2150/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-2150/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 3 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12839158 - PreCommit-HIVE-Build

> Make interval keyword optional while specifying DAY in interval arithmetic
> --
>
> Key: HIVE-13557
> URL: https://issues.apache.org/jira/browse/HIVE-13557
> Project: Hive
>  Issue Type: Sub-task
>  Components: Types
>Reporter: Ashutosh Chauhan
>Assignee: Zoltan Haindrich
> Attachments: HIVE-13557.1.patch, HIVE-13557.1.patch, 
> HIVE-13557.1.patch, HIVE-13557.2.patch, HIVE-13557.3.patch, HIVE-13557.3.patch
>
>
> Currently we support expressions like: {code}
> WHERE SOLD_DATE BETWEEN ((DATE('2000-01-31'))  - INTERVAL '30' DAY) AND 
> DATE('2000-01-31')
> {code}
> We should support:
> {code}
> WHERE SOLD_DATE BETWEEN ((DATE('2000-01-31')) + (-30) DAY) AND 
> DATE('2000-01-31')
> {code}
>   



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13557) Make interval keyword optional while specifying DAY in interval arithmetic

2016-11-16 Thread Zoltan Haindrich (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13557?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltan Haindrich updated HIVE-13557:

Attachment: HIVE-13557.3.patch

> Make interval keyword optional while specifying DAY in interval arithmetic
> --
>
> Key: HIVE-13557
> URL: https://issues.apache.org/jira/browse/HIVE-13557
> Project: Hive
>  Issue Type: Sub-task
>  Components: Types
>Reporter: Ashutosh Chauhan
>Assignee: Zoltan Haindrich
> Attachments: HIVE-13557.1.patch, HIVE-13557.1.patch, 
> HIVE-13557.1.patch, HIVE-13557.2.patch, HIVE-13557.3.patch, HIVE-13557.3.patch
>
>
> Currently we support expressions like: {code}
> WHERE SOLD_DATE BETWEEN ((DATE('2000-01-31'))  - INTERVAL '30' DAY) AND 
> DATE('2000-01-31')
> {code}
> We should support:
> {code}
> WHERE SOLD_DATE BETWEEN ((DATE('2000-01-31')) + (-30) DAY) AND 
> DATE('2000-01-31')
> {code}
>   



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-14582) Add trunc(numeric) udf

2016-11-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-14582?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15670136#comment-15670136
 ] 

Hive QA commented on HIVE-14582:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12839148/HIVE-14582.1.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 7 failed/errored test(s), 10695 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[udf_trunc] (batchId=3)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[udf_trunc_number] 
(batchId=10)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[orc_ppd_schema_evol_3a]
 (batchId=133)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[transform_ppr2] 
(batchId=133)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[join_acid_non_acid]
 (batchId=150)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[union_fast_stats]
 (batchId=145)
org.apache.hive.spark.client.TestSparkClient.testJobSubmission (batchId=272)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/2149/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/2149/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-2149/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 7 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12839148 - PreCommit-HIVE-Build

> Add trunc(numeric) udf
> --
>
> Key: HIVE-14582
> URL: https://issues.apache.org/jira/browse/HIVE-14582
> Project: Hive
>  Issue Type: Sub-task
>  Components: SQL
>Reporter: Ashutosh Chauhan
>Assignee: Chinna Rao Lalam
> Attachments: HIVE-14582.1.patch, HIVE-14582.patch
>
>
> https://docs.oracle.com/cd/B19306_01/server.102/b14200/functions200.htm



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-14582) Add trunc(numeric) udf

2016-11-16 Thread Chinna Rao Lalam (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-14582?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chinna Rao Lalam updated HIVE-14582:

Status: Patch Available  (was: Open)

> Add trunc(numeric) udf
> --
>
> Key: HIVE-14582
> URL: https://issues.apache.org/jira/browse/HIVE-14582
> Project: Hive
>  Issue Type: Sub-task
>  Components: SQL
>Reporter: Ashutosh Chauhan
>Assignee: Chinna Rao Lalam
> Attachments: HIVE-14582.1.patch, HIVE-14582.patch
>
>
> https://docs.oracle.com/cd/B19306_01/server.102/b14200/functions200.htm



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-14582) Add trunc(numeric) udf

2016-11-16 Thread Chinna Rao Lalam (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-14582?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chinna Rao Lalam updated HIVE-14582:

Attachment: HIVE-14582.1.patch

Hi [~ashutoshc], Please take a look at patch.

Now trunc() will accept date or number type arguments and it will behave as 
below

trunc(date, fmt) / trunc(N,D) - Returns 

If input is date returns date with the time portion of the day truncated to the 
unit specified by the format model fmt. 
If you omit fmt, then date is truncated to "the nearest day. It now only 
supports 'MONTH'/'MON'/'MM' and 'YEAR'/''/'YY' as format.

If input is a number group returns N truncated to D decimal places. If D is 
omitted, then N is truncated to 0 places.
D can be negative to truncate (make zero) D digits left of the decimal point.

> Add trunc(numeric) udf
> --
>
> Key: HIVE-14582
> URL: https://issues.apache.org/jira/browse/HIVE-14582
> Project: Hive
>  Issue Type: Sub-task
>  Components: SQL
>Reporter: Ashutosh Chauhan
>Assignee: Chinna Rao Lalam
> Attachments: HIVE-14582.1.patch, HIVE-14582.patch
>
>
> https://docs.oracle.com/cd/B19306_01/server.102/b14200/functions200.htm



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-14582) Add trunc(numeric) udf

2016-11-16 Thread Chinna Rao Lalam (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-14582?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chinna Rao Lalam updated HIVE-14582:

Attachment: (was: HIVE-14582.1.patch)

> Add trunc(numeric) udf
> --
>
> Key: HIVE-14582
> URL: https://issues.apache.org/jira/browse/HIVE-14582
> Project: Hive
>  Issue Type: Sub-task
>  Components: SQL
>Reporter: Ashutosh Chauhan
>Assignee: Chinna Rao Lalam
> Attachments: HIVE-14582.patch
>
>
> https://docs.oracle.com/cd/B19306_01/server.102/b14200/functions200.htm



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15196) LLAP UI: HIVE-14984 broke LLAP UI

2016-11-16 Thread Barna Zsombor Klara (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15196?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Barna Zsombor Klara updated HIVE-15196:
---
Attachment: HIVE-15196.patch

Reuploading patch because pre-commit didn' run.

> LLAP UI: HIVE-14984 broke LLAP UI
> -
>
> Key: HIVE-15196
> URL: https://issues.apache.org/jira/browse/HIVE-15196
> Project: Hive
>  Issue Type: Bug
>  Components: llap, Web UI
>Affects Versions: 2.2.0
>Reporter: Gopal V
>Assignee: Barna Zsombor Klara
> Attachments: HIVE-15196.patch
>
>
> Hard-rewrite rule to the HS2 UI in HttpServer impl
> {code}
> RewriteRegexRule rootRule = new RewriteRegexRule();
> rootRule.setRegex("^/$");
> rootRule.setReplacement("/hiveserver2.jsp");
> rootRule.setTerminating(true);
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HIVE-15196) LLAP UI: HIVE-14984 broke LLAP UI

2016-11-16 Thread Barna Zsombor Klara (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15196?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15669956#comment-15669956
 ] 

Barna Zsombor Klara edited comment on HIVE-15196 at 11/16/16 9:34 AM:
--

Reuploading patch because pre-commit didn't run.


was (Author: zsombor.klara):
Reuploading patch because pre-commit didn' run.

> LLAP UI: HIVE-14984 broke LLAP UI
> -
>
> Key: HIVE-15196
> URL: https://issues.apache.org/jira/browse/HIVE-15196
> Project: Hive
>  Issue Type: Bug
>  Components: llap, Web UI
>Affects Versions: 2.2.0
>Reporter: Gopal V
>Assignee: Barna Zsombor Klara
> Attachments: HIVE-15196.patch
>
>
> Hard-rewrite rule to the HS2 UI in HttpServer impl
> {code}
> RewriteRegexRule rootRule = new RewriteRegexRule();
> rootRule.setRegex("^/$");
> rootRule.setReplacement("/hiveserver2.jsp");
> rootRule.setTerminating(true);
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15196) LLAP UI: HIVE-14984 broke LLAP UI

2016-11-16 Thread Barna Zsombor Klara (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15196?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Barna Zsombor Klara updated HIVE-15196:
---
Attachment: (was: HIVE-15196.patch)

> LLAP UI: HIVE-14984 broke LLAP UI
> -
>
> Key: HIVE-15196
> URL: https://issues.apache.org/jira/browse/HIVE-15196
> Project: Hive
>  Issue Type: Bug
>  Components: llap, Web UI
>Affects Versions: 2.2.0
>Reporter: Gopal V
>Assignee: Barna Zsombor Klara
>
> Hard-rewrite rule to the HS2 UI in HttpServer impl
> {code}
> RewriteRegexRule rootRule = new RewriteRegexRule();
> rootRule.setRegex("^/$");
> rootRule.setReplacement("/hiveserver2.jsp");
> rootRule.setTerminating(true);
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-1478) Non-boolean expression in WHERE clause throws exception

2016-11-16 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-1478?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15669865#comment-15669865
 ] 

Hive QA commented on HIVE-1478:
---



Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12839133/HIVE-1478.1.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 8 failed/errored test(s), 10695 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[annotate_stats_filter] 
(batchId=8)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[vector_empty_where] 
(batchId=21)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[orc_ppd_schema_evol_3a]
 (batchId=133)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[transform_ppr2] 
(batchId=133)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[join_acid_non_acid]
 (batchId=150)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[union_fast_stats]
 (batchId=145)
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver[explainanalyze_2] 
(batchId=91)
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver[explainanalyze_4] 
(batchId=91)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/2148/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/2148/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-2148/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 8 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12839133 - PreCommit-HIVE-Build

> Non-boolean expression in WHERE clause throws exception
> ---
>
> Key: HIVE-1478
> URL: https://issues.apache.org/jira/browse/HIVE-1478
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 0.7.0
>Reporter: Paul Yang
>Assignee: Zoltan Haindrich
>Priority: Minor
> Attachments: HIVE-1478.1.patch
>
>
> If the expression in the where clause does not evaluate to a boolean, the job 
> will fail with the following exception in the task logs:
> Query:
> SELECT key FROM src WHERE 1;
> Exception in mapper:
> 2010-07-21 17:00:31,460 FATAL ExecMapper: 
> org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while 
> processing row {"key":"238","value":"val_238"}
>   at 
> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:417)
>   at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:180)
>   at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50)
>   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:358)
>   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:307)
>   at org.apache.hadoop.mapred.Child.main(Child.java:159)
> Caused by: java.lang.ClassCastException: java.lang.Integer cannot be cast to 
> java.lang.Boolean
>   at 
> org.apache.hadoop.hive.ql.exec.FilterOperator.processOp(FilterOperator.java:84)
>   at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
>   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
>   at 
> org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:45)
>   at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:457)
>   at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:697)
>   at 
> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:400)
>   ... 5 more



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)