[jira] [Commented] (HIVE-22079) Post order walker for iterating over expression tree

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22079?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16926345#comment-16926345
 ] 

Hive QA commented on HIVE-22079:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
49s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
58s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
37s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m 
28s{color} | {color:blue} ql in master has 2246 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
51s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
18s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
58s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
58s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
38s{color} | {color:red} ql: The patch generated 9 new + 16 unchanged - 0 fixed 
= 25 total (was 16) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
41s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
52s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
13s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 21m 51s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-18517/dev-support/hive-personality.sh
 |
| git revision | master / 1dfa2d8 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18517/yetus/diff-checkstyle-ql.txt
 |
| modules | C: ql U: ql |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18517/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Post order walker for iterating over expression tree
> 
>
> Key: HIVE-22079
> URL: https://issues.apache.org/jira/browse/HIVE-22079
> Project: Hive
>  Issue Type: Improvement
>  Components: Logical Optimizer, Physical Optimizer
>Affects Versions: 4.0.0
>Reporter: Vineet Garg
>Assignee: Vineet Garg
>Priority: Major
> Attachments: HIVE-22079.1.patch, HIVE-22079.2.patch, 
> HIVE-22079.3.patch
>
>
> Current {{DefaultGraphWalker}} is used to iterate over an expression tree. 
> This walker uses hash map to keep track of visited/processed nodes. If an 
> expression tree is large this adds significant overhead due to map lookup.
> For an expression trees we can instead use post order traversal and avoid 
> using map.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-20113) Shuffle avoidance: Disable 1-1 edges for sorted shuffle

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-20113?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16926337#comment-16926337
 ] 

Hive QA commented on HIVE-20113:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12979886/HIVE-20113.7.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 3 failed/errored test(s), 16718 tests 
executed
*Failed tests:*
{noformat}
TestDataSourceProviderFactory - did not produce a TEST-*.xml file (likely timed 
out) (batchId=233)
TestObjectStore - did not produce a TEST-*.xml file (likely timed out) 
(batchId=233)
org.apache.hive.minikdc.TestJdbcWithMiniKdcSQLAuthHttp.testAuthorization1 
(batchId=298)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/18516/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/18516/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-18516/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 3 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12979886 - PreCommit-HIVE-Build

> Shuffle avoidance: Disable 1-1 edges for sorted shuffle 
> 
>
> Key: HIVE-20113
> URL: https://issues.apache.org/jira/browse/HIVE-20113
> Project: Hive
>  Issue Type: Bug
>  Components: Tez
>Reporter: Gopal V
>Assignee: Gopal V
>Priority: Major
>  Labels: Branch3Candidate
> Attachments: HIVE-20113.1.patch, HIVE-20113.2.patch, 
> HIVE-20113.3.patch, HIVE-20113.4.patch, HIVE-20113.4.patch, 
> HIVE-20113.5.patch, HIVE-20113.6.patch, HIVE-20113.7.patch
>
>
> The sorted shuffle avoidance can have some issues when the shuffle data gets 
> broken up into multiple chunks on disk.
> The 1-1 edge cannot skip the tez final merge - there's no reason for 1-1 to 
> have a final merge at all, it should open a single compressed file and write 
> a single index entry.
> Until the shuffle issue is resolved & a lot more testing, it is prudent to 
> disable the optimization for sorted shuffle edges and stop rewriting the 
> RS(sorted) = = = RS(sorted) into RS(sorted) = = = RS(FORWARD).



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-20113) Shuffle avoidance: Disable 1-1 edges for sorted shuffle

2019-09-09 Thread Vineet Garg (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-20113?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16926333#comment-16926333
 ] 

Vineet Garg commented on HIVE-20113:


[~jcamachorodriguez] Can you please take a look?

> Shuffle avoidance: Disable 1-1 edges for sorted shuffle 
> 
>
> Key: HIVE-20113
> URL: https://issues.apache.org/jira/browse/HIVE-20113
> Project: Hive
>  Issue Type: Bug
>  Components: Tez
>Reporter: Gopal V
>Assignee: Gopal V
>Priority: Major
>  Labels: Branch3Candidate
> Attachments: HIVE-20113.1.patch, HIVE-20113.2.patch, 
> HIVE-20113.3.patch, HIVE-20113.4.patch, HIVE-20113.4.patch, 
> HIVE-20113.5.patch, HIVE-20113.6.patch, HIVE-20113.7.patch
>
>
> The sorted shuffle avoidance can have some issues when the shuffle data gets 
> broken up into multiple chunks on disk.
> The 1-1 edge cannot skip the tez final merge - there's no reason for 1-1 to 
> have a final merge at all, it should open a single compressed file and write 
> a single index entry.
> Until the shuffle issue is resolved & a lot more testing, it is prudent to 
> disable the optimization for sorted shuffle edges and stop rewriting the 
> RS(sorted) = = = RS(sorted) into RS(sorted) = = = RS(FORWARD).



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-22164) Vectorized Limit operator returns wrong number of results with offset

2019-09-09 Thread Vineet Garg (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22164?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-22164:
---
Fix Version/s: 4.0.0
   Resolution: Fixed
   Status: Resolved  (was: Patch Available)

Pushed to master. Thanks [~rameshkumar]

> Vectorized Limit operator returns wrong number of results with offset
> -
>
> Key: HIVE-22164
> URL: https://issues.apache.org/jira/browse/HIVE-22164
> Project: Hive
>  Issue Type: Bug
>  Components: Hive, llap, Vectorization
>Affects Versions: 4.0.0
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-22164.1.patch
>
>
> Vectorized Limit operator returns wrong number of results with offset



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-22164) Vectorized Limit operator returns wrong number of results with offset

2019-09-09 Thread Vineet Garg (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22164?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16926329#comment-16926329
 ] 

Vineet Garg commented on HIVE-22164:


LGTM +1

> Vectorized Limit operator returns wrong number of results with offset
> -
>
> Key: HIVE-22164
> URL: https://issues.apache.org/jira/browse/HIVE-22164
> Project: Hive
>  Issue Type: Bug
>  Components: Hive, llap, Vectorization
>Affects Versions: 4.0.0
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
> Attachments: HIVE-22164.1.patch
>
>
> Vectorized Limit operator returns wrong number of results with offset



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-22181) Same query but different result occasionally

2019-09-09 Thread qingfa zhou (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22181?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

qingfa zhou updated HIVE-22181:
---
Attachment: image-2019-09-10-12-41-16-717.png

> Same query but different result occasionally
> 
>
> Key: HIVE-22181
> URL: https://issues.apache.org/jira/browse/HIVE-22181
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 2.3.0
> Environment: hadoop:2.7.3.2.6.1.0-129
> hive:2.3.0
> tez:0.8.4
> java:1.8.0_212 
>Reporter: qingfa zhou
>Priority: Major
> Attachments: WX20190910-123604.png, WX20190910-123621.png, 
> WX20190910-123641.png, image-2019-09-10-12-40-34-958.png, 
> image-2019-09-10-12-40-53-338.png, image-2019-09-10-12-41-16-717.png, 
> image-2019-09-10-12-41-35-056.png
>
>
> h3. 1.Phenomenon
>     Running the same SQL(hive on tez) several times will result in 
> inconsistent results . The following results are generated respectively
>     1)
>      OK
>      3951864 808
>     2)
>      OK
>      1822979 353
>     But the first result is true.
> h3. 2.SQL
> select
>  count(1),count(distinct store_code)
>  from 
>  (
>  select
>  store_code,
>  all_pay_id,
>  payable_price,
>  row_number()over(partition by store_code order by payable_price desc) as rw,
>  count(1)over(partition by store_code) as store_user
>  from 
>  (
>  select 
>  store_code,
>  all_pay_id,
>  sum(payable_price) as payable_price
>  from data_promotion.mdw_user_promotion_shopping_behaivor_four_week_detail a
>  where dt='20190904'
>  group by store_code,all_pay_id
>  ) a 
>  order by rw 
>  ) a 
>  where rw/store_user<=0.8
>  ;
>  
> h3. 3.Troubleshooting
>     I lookup task result from tez web ui,give the result as follows:
>            the first: 
> !image-2019-09-10-12-40-34-958.png!
>         the second:
> !image-2019-09-10-12-40-53-338.png!
>       so , The second RECORDS_OUT_INTERMEDIATE_Reducer_4's result is much 
> lower than the first.
>      This is detailed info of the second result from reducer_3 to reducer_4 
> !image-2019-09-10-12-41-16-717.png!   
> This is execution plan of the sql:
> OK
>  Vertex dependency in root stage
>  Reducer 2 <- Map 1 (SIMPLE_EDGE)
>  Reducer 3 <- Reducer 2 (SIMPLE_EDGE)
>  Reducer 4 <- Reducer 3 (SIMPLE_EDGE)
>  Reducer 5 <- Reducer 4 (SIMPLE_EDGE)
>  Reducer 6 <- Reducer 5 (SIMPLE_EDGE)
> Stage-0
>  Fetch Operator
>  limit:-1
>  Stage-1
>  Reducer 6
>  File Output Operator [FS_24]
>  Group By Operator [GBY_22] (rows=1 width=128)
>  Output:["_col0","_col1"],aggregations:["count(VALUE._col0)","count(DISTINCT 
> KEY._col0:0._col0)"]
>  <-Reducer 5 [SIMPLE_EDGE]
>  SHUFFLE [RS_21]
>  Group By Operator [GBY_20] (rows=5619870 width=2328)
>  Output:["_col0","_col1","_col2"],aggregations:["count(1)","count(DISTINCT 
> _col0)"],keys:_col0
>  Select Operator [SEL_17] (rows=5619870 width=2328)
>  Output:["_col0"]
>  <-Reducer 4 [SIMPLE_EDGE]
>  SHUFFLE [RS_16]
>  Select Operator [SEL_14] (rows=5619870 width=2328)
>  Output:["_col0","_col3"]
>  Filter Operator [FIL_25] (rows=5619870 width=2328)
>  predicate:((_col0 / count_window_1) <= 0.8)
>  PTF Operator [PTF_13] (rows=16859610 width=2328)
>  Function definitions:[{},
> {"name:":"windowingtablefunction","order by:":"_col1 ASC NULLS 
> FIRST","partition by:":"_col1"}
> ]
>  Select Operator [SEL_12] (rows=16859610 width=2328)
>  Output:["_col0","_col1"]
>  <-Reducer 3 [SIMPLE_EDGE]
>  SHUFFLE [RS_11]
>  PartitionCols:_col0
>  Select Operator [SEL_10] (rows=16859610 width=2328)
>  Output:["_col0","row_number_window_0"]
>  PTF Operator [PTF_9] (rows=16859610 width=2328)
>  Function definitions:[{},
> {"name:":"windowingtablefunction","order by:":"_col2 DESC NULLS 
> LAST","partition by:":"_col0"}
> ]
>  Select Operator [SEL_8] (rows=16859610 width=2328)
>  Output:["_col0","_col2"]
>  <-Reducer 2 [SIMPLE_EDGE]
>  SHUFFLE [RS_7]
>  PartitionCols:_col0
>  Select Operator [SEL_6] (rows=16859610 width=2328)
>  Output:["_col0","_col2"]
>  Group By Operator [GBY_5] (rows=16859610 width=2328)
>  
> Output:["_col0","_col1","_col2"],aggregations:["sum(VALUE._col0)"],keys:KEY._col0,
>  KEY._col1
>  <-Map 1 [SIMPLE_EDGE]
>  SHUFFLE [RS_4]
>  PartitionCols:_col0, _col1
>  Group By Operator [GBY_3] (rows=33719220 width=2328)
>  
> Output:["_col0","_col1","_col2"],aggregations:["sum(payable_price)"],keys:store_code,
>  all_pay_id
>  Select Operator [SEL_2] (rows=33719220 width=2328)
>  Output:["store_code","all_pay_id","payable_price"]
>  TableScan [TS_0] (rows=33719220 width=2328)
>  
> data_promotion@mdw_user_promotion_shopping_behaivor_four_week_detail,a,Tbl:COMPLETE,Col:NONE,Output:["all_pay_id","store_code","payable_price"]
>  
>     Reducer 4 show PTF , so I don't know if this result has anything to do 
> with PTF? Can you help me.
>  
>  



--
This message was sent by Atlassian Jira

[jira] [Updated] (HIVE-22181) Same query but different result occasionally

2019-09-09 Thread qingfa zhou (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22181?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

qingfa zhou updated HIVE-22181:
---
Description: 
h3. 1.Phenomenon

    Running the same SQL(hive on tez) several times will result in inconsistent 
results . The following results are generated respectively

    1)

     OK

     3951864 808

    2)

     OK

     1822979 353

    But the first result is true.
h3. 2.SQL

select
 count(1),count(distinct store_code)
 from 
 (
 select
 store_code,
 all_pay_id,
 payable_price,
 row_number()over(partition by store_code order by payable_price desc) as rw,
 count(1)over(partition by store_code) as store_user
 from 
 (
 select 
 store_code,
 all_pay_id,
 sum(payable_price) as payable_price
 from data_promotion.mdw_user_promotion_shopping_behaivor_four_week_detail a
 where dt='20190904'
 group by store_code,all_pay_id
 ) a 
 order by rw 
 ) a 
 where rw/store_user<=0.8
 ;

 
h3. 3.Troubleshooting

    I lookup task result from tez web ui,give the result as follows:

           the first: 

!image-2019-09-10-12-40-34-958.png!

        the second:

!image-2019-09-10-12-40-53-338.png!

      so , The second RECORDS_OUT_INTERMEDIATE_Reducer_4's result is much lower 
than the first.

     This is detailed info of the second result from reducer_3 to reducer_4 

!image-2019-09-10-12-41-16-717.png!   

This is execution plan of the sql:

OK
 Vertex dependency in root stage
 Reducer 2 <- Map 1 (SIMPLE_EDGE)
 Reducer 3 <- Reducer 2 (SIMPLE_EDGE)
 Reducer 4 <- Reducer 3 (SIMPLE_EDGE)
 Reducer 5 <- Reducer 4 (SIMPLE_EDGE)
 Reducer 6 <- Reducer 5 (SIMPLE_EDGE)

Stage-0
 Fetch Operator
 limit:-1
 Stage-1
 Reducer 6
 File Output Operator [FS_24]
 Group By Operator [GBY_22] (rows=1 width=128)
 Output:["_col0","_col1"],aggregations:["count(VALUE._col0)","count(DISTINCT 
KEY._col0:0._col0)"]
 <-Reducer 5 [SIMPLE_EDGE]
 SHUFFLE [RS_21]
 Group By Operator [GBY_20] (rows=5619870 width=2328)
 Output:["_col0","_col1","_col2"],aggregations:["count(1)","count(DISTINCT 
_col0)"],keys:_col0
 Select Operator [SEL_17] (rows=5619870 width=2328)
 Output:["_col0"]
 <-Reducer 4 [SIMPLE_EDGE]
 SHUFFLE [RS_16]
 Select Operator [SEL_14] (rows=5619870 width=2328)
 Output:["_col0","_col3"]
 Filter Operator [FIL_25] (rows=5619870 width=2328)
 predicate:((_col0 / count_window_1) <= 0.8)
 PTF Operator [PTF_13] (rows=16859610 width=2328)
 Function definitions:[{},

{"name:":"windowingtablefunction","order by:":"_col1 ASC NULLS 
FIRST","partition by:":"_col1"}

]
 Select Operator [SEL_12] (rows=16859610 width=2328)
 Output:["_col0","_col1"]
 <-Reducer 3 [SIMPLE_EDGE]
 SHUFFLE [RS_11]
 PartitionCols:_col0
 Select Operator [SEL_10] (rows=16859610 width=2328)
 Output:["_col0","row_number_window_0"]
 PTF Operator [PTF_9] (rows=16859610 width=2328)
 Function definitions:[{},

{"name:":"windowingtablefunction","order by:":"_col2 DESC NULLS 
LAST","partition by:":"_col0"}

]
 Select Operator [SEL_8] (rows=16859610 width=2328)
 Output:["_col0","_col2"]
 <-Reducer 2 [SIMPLE_EDGE]
 SHUFFLE [RS_7]
 PartitionCols:_col0
 Select Operator [SEL_6] (rows=16859610 width=2328)
 Output:["_col0","_col2"]
 Group By Operator [GBY_5] (rows=16859610 width=2328)
 
Output:["_col0","_col1","_col2"],aggregations:["sum(VALUE._col0)"],keys:KEY._col0,
 KEY._col1
 <-Map 1 [SIMPLE_EDGE]
 SHUFFLE [RS_4]
 PartitionCols:_col0, _col1
 Group By Operator [GBY_3] (rows=33719220 width=2328)
 
Output:["_col0","_col1","_col2"],aggregations:["sum(payable_price)"],keys:store_code,
 all_pay_id
 Select Operator [SEL_2] (rows=33719220 width=2328)
 Output:["store_code","all_pay_id","payable_price"]
 TableScan [TS_0] (rows=33719220 width=2328)
 
data_promotion@mdw_user_promotion_shopping_behaivor_four_week_detail,a,Tbl:COMPLETE,Col:NONE,Output:["all_pay_id","store_code","payable_price"]

 

    Reducer 4 show PTF , so I don't know if this result has anything to do with 
PTF? Can you help me.

 

 

  was:
h3. 1.Phenomenon

    Running the same SQL(hive on tez) several times will result in inconsistent 
results . The following results are generated respectively

    1)

     OK

     3951864 808

    2)

     OK

     1822979 353

    But the first result is true.
h3. 2.SQL

select
 count(1),count(distinct store_code)
 from 
 (
 select
 store_code,
 all_pay_id,
 payable_price,
 row_number()over(partition by store_code order by payable_price desc) as rw,
 count(1)over(partition by store_code) as store_user
 from 
 (
 select 
 store_code,
 all_pay_id,
 sum(payable_price) as payable_price
 from data_promotion.mdw_user_promotion_shopping_behaivor_four_week_detail a
 where dt='20190904'
 group by store_code,all_pay_id
 ) a 
 order by rw 
 ) a 
 where rw/store_user<=0.8
 ;

 
h3. 3.Troubleshooting

    I lookup task result from tez web ui,give the result as follows:

           the first: 

        the second:

!image-2019-09-10-12-18-08-769.png!

      so , The second 

[jira] [Updated] (HIVE-22181) Same query but different result occasionally

2019-09-09 Thread qingfa zhou (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22181?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

qingfa zhou updated HIVE-22181:
---
Attachment: image-2019-09-10-12-41-35-056.png

> Same query but different result occasionally
> 
>
> Key: HIVE-22181
> URL: https://issues.apache.org/jira/browse/HIVE-22181
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 2.3.0
> Environment: hadoop:2.7.3.2.6.1.0-129
> hive:2.3.0
> tez:0.8.4
> java:1.8.0_212 
>Reporter: qingfa zhou
>Priority: Major
> Attachments: WX20190910-123604.png, WX20190910-123621.png, 
> WX20190910-123641.png, image-2019-09-10-12-40-34-958.png, 
> image-2019-09-10-12-40-53-338.png, image-2019-09-10-12-41-16-717.png, 
> image-2019-09-10-12-41-35-056.png
>
>
> h3. 1.Phenomenon
>     Running the same SQL(hive on tez) several times will result in 
> inconsistent results . The following results are generated respectively
>     1)
>      OK
>      3951864 808
>     2)
>      OK
>      1822979 353
>     But the first result is true.
> h3. 2.SQL
> select
>  count(1),count(distinct store_code)
>  from 
>  (
>  select
>  store_code,
>  all_pay_id,
>  payable_price,
>  row_number()over(partition by store_code order by payable_price desc) as rw,
>  count(1)over(partition by store_code) as store_user
>  from 
>  (
>  select 
>  store_code,
>  all_pay_id,
>  sum(payable_price) as payable_price
>  from data_promotion.mdw_user_promotion_shopping_behaivor_four_week_detail a
>  where dt='20190904'
>  group by store_code,all_pay_id
>  ) a 
>  order by rw 
>  ) a 
>  where rw/store_user<=0.8
>  ;
>  
> h3. 3.Troubleshooting
>     I lookup task result from tez web ui,give the result as follows:
>            the first: 
> !image-2019-09-10-12-40-34-958.png!
>         the second:
> !image-2019-09-10-12-40-53-338.png!
>       so , The second RECORDS_OUT_INTERMEDIATE_Reducer_4's result is much 
> lower than the first.
>      This is detailed info of the second result from reducer_3 to reducer_4 
> !image-2019-09-10-12-41-35-056.png!    
> This is execution plan of the sql:
> OK
>  Vertex dependency in root stage
>  Reducer 2 <- Map 1 (SIMPLE_EDGE)
>  Reducer 3 <- Reducer 2 (SIMPLE_EDGE)
>  Reducer 4 <- Reducer 3 (SIMPLE_EDGE)
>  Reducer 5 <- Reducer 4 (SIMPLE_EDGE)
>  Reducer 6 <- Reducer 5 (SIMPLE_EDGE)
> Stage-0
>  Fetch Operator
>  limit:-1
>  Stage-1
>  Reducer 6
>  File Output Operator [FS_24]
>  Group By Operator [GBY_22] (rows=1 width=128)
>  Output:["_col0","_col1"],aggregations:["count(VALUE._col0)","count(DISTINCT 
> KEY._col0:0._col0)"]
>  <-Reducer 5 [SIMPLE_EDGE]
>  SHUFFLE [RS_21]
>  Group By Operator [GBY_20] (rows=5619870 width=2328)
>  Output:["_col0","_col1","_col2"],aggregations:["count(1)","count(DISTINCT 
> _col0)"],keys:_col0
>  Select Operator [SEL_17] (rows=5619870 width=2328)
>  Output:["_col0"]
>  <-Reducer 4 [SIMPLE_EDGE]
>  SHUFFLE [RS_16]
>  Select Operator [SEL_14] (rows=5619870 width=2328)
>  Output:["_col0","_col3"]
>  Filter Operator [FIL_25] (rows=5619870 width=2328)
>  predicate:((_col0 / count_window_1) <= 0.8)
>  PTF Operator [PTF_13] (rows=16859610 width=2328)
>  Function definitions:[{},
> {"name:":"windowingtablefunction","order by:":"_col1 ASC NULLS 
> FIRST","partition by:":"_col1"}
> ]
>  Select Operator [SEL_12] (rows=16859610 width=2328)
>  Output:["_col0","_col1"]
>  <-Reducer 3 [SIMPLE_EDGE]
>  SHUFFLE [RS_11]
>  PartitionCols:_col0
>  Select Operator [SEL_10] (rows=16859610 width=2328)
>  Output:["_col0","row_number_window_0"]
>  PTF Operator [PTF_9] (rows=16859610 width=2328)
>  Function definitions:[{},
> {"name:":"windowingtablefunction","order by:":"_col2 DESC NULLS 
> LAST","partition by:":"_col0"}
> ]
>  Select Operator [SEL_8] (rows=16859610 width=2328)
>  Output:["_col0","_col2"]
>  <-Reducer 2 [SIMPLE_EDGE]
>  SHUFFLE [RS_7]
>  PartitionCols:_col0
>  Select Operator [SEL_6] (rows=16859610 width=2328)
>  Output:["_col0","_col2"]
>  Group By Operator [GBY_5] (rows=16859610 width=2328)
>  
> Output:["_col0","_col1","_col2"],aggregations:["sum(VALUE._col0)"],keys:KEY._col0,
>  KEY._col1
>  <-Map 1 [SIMPLE_EDGE]
>  SHUFFLE [RS_4]
>  PartitionCols:_col0, _col1
>  Group By Operator [GBY_3] (rows=33719220 width=2328)
>  
> Output:["_col0","_col1","_col2"],aggregations:["sum(payable_price)"],keys:store_code,
>  all_pay_id
>  Select Operator [SEL_2] (rows=33719220 width=2328)
>  Output:["store_code","all_pay_id","payable_price"]
>  TableScan [TS_0] (rows=33719220 width=2328)
>  
> data_promotion@mdw_user_promotion_shopping_behaivor_four_week_detail,a,Tbl:COMPLETE,Col:NONE,Output:["all_pay_id","store_code","payable_price"]
>  
>     Reducer 4 show PTF , so I don't know if this result has anything to do 
> with PTF? Can you help me.
>  
>  



--
This message was sent by Atlassian Jira

[jira] [Updated] (HIVE-22181) Same query but different result occasionally

2019-09-09 Thread qingfa zhou (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22181?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

qingfa zhou updated HIVE-22181:
---
Description: 
h3. 1.Phenomenon

    Running the same SQL(hive on tez) several times will result in inconsistent 
results . The following results are generated respectively

    1)

     OK

     3951864 808

    2)

     OK

     1822979 353

    But the first result is true.
h3. 2.SQL

select
 count(1),count(distinct store_code)
 from 
 (
 select
 store_code,
 all_pay_id,
 payable_price,
 row_number()over(partition by store_code order by payable_price desc) as rw,
 count(1)over(partition by store_code) as store_user
 from 
 (
 select 
 store_code,
 all_pay_id,
 sum(payable_price) as payable_price
 from data_promotion.mdw_user_promotion_shopping_behaivor_four_week_detail a
 where dt='20190904'
 group by store_code,all_pay_id
 ) a 
 order by rw 
 ) a 
 where rw/store_user<=0.8
 ;

 
h3. 3.Troubleshooting

    I lookup task result from tez web ui,give the result as follows:

           the first: 

!image-2019-09-10-12-40-34-958.png!

        the second:

!image-2019-09-10-12-40-53-338.png!

      so , The second RECORDS_OUT_INTERMEDIATE_Reducer_4's result is much lower 
than the first.

     This is detailed info of the second result from reducer_3 to reducer_4 

!image-2019-09-10-12-41-35-056.png!    

This is execution plan of the sql:

OK
 Vertex dependency in root stage
 Reducer 2 <- Map 1 (SIMPLE_EDGE)
 Reducer 3 <- Reducer 2 (SIMPLE_EDGE)
 Reducer 4 <- Reducer 3 (SIMPLE_EDGE)
 Reducer 5 <- Reducer 4 (SIMPLE_EDGE)
 Reducer 6 <- Reducer 5 (SIMPLE_EDGE)

Stage-0
 Fetch Operator
 limit:-1
 Stage-1
 Reducer 6
 File Output Operator [FS_24]
 Group By Operator [GBY_22] (rows=1 width=128)
 Output:["_col0","_col1"],aggregations:["count(VALUE._col0)","count(DISTINCT 
KEY._col0:0._col0)"]
 <-Reducer 5 [SIMPLE_EDGE]
 SHUFFLE [RS_21]
 Group By Operator [GBY_20] (rows=5619870 width=2328)
 Output:["_col0","_col1","_col2"],aggregations:["count(1)","count(DISTINCT 
_col0)"],keys:_col0
 Select Operator [SEL_17] (rows=5619870 width=2328)
 Output:["_col0"]
 <-Reducer 4 [SIMPLE_EDGE]
 SHUFFLE [RS_16]
 Select Operator [SEL_14] (rows=5619870 width=2328)
 Output:["_col0","_col3"]
 Filter Operator [FIL_25] (rows=5619870 width=2328)
 predicate:((_col0 / count_window_1) <= 0.8)
 PTF Operator [PTF_13] (rows=16859610 width=2328)
 Function definitions:[{},

{"name:":"windowingtablefunction","order by:":"_col1 ASC NULLS 
FIRST","partition by:":"_col1"}

]
 Select Operator [SEL_12] (rows=16859610 width=2328)
 Output:["_col0","_col1"]
 <-Reducer 3 [SIMPLE_EDGE]
 SHUFFLE [RS_11]
 PartitionCols:_col0
 Select Operator [SEL_10] (rows=16859610 width=2328)
 Output:["_col0","row_number_window_0"]
 PTF Operator [PTF_9] (rows=16859610 width=2328)
 Function definitions:[{},

{"name:":"windowingtablefunction","order by:":"_col2 DESC NULLS 
LAST","partition by:":"_col0"}

]
 Select Operator [SEL_8] (rows=16859610 width=2328)
 Output:["_col0","_col2"]
 <-Reducer 2 [SIMPLE_EDGE]
 SHUFFLE [RS_7]
 PartitionCols:_col0
 Select Operator [SEL_6] (rows=16859610 width=2328)
 Output:["_col0","_col2"]
 Group By Operator [GBY_5] (rows=16859610 width=2328)
 
Output:["_col0","_col1","_col2"],aggregations:["sum(VALUE._col0)"],keys:KEY._col0,
 KEY._col1
 <-Map 1 [SIMPLE_EDGE]
 SHUFFLE [RS_4]
 PartitionCols:_col0, _col1
 Group By Operator [GBY_3] (rows=33719220 width=2328)
 
Output:["_col0","_col1","_col2"],aggregations:["sum(payable_price)"],keys:store_code,
 all_pay_id
 Select Operator [SEL_2] (rows=33719220 width=2328)
 Output:["store_code","all_pay_id","payable_price"]
 TableScan [TS_0] (rows=33719220 width=2328)
 
data_promotion@mdw_user_promotion_shopping_behaivor_four_week_detail,a,Tbl:COMPLETE,Col:NONE,Output:["all_pay_id","store_code","payable_price"]

 

    Reducer 4 show PTF , so I don't know if this result has anything to do with 
PTF? Can you help me.

 

 

  was:
h3. 1.Phenomenon

    Running the same SQL(hive on tez) several times will result in inconsistent 
results . The following results are generated respectively

    1)

     OK

     3951864 808

    2)

     OK

     1822979 353

    But the first result is true.
h3. 2.SQL

select
 count(1),count(distinct store_code)
 from 
 (
 select
 store_code,
 all_pay_id,
 payable_price,
 row_number()over(partition by store_code order by payable_price desc) as rw,
 count(1)over(partition by store_code) as store_user
 from 
 (
 select 
 store_code,
 all_pay_id,
 sum(payable_price) as payable_price
 from data_promotion.mdw_user_promotion_shopping_behaivor_four_week_detail a
 where dt='20190904'
 group by store_code,all_pay_id
 ) a 
 order by rw 
 ) a 
 where rw/store_user<=0.8
 ;

 
h3. 3.Troubleshooting

    I lookup task result from tez web ui,give the result as follows:

           the first: 

!image-2019-09-10-12-40-34-958.png!

        the second:

!image-2019-09-10-12-40-53-338.png!

      so , The 

[jira] [Updated] (HIVE-22181) Same query but different result occasionally

2019-09-09 Thread qingfa zhou (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22181?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

qingfa zhou updated HIVE-22181:
---
Attachment: image-2019-09-10-12-40-53-338.png

> Same query but different result occasionally
> 
>
> Key: HIVE-22181
> URL: https://issues.apache.org/jira/browse/HIVE-22181
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 2.3.0
> Environment: hadoop:2.7.3.2.6.1.0-129
> hive:2.3.0
> tez:0.8.4
> java:1.8.0_212 
>Reporter: qingfa zhou
>Priority: Major
> Attachments: WX20190910-123604.png, WX20190910-123621.png, 
> WX20190910-123641.png, image-2019-09-10-12-40-34-958.png, 
> image-2019-09-10-12-40-53-338.png
>
>
> h3. 1.Phenomenon
>     Running the same SQL(hive on tez) several times will result in 
> inconsistent results . The following results are generated respectively
>     1)
>      OK
>      3951864 808
>     2)
>      OK
>      1822979 353
>     But the first result is true.
> h3. 2.SQL
> select
>  count(1),count(distinct store_code)
>  from 
>  (
>  select
>  store_code,
>  all_pay_id,
>  payable_price,
>  row_number()over(partition by store_code order by payable_price desc) as rw,
>  count(1)over(partition by store_code) as store_user
>  from 
>  (
>  select 
>  store_code,
>  all_pay_id,
>  sum(payable_price) as payable_price
>  from data_promotion.mdw_user_promotion_shopping_behaivor_four_week_detail a
>  where dt='20190904'
>  group by store_code,all_pay_id
>  ) a 
>  order by rw 
>  ) a 
>  where rw/store_user<=0.8
>  ;
>  
> h3. 3.Troubleshooting
>     I lookup task result from tez web ui,give the result as follows:
>            the first: 
>         the second:
> !image-2019-09-10-12-18-08-769.png!
>       so , The second RECORDS_OUT_INTERMEDIATE_Reducer_4's result is much 
> lower than the first.
>      This is detailed info of the second result from reducer_3 to reducer_4 
> !image-2019-09-10-12-26-41-681.png!  
> This is execution plan of the sql:
> OK
>  Vertex dependency in root stage
>  Reducer 2 <- Map 1 (SIMPLE_EDGE)
>  Reducer 3 <- Reducer 2 (SIMPLE_EDGE)
>  Reducer 4 <- Reducer 3 (SIMPLE_EDGE)
>  Reducer 5 <- Reducer 4 (SIMPLE_EDGE)
>  Reducer 6 <- Reducer 5 (SIMPLE_EDGE)
> Stage-0
>  Fetch Operator
>  limit:-1
>  Stage-1
>  Reducer 6
>  File Output Operator [FS_24]
>  Group By Operator [GBY_22] (rows=1 width=128)
>  Output:["_col0","_col1"],aggregations:["count(VALUE._col0)","count(DISTINCT 
> KEY._col0:0._col0)"]
>  <-Reducer 5 [SIMPLE_EDGE]
>  SHUFFLE [RS_21]
>  Group By Operator [GBY_20] (rows=5619870 width=2328)
>  Output:["_col0","_col1","_col2"],aggregations:["count(1)","count(DISTINCT 
> _col0)"],keys:_col0
>  Select Operator [SEL_17] (rows=5619870 width=2328)
>  Output:["_col0"]
>  <-Reducer 4 [SIMPLE_EDGE]
>  SHUFFLE [RS_16]
>  Select Operator [SEL_14] (rows=5619870 width=2328)
>  Output:["_col0","_col3"]
>  Filter Operator [FIL_25] (rows=5619870 width=2328)
>  predicate:((_col0 / count_window_1) <= 0.8)
>  PTF Operator [PTF_13] (rows=16859610 width=2328)
>  Function definitions:[{},\\{"name:":"windowingtablefunction","order 
> by:":"_col1 ASC NULLS FIRST","partition by:":"_col1"}]
>  Select Operator [SEL_12] (rows=16859610 width=2328)
>  Output:["_col0","_col1"]
>  <-Reducer 3 [SIMPLE_EDGE]
>  SHUFFLE [RS_11]
>  PartitionCols:_col0
>  Select Operator [SEL_10] (rows=16859610 width=2328)
>  Output:["_col0","row_number_window_0"]
>  PTF Operator [PTF_9] (rows=16859610 width=2328)
>  Function definitions:[{},\\{"name:":"windowingtablefunction","order 
> by:":"_col2 DESC NULLS LAST","partition by:":"_col0"}]
>  Select Operator [SEL_8] (rows=16859610 width=2328)
>  Output:["_col0","_col2"]
>  <-Reducer 2 [SIMPLE_EDGE]
>  SHUFFLE [RS_7]
>  PartitionCols:_col0
>  Select Operator [SEL_6] (rows=16859610 width=2328)
>  Output:["_col0","_col2"]
>  Group By Operator [GBY_5] (rows=16859610 width=2328)
>  
> Output:["_col0","_col1","_col2"],aggregations:["sum(VALUE._col0)"],keys:KEY._col0,
>  KEY._col1
>  <-Map 1 [SIMPLE_EDGE]
>  SHUFFLE [RS_4]
>  PartitionCols:_col0, _col1
>  Group By Operator [GBY_3] (rows=33719220 width=2328)
>  
> Output:["_col0","_col1","_col2"],aggregations:["sum(payable_price)"],keys:store_code,
>  all_pay_id
>  Select Operator [SEL_2] (rows=33719220 width=2328)
>  Output:["store_code","all_pay_id","payable_price"]
>  TableScan [TS_0] (rows=33719220 width=2328)
>  
> data_promotion@mdw_user_promotion_shopping_behaivor_four_week_detail,a,Tbl:COMPLETE,Col:NONE,Output:["all_pay_id","store_code","payable_price"]
>  
>     Reducer 4 show PTF , so I don't know if this result has anything to do 
> with PTF? Can you help me.
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-22181) Same query but different result occasionally

2019-09-09 Thread qingfa zhou (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22181?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

qingfa zhou updated HIVE-22181:
---
Attachment: image-2019-09-10-12-40-34-958.png

> Same query but different result occasionally
> 
>
> Key: HIVE-22181
> URL: https://issues.apache.org/jira/browse/HIVE-22181
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 2.3.0
> Environment: hadoop:2.7.3.2.6.1.0-129
> hive:2.3.0
> tez:0.8.4
> java:1.8.0_212 
>Reporter: qingfa zhou
>Priority: Major
> Attachments: WX20190910-123604.png, WX20190910-123621.png, 
> WX20190910-123641.png, image-2019-09-10-12-40-34-958.png, 
> image-2019-09-10-12-40-53-338.png
>
>
> h3. 1.Phenomenon
>     Running the same SQL(hive on tez) several times will result in 
> inconsistent results . The following results are generated respectively
>     1)
>      OK
>      3951864 808
>     2)
>      OK
>      1822979 353
>     But the first result is true.
> h3. 2.SQL
> select
>  count(1),count(distinct store_code)
>  from 
>  (
>  select
>  store_code,
>  all_pay_id,
>  payable_price,
>  row_number()over(partition by store_code order by payable_price desc) as rw,
>  count(1)over(partition by store_code) as store_user
>  from 
>  (
>  select 
>  store_code,
>  all_pay_id,
>  sum(payable_price) as payable_price
>  from data_promotion.mdw_user_promotion_shopping_behaivor_four_week_detail a
>  where dt='20190904'
>  group by store_code,all_pay_id
>  ) a 
>  order by rw 
>  ) a 
>  where rw/store_user<=0.8
>  ;
>  
> h3. 3.Troubleshooting
>     I lookup task result from tez web ui,give the result as follows:
>            the first: 
>         the second:
> !image-2019-09-10-12-18-08-769.png!
>       so , The second RECORDS_OUT_INTERMEDIATE_Reducer_4's result is much 
> lower than the first.
>      This is detailed info of the second result from reducer_3 to reducer_4 
> !image-2019-09-10-12-26-41-681.png!  
> This is execution plan of the sql:
> OK
>  Vertex dependency in root stage
>  Reducer 2 <- Map 1 (SIMPLE_EDGE)
>  Reducer 3 <- Reducer 2 (SIMPLE_EDGE)
>  Reducer 4 <- Reducer 3 (SIMPLE_EDGE)
>  Reducer 5 <- Reducer 4 (SIMPLE_EDGE)
>  Reducer 6 <- Reducer 5 (SIMPLE_EDGE)
> Stage-0
>  Fetch Operator
>  limit:-1
>  Stage-1
>  Reducer 6
>  File Output Operator [FS_24]
>  Group By Operator [GBY_22] (rows=1 width=128)
>  Output:["_col0","_col1"],aggregations:["count(VALUE._col0)","count(DISTINCT 
> KEY._col0:0._col0)"]
>  <-Reducer 5 [SIMPLE_EDGE]
>  SHUFFLE [RS_21]
>  Group By Operator [GBY_20] (rows=5619870 width=2328)
>  Output:["_col0","_col1","_col2"],aggregations:["count(1)","count(DISTINCT 
> _col0)"],keys:_col0
>  Select Operator [SEL_17] (rows=5619870 width=2328)
>  Output:["_col0"]
>  <-Reducer 4 [SIMPLE_EDGE]
>  SHUFFLE [RS_16]
>  Select Operator [SEL_14] (rows=5619870 width=2328)
>  Output:["_col0","_col3"]
>  Filter Operator [FIL_25] (rows=5619870 width=2328)
>  predicate:((_col0 / count_window_1) <= 0.8)
>  PTF Operator [PTF_13] (rows=16859610 width=2328)
>  Function definitions:[{},\\{"name:":"windowingtablefunction","order 
> by:":"_col1 ASC NULLS FIRST","partition by:":"_col1"}]
>  Select Operator [SEL_12] (rows=16859610 width=2328)
>  Output:["_col0","_col1"]
>  <-Reducer 3 [SIMPLE_EDGE]
>  SHUFFLE [RS_11]
>  PartitionCols:_col0
>  Select Operator [SEL_10] (rows=16859610 width=2328)
>  Output:["_col0","row_number_window_0"]
>  PTF Operator [PTF_9] (rows=16859610 width=2328)
>  Function definitions:[{},\\{"name:":"windowingtablefunction","order 
> by:":"_col2 DESC NULLS LAST","partition by:":"_col0"}]
>  Select Operator [SEL_8] (rows=16859610 width=2328)
>  Output:["_col0","_col2"]
>  <-Reducer 2 [SIMPLE_EDGE]
>  SHUFFLE [RS_7]
>  PartitionCols:_col0
>  Select Operator [SEL_6] (rows=16859610 width=2328)
>  Output:["_col0","_col2"]
>  Group By Operator [GBY_5] (rows=16859610 width=2328)
>  
> Output:["_col0","_col1","_col2"],aggregations:["sum(VALUE._col0)"],keys:KEY._col0,
>  KEY._col1
>  <-Map 1 [SIMPLE_EDGE]
>  SHUFFLE [RS_4]
>  PartitionCols:_col0, _col1
>  Group By Operator [GBY_3] (rows=33719220 width=2328)
>  
> Output:["_col0","_col1","_col2"],aggregations:["sum(payable_price)"],keys:store_code,
>  all_pay_id
>  Select Operator [SEL_2] (rows=33719220 width=2328)
>  Output:["store_code","all_pay_id","payable_price"]
>  TableScan [TS_0] (rows=33719220 width=2328)
>  
> data_promotion@mdw_user_promotion_shopping_behaivor_four_week_detail,a,Tbl:COMPLETE,Col:NONE,Output:["all_pay_id","store_code","payable_price"]
>  
>     Reducer 4 show PTF , so I don't know if this result has anything to do 
> with PTF? Can you help me.
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-22181) Same query but different result occasionally

2019-09-09 Thread qingfa zhou (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22181?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

qingfa zhou updated HIVE-22181:
---
Description: 
h3. 1.Phenomenon

    Running the same SQL(hive on tez) several times will result in inconsistent 
results . The following results are generated respectively

    1)

     OK

     3951864 808

    2)

     OK

     1822979 353

    But the first result is true.
h3. 2.SQL

select
 count(1),count(distinct store_code)
 from 
 (
 select
 store_code,
 all_pay_id,
 payable_price,
 row_number()over(partition by store_code order by payable_price desc) as rw,
 count(1)over(partition by store_code) as store_user
 from 
 (
 select 
 store_code,
 all_pay_id,
 sum(payable_price) as payable_price
 from data_promotion.mdw_user_promotion_shopping_behaivor_four_week_detail a
 where dt='20190904'
 group by store_code,all_pay_id
 ) a 
 order by rw 
 ) a 
 where rw/store_user<=0.8
 ;

 
h3. 3.Troubleshooting

    I lookup task result from tez web ui,give the result as follows:

           the first: 

        the second:

!image-2019-09-10-12-18-08-769.png!

      so , The second RECORDS_OUT_INTERMEDIATE_Reducer_4's result is much lower 
than the first.

     This is detailed info of the second result from reducer_3 to reducer_4 

!image-2019-09-10-12-26-41-681.png!  

This is execution plan of the sql:

OK
 Vertex dependency in root stage
 Reducer 2 <- Map 1 (SIMPLE_EDGE)
 Reducer 3 <- Reducer 2 (SIMPLE_EDGE)
 Reducer 4 <- Reducer 3 (SIMPLE_EDGE)
 Reducer 5 <- Reducer 4 (SIMPLE_EDGE)
 Reducer 6 <- Reducer 5 (SIMPLE_EDGE)

Stage-0
 Fetch Operator
 limit:-1
 Stage-1
 Reducer 6
 File Output Operator [FS_24]
 Group By Operator [GBY_22] (rows=1 width=128)
 Output:["_col0","_col1"],aggregations:["count(VALUE._col0)","count(DISTINCT 
KEY._col0:0._col0)"]
 <-Reducer 5 [SIMPLE_EDGE]
 SHUFFLE [RS_21]
 Group By Operator [GBY_20] (rows=5619870 width=2328)
 Output:["_col0","_col1","_col2"],aggregations:["count(1)","count(DISTINCT 
_col0)"],keys:_col0
 Select Operator [SEL_17] (rows=5619870 width=2328)
 Output:["_col0"]
 <-Reducer 4 [SIMPLE_EDGE]
 SHUFFLE [RS_16]
 Select Operator [SEL_14] (rows=5619870 width=2328)
 Output:["_col0","_col3"]
 Filter Operator [FIL_25] (rows=5619870 width=2328)
 predicate:((_col0 / count_window_1) <= 0.8)
 PTF Operator [PTF_13] (rows=16859610 width=2328)
 Function definitions:[{},\\{"name:":"windowingtablefunction","order 
by:":"_col1 ASC NULLS FIRST","partition by:":"_col1"}]
 Select Operator [SEL_12] (rows=16859610 width=2328)
 Output:["_col0","_col1"]
 <-Reducer 3 [SIMPLE_EDGE]
 SHUFFLE [RS_11]
 PartitionCols:_col0
 Select Operator [SEL_10] (rows=16859610 width=2328)
 Output:["_col0","row_number_window_0"]
 PTF Operator [PTF_9] (rows=16859610 width=2328)
 Function definitions:[{},\\{"name:":"windowingtablefunction","order 
by:":"_col2 DESC NULLS LAST","partition by:":"_col0"}]
 Select Operator [SEL_8] (rows=16859610 width=2328)
 Output:["_col0","_col2"]
 <-Reducer 2 [SIMPLE_EDGE]
 SHUFFLE [RS_7]
 PartitionCols:_col0
 Select Operator [SEL_6] (rows=16859610 width=2328)
 Output:["_col0","_col2"]
 Group By Operator [GBY_5] (rows=16859610 width=2328)
 
Output:["_col0","_col1","_col2"],aggregations:["sum(VALUE._col0)"],keys:KEY._col0,
 KEY._col1
 <-Map 1 [SIMPLE_EDGE]
 SHUFFLE [RS_4]
 PartitionCols:_col0, _col1
 Group By Operator [GBY_3] (rows=33719220 width=2328)
 
Output:["_col0","_col1","_col2"],aggregations:["sum(payable_price)"],keys:store_code,
 all_pay_id
 Select Operator [SEL_2] (rows=33719220 width=2328)
 Output:["store_code","all_pay_id","payable_price"]
 TableScan [TS_0] (rows=33719220 width=2328)
 
data_promotion@mdw_user_promotion_shopping_behaivor_four_week_detail,a,Tbl:COMPLETE,Col:NONE,Output:["all_pay_id","store_code","payable_price"]

 

    Reducer 4 show PTF , so I don't know if this result has anything to do with 
PTF? Can you help me.

 

 

  was:
h3. 1.Phenomenon

    Running the same SQL(hive on tez) several times will result in inconsistent 
results . The following results are generated respectively

    1)

     OK

     3951864 808

    2)

     OK

     1822979 353

    But the first result is true.
h3. 2.SQL

select
 count(1),count(distinct store_code)
 from 
 (
 select
 store_code,
 all_pay_id,
 payable_price,
 row_number()over(partition by store_code order by payable_price desc) as rw,
 count(1)over(partition by store_code) as store_user
 from 
 (
 select 
 store_code,
 all_pay_id,
 sum(payable_price) as payable_price
 from data_promotion.mdw_user_promotion_shopping_behaivor_four_week_detail a
 where dt='20190904'
 group by store_code,all_pay_id
 ) a 
 order by rw 
 ) a 
 where rw/store_user<=0.8
 ;

 
h3. 3.Troubleshooting

    I lookup task result from tez web ui,give the result as follows:

           the first: !image-2019-09-10-12-16-45-385.png!

        the second:

!image-2019-09-10-12-18-08-769.png!

      so , The second RECORDS_OUT_INTERMEDIATE_Reducer_4's 

[jira] [Commented] (HIVE-20113) Shuffle avoidance: Disable 1-1 edges for sorted shuffle

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-20113?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16926323#comment-16926323
 ] 

Hive QA commented on HIVE-20113:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
54s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
12s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
46s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  4m 
21s{color} | {color:blue} ql in master has 2246 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
0s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
32s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
9s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m  
9s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
45s{color} | {color:red} ql: The patch generated 1 new + 347 unchanged - 1 
fixed = 348 total (was 348) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  4m 
23s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
2s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 26m  9s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-18516/dev-support/hive-personality.sh
 |
| git revision | master / d244da5 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18516/yetus/diff-checkstyle-ql.txt
 |
| modules | C: ql U: ql |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18516/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Shuffle avoidance: Disable 1-1 edges for sorted shuffle 
> 
>
> Key: HIVE-20113
> URL: https://issues.apache.org/jira/browse/HIVE-20113
> Project: Hive
>  Issue Type: Bug
>  Components: Tez
>Reporter: Gopal V
>Assignee: Gopal V
>Priority: Major
>  Labels: Branch3Candidate
> Attachments: HIVE-20113.1.patch, HIVE-20113.2.patch, 
> HIVE-20113.3.patch, HIVE-20113.4.patch, HIVE-20113.4.patch, 
> HIVE-20113.5.patch, HIVE-20113.6.patch, HIVE-20113.7.patch
>
>
> The sorted shuffle avoidance can have some issues when the shuffle data gets 
> broken up into multiple chunks on disk.
> The 1-1 edge cannot skip the tez final merge - there's no reason for 1-1 to 
> have a final merge at all, it should open a single compressed file and write 
> a single index entry.
> Until the shuffle issue is resolved & a lot more testing, it is prudent to 
> disable the optimization for sorted shuffle edges and stop rewriting the 
> RS(sorted) = = = RS(sorted) into RS(sorted) = = = RS(FORWARD).



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-22079) Post order walker for iterating over expression tree

2019-09-09 Thread Vineet Garg (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22079?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-22079:
---
Status: Open  (was: Patch Available)

> Post order walker for iterating over expression tree
> 
>
> Key: HIVE-22079
> URL: https://issues.apache.org/jira/browse/HIVE-22079
> Project: Hive
>  Issue Type: Improvement
>  Components: Logical Optimizer, Physical Optimizer
>Affects Versions: 4.0.0
>Reporter: Vineet Garg
>Assignee: Vineet Garg
>Priority: Major
> Attachments: HIVE-22079.1.patch, HIVE-22079.2.patch, 
> HIVE-22079.3.patch
>
>
> Current {{DefaultGraphWalker}} is used to iterate over an expression tree. 
> This walker uses hash map to keep track of visited/processed nodes. If an 
> expression tree is large this adds significant overhead due to map lookup.
> For an expression trees we can instead use post order traversal and avoid 
> using map.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-22079) Post order walker for iterating over expression tree

2019-09-09 Thread Vineet Garg (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22079?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-22079:
---
Status: Patch Available  (was: Open)

> Post order walker for iterating over expression tree
> 
>
> Key: HIVE-22079
> URL: https://issues.apache.org/jira/browse/HIVE-22079
> Project: Hive
>  Issue Type: Improvement
>  Components: Logical Optimizer, Physical Optimizer
>Affects Versions: 4.0.0
>Reporter: Vineet Garg
>Assignee: Vineet Garg
>Priority: Major
> Attachments: HIVE-22079.1.patch, HIVE-22079.2.patch, 
> HIVE-22079.3.patch
>
>
> Current {{DefaultGraphWalker}} is used to iterate over an expression tree. 
> This walker uses hash map to keep track of visited/processed nodes. If an 
> expression tree is large this adds significant overhead due to map lookup.
> For an expression trees we can instead use post order traversal and avoid 
> using map.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-22079) Post order walker for iterating over expression tree

2019-09-09 Thread Vineet Garg (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22079?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-22079:
---
Attachment: HIVE-22079.3.patch

> Post order walker for iterating over expression tree
> 
>
> Key: HIVE-22079
> URL: https://issues.apache.org/jira/browse/HIVE-22079
> Project: Hive
>  Issue Type: Improvement
>  Components: Logical Optimizer, Physical Optimizer
>Affects Versions: 4.0.0
>Reporter: Vineet Garg
>Assignee: Vineet Garg
>Priority: Major
> Attachments: HIVE-22079.1.patch, HIVE-22079.2.patch, 
> HIVE-22079.3.patch
>
>
> Current {{DefaultGraphWalker}} is used to iterate over an expression tree. 
> This walker uses hash map to keep track of visited/processed nodes. If an 
> expression tree is large this adds significant overhead due to map lookup.
> For an expression trees we can instead use post order traversal and avoid 
> using map.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-15956) StackOverflowError when drop lots of partitions

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-15956?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16926313#comment-16926313
 ] 

Hive QA commented on HIVE-15956:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12979885/HIVE-15956.2.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 16750 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/18515/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/18515/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-18515/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12979885 - PreCommit-HIVE-Build

> StackOverflowError when drop lots of partitions
> ---
>
> Key: HIVE-15956
> URL: https://issues.apache.org/jira/browse/HIVE-15956
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Affects Versions: 1.3.0, 2.2.0
>Reporter: Niklaus Xiao
>Assignee: Denys Kuzmenko
>Priority: Major
> Attachments: HIVE-15956.2.patch, HIVE-15956.patch
>
>
> Repro steps:
> 1. Create partitioned table and add 1 partitions
> {code}
> create table test_partition(id int) partitioned by (dt int);
> alter table test_partition add partition(dt=1);
> alter table test_partition add partition(dt=3);
> alter table test_partition add partition(dt=4);
> ...
> alter table test_partition add partition(dt=1);
> {code}
> 2. Drop 9000 partitions:
> {code}
> alter table test_partition drop partition(dt<9000);
> {code}
> Step 2 will fail with StackOverflowError:
> {code}
> Exception in thread "pool-7-thread-161" java.lang.StackOverflowError
> at 
> org.datanucleus.query.expression.ExpressionCompiler.isOperator(ExpressionCompiler.java:819)
> at 
> org.datanucleus.query.expression.ExpressionCompiler.compileOrAndExpression(ExpressionCompiler.java:190)
> at 
> org.datanucleus.query.expression.ExpressionCompiler.compileExpression(ExpressionCompiler.java:179)
> at 
> org.datanucleus.query.expression.ExpressionCompiler.compileOrAndExpression(ExpressionCompiler.java:192)
> at 
> org.datanucleus.query.expression.ExpressionCompiler.compileExpression(ExpressionCompiler.java:179)
> at 
> org.datanucleus.query.expression.ExpressionCompiler.compileOrAndExpression(ExpressionCompiler.java:192)
> at 
> org.datanucleus.query.expression.ExpressionCompiler.compileExpression(ExpressionCompiler.java:179)
> {code}
> {code}
> Exception in thread "pool-7-thread-198" java.lang.StackOverflowError
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:83)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-15956) StackOverflowError when drop lots of partitions

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-15956?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16926285#comment-16926285
 ] 

Hive QA commented on HIVE-15956:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
28s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
24s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
19s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  1m  
6s{color} | {color:blue} standalone-metastore/metastore-server in master has 
181 extant Findbugs warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
18s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
28s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
24s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
24s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
19s{color} | {color:red} standalone-metastore/metastore-server: The patch 
generated 1 new + 412 unchanged - 0 fixed = 413 total (was 412) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  1m 
14s{color} | {color:red} standalone-metastore/metastore-server generated 1 new 
+ 179 unchanged - 2 fixed = 180 total (was 181) {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
20s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
12s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 13m 57s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| FindBugs | module:standalone-metastore/metastore-server |
|  |  org.apache.hadoop.hive.metastore.ObjectStore$18.run(List) concatenates 
strings using + in a loop  At ObjectStore.java:in a loop  At 
ObjectStore.java:[line 8771] |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-18515/dev-support/hive-personality.sh
 |
| git revision | master / d244da5 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18515/yetus/diff-checkstyle-standalone-metastore_metastore-server.txt
 |
| findbugs | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18515/yetus/new-findbugs-standalone-metastore_metastore-server.html
 |
| modules | C: standalone-metastore/metastore-server U: 
standalone-metastore/metastore-server |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18515/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> StackOverflowError when drop lots of partitions
> ---
>
> Key: HIVE-15956
> URL: https://issues.apache.org/jira/browse/HIVE-15956
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Affects Versions: 1.3.0, 2.2.0
>Reporter: Niklaus Xiao
>Assignee: Denys Kuzmenko
>Priority: Major
> Attachments: HIVE-15956.2.patch, HIVE-15956.patch
>
>
> Repro steps:
> 1. Create partitioned table and add 1 partitions
> {code}
> create table test_partition(id int) partitioned by (dt int);
> alter table test_partition add partition(dt=1);
> alter table test_partition add partition(dt=3);
> alter table test_partition add partition(dt=4);
> ...
> alter table test_partition add partition(dt=1);
> {code}
> 2. Drop 9000 partitions:
> {code}
> alter table 

[jira] [Commented] (HIVE-22158) HMS Translation layer - Disallow non-ACID MANAGED tables.

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22158?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16926275#comment-16926275
 ] 

Hive QA commented on HIVE-22158:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12979870/HIVE-22158.2.patch

{color:green}SUCCESS:{color} +1 due to 2 test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 16750 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/18514/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/18514/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-18514/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12979870 - PreCommit-HIVE-Build

> HMS Translation layer - Disallow non-ACID MANAGED tables.
> -
>
> Key: HIVE-22158
> URL: https://issues.apache.org/jira/browse/HIVE-22158
> Project: Hive
>  Issue Type: Sub-task
>  Components: Hive
>Affects Versions: 4.0.0
>Reporter: Naveen Gangam
>Assignee: Naveen Gangam
>Priority: Major
> Attachments: HIVE-22158.1.patch, HIVE-22158.1.patch, 
> HIVE-22158.2.patch
>
>
> In the recent commits, we have allowed non-ACID MANAGED tables to be created 
> by clients that have some form of ACID WRITE capabilities. 
> I think it would make sense to disallow this entirely. MANAGED tables should 
> be ACID tables only.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-22158) HMS Translation layer - Disallow non-ACID MANAGED tables.

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22158?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16926241#comment-16926241
 ] 

Hive QA commented on HIVE-22158:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
31s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  6m 
29s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
59s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
35s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  1m  
4s{color} | {color:blue} standalone-metastore/metastore-server in master has 
181 extant Findbugs warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
38s{color} | {color:blue} itests/hive-unit in master has 2 extant Findbugs 
warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
37s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
25s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
 3s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
59s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
59s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
20s{color} | {color:red} standalone-metastore/metastore-server: The patch 
generated 3 new + 402 unchanged - 3 fixed = 405 total (was 405) {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
15s{color} | {color:red} itests/hive-unit: The patch generated 4 new + 151 
unchanged - 0 fixed = 155 total (was 151) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
52s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
38s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
12s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 18m 14s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-18514/dev-support/hive-personality.sh
 |
| git revision | master / d244da5 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18514/yetus/diff-checkstyle-standalone-metastore_metastore-server.txt
 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18514/yetus/diff-checkstyle-itests_hive-unit.txt
 |
| modules | C: standalone-metastore/metastore-server itests/hive-unit U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18514/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> HMS Translation layer - Disallow non-ACID MANAGED tables.
> -
>
> Key: HIVE-22158
> URL: https://issues.apache.org/jira/browse/HIVE-22158
> Project: Hive
>  Issue Type: Sub-task
>  Components: Hive
>Affects Versions: 4.0.0
>Reporter: Naveen Gangam
>Assignee: Naveen Gangam
>Priority: Major
> Attachments: HIVE-22158.1.patch, HIVE-22158.1.patch, 
> HIVE-22158.2.patch
>
>
> In the recent commits, we have allowed non-ACID MANAGED tables to be created 
> by clients 

[jira] [Commented] (HIVE-22126) hive-exec packaging should shade guava

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22126?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16926228#comment-16926228
 ] 

Hive QA commented on HIVE-22126:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12979861/HIVE-22126.01.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 16750 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/18513/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/18513/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-18513/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12979861 - PreCommit-HIVE-Build

> hive-exec packaging should shade guava
> --
>
> Key: HIVE-22126
> URL: https://issues.apache.org/jira/browse/HIVE-22126
> Project: Hive
>  Issue Type: Bug
>Reporter: Vihang Karajgaonkar
>Assignee: Laszlo Pinter
>Priority: Major
> Attachments: HIVE-22126.01.patch
>
>
> The ql/pom.xml includes complete guava library into hive-exec.jar 
> https://github.com/apache/hive/blob/master/ql/pom.xml#L990 This causes a 
> problems for downstream clients of hive which have hive-exec.jar in their 
> classpath since they are pinned to the same guava version as that of hive. 
> We should shade guava classes so that other components which depend on 
> hive-exec can independently use a different version of guava as needed.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-22126) hive-exec packaging should shade guava

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22126?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16926205#comment-16926205
 ] 

Hive QA commented on HIVE-22126:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 10m 
 0s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
2s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
52s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
18s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
56s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
56s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
1s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
54s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
12s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 15m 48s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  xml  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-18513/dev-support/hive-personality.sh
 |
| git revision | master / d244da5 |
| Default Java | 1.8.0_111 |
| modules | C: ql U: ql |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18513/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> hive-exec packaging should shade guava
> --
>
> Key: HIVE-22126
> URL: https://issues.apache.org/jira/browse/HIVE-22126
> Project: Hive
>  Issue Type: Bug
>Reporter: Vihang Karajgaonkar
>Assignee: Laszlo Pinter
>Priority: Major
> Attachments: HIVE-22126.01.patch
>
>
> The ql/pom.xml includes complete guava library into hive-exec.jar 
> https://github.com/apache/hive/blob/master/ql/pom.xml#L990 This causes a 
> problems for downstream clients of hive which have hive-exec.jar in their 
> classpath since they are pinned to the same guava version as that of hive. 
> We should shade guava classes so that other components which depend on 
> hive-exec can independently use a different version of guava as needed.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-21884) Scheduled query support

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-21884?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16926202#comment-16926202
 ] 

Hive QA commented on HIVE-21884:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m  
8s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
16s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  9m 
15s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  4m 
18s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  2m 
16s{color} | {color:blue} standalone-metastore/metastore-common in master has 
32 extant Findbugs warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
31s{color} | {color:blue} common in master has 61 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  1m  
7s{color} | {color:blue} standalone-metastore/metastore-server in master has 
181 extant Findbugs warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m 
40s{color} | {color:blue} ql in master has 2246 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
33s{color} | {color:blue} service in master has 48 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
41s{color} | {color:blue} itests/util in master has 44 extant Findbugs 
warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  9m  
8s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
23s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 10m 
24s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  9m 
20s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  9m 
20s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
13s{color} | {color:red} standalone-metastore/metastore-common: The patch 
generated 1 new + 495 unchanged - 0 fixed = 496 total (was 495) {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
14s{color} | {color:red} common: The patch generated 2 new + 367 unchanged - 0 
fixed = 369 total (was 367) {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
 9s{color} | {color:green} The patch metastore passed checkstyle {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
25s{color} | {color:red} standalone-metastore/metastore-server: The patch 
generated 18 new + 1250 unchanged - 15 fixed = 1268 total (was 1265) {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
39s{color} | {color:red} ql: The patch generated 61 new + 224 unchanged - 0 
fixed = 285 total (was 224) {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
12s{color} | {color:green} service: The patch generated 0 new + 38 unchanged - 
1 fixed = 38 total (was 39) {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  2m 
10s{color} | {color:red} root: The patch generated 82 new + 2421 unchanged - 16 
fixed = 2503 total (was 2437) {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
10s{color} | {color:green} The patch hcatalog-unit passed checkstyle {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
12s{color} | {color:green} The patch util passed checkstyle {color} |
| {color:red}-1{color} | {color:red} whitespace {color} | {color:red}  0m  
0s{color} | {color:red} The patch has 17 line(s) that end in whitespace. Use 
git apply --whitespace=fix <>. Refer 
https://git-scm.com/docs/git-apply {color} |
| {color:red}-1{color} | {color:red} whitespace {color} | {color:red}  0m  
1s{color} | {color:red} The patch 77 line(s) with tabs. {color} |
| {color:green}+1{color} | 

[jira] [Commented] (HIVE-21884) Scheduled query support

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-21884?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16926185#comment-16926185
 ] 

Hive QA commented on HIVE-21884:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12979859/HIVE-21844.07.patch

{color:green}SUCCESS:{color} +1 due to 9 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 65 failed/errored test(s), 16784 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druid_materialized_view_rewrite_ssb]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druid_timeseries]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druid_timestamptz2]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druid_topn] 
(batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_dynamic_partition]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_expressions]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_extractTime]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_floorTime]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_mv] 
(batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_test1]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_test_alter]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_test_insert]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_test_ts]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidKafkaCliDriver.testCliDriver[druidkafkamini_avro]
 (batchId=297)
org.apache.hadoop.hive.cli.TestMiniDruidKafkaCliDriver.testCliDriver[druidkafkamini_basic]
 (batchId=297)
org.apache.hadoop.hive.cli.TestMiniDruidKafkaCliDriver.testCliDriver[druidkafkamini_csv]
 (batchId=297)
org.apache.hadoop.hive.cli.TestMiniDruidKafkaCliDriver.testCliDriver[druidkafkamini_delimited]
 (batchId=297)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[cttl] 
(batchId=157)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[table_nonprintable]
 (batchId=157)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[temp_table_external]
 (batchId=158)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[resourceplan]
 (batchId=175)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[strict_managed_tables_sysdb]
 (batchId=176)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[sysdb] 
(batchId=171)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[sysdb_schq] 
(batchId=176)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[tez_join_hash]
 (batchId=176)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[vector_case_when_2]
 (batchId=176)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[vector_udf_string_to_boolean]
 (batchId=176)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[vectorized_parquet]
 (batchId=176)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[bucket4] 
(batchId=195)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[constprog_partitioner]
 (batchId=195)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[constprog_semijoin]
 (batchId=195)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[infer_bucket_sort_map_operators]
 (batchId=195)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[infer_bucket_sort_merge]
 (batchId=195)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[insert_overwrite_directory2]
 (batchId=195)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[orc_merge4]
 (batchId=195)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[orc_merge_incompat1]
 (batchId=195)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[root_dir_external_table]
 (batchId=195)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning_2]
 (batchId=195)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning_3]
 (batchId=195)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning_4]
 (batchId=195)
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[authorization_schq1]
 (batchId=102)
org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.testCliDriver[ct_noperm_loc]
 (batchId=104)
org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.testCliDriver[ctas_noperm_loc]
 (batchId=104)

[jira] [Assigned] (HIVE-21985) LLAP IO: Log schema evolution incompatibilities at WARN level always

2019-09-09 Thread Mustafa Iman (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21985?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mustafa Iman reassigned HIVE-21985:
---

Assignee: Mustafa Iman

> LLAP IO: Log schema evolution incompatibilities at WARN level always
> 
>
> Key: HIVE-21985
> URL: https://issues.apache.org/jira/browse/HIVE-21985
> Project: Hive
>  Issue Type: Bug
>Reporter: Gopal V
>Assignee: Mustafa Iman
>Priority: Major
>
> When reading incompatible schema, LLAP IO simply skips over the file and does 
> not cache it.
> The logging at WARN level would be useful and simplify the root-cause via 
> logs.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-22076) JDK11: Remove ParallelGC in debug.sh

2019-09-09 Thread Ashutosh Chauhan (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22076?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-22076:

Fix Version/s: 4.0.0
   Resolution: Fixed
   Status: Resolved  (was: Patch Available)

Pushed to master. Thanks, Gopal!

> JDK11: Remove ParallelGC in debug.sh
> 
>
> Key: HIVE-22076
> URL: https://issues.apache.org/jira/browse/HIVE-22076
> Project: Hive
>  Issue Type: Bug
>  Components: Diagnosability
>Affects Versions: 4.0.0
>Reporter: Gopal V
>Assignee: Gopal V
>Priority: Minor
> Fix For: 4.0.0
>
> Attachments: HIVE-22076.1.patch
>
>
> The JDK debug mode no longer depends on ParallelGC 
> This was a workaround for JDK6 bug - 
> https://bugs.java.com/bugdatabase/view_bug.do?bug_id=6862295



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-22076) JDK11: Remove ParallelGC in debug.sh

2019-09-09 Thread Ashutosh Chauhan (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22076?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16926181#comment-16926181
 ] 

Ashutosh Chauhan commented on HIVE-22076:
-

+1

> JDK11: Remove ParallelGC in debug.sh
> 
>
> Key: HIVE-22076
> URL: https://issues.apache.org/jira/browse/HIVE-22076
> Project: Hive
>  Issue Type: Bug
>  Components: Diagnosability
>Affects Versions: 4.0.0
>Reporter: Gopal V
>Assignee: Gopal V
>Priority: Minor
> Attachments: HIVE-22076.1.patch
>
>
> The JDK debug mode no longer depends on ParallelGC 
> This was a workaround for JDK6 bug - 
> https://bugs.java.com/bugdatabase/view_bug.do?bug_id=6862295



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-9223) HiveServer2 on Tez doesn't support concurrent queries within one session

2019-09-09 Thread Tak Lon (Stephen) Wu (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-9223?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16926160#comment-16926160
 ] 

Tak Lon (Stephen) Wu commented on HIVE-9223:


we're also seeing this issue, is this Jira still valid for tracking ?

> HiveServer2 on Tez doesn't support concurrent queries within one session
> 
>
> Key: HIVE-9223
> URL: https://issues.apache.org/jira/browse/HIVE-9223
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2
>Reporter: Pala M Muthaia
>Priority: Major
>
> When a user submits multiple queries in the same HS2 session (using thrift 
> interface) concurrently, the query goes through the same TezSessionState and 
> ends up being submitted to the same Tez AM, and the second query fails with 
> the error "App master already running a DAG"
> Is this by design? I looked into the code, and the comments as well as the 
> code suggest support only for serial execution of queries within the same 
> HiveServer2 session (on tez). 
> This works for CLI environment but in a server, it is plausible that client 
> sends multiple concurrent queries under the same session (e.g: a web app that 
> executes queries for user, such as Cloudera Hue). So shouldn't HS2 on Tez 
> implementation support concurrent queries?



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-22164) Vectorized Limit operator returns wrong number of results with offset

2019-09-09 Thread Richard Zhang (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22164?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16926137#comment-16926137
 ] 

Richard Zhang commented on HIVE-22164:
--

hi Vineet:

Can you do a review of the above pull request?

> Vectorized Limit operator returns wrong number of results with offset
> -
>
> Key: HIVE-22164
> URL: https://issues.apache.org/jira/browse/HIVE-22164
> Project: Hive
>  Issue Type: Bug
>  Components: Hive, llap, Vectorization
>Affects Versions: 4.0.0
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
> Attachments: HIVE-22164.1.patch
>
>
> Vectorized Limit operator returns wrong number of results with offset



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-22059) hive-exec jar doesn't contain (fasterxml) jackson library

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22059?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16926105#comment-16926105
 ] 

Hive QA commented on HIVE-22059:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12979858/HIVE-22059.03.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 632 failed/errored test(s), 16015 tests 
executed
*Failed tests:*
{noformat}
TestSparkCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=113)

[bucketmapjoin4.q,bucket_map_join_spark4.q,union21.q,groupby2_noskew.q,timestamp_2.q,date_join1.q,mergejoins.q,smb_mapjoin_11.q,auto_sortmerge_join_3.q,mapjoin_test_outer.q,vectorization_9.q,merge2.q,groupby6_noskew.q,auto_join_without_localtask.q,multi_join_union.q]
TestSparkCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=114)

[join_cond_pushdown_unqual4.q,union_remove_7.q,join13.q,join_vc.q,groupby_cube1.q,parquet_vectorization_2.q,bucket_map_join_spark2.q,sample3.q,smb_mapjoin_19.q,union23.q,union.q,union31.q,cbo_udf_udaf.q,ptf_decimal.q,bucketmapjoin2.q]
TestSparkCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=115)

[parallel_join1.q,union27.q,union12.q,groupby7_map_multi_single_reducer.q,varchar_join1.q,join7.q,join_reorder4.q,skewjoinopt2.q,bucketsortoptimize_insert_2.q,smb_mapjoin_17.q,script_env_var1.q,groupby7_map.q,bucketsortoptimize_insert_8.q,stats16.q,union20.q]
TestSparkCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=116)

[ptf_general_queries.q,auto_join_reordering_values.q,sample2.q,join1.q,decimal_join.q,mapjoin_subquery2.q,join32_lessSize.q,mapjoin1.q,skewjoinopt18.q,union_remove_18.q,join25.q,groupby3.q,groupby9.q,bucketsortoptimize_insert_6.q,ctas.q]
TestSparkCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=117)

[groupby_map_ppr.q,nullgroup4_multi_distinct.q,join_rc.q,union14.q,order2.q,smb_mapjoin_12.q,vector_cast_constant.q,union_remove_4.q,parquet_vectorization_1.q,auto_join11.q,udaf_collect_set.q,vectorization_12.q,groupby_sort_skew_1_23.q,smb_mapjoin_25.q,skewjoinopt12.q]
TestSparkCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=118)

[skewjoinopt15.q,auto_join18.q,list_bucket_dml_2.q,input1_limit.q,load_dyn_part3.q,union_remove_14.q,auto_sortmerge_join_14.q,auto_sortmerge_join_15.q,union10.q,bucket_map_join_tez2.q,groupby5_map_skew.q,load_dyn_part7.q,join_reorder.q,bucketmapjoin8.q,union34.q]
TestSparkCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=119)

[avro_joins.q,parquet_vectorization_8.q,auto_join14.q,vectorization_14.q,auto_join26.q,stats1.q,cbo_stats.q,union22.q,union_view.q,subquery_views.q,smb_mapjoin_22.q,stats15.q,ptf_matchpath.q,transform_ppr1.q,sample1.q]
TestSparkCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=120)

[limit_pushdown2.q,leftsemijoin_mr.q,parquet_vectorization_0.q,skewjoinopt16.q,bucket3.q,skewjoinopt13.q,auto_sortmerge_join_6.q,bucketmapjoin9.q,auto_join15.q,union_remove_24.q,join22.q,sample4.q,multi_insert_gby.q,join33.q,join_cond_pushdown_unqual2.q]
TestSparkCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=121)

[vector_decimal_aggregate.q,skewjoin_noskew.q,ppd_join3.q,auto_join23.q,join10.q,union_ppr.q,subquery_multi.q,join32.q,input18.q,cbo_simple_select.q,ptf.q,vectorized_nested_mapjoin.q,union18.q,groupby1.q,join_reorder2.q]
TestSparkCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=122)

[order.q,join_merge_multi_expressions.q,skewjoinopt10.q,union_remove_11.q,vectorized_math_funcs.q,vectorization_2.q,skewjoinopt6.q,union_remove_19.q,join14.q,groupby_multi_single_reducer2.q,stats3.q,parquet_join.q,outer_join_ppr.q,join26.q,load_dyn_part10.q]
TestSparkCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=123)

[skewjoinopt3.q,skewjoinopt19.q,timestamp_comparison.q,bucketmapjoin_negative.q,union5.q,insert_into1.q,vectorization_4.q,parquet_vectorization_10.q,vector_left_outer_join.q,decimal_1_1.q,semijoin.q,skewjoinopt9.q,smb_mapjoin_3.q,stats10.q,rcfile_bigdata.q]
TestSparkCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=124)

[parquet_vectorization_limit.q,multi_insert_mixed.q,smb_mapjoin_4.q,join_cond_pushdown_3.q,insert1.q,union_remove_10.q,mapreduce2.q,udf_in_file.q,skewjoinopt5.q,auto_join12.q,skewjoin.q,vectorization_part_project.q,vector_count_distinct.q,nullgroup4.q,parquet_vectorization_12.q]
TestSparkCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=125)


[jira] [Updated] (HIVE-22177) Beeline query for special character like £ does not give back result

2019-09-09 Thread Sanjay Mishra (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22177?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sanjay Mishra updated HIVE-22177:
-
Affects Version/s: 2.3.6

> Beeline query for special character like £ does not give back result
> 
>
> Key: HIVE-22177
> URL: https://issues.apache.org/jira/browse/HIVE-22177
> Project: Hive
>  Issue Type: Bug
>  Components: Beeline
>Affects Versions: 1.1.0, 2.3.6
> Environment: Cloudera 5.15.1 Hive 1.1.0
> beeline and Impala
>Reporter: Sanjay Mishra
>Priority: Major
>
> I have a table in hive with a column having special character like £. When I 
> try to a query as simple as 'SELECT * FROM TABLE WHERE COL = "£"; ', the 
> query does not come back with any result. 
> The table data is UTF-8 encoded. The sql file that contains the above sql and 
> run via beeline is also UTF-8 encoded.
> Interestingly, if I try to create the run the same query using impala against 
> the same table, I get back the columns containing the £ symbol.
> Also when the data from the table is exported (SELECT * FROM THE TABLE and 
> save the output to csv file) the column with special character is shown as 
> NULL.
>  
> Steps to replicate:
>  # create test.sql file with the below content: 
>  # 
> {code:java}
>  CREATE TABLE T1 (test String);
> INSERT INTO T1 VALUES('£');
> {code}
>  # beeline>!run test.sql
>  # create another file test_select.sql with the below content:
>  # 
> {code:java}
> SELECT * FROM T1 WHERE test='£';{code}
>  # beeline>!run test_select.sql
>  # No results will be returned back



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-22059) hive-exec jar doesn't contain (fasterxml) jackson library

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22059?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16926039#comment-16926039
 ] 

Hive QA commented on HIVE-22059:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
37s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
57s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
50s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
16s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
58s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
58s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
1s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
51s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
12s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 12m 54s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  xml  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-18511/dev-support/hive-personality.sh
 |
| git revision | master / 9667f25 |
| Default Java | 1.8.0_111 |
| modules | C: ql U: ql |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18511/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> hive-exec jar doesn't contain (fasterxml) jackson library
> -
>
> Key: HIVE-22059
> URL: https://issues.apache.org/jira/browse/HIVE-22059
> Project: Hive
>  Issue Type: Bug
>Reporter: Zoltan Haindrich
>Assignee: Zoltan Haindrich
>Priority: Major
> Attachments: HIVE-22059.01.patch, HIVE-22059.02.patch, 
> HIVE-22059.03.patch
>
>
> While deploying master branch into a container I've noticed that the jackson 
> libraries are not 100% sure that are available at runtime - this is probably 
> due to the fact that we are still using the "old" codehaus jackson and also 
> the "new" fasterxml one.
> {code:java}
> ]Vertex killed, vertexName=Reducer 2, 
> vertexId=vertex_1564408646590_0005_1_01, diagnostics=[Vertex received Kill in 
> INITED state., Vertex vertex_1564408646590_0005_1_01 [Reducer 2] 
> killed/failed due to:OTHER_VERTEX_FAILURE]DAG did not succeed due to 
> VERTEX_FAILURE. failedVertices:1 killedVertices:1
> INFO : Completed executing 
> command(queryId=vagrant_20190729141949_8d8c7f0d-0ac4-4d76-ba12-6ec01561b040); 
> Time taken: 5.127 seconds
> INFO : Concurrency mode is disabled, not creating a lock manager
> Error: Error while processing statement: FAILED: Execution Error, return code 
> 2 from org.apache.hadoop.hive.ql.exec.tez.TezTask. Vertex failed, 
> vertexName=Map 1, vertexId=vertex_1564408646590_0005_1_00, 
> diagnostics=[Vertex vertex_1564408646590_0005_1_00 [Map 1] killed/failed due 
> to:ROOT_INPUT_INIT_FAILURE, Vertex Input: _dummy_table initializer failed, 
> vertex=vertex_1564408646590_0005_1_00 [Map 1], 
> java.lang.NoClassDefFoundError: com/fasterxml/jackson/databind/ObjectMapper
> at org.apache.hadoop.hive.ql.exec.Utilities.(Utilities.java:226)
> at org.apache.hadoop.hive.ql.io.HiveInputFormat.init(HiveInputFormat.java:428)
> at 
> org.apache.hadoop.hive.ql.io.CombineHiveInputFormat.getSplits(CombineHiveInputFormat.java:508)
> at 
> org.apache.tez.mapreduce.hadoop.MRInputHelpers.generateOldSplits(MRInputHelpers.java:488)
> at 
> 

[jira] [Updated] (HIVE-20113) Shuffle avoidance: Disable 1-1 edges for sorted shuffle

2019-09-09 Thread Vineet Garg (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-20113?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-20113:
---
Attachment: HIVE-20113.7.patch

> Shuffle avoidance: Disable 1-1 edges for sorted shuffle 
> 
>
> Key: HIVE-20113
> URL: https://issues.apache.org/jira/browse/HIVE-20113
> Project: Hive
>  Issue Type: Bug
>  Components: Tez
>Reporter: Gopal V
>Assignee: Gopal V
>Priority: Major
>  Labels: Branch3Candidate
> Attachments: HIVE-20113.1.patch, HIVE-20113.2.patch, 
> HIVE-20113.3.patch, HIVE-20113.4.patch, HIVE-20113.4.patch, 
> HIVE-20113.5.patch, HIVE-20113.6.patch, HIVE-20113.7.patch
>
>
> The sorted shuffle avoidance can have some issues when the shuffle data gets 
> broken up into multiple chunks on disk.
> The 1-1 edge cannot skip the tez final merge - there's no reason for 1-1 to 
> have a final merge at all, it should open a single compressed file and write 
> a single index entry.
> Until the shuffle issue is resolved & a lot more testing, it is prudent to 
> disable the optimization for sorted shuffle edges and stop rewriting the 
> RS(sorted) = = = RS(sorted) into RS(sorted) = = = RS(FORWARD).



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-20113) Shuffle avoidance: Disable 1-1 edges for sorted shuffle

2019-09-09 Thread Vineet Garg (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-20113?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-20113:
---
Status: Open  (was: Patch Available)

> Shuffle avoidance: Disable 1-1 edges for sorted shuffle 
> 
>
> Key: HIVE-20113
> URL: https://issues.apache.org/jira/browse/HIVE-20113
> Project: Hive
>  Issue Type: Bug
>  Components: Tez
>Reporter: Gopal V
>Assignee: Gopal V
>Priority: Major
>  Labels: Branch3Candidate
> Attachments: HIVE-20113.1.patch, HIVE-20113.2.patch, 
> HIVE-20113.3.patch, HIVE-20113.4.patch, HIVE-20113.4.patch, 
> HIVE-20113.5.patch, HIVE-20113.6.patch, HIVE-20113.7.patch
>
>
> The sorted shuffle avoidance can have some issues when the shuffle data gets 
> broken up into multiple chunks on disk.
> The 1-1 edge cannot skip the tez final merge - there's no reason for 1-1 to 
> have a final merge at all, it should open a single compressed file and write 
> a single index entry.
> Until the shuffle issue is resolved & a lot more testing, it is prudent to 
> disable the optimization for sorted shuffle edges and stop rewriting the 
> RS(sorted) = = = RS(sorted) into RS(sorted) = = = RS(FORWARD).



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-20113) Shuffle avoidance: Disable 1-1 edges for sorted shuffle

2019-09-09 Thread Vineet Garg (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-20113?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-20113:
---
Status: Patch Available  (was: Open)

> Shuffle avoidance: Disable 1-1 edges for sorted shuffle 
> 
>
> Key: HIVE-20113
> URL: https://issues.apache.org/jira/browse/HIVE-20113
> Project: Hive
>  Issue Type: Bug
>  Components: Tez
>Reporter: Gopal V
>Assignee: Gopal V
>Priority: Major
>  Labels: Branch3Candidate
> Attachments: HIVE-20113.1.patch, HIVE-20113.2.patch, 
> HIVE-20113.3.patch, HIVE-20113.4.patch, HIVE-20113.4.patch, 
> HIVE-20113.5.patch, HIVE-20113.6.patch, HIVE-20113.7.patch
>
>
> The sorted shuffle avoidance can have some issues when the shuffle data gets 
> broken up into multiple chunks on disk.
> The 1-1 edge cannot skip the tez final merge - there's no reason for 1-1 to 
> have a final merge at all, it should open a single compressed file and write 
> a single index entry.
> Until the shuffle issue is resolved & a lot more testing, it is prudent to 
> disable the optimization for sorted shuffle edges and stop rewriting the 
> RS(sorted) = = = RS(sorted) into RS(sorted) = = = RS(FORWARD).



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-15956) StackOverflowError when drop lots of partitions

2019-09-09 Thread Denys Kuzmenko (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-15956?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Denys Kuzmenko updated HIVE-15956:
--
Attachment: HIVE-15956.2.patch

> StackOverflowError when drop lots of partitions
> ---
>
> Key: HIVE-15956
> URL: https://issues.apache.org/jira/browse/HIVE-15956
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Affects Versions: 1.3.0, 2.2.0
>Reporter: Niklaus Xiao
>Assignee: Denys Kuzmenko
>Priority: Major
> Attachments: HIVE-15956.2.patch, HIVE-15956.patch
>
>
> Repro steps:
> 1. Create partitioned table and add 1 partitions
> {code}
> create table test_partition(id int) partitioned by (dt int);
> alter table test_partition add partition(dt=1);
> alter table test_partition add partition(dt=3);
> alter table test_partition add partition(dt=4);
> ...
> alter table test_partition add partition(dt=1);
> {code}
> 2. Drop 9000 partitions:
> {code}
> alter table test_partition drop partition(dt<9000);
> {code}
> Step 2 will fail with StackOverflowError:
> {code}
> Exception in thread "pool-7-thread-161" java.lang.StackOverflowError
> at 
> org.datanucleus.query.expression.ExpressionCompiler.isOperator(ExpressionCompiler.java:819)
> at 
> org.datanucleus.query.expression.ExpressionCompiler.compileOrAndExpression(ExpressionCompiler.java:190)
> at 
> org.datanucleus.query.expression.ExpressionCompiler.compileExpression(ExpressionCompiler.java:179)
> at 
> org.datanucleus.query.expression.ExpressionCompiler.compileOrAndExpression(ExpressionCompiler.java:192)
> at 
> org.datanucleus.query.expression.ExpressionCompiler.compileExpression(ExpressionCompiler.java:179)
> at 
> org.datanucleus.query.expression.ExpressionCompiler.compileOrAndExpression(ExpressionCompiler.java:192)
> at 
> org.datanucleus.query.expression.ExpressionCompiler.compileExpression(ExpressionCompiler.java:179)
> {code}
> {code}
> Exception in thread "pool-7-thread-198" java.lang.StackOverflowError
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:83)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Assigned] (HIVE-15956) StackOverflowError when drop lots of partitions

2019-09-09 Thread Denys Kuzmenko (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-15956?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Denys Kuzmenko reassigned HIVE-15956:
-

Assignee: Denys Kuzmenko  (was: Niklaus Xiao)

> StackOverflowError when drop lots of partitions
> ---
>
> Key: HIVE-15956
> URL: https://issues.apache.org/jira/browse/HIVE-15956
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Affects Versions: 1.3.0, 2.2.0
>Reporter: Niklaus Xiao
>Assignee: Denys Kuzmenko
>Priority: Major
> Attachments: HIVE-15956.patch
>
>
> Repro steps:
> 1. Create partitioned table and add 1 partitions
> {code}
> create table test_partition(id int) partitioned by (dt int);
> alter table test_partition add partition(dt=1);
> alter table test_partition add partition(dt=3);
> alter table test_partition add partition(dt=4);
> ...
> alter table test_partition add partition(dt=1);
> {code}
> 2. Drop 9000 partitions:
> {code}
> alter table test_partition drop partition(dt<9000);
> {code}
> Step 2 will fail with StackOverflowError:
> {code}
> Exception in thread "pool-7-thread-161" java.lang.StackOverflowError
> at 
> org.datanucleus.query.expression.ExpressionCompiler.isOperator(ExpressionCompiler.java:819)
> at 
> org.datanucleus.query.expression.ExpressionCompiler.compileOrAndExpression(ExpressionCompiler.java:190)
> at 
> org.datanucleus.query.expression.ExpressionCompiler.compileExpression(ExpressionCompiler.java:179)
> at 
> org.datanucleus.query.expression.ExpressionCompiler.compileOrAndExpression(ExpressionCompiler.java:192)
> at 
> org.datanucleus.query.expression.ExpressionCompiler.compileExpression(ExpressionCompiler.java:179)
> at 
> org.datanucleus.query.expression.ExpressionCompiler.compileOrAndExpression(ExpressionCompiler.java:192)
> at 
> org.datanucleus.query.expression.ExpressionCompiler.compileExpression(ExpressionCompiler.java:179)
> {code}
> {code}
> Exception in thread "pool-7-thread-198" java.lang.StackOverflowError
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:83)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> at 
> org.datanucleus.query.expression.DyadicExpression.bind(DyadicExpression.java:87)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-22055) select count gives incorrect result after loading data from text file

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22055?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925926#comment-16925926
 ] 

Hive QA commented on HIVE-22055:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12979856/HIVE-22055.2.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 16751 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/18510/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/18510/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-18510/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12979856 - PreCommit-HIVE-Build

> select count gives incorrect result after loading data from text file
> -
>
> Key: HIVE-22055
> URL: https://issues.apache.org/jira/browse/HIVE-22055
> Project: Hive
>  Issue Type: Task
>  Components: Hive
>Reporter: Attila Magyar
>Assignee: Attila Magyar
>Priority: Major
> Attachments: HIVE-22055.2.patch
>
>
> Add one more load to mm_loaddata.q:
> Load data 3 times (both kv1.txt and kv2.txt contains 500 records)
> {code:java}
> create table load0_mm (key string, value string) stored as textfile 
> tblproperties("transactional"="true", 
> "transactional_properties"="insert_only");
> load data local inpath '../../data/files/kv1.txt' into table load0_mm;
> select count(1) from load0_mm;
> load data local inpath '../../data/files/kv2.txt' into table load0_mm;
> select count(1) from load0_mm;
> load data local inpath '../../data/files/kv2.txt' into table load0_mm;
> select count(1) from load0_mm;{code}
> Expected output
> {code:java}
> PREHOOK: query: load data local inpath '../../data/files/kv2.txt' into table 
> load0_mm
> PREHOOK: type: LOAD
>  A masked pattern was here 
> PREHOOK: Output: default@load0_mm
> POSTHOOK: query: load data local inpath '../../data/files/kv2.txt' into table 
> load0_mm
> POSTHOOK: type: LOAD
>  A masked pattern was here 
> POSTHOOK: Output: default@load0_mm
> PREHOOK: query: select count(1) from load0_mm
> PREHOOK: type: QUERY
> PREHOOK: Input: default@load0_mm
>  A masked pattern was here 
> POSTHOOK: query: select count(1) from load0_mm
> POSTHOOK: type: QUERY
> POSTHOOK: Input: default@load0_mm
>  A masked pattern was here 
> 1500{code}
> Got:
> [ERROR]   TestMiniLlapLocalCliDriver.testCliDriver:59 Client Execution 
> succeeded but contained differences (error code = 1) after executing 
> mm_loaddata.q
> 63c63
> < 1480
> —
> > 1500
>  



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-22055) select count gives incorrect result after loading data from text file

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22055?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925882#comment-16925882
 ] 

Hive QA commented on HIVE-22055:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
16s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  5m 
40s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
6s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
41s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m  
8s{color} | {color:blue} ql in master has 2246 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
34s{color} | {color:blue} llap-server in master has 83 extant Findbugs 
warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
57s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
22s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
26s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
6s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m  
6s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
42s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
53s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
56s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
10s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 22m 31s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-18510/dev-support/hive-personality.sh
 |
| git revision | master / 9667f25 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| modules | C: ql llap-server itests U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18510/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> select count gives incorrect result after loading data from text file
> -
>
> Key: HIVE-22055
> URL: https://issues.apache.org/jira/browse/HIVE-22055
> Project: Hive
>  Issue Type: Task
>  Components: Hive
>Reporter: Attila Magyar
>Assignee: Attila Magyar
>Priority: Major
> Attachments: HIVE-22055.2.patch
>
>
> Add one more load to mm_loaddata.q:
> Load data 3 times (both kv1.txt and kv2.txt contains 500 records)
> {code:java}
> create table load0_mm (key string, value string) stored as textfile 
> tblproperties("transactional"="true", 
> "transactional_properties"="insert_only");
> load data local inpath '../../data/files/kv1.txt' into table load0_mm;
> select count(1) from load0_mm;
> load data local inpath '../../data/files/kv2.txt' into table load0_mm;
> select count(1) from load0_mm;
> load data local inpath '../../data/files/kv2.txt' into table load0_mm;
> select count(1) from load0_mm;{code}
> Expected output
> {code:java}
> PREHOOK: query: load data local inpath '../../data/files/kv2.txt' into table 
> load0_mm
> PREHOOK: type: LOAD
>  A masked 

[jira] [Commented] (HIVE-20683) Add the Ability to push Dynamic Between and Bloom filters to Druid

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-20683?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925857#comment-16925857
 ] 

Hive QA commented on HIVE-20683:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12979857/HIVE-20683.10.patch

{color:green}SUCCESS:{color} +1 due to 4 test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 16751 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/18509/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/18509/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-18509/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12979857 - PreCommit-HIVE-Build

> Add the Ability to push Dynamic Between and Bloom filters to Druid
> --
>
> Key: HIVE-20683
> URL: https://issues.apache.org/jira/browse/HIVE-20683
> Project: Hive
>  Issue Type: New Feature
>  Components: Druid integration
>Reporter: Nishant Bangarwa
>Assignee: Nishant Bangarwa
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-20683.1.patch, HIVE-20683.10.patch, 
> HIVE-20683.2.patch, HIVE-20683.3.patch, HIVE-20683.4.patch, 
> HIVE-20683.5.patch, HIVE-20683.6.patch, HIVE-20683.8.patch, HIVE-20683.patch
>
>  Time Spent: 5h 10m
>  Remaining Estimate: 0h
>
> For optimizing joins, Hive generates BETWEEN filter with min-max and BLOOM 
> filter for filtering one side of semi-join.
> Druid 0.13.0 will have support for Bloom filters (Added via 
> https://github.com/apache/incubator-druid/pull/6222)
> Implementation details - 
> # Hive generates and passes the filters as part of 'filterExpr' in TableScan. 
> # DruidQueryBasedRecordReader gets this filter passed as part of the conf. 
> # During execution phase, before sending the query to druid in 
> DruidQueryBasedRecordReader we will deserialize this filter, translate it 
> into a DruidDimFilter and add it to existing DruidQuery.  Tez executor 
> already ensures that when we start reading results from the record reader, 
> all the dynamic values are initialized. 
> # Explaining a druid query also prints the query sent to druid as 
> {{druid.json.query}}. We also need to make sure to update the druid query 
> with the filters. During explain we do not have the actual values for the 
> dynamic values, so instead of values we will print the dynamic expression 
> itself as part of druid query. 
> Note:- This work needs druid to be updated to version 0.13.0



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-20683) Add the Ability to push Dynamic Between and Bloom filters to Druid

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-20683?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925785#comment-16925785
 ] 

Hive QA commented on HIVE-20683:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m  
5s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  5m 
59s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  5m 
57s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  2m 
15s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m  
8s{color} | {color:blue} ql in master has 2246 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
23s{color} | {color:blue} druid-handler in master has 3 extant Findbugs 
warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
16s{color} | {color:blue} itests/qtest-druid in master has 7 extant Findbugs 
warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  6m  
2s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
20s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
29s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  5m 
57s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  5m 
57s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m  
8s{color} | {color:red} druid-handler: The patch generated 1 new + 0 unchanged 
- 0 fixed = 1 total (was 0) {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  1m 
27s{color} | {color:red} root: The patch generated 1 new + 35 unchanged - 0 
fixed = 36 total (was 35) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
2s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  4m  
2s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  6m  
0s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
10s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 52m  0s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  xml  compile  findbugs  
checkstyle  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-18509/dev-support/hive-personality.sh
 |
| git revision | master / 9667f25 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18509/yetus/diff-checkstyle-druid-handler.txt
 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18509/yetus/diff-checkstyle-root.txt
 |
| modules | C: ql druid-handler . itests itests/qtest-druid U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18509/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Add the Ability to push Dynamic Between and Bloom filters to Druid
> --
>
> Key: HIVE-20683
> URL: https://issues.apache.org/jira/browse/HIVE-20683
> Project: Hive
>  Issue Type: New Feature
>  Components: Druid integration
>Reporter: Nishant Bangarwa
>Assignee: Nishant Bangarwa

[jira] [Updated] (HIVE-22158) HMS Translation layer - Disallow non-ACID MANAGED tables.

2019-09-09 Thread Naveen Gangam (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22158?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Naveen Gangam updated HIVE-22158:
-
Status: Patch Available  (was: Open)

Attached a patch to fix the failing test. The test failure is a resultant of a 
managed table being converted to external table. The replication code is not 
setup with the appropriate config properties to be able to work with external 
table locations.

> HMS Translation layer - Disallow non-ACID MANAGED tables.
> -
>
> Key: HIVE-22158
> URL: https://issues.apache.org/jira/browse/HIVE-22158
> Project: Hive
>  Issue Type: Sub-task
>  Components: Hive
>Affects Versions: 4.0.0
>Reporter: Naveen Gangam
>Assignee: Naveen Gangam
>Priority: Major
> Attachments: HIVE-22158.1.patch, HIVE-22158.1.patch, 
> HIVE-22158.2.patch
>
>
> In the recent commits, we have allowed non-ACID MANAGED tables to be created 
> by clients that have some form of ACID WRITE capabilities. 
> I think it would make sense to disallow this entirely. MANAGED tables should 
> be ACID tables only.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-22158) HMS Translation layer - Disallow non-ACID MANAGED tables.

2019-09-09 Thread Naveen Gangam (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22158?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Naveen Gangam updated HIVE-22158:
-
Attachment: HIVE-22158.2.patch

> HMS Translation layer - Disallow non-ACID MANAGED tables.
> -
>
> Key: HIVE-22158
> URL: https://issues.apache.org/jira/browse/HIVE-22158
> Project: Hive
>  Issue Type: Sub-task
>  Components: Hive
>Affects Versions: 4.0.0
>Reporter: Naveen Gangam
>Assignee: Naveen Gangam
>Priority: Major
> Attachments: HIVE-22158.1.patch, HIVE-22158.1.patch, 
> HIVE-22158.2.patch
>
>
> In the recent commits, we have allowed non-ACID MANAGED tables to be created 
> by clients that have some form of ACID WRITE capabilities. 
> I think it would make sense to disallow this entirely. MANAGED tables should 
> be ACID tables only.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-22158) HMS Translation layer - Disallow non-ACID MANAGED tables.

2019-09-09 Thread Naveen Gangam (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22158?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Naveen Gangam updated HIVE-22158:
-
Status: Open  (was: Patch Available)

> HMS Translation layer - Disallow non-ACID MANAGED tables.
> -
>
> Key: HIVE-22158
> URL: https://issues.apache.org/jira/browse/HIVE-22158
> Project: Hive
>  Issue Type: Sub-task
>  Components: Hive
>Affects Versions: 4.0.0
>Reporter: Naveen Gangam
>Assignee: Naveen Gangam
>Priority: Major
> Attachments: HIVE-22158.1.patch, HIVE-22158.1.patch, 
> HIVE-22158.2.patch
>
>
> In the recent commits, we have allowed non-ACID MANAGED tables to be created 
> by clients that have some form of ACID WRITE capabilities. 
> I think it would make sense to disallow this entirely. MANAGED tables should 
> be ACID tables only.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-22083) Values of tag order cannot be null, so it can be "byte" instead of "Byte"

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22083?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925742#comment-16925742
 ] 

Hive QA commented on HIVE-22083:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12979855/HIVE-22083.4.patch

{color:red}ERROR:{color} -1 due to build exiting with an error

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/18508/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/18508/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-18508/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hiveptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ date '+%Y-%m-%d %T.%3N'
2019-09-09 14:36:27.499
+ [[ -n /usr/lib/jvm/java-8-openjdk-amd64 ]]
+ export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ export 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'MAVEN_OPTS=-Xmx1g '
+ MAVEN_OPTS='-Xmx1g '
+ cd /data/hiveptest/working/
+ tee /data/hiveptest/logs/PreCommit-HIVE-Build-18508/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ git = \s\v\n ]]
+ [[ git = \g\i\t ]]
+ [[ -z master ]]
+ [[ -d apache-github-source-source ]]
+ [[ ! -d apache-github-source-source/.git ]]
+ [[ ! -d apache-github-source-source ]]
+ date '+%Y-%m-%d %T.%3N'
2019-09-09 14:36:27.514
+ cd apache-github-source-source
+ git fetch origin
+ git reset --hard HEAD
HEAD is now at 9667f25 HIVE-21942: Remove useless MetastoreType enum from 
AbstractCliConfig (Imre Molnar via Laszlo Bodor)
+ git clean -f -d
+ git checkout master
Already on 'master'
Your branch is up-to-date with 'origin/master'.
+ git reset --hard origin/master
HEAD is now at 9667f25 HIVE-21942: Remove useless MetastoreType enum from 
AbstractCliConfig (Imre Molnar via Laszlo Bodor)
+ git merge --ff-only origin/master
Already up-to-date.
+ date '+%Y-%m-%d %T.%3N'
2019-09-09 14:36:29.097
+ rm -rf ../yetus_PreCommit-HIVE-Build-18508
+ mkdir ../yetus_PreCommit-HIVE-Build-18508
+ git gc
+ cp -R . ../yetus_PreCommit-HIVE-Build-18508
+ mkdir /data/hiveptest/logs/PreCommit-HIVE-Build-18508/yetus
+ patchCommandPath=/data/hiveptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hiveptest/working/scratch/build.patch
+ [[ -f /data/hiveptest/working/scratch/build.patch ]]
+ chmod +x /data/hiveptest/working/scratch/smart-apply-patch.sh
+ /data/hiveptest/working/scratch/smart-apply-patch.sh 
/data/hiveptest/working/scratch/build.patch
error: a/ql/src/java/org/apache/hadoop/hive/ql/exec/CommonJoinOperator.java: 
does not exist in index
error: a/ql/src/java/org/apache/hadoop/hive/ql/exec/HashTableSinkOperator.java: 
does not exist in index
error: a/ql/src/java/org/apache/hadoop/hive/ql/exec/JoinUtil.java: does not 
exist in index
error: a/ql/src/java/org/apache/hadoop/hive/ql/exec/SkewJoinHandler.java: does 
not exist in index
error: 
a/ql/src/java/org/apache/hadoop/hive/ql/exec/SparkHashTableSinkOperator.java: 
does not exist in index
error: 
a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorMapJoinBaseOperator.java:
 does not exist in index
error: 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/BucketingSortingReduceSinkOptimizer.java:
 does not exist in index
error: 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/FixedBucketPruningOptimizer.java:
 does not exist in index
error: a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/JoinReorder.java: does 
not exist in index
error: a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/MapJoinProcessor.java: 
does not exist in index
error: 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/correlation/CorrelationOptimizer.java:
 does not exist in index
error: 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/BucketingSortingOpProcFactory.java:
 does not exist in index
error: 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/GenMRSkewJoinProcessor.java:
 does not exist in index
error: 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/GenSparkSkewJoinProcessor.java:
 does not exist in index
error: 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/SortMergeJoinTaskDispatcher.java:
 does not exist in index
error: 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java: 
does not exist in index
error: a/ql/src/java/org/apache/hadoop/hive/ql/plan/HashTableSinkDesc.java: 
does not exist in index
error: a/ql/src/java/org/apache/hadoop/hive/ql/plan/JoinDesc.java: does not 
exist in index
error: 

[jira] [Updated] (HIVE-22126) hive-exec packaging should shade guava

2019-09-09 Thread Laszlo Pinter (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22126?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Laszlo Pinter updated HIVE-22126:
-
Status: Patch Available  (was: Open)

> hive-exec packaging should shade guava
> --
>
> Key: HIVE-22126
> URL: https://issues.apache.org/jira/browse/HIVE-22126
> Project: Hive
>  Issue Type: Bug
>Reporter: Vihang Karajgaonkar
>Assignee: Laszlo Pinter
>Priority: Major
> Attachments: HIVE-22126.01.patch
>
>
> The ql/pom.xml includes complete guava library into hive-exec.jar 
> https://github.com/apache/hive/blob/master/ql/pom.xml#L990 This causes a 
> problems for downstream clients of hive which have hive-exec.jar in their 
> classpath since they are pinned to the same guava version as that of hive. 
> We should shade guava classes so that other components which depend on 
> hive-exec can independently use a different version of guava as needed.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-22126) hive-exec packaging should shade guava

2019-09-09 Thread Laszlo Pinter (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22126?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Laszlo Pinter updated HIVE-22126:
-
Attachment: HIVE-22126.01.patch

> hive-exec packaging should shade guava
> --
>
> Key: HIVE-22126
> URL: https://issues.apache.org/jira/browse/HIVE-22126
> Project: Hive
>  Issue Type: Bug
>Reporter: Vihang Karajgaonkar
>Assignee: Laszlo Pinter
>Priority: Major
> Attachments: HIVE-22126.01.patch
>
>
> The ql/pom.xml includes complete guava library into hive-exec.jar 
> https://github.com/apache/hive/blob/master/ql/pom.xml#L990 This causes a 
> problems for downstream clients of hive which have hive-exec.jar in their 
> classpath since they are pinned to the same guava version as that of hive. 
> We should shade guava classes so that other components which depend on 
> hive-exec can independently use a different version of guava as needed.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-21508) ClassCastException when initializing HiveMetaStoreClient on JDK10 or newer

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-21508?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925740#comment-16925740
 ] 

Hive QA commented on HIVE-21508:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12979845/HIVE-21508.1.branch-2.3.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 8 failed/errored test(s), 10587 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[dynamic_semijoin_reduction]
 (batchId=151)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[mergejoin] 
(batchId=153)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[vectorized_dynamic_semijoin_reduction2]
 (batchId=150)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[vectorized_dynamic_semijoin_reduction]
 (batchId=142)
org.apache.hadoop.hive.cli.TestSparkCliDriver.org.apache.hadoop.hive.cli.TestSparkCliDriver
 (batchId=110)
org.apache.hadoop.hive.ql.TestTxnCommands2.testNonAcidToAcidConversion02 
(batchId=264)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdate.testNonAcidToAcidConversion02
 (batchId=276)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testNonAcidToAcidConversion02
 (batchId=273)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/18507/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/18507/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-18507/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 8 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12979845 - PreCommit-HIVE-Build

> ClassCastException when initializing HiveMetaStoreClient on JDK10 or newer
> --
>
> Key: HIVE-21508
> URL: https://issues.apache.org/jira/browse/HIVE-21508
> Project: Hive
>  Issue Type: Bug
>  Components: Clients
>Affects Versions: 3.2.0, 2.3.4
>Reporter: Adar Dembo
>Assignee: Ana Jalba
>Priority: Major
> Fix For: 4.0.0, 2.3.7
>
> Attachments: HIVE-21508.1.branch-2.3.patch, HIVE-21508.patch
>
>
> There's this block of code in {{HiveMetaStoreClient:resolveUris}} (called 
> from the constructor) on master:
> {noformat}
>   private URI metastoreUris[];
>   ...
>   if (MetastoreConf.getVar(conf, 
> ConfVars.THRIFT_URI_SELECTION).equalsIgnoreCase("RANDOM")) {
> List uriList = Arrays.asList(metastoreUris);
> Collections.shuffle(uriList);
> metastoreUris = (URI[]) uriList.toArray();
>   }
> {noformat}
> The cast to {{URI[]}} throws a {{ClassCastException}} beginning with JDK 10, 
> possibly with JDK 9 as well. Note that {{THRIFT_URI_SELECTION}} defaults to 
> {{RANDOM}} so this should affect anyone who creates a 
> {{HiveMetaStoreClient}}. On master this can be overridden with {{SEQUENTIAL}} 
> to avoid the broken case; I'm working against 2.3.4 where there's no such 
> workaround.
> [Here's|https://stackoverflow.com/questions/51372788/array-cast-java-8-vs-java-9]
>  a StackOverflow post that explains the issue in more detail. Interestingly, 
> the author described the issue in the context of the HMS; not sure why there 
> was no follow up with a Hive bug report.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Assigned] (HIVE-22126) hive-exec packaging should shade guava

2019-09-09 Thread Laszlo Pinter (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22126?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Laszlo Pinter reassigned HIVE-22126:


Assignee: Laszlo Pinter

> hive-exec packaging should shade guava
> --
>
> Key: HIVE-22126
> URL: https://issues.apache.org/jira/browse/HIVE-22126
> Project: Hive
>  Issue Type: Bug
>Reporter: Vihang Karajgaonkar
>Assignee: Laszlo Pinter
>Priority: Major
>
> The ql/pom.xml includes complete guava library into hive-exec.jar 
> https://github.com/apache/hive/blob/master/ql/pom.xml#L990 This causes a 
> problems for downstream clients of hive which have hive-exec.jar in their 
> classpath since they are pinned to the same guava version as that of hive. 
> We should shade guava classes so that other components which depend on 
> hive-exec can independently use a different version of guava as needed.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-21884) Scheduled query support

2019-09-09 Thread Zoltan Haindrich (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21884?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltan Haindrich updated HIVE-21884:

Attachment: HIVE-21844.07.patch

> Scheduled query support
> ---
>
> Key: HIVE-21884
> URL: https://issues.apache.org/jira/browse/HIVE-21884
> Project: Hive
>  Issue Type: Improvement
>Reporter: Zoltan Haindrich
>Assignee: Zoltan Haindrich
>Priority: Major
> Attachments: HIVE-21844.04.patch, HIVE-21844.05.patch, 
> HIVE-21844.06.patch, HIVE-21844.07.patch, HIVE-21884.01.patch, 
> HIVE-21884.02.patch, HIVE-21884.03.patch, Scheduled queries2.pdf
>
>
> design document:
> https://docs.google.com/document/d/1mJSFdJi_1cbxJTXC9QvGw2rQ3zzJkNfxOO6b5esmyCE/edit#
> in case the google doc is not reachable:  [^Scheduled queries2.pdf] 



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-22059) hive-exec jar doesn't contain (fasterxml) jackson library

2019-09-09 Thread Zoltan Haindrich (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22059?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltan Haindrich updated HIVE-22059:

Attachment: HIVE-22059.03.patch

> hive-exec jar doesn't contain (fasterxml) jackson library
> -
>
> Key: HIVE-22059
> URL: https://issues.apache.org/jira/browse/HIVE-22059
> Project: Hive
>  Issue Type: Bug
>Reporter: Zoltan Haindrich
>Assignee: Zoltan Haindrich
>Priority: Major
> Attachments: HIVE-22059.01.patch, HIVE-22059.02.patch, 
> HIVE-22059.03.patch
>
>
> While deploying master branch into a container I've noticed that the jackson 
> libraries are not 100% sure that are available at runtime - this is probably 
> due to the fact that we are still using the "old" codehaus jackson and also 
> the "new" fasterxml one.
> {code:java}
> ]Vertex killed, vertexName=Reducer 2, 
> vertexId=vertex_1564408646590_0005_1_01, diagnostics=[Vertex received Kill in 
> INITED state., Vertex vertex_1564408646590_0005_1_01 [Reducer 2] 
> killed/failed due to:OTHER_VERTEX_FAILURE]DAG did not succeed due to 
> VERTEX_FAILURE. failedVertices:1 killedVertices:1
> INFO : Completed executing 
> command(queryId=vagrant_20190729141949_8d8c7f0d-0ac4-4d76-ba12-6ec01561b040); 
> Time taken: 5.127 seconds
> INFO : Concurrency mode is disabled, not creating a lock manager
> Error: Error while processing statement: FAILED: Execution Error, return code 
> 2 from org.apache.hadoop.hive.ql.exec.tez.TezTask. Vertex failed, 
> vertexName=Map 1, vertexId=vertex_1564408646590_0005_1_00, 
> diagnostics=[Vertex vertex_1564408646590_0005_1_00 [Map 1] killed/failed due 
> to:ROOT_INPUT_INIT_FAILURE, Vertex Input: _dummy_table initializer failed, 
> vertex=vertex_1564408646590_0005_1_00 [Map 1], 
> java.lang.NoClassDefFoundError: com/fasterxml/jackson/databind/ObjectMapper
> at org.apache.hadoop.hive.ql.exec.Utilities.(Utilities.java:226)
> at org.apache.hadoop.hive.ql.io.HiveInputFormat.init(HiveInputFormat.java:428)
> at 
> org.apache.hadoop.hive.ql.io.CombineHiveInputFormat.getSplits(CombineHiveInputFormat.java:508)
> at 
> org.apache.tez.mapreduce.hadoop.MRInputHelpers.generateOldSplits(MRInputHelpers.java:488)
> at 
> org.apache.tez.mapreduce.hadoop.MRInputHelpers.generateInputSplitsToMem(MRInputHelpers.java:337)
> at 
> org.apache.tez.mapreduce.common.MRInputAMSplitGenerator.initialize(MRInputAMSplitGenerator.java:122)
> at 
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:278)
> at 
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:269)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:422)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
> at 
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:269)
> at 
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:253)
> at 
> com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:108)
> at 
> com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:41)
> at 
> com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:77)
> at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> at java.lang.Thread.run(Thread.java:748)
> Caused by: java.lang.ClassNotFoundException: 
> com.fasterxml.jackson.databind.ObjectMapper
> at java.net.URLClassLoader.findClass(URLClassLoader.java:382)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:349)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
> ... 19 more
> ]Vertex killed, vertexName=Reducer 2, 
> vertexId=vertex_1564408646590_0005_1_01, diagnostics=[Vertex received Kill in 
> INITED state., Vertex vertex_1564408646590_0005_1_01 [Reducer 2] 
> killed/failed due to:OTHER_VERTEX_FAILURE]DAG did not succeed due to 
> VERTEX_FAILURE. failedVertices:1 killedVertices:1 (state=08S01,code=2)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-22107) Correlated subquery producing wrong schema

2019-09-09 Thread Jira


[ 
https://issues.apache.org/jira/browse/HIVE-22107?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925716#comment-16925716
 ] 

Hankó Gergely commented on HIVE-22107:
--

Hi, [~vgarg], the file of the added test case 
([subquery_notexists.q|https://github.com/apache/hive/pull/755/files#diff-386c2b9c0e9b16261791f06236dbd7cb])
 is not added to any set in 
itests/src/test/resources/testconfiguration.properties.

Is this intentional?

> Correlated subquery producing wrong schema
> --
>
> Key: HIVE-22107
> URL: https://issues.apache.org/jira/browse/HIVE-22107
> Project: Hive
>  Issue Type: Bug
>  Components: Logical Optimizer
>Affects Versions: 4.0.0
>Reporter: Vineet Garg
>Assignee: Vineet Garg
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0
>
> Attachments: HIVE-22107.1.patch, HIVE-22107.2.patch, 
> HIVE-22107.3.patch, HIVE-22107.4.patch, HIVE-22107.5.patch
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> *Repro*
> {code:sql}
> create table test(id int, name string,dept string);
> insert into test values(1,'a','it'),(2,'b','eee'),(NULL, 'c', 'cse');
> select distinct 'empno' as eid, a.id from test a where NOT EXISTS (select 
> c.id from test c where a.id=c.id);
> {code}
> {code}
> +---++
> |  eid  |  a.id  |
> +---++
> | NULL  | empno  |
> +---++
> {code}



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Work logged] (HIVE-20683) Add the Ability to push Dynamic Between and Bloom filters to Druid

2019-09-09 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-20683?focusedWorklogId=308883=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-308883
 ]

ASF GitHub Bot logged work on HIVE-20683:
-

Author: ASF GitHub Bot
Created on: 09/Sep/19 13:57
Start Date: 09/Sep/19 13:57
Worklog Time Spent: 10m 
  Work Description: nishantmonu51 commented on pull request #723: 
[HIVE-20683] Add the Ability to push Dynamic Between and Bloom filters to Druid
URL: https://github.com/apache/hive/pull/723#discussion_r322256584
 
 

 ##
 File path: 
druid-handler/src/java/org/apache/hadoop/hive/druid/DruidStorageHandlerUtils.java
 ##
 @@ -894,4 +945,255 @@ public static IndexSpec getIndexSpec(Configuration jc) {
 ImmutableList aggregatorFactories = 
aggregatorFactoryBuilder.build();
 return Pair.of(dimensions, aggregatorFactories.toArray(new 
AggregatorFactory[0]));
   }
+
+  // Druid only supports String,Long,Float,Double selectors
+  private static Set druidSupportedTypeInfos = 
ImmutableSet.of(
+  TypeInfoFactory.stringTypeInfo, TypeInfoFactory.charTypeInfo,
+  TypeInfoFactory.varcharTypeInfo, TypeInfoFactory.byteTypeInfo,
+  TypeInfoFactory.intTypeInfo, TypeInfoFactory.longTypeInfo,
+  TypeInfoFactory.shortTypeInfo, TypeInfoFactory.doubleTypeInfo
+  );
+
+  private static Set stringTypeInfos = ImmutableSet.of(
+  TypeInfoFactory.stringTypeInfo,
+  TypeInfoFactory.charTypeInfo, TypeInfoFactory.varcharTypeInfo
+  );
+
+
+  public static org.apache.druid.query.Query 
addDynamicFilters(org.apache.druid.query.Query query,
+  ExprNodeGenericFuncDesc filterExpr, Configuration conf, boolean 
resolveDynamicValues
+  ) {
+List virtualColumns = Arrays
+.asList(getVirtualColumns(query).getVirtualColumns());
+org.apache.druid.query.Query rv = query;
+DimFilter joinReductionFilter = toDruidFilter(filterExpr, conf, 
virtualColumns,
+resolveDynamicValues
+);
+if(joinReductionFilter != null) {
+  String type = query.getType();
+  DimFilter filter = new AndDimFilter(joinReductionFilter, 
query.getFilter());
+  switch (type) {
+  case org.apache.druid.query.Query.TIMESERIES:
+rv = Druids.TimeseriesQueryBuilder.copy((TimeseriesQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.TOPN:
+rv = new TopNQueryBuilder((TopNQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.GROUP_BY:
+rv = new GroupByQuery.Builder((GroupByQuery) query)
+.setDimFilter(filter)
+.setVirtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.SCAN:
+rv = ScanQuery.ScanQueryBuilder.copy((ScanQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.SELECT:
+rv = Druids.SelectQueryBuilder.copy((SelectQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  default:
+throw new UnsupportedOperationException("Unsupported Query type " + 
type);
+  }
+}
+return rv;
+  }
+
+  @Nullable
+  private static DimFilter toDruidFilter(ExprNodeDesc filterExpr, 
Configuration configuration,
+  List virtualColumns, boolean resolveDynamicValues
+  ) {
+if(filterExpr == null) {
+  return null;
+}
+Class genericUDFClass = 
getGenericUDFClassFromExprDesc(filterExpr);
+if(FunctionRegistry.isOpAnd(filterExpr)) {
+  Iterator iterator = filterExpr.getChildren().iterator();
+  List delegates = Lists.newArrayList();
+  while (iterator.hasNext()) {
+DimFilter filter = toDruidFilter(iterator.next(), configuration, 
virtualColumns,
+resolveDynamicValues
+);
+if(filter != null) {
+  delegates.add(filter);
+}
+  }
+  if(delegates != null && !delegates.isEmpty()) {
+return new AndDimFilter(delegates);
+  }
+}
+if(FunctionRegistry.isOpOr(filterExpr)) {
+  Iterator iterator = filterExpr.getChildren().iterator();
+  List delegates = Lists.newArrayList();
+  while (iterator.hasNext()) {
+DimFilter filter = toDruidFilter(iterator.next(), configuration, 
virtualColumns,
+resolveDynamicValues
+);
+if(filter != null) {
+  delegates.add(filter);
+}
+  }
+  if(delegates != null) {
+return new OrDimFilter(delegates);
+  }
+} else 

[jira] [Work logged] (HIVE-20683) Add the Ability to push Dynamic Between and Bloom filters to Druid

2019-09-09 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-20683?focusedWorklogId=308884=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-308884
 ]

ASF GitHub Bot logged work on HIVE-20683:
-

Author: ASF GitHub Bot
Created on: 09/Sep/19 13:57
Start Date: 09/Sep/19 13:57
Worklog Time Spent: 10m 
  Work Description: nishantmonu51 commented on issue #723: [HIVE-20683] Add 
the Ability to push Dynamic Between and Bloom filters to Druid
URL: https://github.com/apache/hive/pull/723#issuecomment-529487533
 
 
   @b-slim : Updated patch 
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 308884)
Time Spent: 5h 10m  (was: 5h)

> Add the Ability to push Dynamic Between and Bloom filters to Druid
> --
>
> Key: HIVE-20683
> URL: https://issues.apache.org/jira/browse/HIVE-20683
> Project: Hive
>  Issue Type: New Feature
>  Components: Druid integration
>Reporter: Nishant Bangarwa
>Assignee: Nishant Bangarwa
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-20683.1.patch, HIVE-20683.10.patch, 
> HIVE-20683.2.patch, HIVE-20683.3.patch, HIVE-20683.4.patch, 
> HIVE-20683.5.patch, HIVE-20683.6.patch, HIVE-20683.8.patch, HIVE-20683.patch
>
>  Time Spent: 5h 10m
>  Remaining Estimate: 0h
>
> For optimizing joins, Hive generates BETWEEN filter with min-max and BLOOM 
> filter for filtering one side of semi-join.
> Druid 0.13.0 will have support for Bloom filters (Added via 
> https://github.com/apache/incubator-druid/pull/6222)
> Implementation details - 
> # Hive generates and passes the filters as part of 'filterExpr' in TableScan. 
> # DruidQueryBasedRecordReader gets this filter passed as part of the conf. 
> # During execution phase, before sending the query to druid in 
> DruidQueryBasedRecordReader we will deserialize this filter, translate it 
> into a DruidDimFilter and add it to existing DruidQuery.  Tez executor 
> already ensures that when we start reading results from the record reader, 
> all the dynamic values are initialized. 
> # Explaining a druid query also prints the query sent to druid as 
> {{druid.json.query}}. We also need to make sure to update the druid query 
> with the filters. During explain we do not have the actual values for the 
> dynamic values, so instead of values we will print the dynamic expression 
> itself as part of druid query. 
> Note:- This work needs druid to be updated to version 0.13.0



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Work logged] (HIVE-20683) Add the Ability to push Dynamic Between and Bloom filters to Druid

2019-09-09 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-20683?focusedWorklogId=308882=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-308882
 ]

ASF GitHub Bot logged work on HIVE-20683:
-

Author: ASF GitHub Bot
Created on: 09/Sep/19 13:57
Start Date: 09/Sep/19 13:57
Worklog Time Spent: 10m 
  Work Description: nishantmonu51 commented on pull request #723: 
[HIVE-20683] Add the Ability to push Dynamic Between and Bloom filters to Druid
URL: https://github.com/apache/hive/pull/723#discussion_r322256531
 
 

 ##
 File path: 
druid-handler/src/java/org/apache/hadoop/hive/druid/DruidStorageHandlerUtils.java
 ##
 @@ -894,4 +945,255 @@ public static IndexSpec getIndexSpec(Configuration jc) {
 ImmutableList aggregatorFactories = 
aggregatorFactoryBuilder.build();
 return Pair.of(dimensions, aggregatorFactories.toArray(new 
AggregatorFactory[0]));
   }
+
+  // Druid only supports String,Long,Float,Double selectors
+  private static Set druidSupportedTypeInfos = 
ImmutableSet.of(
+  TypeInfoFactory.stringTypeInfo, TypeInfoFactory.charTypeInfo,
+  TypeInfoFactory.varcharTypeInfo, TypeInfoFactory.byteTypeInfo,
+  TypeInfoFactory.intTypeInfo, TypeInfoFactory.longTypeInfo,
+  TypeInfoFactory.shortTypeInfo, TypeInfoFactory.doubleTypeInfo
+  );
+
+  private static Set stringTypeInfos = ImmutableSet.of(
+  TypeInfoFactory.stringTypeInfo,
+  TypeInfoFactory.charTypeInfo, TypeInfoFactory.varcharTypeInfo
+  );
+
+
+  public static org.apache.druid.query.Query 
addDynamicFilters(org.apache.druid.query.Query query,
+  ExprNodeGenericFuncDesc filterExpr, Configuration conf, boolean 
resolveDynamicValues
+  ) {
+List virtualColumns = Arrays
+.asList(getVirtualColumns(query).getVirtualColumns());
+org.apache.druid.query.Query rv = query;
+DimFilter joinReductionFilter = toDruidFilter(filterExpr, conf, 
virtualColumns,
+resolveDynamicValues
+);
+if(joinReductionFilter != null) {
+  String type = query.getType();
+  DimFilter filter = new AndDimFilter(joinReductionFilter, 
query.getFilter());
+  switch (type) {
+  case org.apache.druid.query.Query.TIMESERIES:
+rv = Druids.TimeseriesQueryBuilder.copy((TimeseriesQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.TOPN:
+rv = new TopNQueryBuilder((TopNQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.GROUP_BY:
+rv = new GroupByQuery.Builder((GroupByQuery) query)
+.setDimFilter(filter)
+.setVirtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.SCAN:
+rv = ScanQuery.ScanQueryBuilder.copy((ScanQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.SELECT:
+rv = Druids.SelectQueryBuilder.copy((SelectQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  default:
+throw new UnsupportedOperationException("Unsupported Query type " + 
type);
+  }
+}
+return rv;
+  }
+
+  @Nullable
+  private static DimFilter toDruidFilter(ExprNodeDesc filterExpr, 
Configuration configuration,
+  List virtualColumns, boolean resolveDynamicValues
+  ) {
+if(filterExpr == null) {
+  return null;
+}
+Class genericUDFClass = 
getGenericUDFClassFromExprDesc(filterExpr);
+if(FunctionRegistry.isOpAnd(filterExpr)) {
+  Iterator iterator = filterExpr.getChildren().iterator();
+  List delegates = Lists.newArrayList();
+  while (iterator.hasNext()) {
+DimFilter filter = toDruidFilter(iterator.next(), configuration, 
virtualColumns,
+resolveDynamicValues
+);
+if(filter != null) {
+  delegates.add(filter);
+}
+  }
+  if(delegates != null && !delegates.isEmpty()) {
+return new AndDimFilter(delegates);
+  }
+}
+if(FunctionRegistry.isOpOr(filterExpr)) {
+  Iterator iterator = filterExpr.getChildren().iterator();
+  List delegates = Lists.newArrayList();
+  while (iterator.hasNext()) {
+DimFilter filter = toDruidFilter(iterator.next(), configuration, 
virtualColumns,
+resolveDynamicValues
+);
+if(filter != null) {
+  delegates.add(filter);
+}
+  }
+  if(delegates != null) {
+return new OrDimFilter(delegates);
+  }
+} else 

[jira] [Work logged] (HIVE-20683) Add the Ability to push Dynamic Between and Bloom filters to Druid

2019-09-09 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-20683?focusedWorklogId=308881=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-308881
 ]

ASF GitHub Bot logged work on HIVE-20683:
-

Author: ASF GitHub Bot
Created on: 09/Sep/19 13:57
Start Date: 09/Sep/19 13:57
Worklog Time Spent: 10m 
  Work Description: nishantmonu51 commented on pull request #723: 
[HIVE-20683] Add the Ability to push Dynamic Between and Bloom filters to Druid
URL: https://github.com/apache/hive/pull/723#discussion_r322256384
 
 

 ##
 File path: 
druid-handler/src/java/org/apache/hadoop/hive/druid/DruidStorageHandlerUtils.java
 ##
 @@ -91,17 +114,37 @@
 import org.apache.hadoop.fs.FileStatus;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.hive.common.io.NonSyncByteArrayInputStream;
 
 Review comment:
   removed.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 308881)
Time Spent: 4h 40m  (was: 4.5h)

> Add the Ability to push Dynamic Between and Bloom filters to Druid
> --
>
> Key: HIVE-20683
> URL: https://issues.apache.org/jira/browse/HIVE-20683
> Project: Hive
>  Issue Type: New Feature
>  Components: Druid integration
>Reporter: Nishant Bangarwa
>Assignee: Nishant Bangarwa
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-20683.1.patch, HIVE-20683.10.patch, 
> HIVE-20683.2.patch, HIVE-20683.3.patch, HIVE-20683.4.patch, 
> HIVE-20683.5.patch, HIVE-20683.6.patch, HIVE-20683.8.patch, HIVE-20683.patch
>
>  Time Spent: 4h 40m
>  Remaining Estimate: 0h
>
> For optimizing joins, Hive generates BETWEEN filter with min-max and BLOOM 
> filter for filtering one side of semi-join.
> Druid 0.13.0 will have support for Bloom filters (Added via 
> https://github.com/apache/incubator-druid/pull/6222)
> Implementation details - 
> # Hive generates and passes the filters as part of 'filterExpr' in TableScan. 
> # DruidQueryBasedRecordReader gets this filter passed as part of the conf. 
> # During execution phase, before sending the query to druid in 
> DruidQueryBasedRecordReader we will deserialize this filter, translate it 
> into a DruidDimFilter and add it to existing DruidQuery.  Tez executor 
> already ensures that when we start reading results from the record reader, 
> all the dynamic values are initialized. 
> # Explaining a druid query also prints the query sent to druid as 
> {{druid.json.query}}. We also need to make sure to update the druid query 
> with the filters. During explain we do not have the actual values for the 
> dynamic values, so instead of values we will print the dynamic expression 
> itself as part of druid query. 
> Note:- This work needs druid to be updated to version 0.13.0



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Work logged] (HIVE-20683) Add the Ability to push Dynamic Between and Bloom filters to Druid

2019-09-09 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-20683?focusedWorklogId=308880=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-308880
 ]

ASF GitHub Bot logged work on HIVE-20683:
-

Author: ASF GitHub Bot
Created on: 09/Sep/19 13:57
Start Date: 09/Sep/19 13:57
Worklog Time Spent: 10m 
  Work Description: nishantmonu51 commented on pull request #723: 
[HIVE-20683] Add the Ability to push Dynamic Between and Bloom filters to Druid
URL: https://github.com/apache/hive/pull/723#discussion_r322256307
 
 

 ##
 File path: 
druid-handler/src/java/org/apache/hadoop/hive/druid/DruidStorageHandlerUtils.java
 ##
 @@ -894,4 +945,255 @@ public static IndexSpec getIndexSpec(Configuration jc) {
 ImmutableList aggregatorFactories = 
aggregatorFactoryBuilder.build();
 return Pair.of(dimensions, aggregatorFactories.toArray(new 
AggregatorFactory[0]));
   }
+
+  // Druid only supports String,Long,Float,Double selectors
+  private static Set druidSupportedTypeInfos = 
ImmutableSet.of(
+  TypeInfoFactory.stringTypeInfo, TypeInfoFactory.charTypeInfo,
+  TypeInfoFactory.varcharTypeInfo, TypeInfoFactory.byteTypeInfo,
+  TypeInfoFactory.intTypeInfo, TypeInfoFactory.longTypeInfo,
+  TypeInfoFactory.shortTypeInfo, TypeInfoFactory.doubleTypeInfo
+  );
+
+  private static Set stringTypeInfos = ImmutableSet.of(
+  TypeInfoFactory.stringTypeInfo,
+  TypeInfoFactory.charTypeInfo, TypeInfoFactory.varcharTypeInfo
+  );
+
+
+  public static org.apache.druid.query.Query 
addDynamicFilters(org.apache.druid.query.Query query,
+  ExprNodeGenericFuncDesc filterExpr, Configuration conf, boolean 
resolveDynamicValues
+  ) {
+List virtualColumns = Arrays
+.asList(getVirtualColumns(query).getVirtualColumns());
+org.apache.druid.query.Query rv = query;
+DimFilter joinReductionFilter = toDruidFilter(filterExpr, conf, 
virtualColumns,
+resolveDynamicValues
+);
+if(joinReductionFilter != null) {
+  String type = query.getType();
+  DimFilter filter = new AndDimFilter(joinReductionFilter, 
query.getFilter());
+  switch (type) {
+  case org.apache.druid.query.Query.TIMESERIES:
+rv = Druids.TimeseriesQueryBuilder.copy((TimeseriesQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.TOPN:
+rv = new TopNQueryBuilder((TopNQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.GROUP_BY:
+rv = new GroupByQuery.Builder((GroupByQuery) query)
+.setDimFilter(filter)
+.setVirtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.SCAN:
+rv = ScanQuery.ScanQueryBuilder.copy((ScanQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.SELECT:
+rv = Druids.SelectQueryBuilder.copy((SelectQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  default:
+throw new UnsupportedOperationException("Unsupported Query type " + 
type);
+  }
+}
+return rv;
+  }
+
+  @Nullable
+  private static DimFilter toDruidFilter(ExprNodeDesc filterExpr, 
Configuration configuration,
+  List virtualColumns, boolean resolveDynamicValues
+  ) {
+if(filterExpr == null) {
+  return null;
+}
+Class genericUDFClass = 
getGenericUDFClassFromExprDesc(filterExpr);
+if(FunctionRegistry.isOpAnd(filterExpr)) {
+  Iterator iterator = filterExpr.getChildren().iterator();
+  List delegates = Lists.newArrayList();
+  while (iterator.hasNext()) {
+DimFilter filter = toDruidFilter(iterator.next(), configuration, 
virtualColumns,
+resolveDynamicValues
+);
+if(filter != null) {
+  delegates.add(filter);
+}
+  }
+  if(delegates != null && !delegates.isEmpty()) {
 
 Review comment:
   removed null check
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 308880)
Time Spent: 4.5h  (was: 4h 20m)

> Add the Ability to push Dynamic Between and Bloom filters to Druid
> 

[jira] [Work logged] (HIVE-20683) Add the Ability to push Dynamic Between and Bloom filters to Druid

2019-09-09 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-20683?focusedWorklogId=308878=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-308878
 ]

ASF GitHub Bot logged work on HIVE-20683:
-

Author: ASF GitHub Bot
Created on: 09/Sep/19 13:56
Start Date: 09/Sep/19 13:56
Worklog Time Spent: 10m 
  Work Description: nishantmonu51 commented on pull request #723: 
[HIVE-20683] Add the Ability to push Dynamic Between and Bloom filters to Druid
URL: https://github.com/apache/hive/pull/723#discussion_r322256133
 
 

 ##
 File path: 
druid-handler/src/java/org/apache/hadoop/hive/druid/DruidStorageHandlerUtils.java
 ##
 @@ -894,4 +945,255 @@ public static IndexSpec getIndexSpec(Configuration jc) {
 ImmutableList aggregatorFactories = 
aggregatorFactoryBuilder.build();
 return Pair.of(dimensions, aggregatorFactories.toArray(new 
AggregatorFactory[0]));
   }
+
+  // Druid only supports String,Long,Float,Double selectors
+  private static Set druidSupportedTypeInfos = 
ImmutableSet.of(
+  TypeInfoFactory.stringTypeInfo, TypeInfoFactory.charTypeInfo,
+  TypeInfoFactory.varcharTypeInfo, TypeInfoFactory.byteTypeInfo,
+  TypeInfoFactory.intTypeInfo, TypeInfoFactory.longTypeInfo,
+  TypeInfoFactory.shortTypeInfo, TypeInfoFactory.doubleTypeInfo
+  );
+
+  private static Set stringTypeInfos = ImmutableSet.of(
+  TypeInfoFactory.stringTypeInfo,
+  TypeInfoFactory.charTypeInfo, TypeInfoFactory.varcharTypeInfo
+  );
+
+
+  public static org.apache.druid.query.Query 
addDynamicFilters(org.apache.druid.query.Query query,
+  ExprNodeGenericFuncDesc filterExpr, Configuration conf, boolean 
resolveDynamicValues
+  ) {
+List virtualColumns = Arrays
+.asList(getVirtualColumns(query).getVirtualColumns());
+org.apache.druid.query.Query rv = query;
+DimFilter joinReductionFilter = toDruidFilter(filterExpr, conf, 
virtualColumns,
+resolveDynamicValues
+);
+if(joinReductionFilter != null) {
+  String type = query.getType();
+  DimFilter filter = new AndDimFilter(joinReductionFilter, 
query.getFilter());
+  switch (type) {
+  case org.apache.druid.query.Query.TIMESERIES:
+rv = Druids.TimeseriesQueryBuilder.copy((TimeseriesQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.TOPN:
+rv = new TopNQueryBuilder((TopNQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.GROUP_BY:
+rv = new GroupByQuery.Builder((GroupByQuery) query)
+.setDimFilter(filter)
+.setVirtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.SCAN:
+rv = ScanQuery.ScanQueryBuilder.copy((ScanQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.SELECT:
+rv = Druids.SelectQueryBuilder.copy((SelectQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  default:
+throw new UnsupportedOperationException("Unsupported Query type " + 
type);
+  }
+}
+return rv;
+  }
+
+  @Nullable
+  private static DimFilter toDruidFilter(ExprNodeDesc filterExpr, 
Configuration configuration,
+  List virtualColumns, boolean resolveDynamicValues
+  ) {
+if(filterExpr == null) {
+  return null;
+}
+Class genericUDFClass = 
getGenericUDFClassFromExprDesc(filterExpr);
+if(FunctionRegistry.isOpAnd(filterExpr)) {
+  Iterator iterator = filterExpr.getChildren().iterator();
+  List delegates = Lists.newArrayList();
+  while (iterator.hasNext()) {
+DimFilter filter = toDruidFilter(iterator.next(), configuration, 
virtualColumns,
+resolveDynamicValues
+);
+if(filter != null) {
+  delegates.add(filter);
+}
+  }
+  if(delegates != null && !delegates.isEmpty()) {
+return new AndDimFilter(delegates);
+  }
+}
+if(FunctionRegistry.isOpOr(filterExpr)) {
+  Iterator iterator = filterExpr.getChildren().iterator();
+  List delegates = Lists.newArrayList();
+  while (iterator.hasNext()) {
+DimFilter filter = toDruidFilter(iterator.next(), configuration, 
virtualColumns,
+resolveDynamicValues
+);
+if(filter != null) {
+  delegates.add(filter);
+}
+  }
+  if(delegates != null) {
+return new OrDimFilter(delegates);
+  }
+} else 

[jira] [Work logged] (HIVE-20683) Add the Ability to push Dynamic Between and Bloom filters to Druid

2019-09-09 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-20683?focusedWorklogId=308876=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-308876
 ]

ASF GitHub Bot logged work on HIVE-20683:
-

Author: ASF GitHub Bot
Created on: 09/Sep/19 13:56
Start Date: 09/Sep/19 13:56
Worklog Time Spent: 10m 
  Work Description: nishantmonu51 commented on pull request #723: 
[HIVE-20683] Add the Ability to push Dynamic Between and Bloom filters to Druid
URL: https://github.com/apache/hive/pull/723#discussion_r322256014
 
 

 ##
 File path: 
ql/src/test/queries/clientpositive/druidmini_semijoin_reduction_all_types.q
 ##
 @@ -0,0 +1,144 @@
+--! qt:dataset:srcpart
+--! qt:dataset:druid_table_alltypesorc
+--! qt:dataset:alltypesorc
+
+set hive.compute.query.using.stats=false;
+set hive.mapred.mode=nonstrict;
+set hive.explain.user=false;
+set hive.optimize.ppd=true;
+set hive.ppd.remove.duplicatefilters=true;
+set hive.tez.dynamic.partition.pruning=true;
+set hive.tez.dynamic.semijoin.reduction=true;
+set hive.optimize.metadataonly=false;
+set hive.optimize.index.filter=true;
+set hive.stats.autogather=true;
+set hive.tez.bigtable.minsize.semijoin.reduction=1;
+set hive.tez.min.bloom.filter.entries=1;
+set hive.stats.fetch.column.stats=true;
+set hive.disable.unsafe.external.table.operations=false;
+set hive.tez.dynamic.semijoin.reduction.for.mapjoin=true;
+
+DROP TABLE IF EXISTS alltypesorc_small;
+CREATE TABLE alltypesorc_small(
+ctinyint TINYINT,
+csmallint SMALLINT,
+cint INT,
+cbigint BIGINT,
+cfloat FLOAT,
+cdouble DOUBLE,
+cstring1 STRING,
+cstring2 STRING,
+ctimestamp1 TIMESTAMP,
+cboolean1 BOOLEAN,
+cboolean2 BOOLEAN)
+STORED AS ORC;
+Insert into table alltypesorc_small
+Select ctinyint, csmallint, cint, cbigint, cfloat, cdouble, cstring1, 
cstring2, cast(`__time` as timestamp), cboolean1, cboolean2 from 
druid_table_alltypesorc where cstring2 like '%a%' and cstring1 like '%a%';
+Select count(*) from alltypesorc_small;
+Select count(*) from druid_table_alltypesorc;
+
+DESCRIBE druid_table_alltypesorc;
+DESCRIBE alltypesorc_small;
+
+-- Test Joins on all column types one by one
+-- String
+set hive.disable.unsafe.external.table.operations=false;
+EXPLAIN select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.cstring1 = druid_table_alltypesorc.cstring1);
+select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.cstring1 = druid_table_alltypesorc.cstring1);
+set hive.disable.unsafe.external.table.operations=true;
+select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.cstring1 = druid_table_alltypesorc.cstring1);
+
+-- tinyint
+set hive.disable.unsafe.external.table.operations=false;
+EXPLAIN select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.ctinyint = druid_table_alltypesorc.ctinyint);
+select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.ctinyint = druid_table_alltypesorc.ctinyint);
+set hive.disable.unsafe.external.table.operations=true;
+select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.ctinyint = druid_table_alltypesorc.ctinyint);
+
+-- smallint
+set hive.disable.unsafe.external.table.operations=false;
+EXPLAIN select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.csmallint = druid_table_alltypesorc.csmallint);
+select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.csmallint = druid_table_alltypesorc.csmallint);
+set hive.disable.unsafe.external.table.operations=true;
+select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.csmallint = druid_table_alltypesorc.csmallint);
+
+-- int
+set hive.disable.unsafe.external.table.operations=false;
+EXPLAIN select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.cint = druid_table_alltypesorc.cint);
+select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.cint = druid_table_alltypesorc.cint);
+set hive.disable.unsafe.external.table.operations=true;
+select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.cint = druid_table_alltypesorc.cint);
+
+-- bigint
+set hive.disable.unsafe.external.table.operations=false;
+EXPLAIN select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.cbigint = druid_table_alltypesorc.cbigint);
+select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.cbigint = druid_table_alltypesorc.cbigint);
+set hive.disable.unsafe.external.table.operations=true;
+select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.cbigint = 

[jira] [Work logged] (HIVE-20683) Add the Ability to push Dynamic Between and Bloom filters to Druid

2019-09-09 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-20683?focusedWorklogId=308879=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-308879
 ]

ASF GitHub Bot logged work on HIVE-20683:
-

Author: ASF GitHub Bot
Created on: 09/Sep/19 13:56
Start Date: 09/Sep/19 13:56
Worklog Time Spent: 10m 
  Work Description: nishantmonu51 commented on pull request #723: 
[HIVE-20683] Add the Ability to push Dynamic Between and Bloom filters to Druid
URL: https://github.com/apache/hive/pull/723#discussion_r322256224
 
 

 ##
 File path: 
druid-handler/src/java/org/apache/hadoop/hive/druid/DruidStorageHandlerUtils.java
 ##
 @@ -894,4 +945,255 @@ public static IndexSpec getIndexSpec(Configuration jc) {
 ImmutableList aggregatorFactories = 
aggregatorFactoryBuilder.build();
 return Pair.of(dimensions, aggregatorFactories.toArray(new 
AggregatorFactory[0]));
   }
+
+  // Druid only supports String,Long,Float,Double selectors
+  private static Set druidSupportedTypeInfos = 
ImmutableSet.of(
+  TypeInfoFactory.stringTypeInfo, TypeInfoFactory.charTypeInfo,
+  TypeInfoFactory.varcharTypeInfo, TypeInfoFactory.byteTypeInfo,
+  TypeInfoFactory.intTypeInfo, TypeInfoFactory.longTypeInfo,
+  TypeInfoFactory.shortTypeInfo, TypeInfoFactory.doubleTypeInfo
+  );
+
+  private static Set stringTypeInfos = ImmutableSet.of(
+  TypeInfoFactory.stringTypeInfo,
+  TypeInfoFactory.charTypeInfo, TypeInfoFactory.varcharTypeInfo
+  );
+
+
+  public static org.apache.druid.query.Query 
addDynamicFilters(org.apache.druid.query.Query query,
+  ExprNodeGenericFuncDesc filterExpr, Configuration conf, boolean 
resolveDynamicValues
+  ) {
+List virtualColumns = Arrays
+.asList(getVirtualColumns(query).getVirtualColumns());
+org.apache.druid.query.Query rv = query;
+DimFilter joinReductionFilter = toDruidFilter(filterExpr, conf, 
virtualColumns,
+resolveDynamicValues
+);
+if(joinReductionFilter != null) {
+  String type = query.getType();
+  DimFilter filter = new AndDimFilter(joinReductionFilter, 
query.getFilter());
+  switch (type) {
+  case org.apache.druid.query.Query.TIMESERIES:
+rv = Druids.TimeseriesQueryBuilder.copy((TimeseriesQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.TOPN:
+rv = new TopNQueryBuilder((TopNQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.GROUP_BY:
+rv = new GroupByQuery.Builder((GroupByQuery) query)
+.setDimFilter(filter)
+.setVirtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.SCAN:
+rv = ScanQuery.ScanQueryBuilder.copy((ScanQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.SELECT:
+rv = Druids.SelectQueryBuilder.copy((SelectQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  default:
+throw new UnsupportedOperationException("Unsupported Query type " + 
type);
+  }
+}
+return rv;
+  }
+
+  @Nullable
+  private static DimFilter toDruidFilter(ExprNodeDesc filterExpr, 
Configuration configuration,
+  List virtualColumns, boolean resolveDynamicValues
+  ) {
+if(filterExpr == null) {
+  return null;
+}
+Class genericUDFClass = 
getGenericUDFClassFromExprDesc(filterExpr);
+if(FunctionRegistry.isOpAnd(filterExpr)) {
+  Iterator iterator = filterExpr.getChildren().iterator();
+  List delegates = Lists.newArrayList();
+  while (iterator.hasNext()) {
+DimFilter filter = toDruidFilter(iterator.next(), configuration, 
virtualColumns,
+resolveDynamicValues
+);
+if(filter != null) {
+  delegates.add(filter);
+}
+  }
+  if(delegates != null && !delegates.isEmpty()) {
+return new AndDimFilter(delegates);
+  }
+}
+if(FunctionRegistry.isOpOr(filterExpr)) {
+  Iterator iterator = filterExpr.getChildren().iterator();
+  List delegates = Lists.newArrayList();
+  while (iterator.hasNext()) {
+DimFilter filter = toDruidFilter(iterator.next(), configuration, 
virtualColumns,
+resolveDynamicValues
+);
+if(filter != null) {
+  delegates.add(filter);
+}
+  }
+  if(delegates != null) {
 
 Review comment:
   fixed.
 

[jira] [Updated] (HIVE-20683) Add the Ability to push Dynamic Between and Bloom filters to Druid

2019-09-09 Thread Nishant Bangarwa (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-20683?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nishant Bangarwa updated HIVE-20683:

Attachment: HIVE-20683.10.patch

> Add the Ability to push Dynamic Between and Bloom filters to Druid
> --
>
> Key: HIVE-20683
> URL: https://issues.apache.org/jira/browse/HIVE-20683
> Project: Hive
>  Issue Type: New Feature
>  Components: Druid integration
>Reporter: Nishant Bangarwa
>Assignee: Nishant Bangarwa
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-20683.1.patch, HIVE-20683.10.patch, 
> HIVE-20683.2.patch, HIVE-20683.3.patch, HIVE-20683.4.patch, 
> HIVE-20683.5.patch, HIVE-20683.6.patch, HIVE-20683.8.patch, HIVE-20683.patch
>
>  Time Spent: 3h 40m
>  Remaining Estimate: 0h
>
> For optimizing joins, Hive generates BETWEEN filter with min-max and BLOOM 
> filter for filtering one side of semi-join.
> Druid 0.13.0 will have support for Bloom filters (Added via 
> https://github.com/apache/incubator-druid/pull/6222)
> Implementation details - 
> # Hive generates and passes the filters as part of 'filterExpr' in TableScan. 
> # DruidQueryBasedRecordReader gets this filter passed as part of the conf. 
> # During execution phase, before sending the query to druid in 
> DruidQueryBasedRecordReader we will deserialize this filter, translate it 
> into a DruidDimFilter and add it to existing DruidQuery.  Tez executor 
> already ensures that when we start reading results from the record reader, 
> all the dynamic values are initialized. 
> # Explaining a druid query also prints the query sent to druid as 
> {{druid.json.query}}. We also need to make sure to update the druid query 
> with the filters. During explain we do not have the actual values for the 
> dynamic values, so instead of values we will print the dynamic expression 
> itself as part of druid query. 
> Note:- This work needs druid to be updated to version 0.13.0



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Work logged] (HIVE-20683) Add the Ability to push Dynamic Between and Bloom filters to Druid

2019-09-09 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-20683?focusedWorklogId=308877=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-308877
 ]

ASF GitHub Bot logged work on HIVE-20683:
-

Author: ASF GitHub Bot
Created on: 09/Sep/19 13:56
Start Date: 09/Sep/19 13:56
Worklog Time Spent: 10m 
  Work Description: nishantmonu51 commented on pull request #723: 
[HIVE-20683] Add the Ability to push Dynamic Between and Bloom filters to Druid
URL: https://github.com/apache/hive/pull/723#discussion_r322256075
 
 

 ##
 File path: 
druid-handler/src/java/org/apache/hadoop/hive/druid/DruidStorageHandlerUtils.java
 ##
 @@ -894,4 +945,255 @@ public static IndexSpec getIndexSpec(Configuration jc) {
 ImmutableList aggregatorFactories = 
aggregatorFactoryBuilder.build();
 return Pair.of(dimensions, aggregatorFactories.toArray(new 
AggregatorFactory[0]));
   }
+
+  // Druid only supports String,Long,Float,Double selectors
+  private static Set druidSupportedTypeInfos = 
ImmutableSet.of(
+  TypeInfoFactory.stringTypeInfo, TypeInfoFactory.charTypeInfo,
+  TypeInfoFactory.varcharTypeInfo, TypeInfoFactory.byteTypeInfo,
+  TypeInfoFactory.intTypeInfo, TypeInfoFactory.longTypeInfo,
+  TypeInfoFactory.shortTypeInfo, TypeInfoFactory.doubleTypeInfo
+  );
+
+  private static Set stringTypeInfos = ImmutableSet.of(
+  TypeInfoFactory.stringTypeInfo,
+  TypeInfoFactory.charTypeInfo, TypeInfoFactory.varcharTypeInfo
+  );
+
+
+  public static org.apache.druid.query.Query 
addDynamicFilters(org.apache.druid.query.Query query,
+  ExprNodeGenericFuncDesc filterExpr, Configuration conf, boolean 
resolveDynamicValues
+  ) {
+List virtualColumns = Arrays
+.asList(getVirtualColumns(query).getVirtualColumns());
+org.apache.druid.query.Query rv = query;
+DimFilter joinReductionFilter = toDruidFilter(filterExpr, conf, 
virtualColumns,
+resolveDynamicValues
+);
+if(joinReductionFilter != null) {
+  String type = query.getType();
+  DimFilter filter = new AndDimFilter(joinReductionFilter, 
query.getFilter());
+  switch (type) {
+  case org.apache.druid.query.Query.TIMESERIES:
+rv = Druids.TimeseriesQueryBuilder.copy((TimeseriesQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.TOPN:
+rv = new TopNQueryBuilder((TopNQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.GROUP_BY:
+rv = new GroupByQuery.Builder((GroupByQuery) query)
+.setDimFilter(filter)
+.setVirtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.SCAN:
+rv = ScanQuery.ScanQueryBuilder.copy((ScanQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.SELECT:
+rv = Druids.SelectQueryBuilder.copy((SelectQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  default:
+throw new UnsupportedOperationException("Unsupported Query type " + 
type);
+  }
+}
+return rv;
+  }
+
+  @Nullable
+  private static DimFilter toDruidFilter(ExprNodeDesc filterExpr, 
Configuration configuration,
+  List virtualColumns, boolean resolveDynamicValues
+  ) {
+if(filterExpr == null) {
+  return null;
+}
+Class genericUDFClass = 
getGenericUDFClassFromExprDesc(filterExpr);
+if(FunctionRegistry.isOpAnd(filterExpr)) {
+  Iterator iterator = filterExpr.getChildren().iterator();
+  List delegates = Lists.newArrayList();
+  while (iterator.hasNext()) {
+DimFilter filter = toDruidFilter(iterator.next(), configuration, 
virtualColumns,
+resolveDynamicValues
+);
+if(filter != null) {
+  delegates.add(filter);
+}
+  }
+  if(delegates != null && !delegates.isEmpty()) {
+return new AndDimFilter(delegates);
+  }
+}
+if(FunctionRegistry.isOpOr(filterExpr)) {
+  Iterator iterator = filterExpr.getChildren().iterator();
+  List delegates = Lists.newArrayList();
+  while (iterator.hasNext()) {
+DimFilter filter = toDruidFilter(iterator.next(), configuration, 
virtualColumns,
+resolveDynamicValues
+);
+if(filter != null) {
+  delegates.add(filter);
+}
+  }
+  if(delegates != null) {
+return new OrDimFilter(delegates);
+  }
+} else 

[jira] [Updated] (HIVE-22055) select count gives incorrect result after loading data from text file

2019-09-09 Thread Attila Magyar (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22055?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Attila Magyar updated HIVE-22055:
-
Status: Patch Available  (was: Open)

> select count gives incorrect result after loading data from text file
> -
>
> Key: HIVE-22055
> URL: https://issues.apache.org/jira/browse/HIVE-22055
> Project: Hive
>  Issue Type: Task
>  Components: Hive
>Reporter: Attila Magyar
>Assignee: Attila Magyar
>Priority: Major
> Attachments: HIVE-22055.2.patch
>
>
> Add one more load to mm_loaddata.q:
> Load data 3 times (both kv1.txt and kv2.txt contains 500 records)
> {code:java}
> create table load0_mm (key string, value string) stored as textfile 
> tblproperties("transactional"="true", 
> "transactional_properties"="insert_only");
> load data local inpath '../../data/files/kv1.txt' into table load0_mm;
> select count(1) from load0_mm;
> load data local inpath '../../data/files/kv2.txt' into table load0_mm;
> select count(1) from load0_mm;
> load data local inpath '../../data/files/kv2.txt' into table load0_mm;
> select count(1) from load0_mm;{code}
> Expected output
> {code:java}
> PREHOOK: query: load data local inpath '../../data/files/kv2.txt' into table 
> load0_mm
> PREHOOK: type: LOAD
>  A masked pattern was here 
> PREHOOK: Output: default@load0_mm
> POSTHOOK: query: load data local inpath '../../data/files/kv2.txt' into table 
> load0_mm
> POSTHOOK: type: LOAD
>  A masked pattern was here 
> POSTHOOK: Output: default@load0_mm
> PREHOOK: query: select count(1) from load0_mm
> PREHOOK: type: QUERY
> PREHOOK: Input: default@load0_mm
>  A masked pattern was here 
> POSTHOOK: query: select count(1) from load0_mm
> POSTHOOK: type: QUERY
> POSTHOOK: Input: default@load0_mm
>  A masked pattern was here 
> 1500{code}
> Got:
> [ERROR]   TestMiniLlapLocalCliDriver.testCliDriver:59 Client Execution 
> succeeded but contained differences (error code = 1) after executing 
> mm_loaddata.q
> 63c63
> < 1480
> —
> > 1500
>  



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-22055) select count gives incorrect result after loading data from text file

2019-09-09 Thread Attila Magyar (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22055?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Attila Magyar updated HIVE-22055:
-
Attachment: (was: HIVE-22055.1.patch)

> select count gives incorrect result after loading data from text file
> -
>
> Key: HIVE-22055
> URL: https://issues.apache.org/jira/browse/HIVE-22055
> Project: Hive
>  Issue Type: Task
>  Components: Hive
>Reporter: Attila Magyar
>Assignee: Attila Magyar
>Priority: Major
> Attachments: HIVE-22055.2.patch
>
>
> Add one more load to mm_loaddata.q:
> Load data 3 times (both kv1.txt and kv2.txt contains 500 records)
> {code:java}
> create table load0_mm (key string, value string) stored as textfile 
> tblproperties("transactional"="true", 
> "transactional_properties"="insert_only");
> load data local inpath '../../data/files/kv1.txt' into table load0_mm;
> select count(1) from load0_mm;
> load data local inpath '../../data/files/kv2.txt' into table load0_mm;
> select count(1) from load0_mm;
> load data local inpath '../../data/files/kv2.txt' into table load0_mm;
> select count(1) from load0_mm;{code}
> Expected output
> {code:java}
> PREHOOK: query: load data local inpath '../../data/files/kv2.txt' into table 
> load0_mm
> PREHOOK: type: LOAD
>  A masked pattern was here 
> PREHOOK: Output: default@load0_mm
> POSTHOOK: query: load data local inpath '../../data/files/kv2.txt' into table 
> load0_mm
> POSTHOOK: type: LOAD
>  A masked pattern was here 
> POSTHOOK: Output: default@load0_mm
> PREHOOK: query: select count(1) from load0_mm
> PREHOOK: type: QUERY
> PREHOOK: Input: default@load0_mm
>  A masked pattern was here 
> POSTHOOK: query: select count(1) from load0_mm
> POSTHOOK: type: QUERY
> POSTHOOK: Input: default@load0_mm
>  A masked pattern was here 
> 1500{code}
> Got:
> [ERROR]   TestMiniLlapLocalCliDriver.testCliDriver:59 Client Execution 
> succeeded but contained differences (error code = 1) after executing 
> mm_loaddata.q
> 63c63
> < 1480
> —
> > 1500
>  



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-22055) select count gives incorrect result after loading data from text file

2019-09-09 Thread Attila Magyar (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22055?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Attila Magyar updated HIVE-22055:
-
Attachment: HIVE-22055.2.patch

> select count gives incorrect result after loading data from text file
> -
>
> Key: HIVE-22055
> URL: https://issues.apache.org/jira/browse/HIVE-22055
> Project: Hive
>  Issue Type: Task
>  Components: Hive
>Reporter: Attila Magyar
>Assignee: Attila Magyar
>Priority: Major
> Attachments: HIVE-22055.2.patch
>
>
> Add one more load to mm_loaddata.q:
> Load data 3 times (both kv1.txt and kv2.txt contains 500 records)
> {code:java}
> create table load0_mm (key string, value string) stored as textfile 
> tblproperties("transactional"="true", 
> "transactional_properties"="insert_only");
> load data local inpath '../../data/files/kv1.txt' into table load0_mm;
> select count(1) from load0_mm;
> load data local inpath '../../data/files/kv2.txt' into table load0_mm;
> select count(1) from load0_mm;
> load data local inpath '../../data/files/kv2.txt' into table load0_mm;
> select count(1) from load0_mm;{code}
> Expected output
> {code:java}
> PREHOOK: query: load data local inpath '../../data/files/kv2.txt' into table 
> load0_mm
> PREHOOK: type: LOAD
>  A masked pattern was here 
> PREHOOK: Output: default@load0_mm
> POSTHOOK: query: load data local inpath '../../data/files/kv2.txt' into table 
> load0_mm
> POSTHOOK: type: LOAD
>  A masked pattern was here 
> POSTHOOK: Output: default@load0_mm
> PREHOOK: query: select count(1) from load0_mm
> PREHOOK: type: QUERY
> PREHOOK: Input: default@load0_mm
>  A masked pattern was here 
> POSTHOOK: query: select count(1) from load0_mm
> POSTHOOK: type: QUERY
> POSTHOOK: Input: default@load0_mm
>  A masked pattern was here 
> 1500{code}
> Got:
> [ERROR]   TestMiniLlapLocalCliDriver.testCliDriver:59 Client Execution 
> succeeded but contained differences (error code = 1) after executing 
> mm_loaddata.q
> 63c63
> < 1480
> —
> > 1500
>  



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-22055) select count gives incorrect result after loading data from text file

2019-09-09 Thread Attila Magyar (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22055?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Attila Magyar updated HIVE-22055:
-
Status: Open  (was: Patch Available)

> select count gives incorrect result after loading data from text file
> -
>
> Key: HIVE-22055
> URL: https://issues.apache.org/jira/browse/HIVE-22055
> Project: Hive
>  Issue Type: Task
>  Components: Hive
>Reporter: Attila Magyar
>Assignee: Attila Magyar
>Priority: Major
> Attachments: HIVE-22055.1.patch
>
>
> Add one more load to mm_loaddata.q:
> Load data 3 times (both kv1.txt and kv2.txt contains 500 records)
> {code:java}
> create table load0_mm (key string, value string) stored as textfile 
> tblproperties("transactional"="true", 
> "transactional_properties"="insert_only");
> load data local inpath '../../data/files/kv1.txt' into table load0_mm;
> select count(1) from load0_mm;
> load data local inpath '../../data/files/kv2.txt' into table load0_mm;
> select count(1) from load0_mm;
> load data local inpath '../../data/files/kv2.txt' into table load0_mm;
> select count(1) from load0_mm;{code}
> Expected output
> {code:java}
> PREHOOK: query: load data local inpath '../../data/files/kv2.txt' into table 
> load0_mm
> PREHOOK: type: LOAD
>  A masked pattern was here 
> PREHOOK: Output: default@load0_mm
> POSTHOOK: query: load data local inpath '../../data/files/kv2.txt' into table 
> load0_mm
> POSTHOOK: type: LOAD
>  A masked pattern was here 
> POSTHOOK: Output: default@load0_mm
> PREHOOK: query: select count(1) from load0_mm
> PREHOOK: type: QUERY
> PREHOOK: Input: default@load0_mm
>  A masked pattern was here 
> POSTHOOK: query: select count(1) from load0_mm
> POSTHOOK: type: QUERY
> POSTHOOK: Input: default@load0_mm
>  A masked pattern was here 
> 1500{code}
> Got:
> [ERROR]   TestMiniLlapLocalCliDriver.testCliDriver:59 Client Execution 
> succeeded but contained differences (error code = 1) after executing 
> mm_loaddata.q
> 63c63
> < 1480
> —
> > 1500
>  



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-21508) ClassCastException when initializing HiveMetaStoreClient on JDK10 or newer

2019-09-09 Thread Ana Jalba (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-21508?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925699#comment-16925699
 ] 

Ana Jalba commented on HIVE-21508:
--

Local tests pass for hive-cli on branch-2.3 (with the change made) so I'm not 
sure what's happening with that build. 

I'll wait before I add another patch. Maybe it's better to upload a patch, wait 
for results, merge and then upload another patch. Do you think that's 
reasonable, [~pvary]?

> ClassCastException when initializing HiveMetaStoreClient on JDK10 or newer
> --
>
> Key: HIVE-21508
> URL: https://issues.apache.org/jira/browse/HIVE-21508
> Project: Hive
>  Issue Type: Bug
>  Components: Clients
>Affects Versions: 3.2.0, 2.3.4
>Reporter: Adar Dembo
>Assignee: Ana Jalba
>Priority: Major
> Fix For: 4.0.0, 2.3.7
>
> Attachments: HIVE-21508.1.branch-2.3.patch, HIVE-21508.patch
>
>
> There's this block of code in {{HiveMetaStoreClient:resolveUris}} (called 
> from the constructor) on master:
> {noformat}
>   private URI metastoreUris[];
>   ...
>   if (MetastoreConf.getVar(conf, 
> ConfVars.THRIFT_URI_SELECTION).equalsIgnoreCase("RANDOM")) {
> List uriList = Arrays.asList(metastoreUris);
> Collections.shuffle(uriList);
> metastoreUris = (URI[]) uriList.toArray();
>   }
> {noformat}
> The cast to {{URI[]}} throws a {{ClassCastException}} beginning with JDK 10, 
> possibly with JDK 9 as well. Note that {{THRIFT_URI_SELECTION}} defaults to 
> {{RANDOM}} so this should affect anyone who creates a 
> {{HiveMetaStoreClient}}. On master this can be overridden with {{SEQUENTIAL}} 
> to avoid the broken case; I'm working against 2.3.4 where there's no such 
> workaround.
> [Here's|https://stackoverflow.com/questions/51372788/array-cast-java-8-vs-java-9]
>  a StackOverflow post that explains the issue in more detail. Interestingly, 
> the author described the issue in the context of the HMS; not sure why there 
> was no follow up with a Hive bug report.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-21604) preCommit job should not be triggered on non-patch attachments

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-21604?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925696#comment-16925696
 ] 

Hive QA commented on HIVE-21604:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12979841/files.txt

{color:red}ERROR:{color} -1 due to build exiting with an error

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/18506/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/18506/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-18506/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hiveptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ date '+%Y-%m-%d %T.%3N'
2019-09-09 13:36:25.459
+ [[ -n /usr/lib/jvm/java-8-openjdk-amd64 ]]
+ export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ export 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'MAVEN_OPTS=-Xmx1g '
+ MAVEN_OPTS='-Xmx1g '
+ cd /data/hiveptest/working/
+ tee /data/hiveptest/logs/PreCommit-HIVE-Build-18506/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ git = \s\v\n ]]
+ [[ git = \g\i\t ]]
+ [[ -z master ]]
+ [[ -d apache-github-source-source ]]
+ [[ ! -d apache-github-source-source/.git ]]
+ [[ ! -d apache-github-source-source ]]
+ date '+%Y-%m-%d %T.%3N'
2019-09-09 13:36:25.461
+ cd apache-github-source-source
+ git fetch origin
>From https://github.com/apache/hive
   6dbc115..9667f25  master -> origin/master
+ git reset --hard HEAD
HEAD is now at 6dbc115 HIVE-21397: BloomFilter for hive Managed [ACID] table 
does not work as expected (Denys Kuzmenko, reviewed by Gopal Vijayaraghavan and 
Adam Szita)
+ git clean -f -d
Removing standalone-metastore/metastore-server/src/gen/
+ git checkout master
Already on 'master'
Your branch is behind 'origin/master' by 1 commit, and can be fast-forwarded.
  (use "git pull" to update your local branch)
+ git reset --hard origin/master
HEAD is now at 9667f25 HIVE-21942: Remove useless MetastoreType enum from 
AbstractCliConfig (Imre Molnar via Laszlo Bodor)
+ git merge --ff-only origin/master
Already up-to-date.
+ date '+%Y-%m-%d %T.%3N'
2019-09-09 13:36:27.016
+ rm -rf ../yetus_PreCommit-HIVE-Build-18506
+ mkdir ../yetus_PreCommit-HIVE-Build-18506
+ git gc
+ cp -R . ../yetus_PreCommit-HIVE-Build-18506
+ mkdir /data/hiveptest/logs/PreCommit-HIVE-Build-18506/yetus
+ patchCommandPath=/data/hiveptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hiveptest/working/scratch/build.patch
+ [[ -f /data/hiveptest/working/scratch/build.patch ]]
+ chmod +x /data/hiveptest/working/scratch/smart-apply-patch.sh
+ /data/hiveptest/working/scratch/smart-apply-patch.sh 
/data/hiveptest/working/scratch/build.patch
fatal: unrecognized input
fatal: unrecognized input
fatal: unrecognized input
The patch does not appear to apply with p0, p1, or p2
+ result=1
+ '[' 1 -ne 0 ']'
+ rm -rf yetus_PreCommit-HIVE-Build-18506
+ exit 1
'
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12979841 - PreCommit-HIVE-Build

> preCommit job should not be triggered on non-patch attachments
> --
>
> Key: HIVE-21604
> URL: https://issues.apache.org/jira/browse/HIVE-21604
> Project: Hive
>  Issue Type: Bug
>  Components: Testing Infrastructure
>Reporter: Laszlo Bodor
>Assignee: Laszlo Bodor
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-21604.01.patch, files.txt
>
>
> latest example: HIVE-21604
> https://issues.apache.org/jira/browse/HIVE-14669?focusedCommentId=16815520=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#comment-16815520
> I think one should be able to upload any kind of attachments (e.g. 
> screenshot) without triggering the precommit job
> 2 possible ways:
> 1. strict: enable only .patch (should work)
> (2. lenient: introduce blacklist, .png ...)



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-21884) Scheduled query support

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-21884?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925692#comment-16925692
 ] 

Hive QA commented on HIVE-21884:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
47s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  6m 
15s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  7m 
48s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  3m 
43s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  1m 
51s{color} | {color:blue} standalone-metastore/metastore-common in master has 
32 extant Findbugs warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
26s{color} | {color:blue} common in master has 61 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
54s{color} | {color:blue} standalone-metastore/metastore-server in master has 
181 extant Findbugs warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m  
3s{color} | {color:blue} ql in master has 2246 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
29s{color} | {color:blue} service in master has 48 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
35s{color} | {color:blue} itests/util in master has 44 extant Findbugs 
warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  7m 
46s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
20s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
49s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  7m 
48s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  7m 
48s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
12s{color} | {color:red} standalone-metastore/metastore-common: The patch 
generated 1 new + 495 unchanged - 0 fixed = 496 total (was 495) {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
12s{color} | {color:red} common: The patch generated 2 new + 367 unchanged - 0 
fixed = 369 total (was 367) {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
 5s{color} | {color:green} The patch service-rpc passed checkstyle {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
 7s{color} | {color:green} The patch metastore passed checkstyle {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
23s{color} | {color:red} standalone-metastore/metastore-server: The patch 
generated 18 new + 1250 unchanged - 15 fixed = 1268 total (was 1265) {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
33s{color} | {color:red} ql: The patch generated 61 new + 224 unchanged - 0 
fixed = 285 total (was 224) {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
10s{color} | {color:green} service: The patch generated 0 new + 38 unchanged - 
1 fixed = 38 total (was 39) {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  1m 
48s{color} | {color:red} root: The patch generated 82 new + 2421 unchanged - 16 
fixed = 2503 total (was 2437) {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
 8s{color} | {color:green} The patch hcatalog-unit passed checkstyle {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
11s{color} | {color:green} The patch util passed checkstyle {color} |
| {color:red}-1{color} | {color:red} whitespace {color} | {color:red}  0m  
1s{color} | {color:red} The patch has 17 line(s) that end in whitespace. Use 
git apply --whitespace=fix <>. Refer 
https://git-scm.com/docs/git-apply {color} |
| 

[jira] [Commented] (HIVE-21884) Scheduled query support

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-21884?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925683#comment-16925683
 ] 

Hive QA commented on HIVE-21884:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12979839/HIVE-21844.06.patch

{color:green}SUCCESS:{color} +1 due to 9 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 691 failed/errored test(s), 16784 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druid_materialized_view_rewrite_ssb]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druid_timeseries]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druid_timestamptz2]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druid_topn] 
(batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_dynamic_partition]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_expressions]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_extractTime]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_floorTime]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_mv] 
(batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_test1]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_test_alter]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_test_insert]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_test_ts]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidKafkaCliDriver.testCliDriver[druidkafkamini_avro]
 (batchId=297)
org.apache.hadoop.hive.cli.TestMiniDruidKafkaCliDriver.testCliDriver[druidkafkamini_basic]
 (batchId=297)
org.apache.hadoop.hive.cli.TestMiniDruidKafkaCliDriver.testCliDriver[druidkafkamini_csv]
 (batchId=297)
org.apache.hadoop.hive.cli.TestMiniDruidKafkaCliDriver.testCliDriver[druidkafkamini_delimited]
 (batchId=297)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[cttl] 
(batchId=157)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[table_nonprintable]
 (batchId=157)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[temp_table_external]
 (batchId=158)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[resourceplan]
 (batchId=175)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[strict_managed_tables_sysdb]
 (batchId=176)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[sysdb] 
(batchId=171)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[sysdb_schq] 
(batchId=176)
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[authorization_schq1]
 (batchId=102)
org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.testCliDriver[ct_noperm_loc]
 (batchId=104)
org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.testCliDriver[ctas_noperm_loc]
 (batchId=104)
org.apache.hadoop.hive.cli.TestNegativeMinimrCliDriver.testCliDriver[table_nonprintable_negative]
 (batchId=104)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[add_part_multiple] 
(batchId=145)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[alter_merge_orc] 
(batchId=143)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[alter_merge_stats_orc]
 (batchId=131)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[annotate_stats_join]
 (batchId=137)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[auto_join0] 
(batchId=153)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[auto_join10] 
(batchId=129)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[auto_join11] 
(batchId=117)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[auto_join12] 
(batchId=124)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[auto_join13] 
(batchId=150)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[auto_join14] 
(batchId=119)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[auto_join15] 
(batchId=120)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[auto_join16] 
(batchId=131)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[auto_join17] 
(batchId=151)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[auto_join18] 
(batchId=118)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[auto_join18_multi_distinct]
 (batchId=125)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[auto_join19] 
(batchId=143)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[auto_join1] 
(batchId=149)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[auto_join20] 
(batchId=154)

[jira] [Updated] (HIVE-22083) Values of tag order cannot be null, so it can be "byte" instead of "Byte"

2019-09-09 Thread Ivan Suller (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22083?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ivan Suller updated HIVE-22083:
---
Attachment: HIVE-22083.4.patch

> Values of tag order cannot be null, so it can be "byte" instead of "Byte"
> -
>
> Key: HIVE-22083
> URL: https://issues.apache.org/jira/browse/HIVE-22083
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Reporter: Ivan Suller
>Assignee: Ivan Suller
>Priority: Minor
> Attachments: HIVE-22083.1.patch, HIVE-22083.2.patch, 
> HIVE-22083.3.patch, HIVE-22083.4.patch, HIVE-22083.4.patch, HIVE-22083.4.patch
>
>
> Values of tag order cannot be null, so it can be "byte" instead of "Byte". 
> Switching between Byte and byte is "cheap" - the Byte objects are cached by 
> the JVM - but it still costs a bit more memory and CPU usage.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-21942) Remove useless MetastoreType enum from AbstractCliConfig

2019-09-09 Thread Laszlo Bodor (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-21942?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925679#comment-16925679
 ] 

Laszlo Bodor commented on HIVE-21942:
-

pushed to master, thanks [~imolnar] for the patch!

> Remove useless MetastoreType enum from AbstractCliConfig
> 
>
> Key: HIVE-21942
> URL: https://issues.apache.org/jira/browse/HIVE-21942
> Project: Hive
>  Issue Type: Bug
>  Components: Testing Infrastructure
>Reporter: Laszlo Bodor
>Assignee: Imre Molnar
>Priority: Major
>  Labels: newbie
> Fix For: 4.0.0
>
> Attachments: HIVE-21942.1.patch
>
>
> I'm not 100% aware of historical reasons, but this single-value enum seems 
> totally useless at the moment.
> https://github.com/apache/hive/blob/86a15600ae610e1c6017019883ef9344585572ef/itests/util/src/main/java/org/apache/hadoop/hive/cli/control/AbstractCliConfig.java#L50-L52
> According to HIVE-14854, it may have made sense when there was an initiative 
> of hbase metastore, but it's not in the scope anymore AFAIK.
> https://github.com/apache/hive/commit/474425aa62e3f25b119419439373aa684c6c2121



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-21942) Remove useless MetastoreType enum from AbstractCliConfig

2019-09-09 Thread Laszlo Bodor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21942?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Laszlo Bodor updated HIVE-21942:

Resolution: Fixed
Status: Resolved  (was: Patch Available)

> Remove useless MetastoreType enum from AbstractCliConfig
> 
>
> Key: HIVE-21942
> URL: https://issues.apache.org/jira/browse/HIVE-21942
> Project: Hive
>  Issue Type: Bug
>  Components: Testing Infrastructure
>Reporter: Laszlo Bodor
>Assignee: Imre Molnar
>Priority: Major
>  Labels: newbie
> Fix For: 4.0.0
>
> Attachments: HIVE-21942.1.patch
>
>
> I'm not 100% aware of historical reasons, but this single-value enum seems 
> totally useless at the moment.
> https://github.com/apache/hive/blob/86a15600ae610e1c6017019883ef9344585572ef/itests/util/src/main/java/org/apache/hadoop/hive/cli/control/AbstractCliConfig.java#L50-L52
> According to HIVE-14854, it may have made sense when there was an initiative 
> of hbase metastore, but it's not in the scope anymore AFAIK.
> https://github.com/apache/hive/commit/474425aa62e3f25b119419439373aa684c6c2121



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-21397) BloomFilter for hive Managed [ACID] table does not work as expected

2019-09-09 Thread Denys Kuzmenko (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-21397?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925633#comment-16925633
 ] 

Denys Kuzmenko commented on HIVE-21397:
---

Thank you, [~szita]!

> BloomFilter for hive Managed [ACID] table does not work as expected
> ---
>
> Key: HIVE-21397
> URL: https://issues.apache.org/jira/browse/HIVE-21397
> Project: Hive
>  Issue Type: Bug
>  Components: Hive, HiveServer2, Transactions
>Affects Versions: 3.1.1
>Reporter: vaibhav
>Assignee: Denys Kuzmenko
>Priority: Blocker
> Fix For: 4.0.0
>
> Attachments: HIVE-21397.1.patch, HIVE-21397.2.patch, 
> HIVE-21397.3.patch, HIVE-21397.4.patch, HIVE-21397.5.patch, HIVE-21397.patch
>
>
> Steps to Reproduce this issue : 
> - 
> 1. Create a HIveManaged table as below : 
> - 
> {code:java}
> CREATE TABLE `bloomTest`( 
>    `msisdn` string, 
>    `imsi` varchar(20), 
>    `imei` bigint, 
>    `cell_id` bigint) 
>  ROW FORMAT SERDE 
>    'org.apache.hadoop.hive.ql.io.orc.OrcSerde' 
>  STORED AS INPUTFORMAT 
>    'org.apache.hadoop.hive.ql.io.orc.OrcInputFormat' 
>  OUTPUTFORMAT 
>    'org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat' 
>  LOCATION 
>    
> 'hdfs://c1162-node2.squadron-labs.com:8020/warehouse/tablespace/managed/hive/bloomTest;
>  
>  TBLPROPERTIES ( 
>    'bucketing_version'='2', 
>    'orc.bloom.filter.columns'='msisdn,cell_id,imsi', 
>    'orc.bloom.filter.fpp'='0.02', 
>    'transactional'='true', 
>    'transactional_properties'='default', 
>    'transient_lastDdlTime'='1551206683') {code}
> - 
> 2. Insert a few rows. 
> - 
> - 
> 3. Check if bloom filter or active : [ It does not show bloom filters for 
> hive managed tables ] 
> - 
> {code:java}
> [hive@c1162-node2 root]$ hive --orcfiledump 
> hdfs://c1162-node2.squadron-labs.com:8020/warehouse/tablespace/managed/hive/bloomTest/delta_001_001_
>  | grep -i bloom 
> SLF4J: Class path contains multiple SLF4J bindings. 
> SLF4J: Found binding in 
> [jar:file:/usr/hdp/3.1.0.0-78/hive/lib/log4j-slf4j-impl-2.10.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>  
> SLF4J: Found binding in 
> [jar:file:/usr/hdp/3.1.0.0-78/hadoop/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>  
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation. 
> SLF4J: Actual binding is of type 
> [org.apache.logging.slf4j.Log4jLoggerFactory] 
> Processing data file 
> hdfs://c1162-node2.squadron-labs.com:8020/warehouse/tablespace/managed/hive/bloomTest/delta_001_001_/bucket_0
>  [length: 791] 
> Structure for 
> hdfs://c1162-node2.squadron-labs.com:8020/warehouse/tablespace/managed/hive/bloomTest/delta_001_001_/bucket_0
>  {code}
> - 
> On Another hand: For hive External tables it works : 
> - 
> {code:java}
> CREATE external TABLE `ext_bloomTest`( 
>    `msisdn` string, 
>    `imsi` varchar(20), 
>    `imei` bigint, 
>    `cell_id` bigint) 
>  ROW FORMAT SERDE 
>    'org.apache.hadoop.hive.ql.io.orc.OrcSerde' 
>  STORED AS INPUTFORMAT 
>    'org.apache.hadoop.hive.ql.io.orc.OrcInputFormat' 
>  OUTPUTFORMAT 
>    'org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat' 
>  TBLPROPERTIES ( 
>    'bucketing_version'='2', 
>    'orc.bloom.filter.columns'='msisdn,cell_id,imsi', 
>    'orc.bloom.filter.fpp'='0.02') {code}
> - 
> {code:java}
> [hive@c1162-node2 root]$ hive --orcfiledump 
> hdfs://c1162-node2.squadron-labs.com:8020/warehouse/tablespace/external/hive/ext_bloomTest/00_0
>  | grep -i bloom 
> SLF4J: Class path contains multiple SLF4J bindings. 
> SLF4J: Found binding in 
> [jar:file:/usr/hdp/3.1.0.0-78/hive/lib/log4j-slf4j-impl-2.10.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>  
> SLF4J: Found binding in 
> [jar:file:/usr/hdp/3.1.0.0-78/hadoop/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>  
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation. 
> SLF4J: Actual binding is of type 
> [org.apache.logging.slf4j.Log4jLoggerFactory] 
> Processing data file 
> hdfs://c1162-node2.squadron-labs.com:8020/warehouse/tablespace/external/hive/ext_bloomTest/00_0
>  [length: 755] 
> Structure for 
> hdfs://c1162-node2.squadron-labs.com:8020/warehouse/tablespace/external/hive/ext_bloomTest/00_0
>  
> Stream: column 1 section BLOOM_FILTER_UTF8 start: 41 length 110 
> Stream: column 2 section BLOOM_FILTER_UTF8 start: 178 length 114 

[jira] [Updated] (HIVE-21397) BloomFilter for hive Managed [ACID] table does not work as expected

2019-09-09 Thread Adam Szita (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21397?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Adam Szita updated HIVE-21397:
--
Fix Version/s: 4.0.0
   Resolution: Fixed
   Status: Resolved  (was: Patch Available)

Committed to master, thanks [~dkuzmenko].

> BloomFilter for hive Managed [ACID] table does not work as expected
> ---
>
> Key: HIVE-21397
> URL: https://issues.apache.org/jira/browse/HIVE-21397
> Project: Hive
>  Issue Type: Bug
>  Components: Hive, HiveServer2, Transactions
>Affects Versions: 3.1.1
>Reporter: vaibhav
>Assignee: Denys Kuzmenko
>Priority: Blocker
> Fix For: 4.0.0
>
> Attachments: HIVE-21397.1.patch, HIVE-21397.2.patch, 
> HIVE-21397.3.patch, HIVE-21397.4.patch, HIVE-21397.5.patch, HIVE-21397.patch
>
>
> Steps to Reproduce this issue : 
> - 
> 1. Create a HIveManaged table as below : 
> - 
> {code:java}
> CREATE TABLE `bloomTest`( 
>    `msisdn` string, 
>    `imsi` varchar(20), 
>    `imei` bigint, 
>    `cell_id` bigint) 
>  ROW FORMAT SERDE 
>    'org.apache.hadoop.hive.ql.io.orc.OrcSerde' 
>  STORED AS INPUTFORMAT 
>    'org.apache.hadoop.hive.ql.io.orc.OrcInputFormat' 
>  OUTPUTFORMAT 
>    'org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat' 
>  LOCATION 
>    
> 'hdfs://c1162-node2.squadron-labs.com:8020/warehouse/tablespace/managed/hive/bloomTest;
>  
>  TBLPROPERTIES ( 
>    'bucketing_version'='2', 
>    'orc.bloom.filter.columns'='msisdn,cell_id,imsi', 
>    'orc.bloom.filter.fpp'='0.02', 
>    'transactional'='true', 
>    'transactional_properties'='default', 
>    'transient_lastDdlTime'='1551206683') {code}
> - 
> 2. Insert a few rows. 
> - 
> - 
> 3. Check if bloom filter or active : [ It does not show bloom filters for 
> hive managed tables ] 
> - 
> {code:java}
> [hive@c1162-node2 root]$ hive --orcfiledump 
> hdfs://c1162-node2.squadron-labs.com:8020/warehouse/tablespace/managed/hive/bloomTest/delta_001_001_
>  | grep -i bloom 
> SLF4J: Class path contains multiple SLF4J bindings. 
> SLF4J: Found binding in 
> [jar:file:/usr/hdp/3.1.0.0-78/hive/lib/log4j-slf4j-impl-2.10.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>  
> SLF4J: Found binding in 
> [jar:file:/usr/hdp/3.1.0.0-78/hadoop/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>  
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation. 
> SLF4J: Actual binding is of type 
> [org.apache.logging.slf4j.Log4jLoggerFactory] 
> Processing data file 
> hdfs://c1162-node2.squadron-labs.com:8020/warehouse/tablespace/managed/hive/bloomTest/delta_001_001_/bucket_0
>  [length: 791] 
> Structure for 
> hdfs://c1162-node2.squadron-labs.com:8020/warehouse/tablespace/managed/hive/bloomTest/delta_001_001_/bucket_0
>  {code}
> - 
> On Another hand: For hive External tables it works : 
> - 
> {code:java}
> CREATE external TABLE `ext_bloomTest`( 
>    `msisdn` string, 
>    `imsi` varchar(20), 
>    `imei` bigint, 
>    `cell_id` bigint) 
>  ROW FORMAT SERDE 
>    'org.apache.hadoop.hive.ql.io.orc.OrcSerde' 
>  STORED AS INPUTFORMAT 
>    'org.apache.hadoop.hive.ql.io.orc.OrcInputFormat' 
>  OUTPUTFORMAT 
>    'org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat' 
>  TBLPROPERTIES ( 
>    'bucketing_version'='2', 
>    'orc.bloom.filter.columns'='msisdn,cell_id,imsi', 
>    'orc.bloom.filter.fpp'='0.02') {code}
> - 
> {code:java}
> [hive@c1162-node2 root]$ hive --orcfiledump 
> hdfs://c1162-node2.squadron-labs.com:8020/warehouse/tablespace/external/hive/ext_bloomTest/00_0
>  | grep -i bloom 
> SLF4J: Class path contains multiple SLF4J bindings. 
> SLF4J: Found binding in 
> [jar:file:/usr/hdp/3.1.0.0-78/hive/lib/log4j-slf4j-impl-2.10.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>  
> SLF4J: Found binding in 
> [jar:file:/usr/hdp/3.1.0.0-78/hadoop/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>  
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation. 
> SLF4J: Actual binding is of type 
> [org.apache.logging.slf4j.Log4jLoggerFactory] 
> Processing data file 
> hdfs://c1162-node2.squadron-labs.com:8020/warehouse/tablespace/external/hive/ext_bloomTest/00_0
>  [length: 755] 
> Structure for 
> hdfs://c1162-node2.squadron-labs.com:8020/warehouse/tablespace/external/hive/ext_bloomTest/00_0
>  
> Stream: column 1 section BLOOM_FILTER_UTF8 start: 41 length 

[jira] [Commented] (HIVE-21508) ClassCastException when initializing HiveMetaStoreClient on JDK10 or newer

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-21508?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925590#comment-16925590
 ] 

Hive QA commented on HIVE-21508:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12979838/HIVE-21508.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 16749 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[llap_acid2] 
(batchId=163)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/18504/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/18504/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-18504/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12979838 - PreCommit-HIVE-Build

> ClassCastException when initializing HiveMetaStoreClient on JDK10 or newer
> --
>
> Key: HIVE-21508
> URL: https://issues.apache.org/jira/browse/HIVE-21508
> Project: Hive
>  Issue Type: Bug
>  Components: Clients
>Affects Versions: 3.2.0, 2.3.4
>Reporter: Adar Dembo
>Assignee: Ana Jalba
>Priority: Major
> Fix For: 4.0.0, 2.3.7
>
> Attachments: HIVE-21508.1.branch-2.3.patch, HIVE-21508.patch
>
>
> There's this block of code in {{HiveMetaStoreClient:resolveUris}} (called 
> from the constructor) on master:
> {noformat}
>   private URI metastoreUris[];
>   ...
>   if (MetastoreConf.getVar(conf, 
> ConfVars.THRIFT_URI_SELECTION).equalsIgnoreCase("RANDOM")) {
> List uriList = Arrays.asList(metastoreUris);
> Collections.shuffle(uriList);
> metastoreUris = (URI[]) uriList.toArray();
>   }
> {noformat}
> The cast to {{URI[]}} throws a {{ClassCastException}} beginning with JDK 10, 
> possibly with JDK 9 as well. Note that {{THRIFT_URI_SELECTION}} defaults to 
> {{RANDOM}} so this should affect anyone who creates a 
> {{HiveMetaStoreClient}}. On master this can be overridden with {{SEQUENTIAL}} 
> to avoid the broken case; I'm working against 2.3.4 where there's no such 
> workaround.
> [Here's|https://stackoverflow.com/questions/51372788/array-cast-java-8-vs-java-9]
>  a StackOverflow post that explains the issue in more detail. Interestingly, 
> the author described the issue in the context of the HMS; not sure why there 
> was no follow up with a Hive bug report.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-21508) ClassCastException when initializing HiveMetaStoreClient on JDK10 or newer

2019-09-09 Thread Ana Jalba (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21508?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ana Jalba updated HIVE-21508:
-
Fix Version/s: 2.3.7

> ClassCastException when initializing HiveMetaStoreClient on JDK10 or newer
> --
>
> Key: HIVE-21508
> URL: https://issues.apache.org/jira/browse/HIVE-21508
> Project: Hive
>  Issue Type: Bug
>  Components: Clients
>Affects Versions: 3.2.0, 2.3.4
>Reporter: Adar Dembo
>Assignee: Ana Jalba
>Priority: Major
> Fix For: 4.0.0, 2.3.7
>
> Attachments: HIVE-21508.1.branch-2.3.patch, HIVE-21508.patch
>
>
> There's this block of code in {{HiveMetaStoreClient:resolveUris}} (called 
> from the constructor) on master:
> {noformat}
>   private URI metastoreUris[];
>   ...
>   if (MetastoreConf.getVar(conf, 
> ConfVars.THRIFT_URI_SELECTION).equalsIgnoreCase("RANDOM")) {
> List uriList = Arrays.asList(metastoreUris);
> Collections.shuffle(uriList);
> metastoreUris = (URI[]) uriList.toArray();
>   }
> {noformat}
> The cast to {{URI[]}} throws a {{ClassCastException}} beginning with JDK 10, 
> possibly with JDK 9 as well. Note that {{THRIFT_URI_SELECTION}} defaults to 
> {{RANDOM}} so this should affect anyone who creates a 
> {{HiveMetaStoreClient}}. On master this can be overridden with {{SEQUENTIAL}} 
> to avoid the broken case; I'm working against 2.3.4 where there's no such 
> workaround.
> [Here's|https://stackoverflow.com/questions/51372788/array-cast-java-8-vs-java-9]
>  a StackOverflow post that explains the issue in more detail. Interestingly, 
> the author described the issue in the context of the HMS; not sure why there 
> was no follow up with a Hive bug report.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-21508) ClassCastException when initializing HiveMetaStoreClient on JDK10 or newer

2019-09-09 Thread Ana Jalba (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21508?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ana Jalba updated HIVE-21508:
-
Attachment: HIVE-21508.1.branch-2.3.patch

> ClassCastException when initializing HiveMetaStoreClient on JDK10 or newer
> --
>
> Key: HIVE-21508
> URL: https://issues.apache.org/jira/browse/HIVE-21508
> Project: Hive
>  Issue Type: Bug
>  Components: Clients
>Affects Versions: 3.2.0, 2.3.4
>Reporter: Adar Dembo
>Assignee: Ana Jalba
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-21508.1.branch-2.3.patch, HIVE-21508.patch
>
>
> There's this block of code in {{HiveMetaStoreClient:resolveUris}} (called 
> from the constructor) on master:
> {noformat}
>   private URI metastoreUris[];
>   ...
>   if (MetastoreConf.getVar(conf, 
> ConfVars.THRIFT_URI_SELECTION).equalsIgnoreCase("RANDOM")) {
> List uriList = Arrays.asList(metastoreUris);
> Collections.shuffle(uriList);
> metastoreUris = (URI[]) uriList.toArray();
>   }
> {noformat}
> The cast to {{URI[]}} throws a {{ClassCastException}} beginning with JDK 10, 
> possibly with JDK 9 as well. Note that {{THRIFT_URI_SELECTION}} defaults to 
> {{RANDOM}} so this should affect anyone who creates a 
> {{HiveMetaStoreClient}}. On master this can be overridden with {{SEQUENTIAL}} 
> to avoid the broken case; I'm working against 2.3.4 where there's no such 
> workaround.
> [Here's|https://stackoverflow.com/questions/51372788/array-cast-java-8-vs-java-9]
>  a StackOverflow post that explains the issue in more detail. Interestingly, 
> the author described the issue in the context of the HMS; not sure why there 
> was no follow up with a Hive bug report.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-21508) ClassCastException when initializing HiveMetaStoreClient on JDK10 or newer

2019-09-09 Thread Ana Jalba (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-21508?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925568#comment-16925568
 ] 

Ana Jalba commented on HIVE-21508:
--

Looks like the master patch was ok. Can this be merged?

> ClassCastException when initializing HiveMetaStoreClient on JDK10 or newer
> --
>
> Key: HIVE-21508
> URL: https://issues.apache.org/jira/browse/HIVE-21508
> Project: Hive
>  Issue Type: Bug
>  Components: Clients
>Affects Versions: 3.2.0, 2.3.4
>Reporter: Adar Dembo
>Assignee: Ana Jalba
>Priority: Major
> Fix For: 3.2.0
>
> Attachments: HIVE-21508.patch
>
>
> There's this block of code in {{HiveMetaStoreClient:resolveUris}} (called 
> from the constructor) on master:
> {noformat}
>   private URI metastoreUris[];
>   ...
>   if (MetastoreConf.getVar(conf, 
> ConfVars.THRIFT_URI_SELECTION).equalsIgnoreCase("RANDOM")) {
> List uriList = Arrays.asList(metastoreUris);
> Collections.shuffle(uriList);
> metastoreUris = (URI[]) uriList.toArray();
>   }
> {noformat}
> The cast to {{URI[]}} throws a {{ClassCastException}} beginning with JDK 10, 
> possibly with JDK 9 as well. Note that {{THRIFT_URI_SELECTION}} defaults to 
> {{RANDOM}} so this should affect anyone who creates a 
> {{HiveMetaStoreClient}}. On master this can be overridden with {{SEQUENTIAL}} 
> to avoid the broken case; I'm working against 2.3.4 where there's no such 
> workaround.
> [Here's|https://stackoverflow.com/questions/51372788/array-cast-java-8-vs-java-9]
>  a StackOverflow post that explains the issue in more detail. Interestingly, 
> the author described the issue in the context of the HMS; not sure why there 
> was no follow up with a Hive bug report.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-21508) ClassCastException when initializing HiveMetaStoreClient on JDK10 or newer

2019-09-09 Thread Ana Jalba (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21508?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ana Jalba updated HIVE-21508:
-
Fix Version/s: (was: 3.2.0)
   4.0.0

> ClassCastException when initializing HiveMetaStoreClient on JDK10 or newer
> --
>
> Key: HIVE-21508
> URL: https://issues.apache.org/jira/browse/HIVE-21508
> Project: Hive
>  Issue Type: Bug
>  Components: Clients
>Affects Versions: 3.2.0, 2.3.4
>Reporter: Adar Dembo
>Assignee: Ana Jalba
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-21508.patch
>
>
> There's this block of code in {{HiveMetaStoreClient:resolveUris}} (called 
> from the constructor) on master:
> {noformat}
>   private URI metastoreUris[];
>   ...
>   if (MetastoreConf.getVar(conf, 
> ConfVars.THRIFT_URI_SELECTION).equalsIgnoreCase("RANDOM")) {
> List uriList = Arrays.asList(metastoreUris);
> Collections.shuffle(uriList);
> metastoreUris = (URI[]) uriList.toArray();
>   }
> {noformat}
> The cast to {{URI[]}} throws a {{ClassCastException}} beginning with JDK 10, 
> possibly with JDK 9 as well. Note that {{THRIFT_URI_SELECTION}} defaults to 
> {{RANDOM}} so this should affect anyone who creates a 
> {{HiveMetaStoreClient}}. On master this can be overridden with {{SEQUENTIAL}} 
> to avoid the broken case; I'm working against 2.3.4 where there's no such 
> workaround.
> [Here's|https://stackoverflow.com/questions/51372788/array-cast-java-8-vs-java-9]
>  a StackOverflow post that explains the issue in more detail. Interestingly, 
> the author described the issue in the context of the HMS; not sure why there 
> was no follow up with a Hive bug report.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-22178) Parquet FilterPredicate throws CastException after SchemaEvolution.

2019-09-09 Thread Sankar Hariappan (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22178?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sankar Hariappan updated HIVE-22178:

Fix Version/s: 4.0.0
   Resolution: Fixed
   Status: Resolved  (was: Patch Available)

Committed to master!
Thanks [~nareshpr] for the patch!

> Parquet FilterPredicate throws CastException after SchemaEvolution.
> ---
>
> Key: HIVE-22178
> URL: https://issues.apache.org/jira/browse/HIVE-22178
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 3.1.0
>Reporter: Naresh P R
>Assignee: Naresh P R
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-22178.1.patch, HIVE-22178.2.patch, 
> HIVE-22178.3.patch
>
>
> Below are the repro steps.
> {code:java}
> create table parq_test(age int, name string) stored as parquet;
> insert into parq_test values(1, '');
> alter table parq_test change age age string;
> insert into parq_test values('b', '');
> select * from parq_test where age='b';{code}
> Exception thrown after changing column datatype is below
> {code:java}
> Caused by: java.lang.IllegalArgumentException: FilterPredicate column: age's 
> declared type (org.apache.parquet.io.api.Binary) does not match the schema 
> found in file metadata. Column age is of type: INT32
> Valid types for this column are: [class java.lang.Integer]
>  at 
> org.apache.parquet.filter2.predicate.ValidTypeMap.assertTypeValid(ValidTypeMap.java:126)
>  at 
> org.apache.parquet.filter2.predicate.SchemaCompatibilityValidator.validateColumn(SchemaCompatibilityValidator.java:181)
>  at 
> org.apache.parquet.filter2.predicate.SchemaCompatibilityValidator.validateColumnFilterPredicate(SchemaCompatibilityValidator.java:151)
>  at 
> org.apache.parquet.filter2.predicate.SchemaCompatibilityValidator.visit(SchemaCompatibilityValidator.java:85)
>  at 
> org.apache.parquet.filter2.predicate.SchemaCompatibilityValidator.visit(SchemaCompatibilityValidator.java:58)
>  at 
> org.apache.parquet.filter2.predicate.Operators$Eq.accept(Operators.java:181)
>  at 
> org.apache.parquet.filter2.predicate.SchemaCompatibilityValidator.validate(SchemaCompatibilityValidator.java:63)
>  at 
> org.apache.parquet.filter2.compat.RowGroupFilter.visit(RowGroupFilter.java:92)
>  at 
> org.apache.parquet.filter2.compat.RowGroupFilter.visit(RowGroupFilter.java:43)
>  at 
> org.apache.parquet.filter2.compat.FilterCompat$FilterPredicateCompat.accept(FilterCompat.java:137)
>  at 
> org.apache.parquet.filter2.compat.RowGroupFilter.filterRowGroups(RowGroupFilter.java:64)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.ParquetRecordReaderBase.getSplit(ParquetRecordReaderBase.java:111)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.(VectorizedParquetRecordReader.java:147)
>  ... 31 more{code}



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-22178) Parquet FilterPredicate throws CastException after SchemaEvolution.

2019-09-09 Thread Sankar Hariappan (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22178?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925551#comment-16925551
 ] 

Sankar Hariappan commented on HIVE-22178:
-

+1, LGTM

> Parquet FilterPredicate throws CastException after SchemaEvolution.
> ---
>
> Key: HIVE-22178
> URL: https://issues.apache.org/jira/browse/HIVE-22178
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 3.1.0
>Reporter: Naresh P R
>Assignee: Naresh P R
>Priority: Major
> Attachments: HIVE-22178.1.patch, HIVE-22178.2.patch, 
> HIVE-22178.3.patch
>
>
> Below are the repro steps.
> {code:java}
> create table parq_test(age int, name string) stored as parquet;
> insert into parq_test values(1, '');
> alter table parq_test change age age string;
> insert into parq_test values('b', '');
> select * from parq_test where age='b';{code}
> Exception thrown after changing column datatype is below
> {code:java}
> Caused by: java.lang.IllegalArgumentException: FilterPredicate column: age's 
> declared type (org.apache.parquet.io.api.Binary) does not match the schema 
> found in file metadata. Column age is of type: INT32
> Valid types for this column are: [class java.lang.Integer]
>  at 
> org.apache.parquet.filter2.predicate.ValidTypeMap.assertTypeValid(ValidTypeMap.java:126)
>  at 
> org.apache.parquet.filter2.predicate.SchemaCompatibilityValidator.validateColumn(SchemaCompatibilityValidator.java:181)
>  at 
> org.apache.parquet.filter2.predicate.SchemaCompatibilityValidator.validateColumnFilterPredicate(SchemaCompatibilityValidator.java:151)
>  at 
> org.apache.parquet.filter2.predicate.SchemaCompatibilityValidator.visit(SchemaCompatibilityValidator.java:85)
>  at 
> org.apache.parquet.filter2.predicate.SchemaCompatibilityValidator.visit(SchemaCompatibilityValidator.java:58)
>  at 
> org.apache.parquet.filter2.predicate.Operators$Eq.accept(Operators.java:181)
>  at 
> org.apache.parquet.filter2.predicate.SchemaCompatibilityValidator.validate(SchemaCompatibilityValidator.java:63)
>  at 
> org.apache.parquet.filter2.compat.RowGroupFilter.visit(RowGroupFilter.java:92)
>  at 
> org.apache.parquet.filter2.compat.RowGroupFilter.visit(RowGroupFilter.java:43)
>  at 
> org.apache.parquet.filter2.compat.FilterCompat$FilterPredicateCompat.accept(FilterCompat.java:137)
>  at 
> org.apache.parquet.filter2.compat.RowGroupFilter.filterRowGroups(RowGroupFilter.java:64)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.ParquetRecordReaderBase.getSplit(ParquetRecordReaderBase.java:111)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.(VectorizedParquetRecordReader.java:147)
>  ... 31 more{code}



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Assigned] (HIVE-21604) preCommit job should not be triggered on non-patch attachments

2019-09-09 Thread Laszlo Bodor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21604?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Laszlo Bodor reassigned HIVE-21604:
---

Assignee: Laszlo Bodor

> preCommit job should not be triggered on non-patch attachments
> --
>
> Key: HIVE-21604
> URL: https://issues.apache.org/jira/browse/HIVE-21604
> Project: Hive
>  Issue Type: Bug
>  Components: Testing Infrastructure
>Reporter: Laszlo Bodor
>Assignee: Laszlo Bodor
>Priority: Major
> Attachments: HIVE-21604.01.patch, files.txt
>
>
> latest example: HIVE-21604
> https://issues.apache.org/jira/browse/HIVE-14669?focusedCommentId=16815520=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#comment-16815520
> I think one should be able to upload any kind of attachments (e.g. 
> screenshot) without triggering the precommit job
> 2 possible ways:
> 1. strict: enable only .patch (should work)
> (2. lenient: introduce blacklist, .png ...)



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Comment Edited] (HIVE-21604) preCommit job should not be triggered on non-patch attachments

2019-09-09 Thread Laszlo Bodor (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-21604?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925542#comment-16925542
 ] 

Laszlo Bodor edited comment on HIVE-21604 at 9/9/19 10:02 AM:
--

tested with attached files.txt
original script: grep -o '"/jira/secure/attachment/[0-9]*/[^"]*' files.txt | 
grep -v -e 'htm[l]*$' | sort | grep -o '/jira/secure/attachment/[0-9]*/[^"]*'

{code}
/jira/secure/attachment/12964860/files.txt
/jira/secure/attachment/12964869/01_mvn_out.png
/jira/secure/attachment/12964870/02_hive_log.png
/jira/secure/attachment/12965605/Screen%20Shot%202019-04-11%20at%204.54.41%20PM.png
/jira/secure/attachment/12965681/HIVE-14469.01.patch
/jira/secure/attachment/12965721/HIVE-14469.02.patch
/jira/secure/attachment/12967773/HIVE-14469.03.patch
{code}

modified script: grep -o '"/jira/secure/attachment/[0-9]*/[^"]*\.patch*' 
files.txt | sort | grep -o '/jira/secure/attachment/[0-9]*/[^"]*\.patch*' 
{code}
/jira/secure/attachment/12965681/HIVE-14469.01.patch
/jira/secure/attachment/12965721/HIVE-14469.02.patch
/jira/secure/attachment/12967773/HIVE-14469.03.patch
{code}

I think the original "grep -v" wasn't satisfying (only filtered html files)

with the current solution, it doesn't seem too strict to force devs to upload 
only .patch files for patches

[~vihangk1]: could you please review the patch?


was (Author: abstractdog):
tested with attached files.txt
original script: grep -o '"/jira/secure/attachment/[0-9]*/[^"]*' files.txt | 
grep -v -e 'htm[l]*$' | sort | grep -o '/jira/secure/attachment/[0-9]*/[^"]*'

{code}
/jira/secure/attachment/12964860/files.txt
/jira/secure/attachment/12964869/01_mvn_out.png
/jira/secure/attachment/12964870/02_hive_log.png
/jira/secure/attachment/12965605/Screen%20Shot%202019-04-11%20at%204.54.41%20PM.png
/jira/secure/attachment/12965681/HIVE-14469.01.patch
/jira/secure/attachment/12965721/HIVE-14469.02.patch
/jira/secure/attachment/12967773/HIVE-14469.03.patch
{code}

modified script: grep -o '"/jira/secure/attachment/[0-9]*/[^"]*\.patch*' 
files.txt | sort | grep -o '/jira/secure/attachment/[0-9]*/[^"]*\.patch*' 
{code}
/jira/secure/attachment/12965681/HIVE-14469.01.patch
/jira/secure/attachment/12965721/HIVE-14469.02.patch
/jira/secure/attachment/12967773/HIVE-14469.03.patch
{code}

I think the original "grep -v" wasn't satisfying (only filtered html files), as 
it doesn't seem too strict to force devs to upload .patch files

[~vihangk1]: could you please review the patch?

> preCommit job should not be triggered on non-patch attachments
> --
>
> Key: HIVE-21604
> URL: https://issues.apache.org/jira/browse/HIVE-21604
> Project: Hive
>  Issue Type: Bug
>  Components: Testing Infrastructure
>Reporter: Laszlo Bodor
>Priority: Major
> Attachments: HIVE-21604.01.patch, files.txt
>
>
> latest example: HIVE-21604
> https://issues.apache.org/jira/browse/HIVE-14669?focusedCommentId=16815520=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#comment-16815520
> I think one should be able to upload any kind of attachments (e.g. 
> screenshot) without triggering the precommit job
> 2 possible ways:
> 1. strict: enable only .patch (should work)
> (2. lenient: introduce blacklist, .png ...)



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-21604) preCommit job should not be triggered on non-patch attachments

2019-09-09 Thread Laszlo Bodor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21604?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Laszlo Bodor updated HIVE-21604:

Status: Patch Available  (was: Open)

> preCommit job should not be triggered on non-patch attachments
> --
>
> Key: HIVE-21604
> URL: https://issues.apache.org/jira/browse/HIVE-21604
> Project: Hive
>  Issue Type: Bug
>  Components: Testing Infrastructure
>Reporter: Laszlo Bodor
>Assignee: Laszlo Bodor
>Priority: Major
> Attachments: HIVE-21604.01.patch, files.txt
>
>
> latest example: HIVE-21604
> https://issues.apache.org/jira/browse/HIVE-14669?focusedCommentId=16815520=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#comment-16815520
> I think one should be able to upload any kind of attachments (e.g. 
> screenshot) without triggering the precommit job
> 2 possible ways:
> 1. strict: enable only .patch (should work)
> (2. lenient: introduce blacklist, .png ...)



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-21604) preCommit job should not be triggered on non-patch attachments

2019-09-09 Thread Laszlo Bodor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21604?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Laszlo Bodor updated HIVE-21604:

Fix Version/s: 4.0.0

> preCommit job should not be triggered on non-patch attachments
> --
>
> Key: HIVE-21604
> URL: https://issues.apache.org/jira/browse/HIVE-21604
> Project: Hive
>  Issue Type: Bug
>  Components: Testing Infrastructure
>Reporter: Laszlo Bodor
>Assignee: Laszlo Bodor
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-21604.01.patch, files.txt
>
>
> latest example: HIVE-21604
> https://issues.apache.org/jira/browse/HIVE-14669?focusedCommentId=16815520=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#comment-16815520
> I think one should be able to upload any kind of attachments (e.g. 
> screenshot) without triggering the precommit job
> 2 possible ways:
> 1. strict: enable only .patch (should work)
> (2. lenient: introduce blacklist, .png ...)



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Commented] (HIVE-21508) ClassCastException when initializing HiveMetaStoreClient on JDK10 or newer

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-21508?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925543#comment-16925543
 ] 

Hive QA commented on HIVE-21508:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  6m 
47s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
23s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
10s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  1m 
51s{color} | {color:blue} standalone-metastore/metastore-common in master has 
32 extant Findbugs warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
42s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
26s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
23s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
23s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
10s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
58s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
43s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
10s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 14m  1s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-18504/dev-support/hive-personality.sh
 |
| git revision | master / 1b04924 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| modules | C: standalone-metastore/metastore-common U: 
standalone-metastore/metastore-common |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18504/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> ClassCastException when initializing HiveMetaStoreClient on JDK10 or newer
> --
>
> Key: HIVE-21508
> URL: https://issues.apache.org/jira/browse/HIVE-21508
> Project: Hive
>  Issue Type: Bug
>  Components: Clients
>Affects Versions: 3.2.0, 2.3.4
>Reporter: Adar Dembo
>Assignee: Ana Jalba
>Priority: Major
> Fix For: 3.2.0
>
> Attachments: HIVE-21508.patch
>
>
> There's this block of code in {{HiveMetaStoreClient:resolveUris}} (called 
> from the constructor) on master:
> {noformat}
>   private URI metastoreUris[];
>   ...
>   if (MetastoreConf.getVar(conf, 
> ConfVars.THRIFT_URI_SELECTION).equalsIgnoreCase("RANDOM")) {
> List uriList = Arrays.asList(metastoreUris);
> Collections.shuffle(uriList);
> metastoreUris = (URI[]) uriList.toArray();
>   }
> {noformat}
> The cast to {{URI[]}} throws a {{ClassCastException}} beginning with JDK 10, 
> possibly with JDK 9 as well. Note that {{THRIFT_URI_SELECTION}} defaults to 
> {{RANDOM}} so this should affect anyone who creates a 
> {{HiveMetaStoreClient}}. On master this can be overridden with {{SEQUENTIAL}} 
> to avoid the broken case; I'm working against 2.3.4 where there's no such 
> workaround.
> [Here's|https://stackoverflow.com/questions/51372788/array-cast-java-8-vs-java-9]
>  a StackOverflow post that explains the issue in more detail. Interestingly, 
> the author described the issue in the context of the HMS; not sure why 

[jira] [Commented] (HIVE-21604) preCommit job should not be triggered on non-patch attachments

2019-09-09 Thread Laszlo Bodor (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-21604?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925542#comment-16925542
 ] 

Laszlo Bodor commented on HIVE-21604:
-

tested with attached files.txt
original script: grep -o '"/jira/secure/attachment/[0-9]*/[^"]*' files.txt | 
grep -v -e 'htm[l]*$' | sort | grep -o '/jira/secure/attachment/[0-9]*/[^"]*'

{code}
/jira/secure/attachment/12964860/files.txt
/jira/secure/attachment/12964869/01_mvn_out.png
/jira/secure/attachment/12964870/02_hive_log.png
/jira/secure/attachment/12965605/Screen%20Shot%202019-04-11%20at%204.54.41%20PM.png
/jira/secure/attachment/12965681/HIVE-14469.01.patch
/jira/secure/attachment/12965721/HIVE-14469.02.patch
/jira/secure/attachment/12967773/HIVE-14469.03.patch
{code}

modified script: grep -o '"/jira/secure/attachment/[0-9]*/[^"]*\.patch*' 
files.txt | sort | grep -o '/jira/secure/attachment/[0-9]*/[^"]*\.patch*' 
{code}
/jira/secure/attachment/12965681/HIVE-14469.01.patch
/jira/secure/attachment/12965721/HIVE-14469.02.patch
/jira/secure/attachment/12967773/HIVE-14469.03.patch
{code}

I think the original "grep -v" wasn't satisfying (only filtered html files), as 
it doesn't seem too strict to force devs to upload .patch files

[~vihangk1]: could you please review the patch

> preCommit job should not be triggered on non-patch attachments
> --
>
> Key: HIVE-21604
> URL: https://issues.apache.org/jira/browse/HIVE-21604
> Project: Hive
>  Issue Type: Bug
>  Components: Testing Infrastructure
>Reporter: Laszlo Bodor
>Priority: Major
> Attachments: HIVE-21604.01.patch, files.txt
>
>
> latest example: HIVE-21604
> https://issues.apache.org/jira/browse/HIVE-14669?focusedCommentId=16815520=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#comment-16815520
> I think one should be able to upload any kind of attachments (e.g. 
> screenshot) without triggering the precommit job
> 2 possible ways:
> 1. strict: enable only .patch (should work)
> (2. lenient: introduce blacklist, .png ...)



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Comment Edited] (HIVE-21604) preCommit job should not be triggered on non-patch attachments

2019-09-09 Thread Laszlo Bodor (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-21604?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925542#comment-16925542
 ] 

Laszlo Bodor edited comment on HIVE-21604 at 9/9/19 10:00 AM:
--

tested with attached files.txt
original script: grep -o '"/jira/secure/attachment/[0-9]*/[^"]*' files.txt | 
grep -v -e 'htm[l]*$' | sort | grep -o '/jira/secure/attachment/[0-9]*/[^"]*'

{code}
/jira/secure/attachment/12964860/files.txt
/jira/secure/attachment/12964869/01_mvn_out.png
/jira/secure/attachment/12964870/02_hive_log.png
/jira/secure/attachment/12965605/Screen%20Shot%202019-04-11%20at%204.54.41%20PM.png
/jira/secure/attachment/12965681/HIVE-14469.01.patch
/jira/secure/attachment/12965721/HIVE-14469.02.patch
/jira/secure/attachment/12967773/HIVE-14469.03.patch
{code}

modified script: grep -o '"/jira/secure/attachment/[0-9]*/[^"]*\.patch*' 
files.txt | sort | grep -o '/jira/secure/attachment/[0-9]*/[^"]*\.patch*' 
{code}
/jira/secure/attachment/12965681/HIVE-14469.01.patch
/jira/secure/attachment/12965721/HIVE-14469.02.patch
/jira/secure/attachment/12967773/HIVE-14469.03.patch
{code}

I think the original "grep -v" wasn't satisfying (only filtered html files), as 
it doesn't seem too strict to force devs to upload .patch files

[~vihangk1]: could you please review the patch?


was (Author: abstractdog):
tested with attached files.txt
original script: grep -o '"/jira/secure/attachment/[0-9]*/[^"]*' files.txt | 
grep -v -e 'htm[l]*$' | sort | grep -o '/jira/secure/attachment/[0-9]*/[^"]*'

{code}
/jira/secure/attachment/12964860/files.txt
/jira/secure/attachment/12964869/01_mvn_out.png
/jira/secure/attachment/12964870/02_hive_log.png
/jira/secure/attachment/12965605/Screen%20Shot%202019-04-11%20at%204.54.41%20PM.png
/jira/secure/attachment/12965681/HIVE-14469.01.patch
/jira/secure/attachment/12965721/HIVE-14469.02.patch
/jira/secure/attachment/12967773/HIVE-14469.03.patch
{code}

modified script: grep -o '"/jira/secure/attachment/[0-9]*/[^"]*\.patch*' 
files.txt | sort | grep -o '/jira/secure/attachment/[0-9]*/[^"]*\.patch*' 
{code}
/jira/secure/attachment/12965681/HIVE-14469.01.patch
/jira/secure/attachment/12965721/HIVE-14469.02.patch
/jira/secure/attachment/12967773/HIVE-14469.03.patch
{code}

I think the original "grep -v" wasn't satisfying (only filtered html files), as 
it doesn't seem too strict to force devs to upload .patch files

[~vihangk1]: could you please review the patch

> preCommit job should not be triggered on non-patch attachments
> --
>
> Key: HIVE-21604
> URL: https://issues.apache.org/jira/browse/HIVE-21604
> Project: Hive
>  Issue Type: Bug
>  Components: Testing Infrastructure
>Reporter: Laszlo Bodor
>Priority: Major
> Attachments: HIVE-21604.01.patch, files.txt
>
>
> latest example: HIVE-21604
> https://issues.apache.org/jira/browse/HIVE-14669?focusedCommentId=16815520=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#comment-16815520
> I think one should be able to upload any kind of attachments (e.g. 
> screenshot) without triggering the precommit job
> 2 possible ways:
> 1. strict: enable only .patch (should work)
> (2. lenient: introduce blacklist, .png ...)



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-21604) preCommit job should not be triggered on non-patch attachments

2019-09-09 Thread Laszlo Bodor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21604?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Laszlo Bodor updated HIVE-21604:

Attachment: HIVE-21604.01.patch

> preCommit job should not be triggered on non-patch attachments
> --
>
> Key: HIVE-21604
> URL: https://issues.apache.org/jira/browse/HIVE-21604
> Project: Hive
>  Issue Type: Bug
>  Components: Testing Infrastructure
>Reporter: Laszlo Bodor
>Priority: Major
> Attachments: HIVE-21604.01.patch, files.txt
>
>
> latest example: HIVE-21604
> https://issues.apache.org/jira/browse/HIVE-14669?focusedCommentId=16815520=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#comment-16815520
> I think one should be able to upload any kind of attachments (e.g. 
> screenshot) without triggering the precommit job
> 2 possible ways:
> 1. strict: enable only .patch (should work)
> (2. lenient: introduce blacklist, .png ...)



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-21604) preCommit job should not be triggered on non-patch attachments

2019-09-09 Thread Laszlo Bodor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21604?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Laszlo Bodor updated HIVE-21604:

Attachment: files.txt

> preCommit job should not be triggered on non-patch attachments
> --
>
> Key: HIVE-21604
> URL: https://issues.apache.org/jira/browse/HIVE-21604
> Project: Hive
>  Issue Type: Bug
>  Components: Testing Infrastructure
>Reporter: Laszlo Bodor
>Priority: Major
> Attachments: HIVE-21604.01.patch, files.txt
>
>
> latest example: HIVE-21604
> https://issues.apache.org/jira/browse/HIVE-14669?focusedCommentId=16815520=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#comment-16815520
> I think one should be able to upload any kind of attachments (e.g. 
> screenshot) without triggering the precommit job
> 2 possible ways:
> 1. strict: enable only .patch (should work)
> (2. lenient: introduce blacklist, .png ...)



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Work logged] (HIVE-20683) Add the Ability to push Dynamic Between and Bloom filters to Druid

2019-09-09 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-20683?focusedWorklogId=308712=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-308712
 ]

ASF GitHub Bot logged work on HIVE-20683:
-

Author: ASF GitHub Bot
Created on: 09/Sep/19 09:50
Start Date: 09/Sep/19 09:50
Worklog Time Spent: 10m 
  Work Description: b-slim commented on pull request #723: [HIVE-20683] Add 
the Ability to push Dynamic Between and Bloom filters to Druid
URL: https://github.com/apache/hive/pull/723#discussion_r322155098
 
 

 ##
 File path: 
ql/src/test/queries/clientpositive/druidmini_semijoin_reduction_all_types.q
 ##
 @@ -0,0 +1,144 @@
+--! qt:dataset:srcpart
+--! qt:dataset:druid_table_alltypesorc
+--! qt:dataset:alltypesorc
+
+set hive.compute.query.using.stats=false;
+set hive.mapred.mode=nonstrict;
+set hive.explain.user=false;
+set hive.optimize.ppd=true;
+set hive.ppd.remove.duplicatefilters=true;
+set hive.tez.dynamic.partition.pruning=true;
+set hive.tez.dynamic.semijoin.reduction=true;
+set hive.optimize.metadataonly=false;
+set hive.optimize.index.filter=true;
+set hive.stats.autogather=true;
+set hive.tez.bigtable.minsize.semijoin.reduction=1;
+set hive.tez.min.bloom.filter.entries=1;
+set hive.stats.fetch.column.stats=true;
+set hive.disable.unsafe.external.table.operations=false;
+set hive.tez.dynamic.semijoin.reduction.for.mapjoin=true;
+
+DROP TABLE IF EXISTS alltypesorc_small;
+CREATE TABLE alltypesorc_small(
+ctinyint TINYINT,
+csmallint SMALLINT,
+cint INT,
+cbigint BIGINT,
+cfloat FLOAT,
+cdouble DOUBLE,
+cstring1 STRING,
+cstring2 STRING,
+ctimestamp1 TIMESTAMP,
+cboolean1 BOOLEAN,
+cboolean2 BOOLEAN)
+STORED AS ORC;
+Insert into table alltypesorc_small
+Select ctinyint, csmallint, cint, cbigint, cfloat, cdouble, cstring1, 
cstring2, cast(`__time` as timestamp), cboolean1, cboolean2 from 
druid_table_alltypesorc where cstring2 like '%a%' and cstring1 like '%a%';
+Select count(*) from alltypesorc_small;
+Select count(*) from druid_table_alltypesorc;
+
+DESCRIBE druid_table_alltypesorc;
+DESCRIBE alltypesorc_small;
+
+-- Test Joins on all column types one by one
+-- String
+set hive.disable.unsafe.external.table.operations=false;
+EXPLAIN select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.cstring1 = druid_table_alltypesorc.cstring1);
+select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.cstring1 = druid_table_alltypesorc.cstring1);
+set hive.disable.unsafe.external.table.operations=true;
+select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.cstring1 = druid_table_alltypesorc.cstring1);
+
+-- tinyint
+set hive.disable.unsafe.external.table.operations=false;
+EXPLAIN select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.ctinyint = druid_table_alltypesorc.ctinyint);
+select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.ctinyint = druid_table_alltypesorc.ctinyint);
+set hive.disable.unsafe.external.table.operations=true;
+select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.ctinyint = druid_table_alltypesorc.ctinyint);
+
+-- smallint
+set hive.disable.unsafe.external.table.operations=false;
+EXPLAIN select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.csmallint = druid_table_alltypesorc.csmallint);
+select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.csmallint = druid_table_alltypesorc.csmallint);
+set hive.disable.unsafe.external.table.operations=true;
+select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.csmallint = druid_table_alltypesorc.csmallint);
+
+-- int
+set hive.disable.unsafe.external.table.operations=false;
+EXPLAIN select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.cint = druid_table_alltypesorc.cint);
+select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.cint = druid_table_alltypesorc.cint);
+set hive.disable.unsafe.external.table.operations=true;
+select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.cint = druid_table_alltypesorc.cint);
+
+-- bigint
+set hive.disable.unsafe.external.table.operations=false;
+EXPLAIN select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.cbigint = druid_table_alltypesorc.cbigint);
+select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.cbigint = druid_table_alltypesorc.cbigint);
+set hive.disable.unsafe.external.table.operations=true;
+select count(*) from alltypesorc_small join druid_table_alltypesorc on 
(alltypesorc_small.cbigint = druid_table_alltypesorc.cbigint);
+
+-- 

[jira] [Work logged] (HIVE-20683) Add the Ability to push Dynamic Between and Bloom filters to Druid

2019-09-09 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-20683?focusedWorklogId=308708=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-308708
 ]

ASF GitHub Bot logged work on HIVE-20683:
-

Author: ASF GitHub Bot
Created on: 09/Sep/19 09:40
Start Date: 09/Sep/19 09:40
Worklog Time Spent: 10m 
  Work Description: b-slim commented on pull request #723: [HIVE-20683] Add 
the Ability to push Dynamic Between and Bloom filters to Druid
URL: https://github.com/apache/hive/pull/723#discussion_r322150545
 
 

 ##
 File path: 
druid-handler/src/java/org/apache/hadoop/hive/druid/DruidStorageHandlerUtils.java
 ##
 @@ -894,4 +945,255 @@ public static IndexSpec getIndexSpec(Configuration jc) {
 ImmutableList aggregatorFactories = 
aggregatorFactoryBuilder.build();
 return Pair.of(dimensions, aggregatorFactories.toArray(new 
AggregatorFactory[0]));
   }
+
+  // Druid only supports String,Long,Float,Double selectors
+  private static Set druidSupportedTypeInfos = 
ImmutableSet.of(
+  TypeInfoFactory.stringTypeInfo, TypeInfoFactory.charTypeInfo,
+  TypeInfoFactory.varcharTypeInfo, TypeInfoFactory.byteTypeInfo,
+  TypeInfoFactory.intTypeInfo, TypeInfoFactory.longTypeInfo,
+  TypeInfoFactory.shortTypeInfo, TypeInfoFactory.doubleTypeInfo
+  );
+
+  private static Set stringTypeInfos = ImmutableSet.of(
+  TypeInfoFactory.stringTypeInfo,
+  TypeInfoFactory.charTypeInfo, TypeInfoFactory.varcharTypeInfo
+  );
+
+
+  public static org.apache.druid.query.Query 
addDynamicFilters(org.apache.druid.query.Query query,
+  ExprNodeGenericFuncDesc filterExpr, Configuration conf, boolean 
resolveDynamicValues
+  ) {
+List virtualColumns = Arrays
+.asList(getVirtualColumns(query).getVirtualColumns());
+org.apache.druid.query.Query rv = query;
+DimFilter joinReductionFilter = toDruidFilter(filterExpr, conf, 
virtualColumns,
+resolveDynamicValues
+);
+if(joinReductionFilter != null) {
+  String type = query.getType();
+  DimFilter filter = new AndDimFilter(joinReductionFilter, 
query.getFilter());
+  switch (type) {
+  case org.apache.druid.query.Query.TIMESERIES:
+rv = Druids.TimeseriesQueryBuilder.copy((TimeseriesQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.TOPN:
+rv = new TopNQueryBuilder((TopNQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.GROUP_BY:
+rv = new GroupByQuery.Builder((GroupByQuery) query)
+.setDimFilter(filter)
+.setVirtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.SCAN:
+rv = ScanQuery.ScanQueryBuilder.copy((ScanQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.SELECT:
+rv = Druids.SelectQueryBuilder.copy((SelectQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  default:
+throw new UnsupportedOperationException("Unsupported Query type " + 
type);
+  }
+}
+return rv;
+  }
+
+  @Nullable
+  private static DimFilter toDruidFilter(ExprNodeDesc filterExpr, 
Configuration configuration,
+  List virtualColumns, boolean resolveDynamicValues
+  ) {
+if(filterExpr == null) {
+  return null;
+}
+Class genericUDFClass = 
getGenericUDFClassFromExprDesc(filterExpr);
+if(FunctionRegistry.isOpAnd(filterExpr)) {
+  Iterator iterator = filterExpr.getChildren().iterator();
+  List delegates = Lists.newArrayList();
+  while (iterator.hasNext()) {
+DimFilter filter = toDruidFilter(iterator.next(), configuration, 
virtualColumns,
+resolveDynamicValues
+);
+if(filter != null) {
+  delegates.add(filter);
+}
+  }
+  if(delegates != null && !delegates.isEmpty()) {
+return new AndDimFilter(delegates);
+  }
+}
+if(FunctionRegistry.isOpOr(filterExpr)) {
+  Iterator iterator = filterExpr.getChildren().iterator();
+  List delegates = Lists.newArrayList();
+  while (iterator.hasNext()) {
+DimFilter filter = toDruidFilter(iterator.next(), configuration, 
virtualColumns,
+resolveDynamicValues
+);
+if(filter != null) {
+  delegates.add(filter);
+}
+  }
+  if(delegates != null) {
+return new OrDimFilter(delegates);
+  }
+} else if(GenericUDFBetween.class 

[jira] [Updated] (HIVE-21884) Scheduled query support

2019-09-09 Thread Zoltan Haindrich (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21884?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltan Haindrich updated HIVE-21884:

Attachment: HIVE-21844.06.patch

> Scheduled query support
> ---
>
> Key: HIVE-21884
> URL: https://issues.apache.org/jira/browse/HIVE-21884
> Project: Hive
>  Issue Type: Improvement
>Reporter: Zoltan Haindrich
>Assignee: Zoltan Haindrich
>Priority: Major
> Attachments: HIVE-21844.04.patch, HIVE-21844.05.patch, 
> HIVE-21844.06.patch, HIVE-21884.01.patch, HIVE-21884.02.patch, 
> HIVE-21884.03.patch, Scheduled queries2.pdf
>
>
> design document:
> https://docs.google.com/document/d/1mJSFdJi_1cbxJTXC9QvGw2rQ3zzJkNfxOO6b5esmyCE/edit#
> in case the google doc is not reachable:  [^Scheduled queries2.pdf] 



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Updated] (HIVE-21508) ClassCastException when initializing HiveMetaStoreClient on JDK10 or newer

2019-09-09 Thread Ana Jalba (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21508?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ana Jalba updated HIVE-21508:
-
   Attachment: HIVE-21508.patch
Fix Version/s: (was: 2.3.7)
   (was: 2.4.0)
   Status: Patch Available  (was: In Progress)

> ClassCastException when initializing HiveMetaStoreClient on JDK10 or newer
> --
>
> Key: HIVE-21508
> URL: https://issues.apache.org/jira/browse/HIVE-21508
> Project: Hive
>  Issue Type: Bug
>  Components: Clients
>Affects Versions: 2.3.4, 3.2.0
>Reporter: Adar Dembo
>Assignee: Ana Jalba
>Priority: Major
> Fix For: 3.2.0
>
> Attachments: HIVE-21508.patch
>
>
> There's this block of code in {{HiveMetaStoreClient:resolveUris}} (called 
> from the constructor) on master:
> {noformat}
>   private URI metastoreUris[];
>   ...
>   if (MetastoreConf.getVar(conf, 
> ConfVars.THRIFT_URI_SELECTION).equalsIgnoreCase("RANDOM")) {
> List uriList = Arrays.asList(metastoreUris);
> Collections.shuffle(uriList);
> metastoreUris = (URI[]) uriList.toArray();
>   }
> {noformat}
> The cast to {{URI[]}} throws a {{ClassCastException}} beginning with JDK 10, 
> possibly with JDK 9 as well. Note that {{THRIFT_URI_SELECTION}} defaults to 
> {{RANDOM}} so this should affect anyone who creates a 
> {{HiveMetaStoreClient}}. On master this can be overridden with {{SEQUENTIAL}} 
> to avoid the broken case; I'm working against 2.3.4 where there's no such 
> workaround.
> [Here's|https://stackoverflow.com/questions/51372788/array-cast-java-8-vs-java-9]
>  a StackOverflow post that explains the issue in more detail. Interestingly, 
> the author described the issue in the context of the HMS; not sure why there 
> was no follow up with a Hive bug report.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


[jira] [Work logged] (HIVE-20683) Add the Ability to push Dynamic Between and Bloom filters to Druid

2019-09-09 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-20683?focusedWorklogId=308707=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-308707
 ]

ASF GitHub Bot logged work on HIVE-20683:
-

Author: ASF GitHub Bot
Created on: 09/Sep/19 09:36
Start Date: 09/Sep/19 09:36
Worklog Time Spent: 10m 
  Work Description: b-slim commented on pull request #723: [HIVE-20683] Add 
the Ability to push Dynamic Between and Bloom filters to Druid
URL: https://github.com/apache/hive/pull/723#discussion_r322148915
 
 

 ##
 File path: 
druid-handler/src/java/org/apache/hadoop/hive/druid/DruidStorageHandlerUtils.java
 ##
 @@ -894,4 +945,255 @@ public static IndexSpec getIndexSpec(Configuration jc) {
 ImmutableList aggregatorFactories = 
aggregatorFactoryBuilder.build();
 return Pair.of(dimensions, aggregatorFactories.toArray(new 
AggregatorFactory[0]));
   }
+
+  // Druid only supports String,Long,Float,Double selectors
+  private static Set druidSupportedTypeInfos = 
ImmutableSet.of(
+  TypeInfoFactory.stringTypeInfo, TypeInfoFactory.charTypeInfo,
+  TypeInfoFactory.varcharTypeInfo, TypeInfoFactory.byteTypeInfo,
+  TypeInfoFactory.intTypeInfo, TypeInfoFactory.longTypeInfo,
+  TypeInfoFactory.shortTypeInfo, TypeInfoFactory.doubleTypeInfo
+  );
+
+  private static Set stringTypeInfos = ImmutableSet.of(
+  TypeInfoFactory.stringTypeInfo,
+  TypeInfoFactory.charTypeInfo, TypeInfoFactory.varcharTypeInfo
+  );
+
+
+  public static org.apache.druid.query.Query 
addDynamicFilters(org.apache.druid.query.Query query,
+  ExprNodeGenericFuncDesc filterExpr, Configuration conf, boolean 
resolveDynamicValues
+  ) {
+List virtualColumns = Arrays
+.asList(getVirtualColumns(query).getVirtualColumns());
+org.apache.druid.query.Query rv = query;
+DimFilter joinReductionFilter = toDruidFilter(filterExpr, conf, 
virtualColumns,
+resolveDynamicValues
+);
+if(joinReductionFilter != null) {
+  String type = query.getType();
+  DimFilter filter = new AndDimFilter(joinReductionFilter, 
query.getFilter());
+  switch (type) {
+  case org.apache.druid.query.Query.TIMESERIES:
+rv = Druids.TimeseriesQueryBuilder.copy((TimeseriesQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.TOPN:
+rv = new TopNQueryBuilder((TopNQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.GROUP_BY:
+rv = new GroupByQuery.Builder((GroupByQuery) query)
+.setDimFilter(filter)
+.setVirtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.SCAN:
+rv = ScanQuery.ScanQueryBuilder.copy((ScanQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.SELECT:
+rv = Druids.SelectQueryBuilder.copy((SelectQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  default:
+throw new UnsupportedOperationException("Unsupported Query type " + 
type);
+  }
+}
+return rv;
+  }
+
+  @Nullable
+  private static DimFilter toDruidFilter(ExprNodeDesc filterExpr, 
Configuration configuration,
+  List virtualColumns, boolean resolveDynamicValues
+  ) {
+if(filterExpr == null) {
+  return null;
+}
+Class genericUDFClass = 
getGenericUDFClassFromExprDesc(filterExpr);
+if(FunctionRegistry.isOpAnd(filterExpr)) {
+  Iterator iterator = filterExpr.getChildren().iterator();
+  List delegates = Lists.newArrayList();
+  while (iterator.hasNext()) {
+DimFilter filter = toDruidFilter(iterator.next(), configuration, 
virtualColumns,
+resolveDynamicValues
+);
+if(filter != null) {
+  delegates.add(filter);
+}
+  }
+  if(delegates != null && !delegates.isEmpty()) {
+return new AndDimFilter(delegates);
+  }
+}
+if(FunctionRegistry.isOpOr(filterExpr)) {
+  Iterator iterator = filterExpr.getChildren().iterator();
+  List delegates = Lists.newArrayList();
+  while (iterator.hasNext()) {
+DimFilter filter = toDruidFilter(iterator.next(), configuration, 
virtualColumns,
+resolveDynamicValues
+);
+if(filter != null) {
+  delegates.add(filter);
+}
+  }
+  if(delegates != null) {
+return new OrDimFilter(delegates);
+  }
+} else if(GenericUDFBetween.class 

[jira] [Work logged] (HIVE-20683) Add the Ability to push Dynamic Between and Bloom filters to Druid

2019-09-09 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-20683?focusedWorklogId=308706=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-308706
 ]

ASF GitHub Bot logged work on HIVE-20683:
-

Author: ASF GitHub Bot
Created on: 09/Sep/19 09:36
Start Date: 09/Sep/19 09:36
Worklog Time Spent: 10m 
  Work Description: b-slim commented on pull request #723: [HIVE-20683] Add 
the Ability to push Dynamic Between and Bloom filters to Druid
URL: https://github.com/apache/hive/pull/723#discussion_r322148627
 
 

 ##
 File path: 
druid-handler/src/java/org/apache/hadoop/hive/druid/DruidStorageHandlerUtils.java
 ##
 @@ -894,4 +945,255 @@ public static IndexSpec getIndexSpec(Configuration jc) {
 ImmutableList aggregatorFactories = 
aggregatorFactoryBuilder.build();
 return Pair.of(dimensions, aggregatorFactories.toArray(new 
AggregatorFactory[0]));
   }
+
+  // Druid only supports String,Long,Float,Double selectors
+  private static Set druidSupportedTypeInfos = 
ImmutableSet.of(
+  TypeInfoFactory.stringTypeInfo, TypeInfoFactory.charTypeInfo,
+  TypeInfoFactory.varcharTypeInfo, TypeInfoFactory.byteTypeInfo,
+  TypeInfoFactory.intTypeInfo, TypeInfoFactory.longTypeInfo,
+  TypeInfoFactory.shortTypeInfo, TypeInfoFactory.doubleTypeInfo
+  );
+
+  private static Set stringTypeInfos = ImmutableSet.of(
+  TypeInfoFactory.stringTypeInfo,
+  TypeInfoFactory.charTypeInfo, TypeInfoFactory.varcharTypeInfo
+  );
+
+
+  public static org.apache.druid.query.Query 
addDynamicFilters(org.apache.druid.query.Query query,
+  ExprNodeGenericFuncDesc filterExpr, Configuration conf, boolean 
resolveDynamicValues
+  ) {
+List virtualColumns = Arrays
+.asList(getVirtualColumns(query).getVirtualColumns());
+org.apache.druid.query.Query rv = query;
+DimFilter joinReductionFilter = toDruidFilter(filterExpr, conf, 
virtualColumns,
+resolveDynamicValues
+);
+if(joinReductionFilter != null) {
+  String type = query.getType();
+  DimFilter filter = new AndDimFilter(joinReductionFilter, 
query.getFilter());
+  switch (type) {
+  case org.apache.druid.query.Query.TIMESERIES:
+rv = Druids.TimeseriesQueryBuilder.copy((TimeseriesQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.TOPN:
+rv = new TopNQueryBuilder((TopNQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.GROUP_BY:
+rv = new GroupByQuery.Builder((GroupByQuery) query)
+.setDimFilter(filter)
+.setVirtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.SCAN:
+rv = ScanQuery.ScanQueryBuilder.copy((ScanQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.SELECT:
+rv = Druids.SelectQueryBuilder.copy((SelectQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  default:
+throw new UnsupportedOperationException("Unsupported Query type " + 
type);
+  }
+}
+return rv;
+  }
+
+  @Nullable
+  private static DimFilter toDruidFilter(ExprNodeDesc filterExpr, 
Configuration configuration,
+  List virtualColumns, boolean resolveDynamicValues
+  ) {
+if(filterExpr == null) {
+  return null;
+}
+Class genericUDFClass = 
getGenericUDFClassFromExprDesc(filterExpr);
+if(FunctionRegistry.isOpAnd(filterExpr)) {
+  Iterator iterator = filterExpr.getChildren().iterator();
+  List delegates = Lists.newArrayList();
+  while (iterator.hasNext()) {
+DimFilter filter = toDruidFilter(iterator.next(), configuration, 
virtualColumns,
+resolveDynamicValues
+);
+if(filter != null) {
+  delegates.add(filter);
+}
+  }
+  if(delegates != null && !delegates.isEmpty()) {
+return new AndDimFilter(delegates);
+  }
+}
+if(FunctionRegistry.isOpOr(filterExpr)) {
+  Iterator iterator = filterExpr.getChildren().iterator();
+  List delegates = Lists.newArrayList();
+  while (iterator.hasNext()) {
+DimFilter filter = toDruidFilter(iterator.next(), configuration, 
virtualColumns,
+resolveDynamicValues
+);
+if(filter != null) {
+  delegates.add(filter);
+}
+  }
+  if(delegates != null) {
 
 Review comment:
   same here this can not be null.
 

[jira] [Commented] (HIVE-21884) Scheduled query support

2019-09-09 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-21884?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925524#comment-16925524
 ] 

Hive QA commented on HIVE-21884:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12979836/HIVE-21844.05.patch

{color:red}ERROR:{color} -1 due to build exiting with an error

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/18502/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/18502/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-18502/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hiveptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ date '+%Y-%m-%d %T.%3N'
2019-09-09 09:33:43.402
+ [[ -n /usr/lib/jvm/java-8-openjdk-amd64 ]]
+ export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ export 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'MAVEN_OPTS=-Xmx1g '
+ MAVEN_OPTS='-Xmx1g '
+ cd /data/hiveptest/working/
+ tee /data/hiveptest/logs/PreCommit-HIVE-Build-18502/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ git = \s\v\n ]]
+ [[ git = \g\i\t ]]
+ [[ -z master ]]
+ [[ -d apache-github-source-source ]]
+ [[ ! -d apache-github-source-source/.git ]]
+ [[ ! -d apache-github-source-source ]]
+ date '+%Y-%m-%d %T.%3N'
2019-09-09 09:33:43.405
+ cd apache-github-source-source
+ git fetch origin
>From https://github.com/apache/hive
   5565311..1b04924  master -> origin/master
+ git reset --hard HEAD
HEAD is now at 5565311 HIVE-22143 Break up DDLSemanticAnalyzer - extract 
Database related analyzers (Miklos Gergely reviewd by Jesus Camacho Rodriguez)
+ git clean -f -d
Removing ${project.basedir}/
Removing itests/${project.basedir}/
Removing standalone-metastore/metastore-server/src/gen/
+ git checkout master
Already on 'master'
Your branch is behind 'origin/master' by 3 commits, and can be fast-forwarded.
  (use "git pull" to update your local branch)
+ git reset --hard origin/master
HEAD is now at 1b04924 HIVE-22106: Remove cross-query synchronization for the 
partition-eval (Slim B via Gopal V)
+ git merge --ff-only origin/master
Already up-to-date.
+ date '+%Y-%m-%d %T.%3N'
2019-09-09 09:33:44.823
+ rm -rf ../yetus_PreCommit-HIVE-Build-18502
+ mkdir ../yetus_PreCommit-HIVE-Build-18502
+ git gc
+ cp -R . ../yetus_PreCommit-HIVE-Build-18502
+ mkdir /data/hiveptest/logs/PreCommit-HIVE-Build-18502/yetus
+ patchCommandPath=/data/hiveptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hiveptest/working/scratch/build.patch
+ [[ -f /data/hiveptest/working/scratch/build.patch ]]
+ chmod +x /data/hiveptest/working/scratch/smart-apply-patch.sh
+ /data/hiveptest/working/scratch/smart-apply-patch.sh 
/data/hiveptest/working/scratch/build.patch
Going to apply patch with: git apply -p0
/data/hiveptest/working/scratch/build.patch:27: trailing whitespace.
# +-- X: p#1 
/data/hiveptest/working/scratch/build.patch:249: trailing whitespace.
SELECT 'abc123' AS ORIG_VALUE, CAST('abc123' AS BIGINT) AS CAST_AS_BIGINT, CASE 
WHEN CAST('abc123' AS BIGINT) IS NULL THEN 'YES' ELSE 'NO' END AS IS_IT_NULL; 
/data/hiveptest/working/scratch/build.patch:815: trailing whitespace.

/data/hiveptest/working/scratch/build.patch:825: trailing whitespace.

/data/hiveptest/working/scratch/build.patch:836: trailing whitespace.

warning: squelched 177 whitespace errors
warning: 182 lines add whitespace errors.
+ [[ maven == \m\a\v\e\n ]]
+ rm -rf /data/hiveptest/working/maven/org/apache/hive
+ mvn -B clean install -DskipTests -T 4 -q 
-Dmaven.repo.local=/data/hiveptest/working/maven
protoc-jar: executing: [/tmp/protoc5868761099962570501.exe, --version]
libprotoc 2.5.0
protoc-jar: executing: [/tmp/protoc5868761099962570501.exe, 
-I/data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-common/src/main/protobuf/org/apache/hadoop/hive/metastore,
 
--java_out=/data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-common/target/generated-sources,
 
/data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-common/src/main/protobuf/org/apache/hadoop/hive/metastore/metastore.proto]
ANTLR Parser Generator  Version 3.5.2
protoc-jar: executing: [/tmp/protoc1133043159592088760.exe, --version]
libprotoc 2.5.0
ANTLR Parser Generator  Version 3.5.2
Output file 

[jira] [Work logged] (HIVE-20683) Add the Ability to push Dynamic Between and Bloom filters to Druid

2019-09-09 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-20683?focusedWorklogId=308705=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-308705
 ]

ASF GitHub Bot logged work on HIVE-20683:
-

Author: ASF GitHub Bot
Created on: 09/Sep/19 09:35
Start Date: 09/Sep/19 09:35
Worklog Time Spent: 10m 
  Work Description: b-slim commented on pull request #723: [HIVE-20683] Add 
the Ability to push Dynamic Between and Bloom filters to Druid
URL: https://github.com/apache/hive/pull/723#discussion_r322148428
 
 

 ##
 File path: 
druid-handler/src/java/org/apache/hadoop/hive/druid/DruidStorageHandlerUtils.java
 ##
 @@ -894,4 +945,255 @@ public static IndexSpec getIndexSpec(Configuration jc) {
 ImmutableList aggregatorFactories = 
aggregatorFactoryBuilder.build();
 return Pair.of(dimensions, aggregatorFactories.toArray(new 
AggregatorFactory[0]));
   }
+
+  // Druid only supports String,Long,Float,Double selectors
+  private static Set druidSupportedTypeInfos = 
ImmutableSet.of(
+  TypeInfoFactory.stringTypeInfo, TypeInfoFactory.charTypeInfo,
+  TypeInfoFactory.varcharTypeInfo, TypeInfoFactory.byteTypeInfo,
+  TypeInfoFactory.intTypeInfo, TypeInfoFactory.longTypeInfo,
+  TypeInfoFactory.shortTypeInfo, TypeInfoFactory.doubleTypeInfo
+  );
+
+  private static Set stringTypeInfos = ImmutableSet.of(
+  TypeInfoFactory.stringTypeInfo,
+  TypeInfoFactory.charTypeInfo, TypeInfoFactory.varcharTypeInfo
+  );
+
+
+  public static org.apache.druid.query.Query 
addDynamicFilters(org.apache.druid.query.Query query,
+  ExprNodeGenericFuncDesc filterExpr, Configuration conf, boolean 
resolveDynamicValues
+  ) {
+List virtualColumns = Arrays
+.asList(getVirtualColumns(query).getVirtualColumns());
+org.apache.druid.query.Query rv = query;
+DimFilter joinReductionFilter = toDruidFilter(filterExpr, conf, 
virtualColumns,
+resolveDynamicValues
+);
+if(joinReductionFilter != null) {
+  String type = query.getType();
+  DimFilter filter = new AndDimFilter(joinReductionFilter, 
query.getFilter());
+  switch (type) {
+  case org.apache.druid.query.Query.TIMESERIES:
+rv = Druids.TimeseriesQueryBuilder.copy((TimeseriesQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.TOPN:
+rv = new TopNQueryBuilder((TopNQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.GROUP_BY:
+rv = new GroupByQuery.Builder((GroupByQuery) query)
+.setDimFilter(filter)
+.setVirtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.SCAN:
+rv = ScanQuery.ScanQueryBuilder.copy((ScanQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  case org.apache.druid.query.Query.SELECT:
+rv = Druids.SelectQueryBuilder.copy((SelectQuery) query)
+.filters(filter)
+.virtualColumns(VirtualColumns.create(virtualColumns))
+.build();
+break;
+  default:
+throw new UnsupportedOperationException("Unsupported Query type " + 
type);
+  }
+}
+return rv;
+  }
+
+  @Nullable
+  private static DimFilter toDruidFilter(ExprNodeDesc filterExpr, 
Configuration configuration,
+  List virtualColumns, boolean resolveDynamicValues
+  ) {
+if(filterExpr == null) {
+  return null;
+}
+Class genericUDFClass = 
getGenericUDFClassFromExprDesc(filterExpr);
+if(FunctionRegistry.isOpAnd(filterExpr)) {
+  Iterator iterator = filterExpr.getChildren().iterator();
+  List delegates = Lists.newArrayList();
+  while (iterator.hasNext()) {
+DimFilter filter = toDruidFilter(iterator.next(), configuration, 
virtualColumns,
+resolveDynamicValues
+);
+if(filter != null) {
+  delegates.add(filter);
+}
+  }
+  if(delegates != null && !delegates.isEmpty()) {
 
 Review comment:
   how this can be null ?
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 308705)
Time Spent: 3h  (was: 2h 50m)

> Add the Ability to push Dynamic Between and Bloom filters to Druid
> 

[jira] [Work logged] (HIVE-20683) Add the Ability to push Dynamic Between and Bloom filters to Druid

2019-09-09 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-20683?focusedWorklogId=308703=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-308703
 ]

ASF GitHub Bot logged work on HIVE-20683:
-

Author: ASF GitHub Bot
Created on: 09/Sep/19 09:34
Start Date: 09/Sep/19 09:34
Worklog Time Spent: 10m 
  Work Description: b-slim commented on pull request #723: [HIVE-20683] Add 
the Ability to push Dynamic Between and Bloom filters to Druid
URL: https://github.com/apache/hive/pull/723#discussion_r322147955
 
 

 ##
 File path: 
druid-handler/src/java/org/apache/hadoop/hive/druid/DruidStorageHandlerUtils.java
 ##
 @@ -91,17 +114,37 @@
 import org.apache.hadoop.fs.FileStatus;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.hive.common.io.NonSyncByteArrayInputStream;
 
 Review comment:
   this is not used.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 308703)
Time Spent: 2h 50m  (was: 2h 40m)

> Add the Ability to push Dynamic Between and Bloom filters to Druid
> --
>
> Key: HIVE-20683
> URL: https://issues.apache.org/jira/browse/HIVE-20683
> Project: Hive
>  Issue Type: New Feature
>  Components: Druid integration
>Reporter: Nishant Bangarwa
>Assignee: Nishant Bangarwa
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-20683.1.patch, HIVE-20683.2.patch, 
> HIVE-20683.3.patch, HIVE-20683.4.patch, HIVE-20683.5.patch, 
> HIVE-20683.6.patch, HIVE-20683.8.patch, HIVE-20683.patch
>
>  Time Spent: 2h 50m
>  Remaining Estimate: 0h
>
> For optimizing joins, Hive generates BETWEEN filter with min-max and BLOOM 
> filter for filtering one side of semi-join.
> Druid 0.13.0 will have support for Bloom filters (Added via 
> https://github.com/apache/incubator-druid/pull/6222)
> Implementation details - 
> # Hive generates and passes the filters as part of 'filterExpr' in TableScan. 
> # DruidQueryBasedRecordReader gets this filter passed as part of the conf. 
> # During execution phase, before sending the query to druid in 
> DruidQueryBasedRecordReader we will deserialize this filter, translate it 
> into a DruidDimFilter and add it to existing DruidQuery.  Tez executor 
> already ensures that when we start reading results from the record reader, 
> all the dynamic values are initialized. 
> # Explaining a druid query also prints the query sent to druid as 
> {{druid.json.query}}. We also need to make sure to update the druid query 
> with the filters. During explain we do not have the actual values for the 
> dynamic values, so instead of values we will print the dynamic expression 
> itself as part of druid query. 
> Note:- This work needs druid to be updated to version 0.13.0



--
This message was sent by Atlassian Jira
(v8.3.2#803003)


  1   2   >