[jira] [Commented] (HIVE-22736) Support replication across multiple encryption zones

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22736?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024928#comment-17024928
 ] 

Hive QA commented on HIVE-22736:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12991975/HIVE-22736.patch

{color:green}SUCCESS:{color} +1 due to 5 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 17966 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.metastore.TestGetPartitionsUsingProjectionAndFilterSpecs.testGetPartitionsUsingValues
 (batchId=231)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/20374/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/20374/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-20374/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12991975 - PreCommit-HIVE-Build

> Support replication across multiple encryption zones
> 
>
> Key: HIVE-22736
> URL: https://issues.apache.org/jira/browse/HIVE-22736
> Project: Hive
>  Issue Type: Task
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch
>
>  Time Spent: 2h 20m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22736) Support replication across multiple encryption zones

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22736?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024915#comment-17024915
 ] 

Hive QA commented on HIVE-22736:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m  
5s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
41s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  3m 
28s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  2m 
 2s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
19s{color} | {color:blue} shims/common in master has 6 extant Findbugs 
warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
21s{color} | {color:blue} shims/0.23 in master has 7 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  2m 
34s{color} | {color:blue} standalone-metastore/metastore-common in master has 
35 extant Findbugs warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
34s{color} | {color:blue} common in master has 63 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  1m 
11s{color} | {color:blue} standalone-metastore/metastore-server in master has 
181 extant Findbugs warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  4m  
1s{color} | {color:blue} ql in master has 1538 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
40s{color} | {color:blue} itests/hive-unit in master has 2 extant Findbugs 
warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  3m 
15s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
25s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  3m 
55s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  3m 
24s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  3m 
24s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
 9s{color} | {color:green} shims/common: The patch generated 0 new + 93 
unchanged - 1 fixed = 93 total (was 94) {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
 9s{color} | {color:green} The patch 0.23 passed checkstyle {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
11s{color} | {color:green} The patch metastore-common passed checkstyle {color} 
|
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
15s{color} | {color:green} The patch common passed checkstyle {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
20s{color} | {color:green} The patch metastore-server passed checkstyle {color} 
|
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
40s{color} | {color:green} The patch ql passed checkstyle {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
19s{color} | {color:green} The patch hive-unit passed checkstyle {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
2s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 10m 
27s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  3m  
9s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
13s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| 

[jira] [Updated] (HIVE-22739) Schematool should check if upgradeFrom version is identical to current DB schema version

2020-01-27 Thread Naveen Gangam (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22739?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Naveen Gangam updated HIVE-22739:
-
Fix Version/s: 4.0.0
   Resolution: Fixed
   Status: Resolved  (was: Patch Available)

Fix has been pushed to master. Thank you for your contribution [~afan]

> Schematool should check if upgradeFrom version is identical to current DB 
> schema version
> 
>
> Key: HIVE-22739
> URL: https://issues.apache.org/jira/browse/HIVE-22739
> Project: Hive
>  Issue Type: Bug
>  Components: Standalone Metastore
>Affects Versions: 4.0.0
>Reporter: Alice Fan
>Assignee: Alice Fan
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-22739.1.patch
>
>
> Schematool should check if upgradeFrom version is identical to current DB 
> schema version.
> https://github.com/apache/hive/blob/master/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/tools/schematool/SchemaToolTaskUpgrade.java#L41
> We should check if the user input fromVersion is identical to current DB 
> version.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22538) RS deduplication does not always enforce hive.optimize.reducededuplication.min.reducer

2020-01-27 Thread Krisztian Kasa (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22538?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Krisztian Kasa updated HIVE-22538:
--
Attachment: HIVE-22538.8.patch

> RS deduplication does not always enforce 
> hive.optimize.reducededuplication.min.reducer
> --
>
> Key: HIVE-22538
> URL: https://issues.apache.org/jira/browse/HIVE-22538
> Project: Hive
>  Issue Type: Bug
>  Components: Physical Optimizer
>Reporter: Jesus Camacho Rodriguez
>Assignee: Krisztian Kasa
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22538.2.patch, HIVE-22538.3.patch, 
> HIVE-22538.4.patch, HIVE-22538.5.patch, HIVE-22538.6.patch, 
> HIVE-22538.6.patch, HIVE-22538.7.patch, HIVE-22538.8.patch, 
> HIVE-22538.8.patch, HIVE-22538.patch
>
>  Time Spent: 1.5h
>  Remaining Estimate: 0h
>
> For transactional tables, that property might be overriden to 1, which can 
> lead to merging final aggregation into a single stage (hence leading to 
> performance degradation). For instance, when autogather column stats is 
> enabled, this can happen for the following query:
> {code}
> set hive.support.concurrency=true;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
> EXPLAIN
> CREATE TABLE x STORED AS ORC TBLPROPERTIES('transactional'='true') AS
> SELECT * FROM SRC x CLUSTER BY x.key;
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22538) RS deduplication does not always enforce hive.optimize.reducededuplication.min.reducer

2020-01-27 Thread Krisztian Kasa (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22538?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Krisztian Kasa updated HIVE-22538:
--
Status: Patch Available  (was: Open)

> RS deduplication does not always enforce 
> hive.optimize.reducededuplication.min.reducer
> --
>
> Key: HIVE-22538
> URL: https://issues.apache.org/jira/browse/HIVE-22538
> Project: Hive
>  Issue Type: Bug
>  Components: Physical Optimizer
>Reporter: Jesus Camacho Rodriguez
>Assignee: Krisztian Kasa
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22538.2.patch, HIVE-22538.3.patch, 
> HIVE-22538.4.patch, HIVE-22538.5.patch, HIVE-22538.6.patch, 
> HIVE-22538.6.patch, HIVE-22538.7.patch, HIVE-22538.8.patch, 
> HIVE-22538.8.patch, HIVE-22538.patch
>
>  Time Spent: 1.5h
>  Remaining Estimate: 0h
>
> For transactional tables, that property might be overriden to 1, which can 
> lead to merging final aggregation into a single stage (hence leading to 
> performance degradation). For instance, when autogather column stats is 
> enabled, this can happen for the following query:
> {code}
> set hive.support.concurrency=true;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
> EXPLAIN
> CREATE TABLE x STORED AS ORC TBLPROPERTIES('transactional'='true') AS
> SELECT * FROM SRC x CLUSTER BY x.key;
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22538) RS deduplication does not always enforce hive.optimize.reducededuplication.min.reducer

2020-01-27 Thread Krisztian Kasa (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22538?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Krisztian Kasa updated HIVE-22538:
--
Status: Open  (was: Patch Available)

> RS deduplication does not always enforce 
> hive.optimize.reducededuplication.min.reducer
> --
>
> Key: HIVE-22538
> URL: https://issues.apache.org/jira/browse/HIVE-22538
> Project: Hive
>  Issue Type: Bug
>  Components: Physical Optimizer
>Reporter: Jesus Camacho Rodriguez
>Assignee: Krisztian Kasa
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22538.2.patch, HIVE-22538.3.patch, 
> HIVE-22538.4.patch, HIVE-22538.5.patch, HIVE-22538.6.patch, 
> HIVE-22538.6.patch, HIVE-22538.7.patch, HIVE-22538.8.patch, 
> HIVE-22538.8.patch, HIVE-22538.patch
>
>  Time Spent: 1.5h
>  Remaining Estimate: 0h
>
> For transactional tables, that property might be overriden to 1, which can 
> lead to merging final aggregation into a single stage (hence leading to 
> performance degradation). For instance, when autogather column stats is 
> enabled, this can happen for the following query:
> {code}
> set hive.support.concurrency=true;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
> EXPLAIN
> CREATE TABLE x STORED AS ORC TBLPROPERTIES('transactional'='true') AS
> SELECT * FROM SRC x CLUSTER BY x.key;
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22736) Support replication across multiple encryption zones

2020-01-27 Thread Aasha Medhi (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22736?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aasha Medhi updated HIVE-22736:
---
Status: In Progress  (was: Patch Available)

> Support replication across multiple encryption zones
> 
>
> Key: HIVE-22736
> URL: https://issues.apache.org/jira/browse/HIVE-22736
> Project: Hive
>  Issue Type: Task
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch
>
>  Time Spent: 2h 20m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22736) Support replication across multiple encryption zones

2020-01-27 Thread Aasha Medhi (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22736?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aasha Medhi updated HIVE-22736:
---
Attachment: HIVE-22736.patch
Status: Patch Available  (was: In Progress)

> Support replication across multiple encryption zones
> 
>
> Key: HIVE-22736
> URL: https://issues.apache.org/jira/browse/HIVE-22736
> Project: Hive
>  Issue Type: Task
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch
>
>  Time Spent: 2h 20m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22784) Boundary condition to check if there is nothing to truncate in StringExpr functions

2020-01-27 Thread Gopal Vijayaraghavan (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22784?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024896#comment-17024896
 ] 

Gopal Vijayaraghavan commented on HIVE-22784:
-

Looks like rightTrimAndTruncate has an off-by-1 error in it?

> Boundary condition to check if there is nothing to truncate in StringExpr 
> functions
> ---
>
> Key: HIVE-22784
> URL: https://issues.apache.org/jira/browse/HIVE-22784
> Project: Hive
>  Issue Type: Bug
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
> Attachments: HIVE-22784.1.patch
>
>
> Add a boundary condition to check if there is nothing to truncate in 
> StringExpr functions
> truncate()
> rightTrimAndTruncate()



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22736) Support replication across multiple encryption zones

2020-01-27 Thread Aasha Medhi (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22736?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aasha Medhi updated HIVE-22736:
---
Attachment: HIVE-22736.patch
Status: Patch Available  (was: In Progress)

> Support replication across multiple encryption zones
> 
>
> Key: HIVE-22736
> URL: https://issues.apache.org/jira/browse/HIVE-22736
> Project: Hive
>  Issue Type: Task
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch
>
>  Time Spent: 2h 20m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22736) Support replication across multiple encryption zones

2020-01-27 Thread Aasha Medhi (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22736?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aasha Medhi updated HIVE-22736:
---
Status: In Progress  (was: Patch Available)

> Support replication across multiple encryption zones
> 
>
> Key: HIVE-22736
> URL: https://issues.apache.org/jira/browse/HIVE-22736
> Project: Hive
>  Issue Type: Task
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch
>
>  Time Spent: 2h 20m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-22736) Support replication across multiple encryption zones

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22736?focusedWorklogId=378047=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-378047
 ]

ASF GitHub Bot logged work on HIVE-22736:
-

Author: ASF GitHub Bot
Created on: 28/Jan/20 05:53
Start Date: 28/Jan/20 05:53
Worklog Time Spent: 10m 
  Work Description: aasha commented on pull request #883: HIVE-22736 
Support multiple encryption zones in replication
URL: https://github.com/apache/hive/pull/883#discussion_r371620277
 
 

 ##
 File path: 
standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/ReplChangeManager.java
 ##
 @@ -150,16 +152,23 @@ private ReplChangeManager(Configuration conf) throws 
MetaException {
 if (MetastoreConf.getBoolVar(conf, ConfVars.REPLCMENABLED)) {
   ReplChangeManager.enabled = true;
   ReplChangeManager.conf = conf;
-
+  cmRootDir = MetastoreConf.getVar(conf, ConfVars.REPLCMDIR);
+  encryptedCmRootDir = MetastoreConf.getVar(conf, 
ConfVars.REPLCMENCRYPTEDDIR);
   //Create default cm root
-  Path cmroot = new Path(MetastoreConf.getVar(conf, 
ConfVars.REPLCMDIR));
+  Path cmroot = new Path(cmRootDir);
+  HdfsEncryptionShim pathEncryptionShim = hadoopShims
+  .createHdfsEncryptionShim(cmroot.getFileSystem(conf), conf);
+  if (pathEncryptionShim.isPathEncrypted(cmroot)) {
+LOG.warn(ConfVars.REPLCMDIR + " should not be encrypted. To pass 
cm dir for encrypted path use "
 
 Review comment:
   Yes this case needs to be handled
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 378047)
Time Spent: 2h 10m  (was: 2h)

> Support replication across multiple encryption zones
> 
>
> Key: HIVE-22736
> URL: https://issues.apache.org/jira/browse/HIVE-22736
> Project: Hive
>  Issue Type: Task
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch
>
>  Time Spent: 2h 10m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-22736) Support replication across multiple encryption zones

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22736?focusedWorklogId=378048=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-378048
 ]

ASF GitHub Bot logged work on HIVE-22736:
-

Author: ASF GitHub Bot
Created on: 28/Jan/20 05:53
Start Date: 28/Jan/20 05:53
Worklog Time Spent: 10m 
  Work Description: aasha commented on pull request #883: HIVE-22736 
Support multiple encryption zones in replication
URL: https://github.com/apache/hive/pull/883#discussion_r371620301
 
 

 ##
 File path: 
standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/ReplChangeManager.java
 ##
 @@ -442,32 +451,39 @@ public static boolean isCMFileUri(Path fromPath) {
 public void run() {
   try {
 LOG.info("CMClearer started");
-
-long now = System.currentTimeMillis();
-FileSystem fs = cmroot.getFileSystem(conf);
-FileStatus[] files = fs.listStatus(cmroot);
-
-for (FileStatus file : files) {
-  long modifiedTime = file.getModificationTime();
-  if (now - modifiedTime > secRetain*1000) {
-try {
-  if 
(fs.getXAttrs(file.getPath()).containsKey(REMAIN_IN_TRASH_TAG)) {
-boolean succ = Trash.moveToAppropriateTrash(fs, 
file.getPath(), conf);
-if (succ) {
-  LOG.debug("Move " + file.toString() + " to trash");
-} else {
-  LOG.warn("Fail to move " + file.toString() + " to trash");
-}
-  } else {
-boolean succ = fs.delete(file.getPath(), false);
-if (succ) {
-  LOG.debug("Remove " + file.toString());
+for (String encryptionZone : encryptionZones) {
+  Path cmroot;
 
 Review comment:
   Yes
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 378048)
Time Spent: 2h 20m  (was: 2h 10m)

> Support replication across multiple encryption zones
> 
>
> Key: HIVE-22736
> URL: https://issues.apache.org/jira/browse/HIVE-22736
> Project: Hive
>  Issue Type: Task
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch
>
>  Time Spent: 2h 20m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22777) Sorted dynamic partition optimization doesn't work if plan require implicit cast

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22777?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024864#comment-17024864
 ] 

Hive QA commented on HIVE-22777:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12991972/HIVE-22777.2.patch

{color:red}ERROR:{color} -1 due to build exiting with an error

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/20373/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/20373/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-20373/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hiveptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ date '+%Y-%m-%d %T.%3N'
2020-01-28 04:52:10.159
+ [[ -n /usr/lib/jvm/java-8-openjdk-amd64 ]]
+ export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ export 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'MAVEN_OPTS=-Xmx1g '
+ MAVEN_OPTS='-Xmx1g '
+ cd /data/hiveptest/working/
+ tee /data/hiveptest/logs/PreCommit-HIVE-Build-20373/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ git = \s\v\n ]]
+ [[ git = \g\i\t ]]
+ [[ -z master ]]
+ [[ -d apache-github-source-source ]]
+ [[ ! -d apache-github-source-source/.git ]]
+ [[ ! -d apache-github-source-source ]]
+ date '+%Y-%m-%d %T.%3N'
2020-01-28 04:52:10.162
+ cd apache-github-source-source
+ git fetch origin
+ git reset --hard HEAD
HEAD is now at 6a5c0cd HIVE-22753: Fix gradual mem leak: Operationlog related 
appenders should be cleared up on errors (Rajesh Balamohan, reviewed by 
Ashutosh Chauhan)
+ git clean -f -d
Removing standalone-metastore/metastore-server/src/gen/
+ git checkout master
Already on 'master'
Your branch is up-to-date with 'origin/master'.
+ git reset --hard origin/master
HEAD is now at 6a5c0cd HIVE-22753: Fix gradual mem leak: Operationlog related 
appenders should be cleared up on errors (Rajesh Balamohan, reviewed by 
Ashutosh Chauhan)
+ git merge --ff-only origin/master
Already up-to-date.
+ date '+%Y-%m-%d %T.%3N'
2020-01-28 04:52:11.235
+ rm -rf ../yetus_PreCommit-HIVE-Build-20373
+ mkdir ../yetus_PreCommit-HIVE-Build-20373
+ git gc
+ cp -R . ../yetus_PreCommit-HIVE-Build-20373
+ mkdir /data/hiveptest/logs/PreCommit-HIVE-Build-20373/yetus
+ patchCommandPath=/data/hiveptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hiveptest/working/scratch/build.patch
+ [[ -f /data/hiveptest/working/scratch/build.patch ]]
+ chmod +x /data/hiveptest/working/scratch/smart-apply-patch.sh
+ /data/hiveptest/working/scratch/smart-apply-patch.sh 
/data/hiveptest/working/scratch/build.patch
Trying to apply the patch with -p0
error: a/itests/src/test/resources/testconfiguration.properties: does not exist 
in index
error: 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/SortedDynPartitionOptimizer.java:
 does not exist in index
error: a/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java: 
does not exist in index
error: a/ql/src/test/queries/clientpositive/dynpart_sort_opt_bucketing.q: does 
not exist in index
error: a/ql/src/test/results/clientpositive/bucket1.q.out: does not exist in 
index
error: a/ql/src/test/results/clientpositive/bucket2.q.out: does not exist in 
index
error: a/ql/src/test/results/clientpositive/bucket3.q.out: does not exist in 
index
error: a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out: 
does not exist in index
error: a/ql/src/test/results/clientpositive/dynpart_sort_opt_bucketing.q.out: 
does not exist in index
error: a/ql/src/test/results/clientpositive/llap/bucket2.q.out: does not exist 
in index
error: a/ql/src/test/results/clientpositive/llap/bucket3.q.out: does not exist 
in index
error: a/ql/src/test/results/clientpositive/llap/bucket4.q.out: does not exist 
in index
error: a/ql/src/test/results/clientpositive/llap/bucket5.q.out: does not exist 
in index
error: a/ql/src/test/results/clientpositive/llap/bucket_many.q.out: does not 
exist in index
error: a/ql/src/test/results/clientpositive/llap/bucket_num_reducers.q.out: 
does not exist in index
error: a/ql/src/test/results/clientpositive/llap/bucket_num_reducers2.q.out: 
does not exist in index
error: a/ql/src/test/results/clientpositive/llap/check_constraint.q.out: does 
not exist in index
error: a/ql/src/test/results/clientpositive/llap/default_constraint.q.out: does 
not exist in index
error: 
a/ql/src/test/results/clientpositive/llap/disable_merge_for_bucketing.q.out: 
does not exist in 

[jira] [Commented] (HIVE-22786) Agg with distinct can be optimised in HASH mode

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22786?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024861#comment-17024861
 ] 

Hive QA commented on HIVE-22786:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12991971/HIVE-22786.1.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 38 failed/errored test(s), 17953 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[auto_join18] (batchId=14)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[auto_join18_multi_distinct]
 (batchId=30)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[auto_join32] (batchId=98)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[distinct_stats] 
(batchId=82)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[groupby8_map_skew] 
(batchId=57)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[groupby9] (batchId=7)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[groupby_distinct_samekey]
 (batchId=65)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[groupby_multi_insert_common_distinct]
 (batchId=85)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[groupby_nocolumnalign] 
(batchId=63)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[groupby_position] 
(batchId=45)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[join18] (batchId=101)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[join18_multi_distinct] 
(batchId=32)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[multi_insert_distinct] 
(batchId=9)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[multi_insert_gby3] 
(batchId=86)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[nullgroup4] (batchId=28)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[nullgroup4_multi_distinct]
 (batchId=12)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[parquet_vectorization_limit]
 (batchId=28)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[vector_count] 
(batchId=15)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[vectorization_limit] 
(batchId=41)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[explainuser_2] 
(batchId=162)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[count] 
(batchId=171)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[metadataonly1]
 (batchId=183)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[tez_union_multiinsert]
 (batchId=174)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[vector_count]
 (batchId=167)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[vector_groupby_cube1]
 (batchId=178)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[vector_groupby_rollup1]
 (batchId=179)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[vector_groupby_sort_11]
 (batchId=188)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[vectorized_distinct_gby]
 (batchId=185)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[count] (batchId=128)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[groupby8_map_skew] 
(batchId=139)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[groupby9] 
(batchId=119)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[groupby_multi_insert_common_distinct]
 (batchId=150)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[groupby_position] 
(batchId=134)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[multi_insert_gby3] 
(batchId=151)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[nullgroup4] 
(batchId=127)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[nullgroup4_multi_distinct]
 (batchId=120)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[union17] 
(batchId=149)
org.apache.hadoop.hive.ql.TestTxnAddPartition.addPartition (batchId=355)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/20372/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/20372/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-20372/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 38 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12991971 - PreCommit-HIVE-Build

> Agg with distinct can be optimised in HASH mode
> ---
>
> Key: HIVE-22786
> URL: https://issues.apache.org/jira/browse/HIVE-22786
> Project: Hive
>  

[jira] [Work logged] (HIVE-22736) Support replication across multiple encryption zones

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22736?focusedWorklogId=378034=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-378034
 ]

ASF GitHub Bot logged work on HIVE-22736:
-

Author: ASF GitHub Bot
Created on: 28/Jan/20 04:29
Start Date: 28/Jan/20 04:29
Worklog Time Spent: 10m 
  Work Description: aasha commented on pull request #883: HIVE-22736 
Support multiple encryption zones in replication
URL: https://github.com/apache/hive/pull/883#discussion_r371604799
 
 

 ##
 File path: 
itests/hive-unit/src/test/java/org/apache/hadoop/hive/metastore/cache/TestCachedStoreUpdateUsingEvents.java
 ##
 @@ -48,6 +48,8 @@ public void setUp() throws Exception {
 MetastoreConf.setBoolVar(conf, ConfVars.METASTORE_CACHE_CAN_USE_EVENT, 
true);
 MetastoreConf.setBoolVar(conf, ConfVars.HIVE_TXN_STATS_ENABLED, true);
 MetastoreConf.setBoolVar(conf, ConfVars.AGGREGATE_STATS_CACHE_ENABLED, 
false);
+MetastoreConf.setBoolVar(conf, ConfVars.REPLCMENABLED, true);
+MetastoreConf.setVar(conf, ConfVars.REPLCMDIR, "cmroot");
 
 Review comment:
   REPLCMDIR default value is /user/${system:user.name}/repl/functions/
   UTs are not able to resolve ${system:user.name}
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 378034)
Time Spent: 2h  (was: 1h 50m)

> Support replication across multiple encryption zones
> 
>
> Key: HIVE-22736
> URL: https://issues.apache.org/jira/browse/HIVE-22736
> Project: Hive
>  Issue Type: Task
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch
>
>  Time Spent: 2h
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22777) Sorted dynamic partition optimization doesn't work if plan require implicit cast

2020-01-27 Thread Vineet Garg (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22777?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-22777:
---
Status: Patch Available  (was: Open)

> Sorted dynamic partition optimization doesn't work if plan require implicit 
> cast
> 
>
> Key: HIVE-22777
> URL: https://issues.apache.org/jira/browse/HIVE-22777
> Project: Hive
>  Issue Type: Bug
>  Components: Query Planning
>Reporter: Vineet Garg
>Assignee: Vineet Garg
>Priority: Major
> Attachments: HIVE-22777.1.patch, HIVE-22777.2.patch
>
>
> *Repro*
> set hive.stats.autogather=false;
> set hive.optimize.sort.dynamic.partition.threshold=1;
> set hive.optimize.bucketingsorting = true;
> {code:sql}
> drop table if exists t1_staging;
> create table t1_staging(
>a string,
>b int,
>c int,
>d string)
> partitioned by (e  decimal(18,0))
> clustered by(a)
> into 256 buckets STORED AS TEXTFILE;
> load data local inpath '../../data/files/sortdp/00_0' overwrite into 
> table t1_staging partition (e=100);
> drop table t1_n147;
> create table t1_n147(
> a string,
> b decimal(6,0),
> c int,
> d string)
> partitioned by (e decimal(3,0))
> clustered by(a,b)
> into 10 buckets STORED AS ORC TBLPROPERTIES ('transactional'='true');
> set hive.stats.autogather=false;
> set hive.optimize.bucketingsorting = true;
> explain insert overwrite table t1_n147 partition(e) select a,b,c,d,e  from 
> t1_staging;
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22777) Sorted dynamic partition optimization doesn't work if plan require implicit cast

2020-01-27 Thread Vineet Garg (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22777?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-22777:
---
Status: Open  (was: Patch Available)

> Sorted dynamic partition optimization doesn't work if plan require implicit 
> cast
> 
>
> Key: HIVE-22777
> URL: https://issues.apache.org/jira/browse/HIVE-22777
> Project: Hive
>  Issue Type: Bug
>  Components: Query Planning
>Reporter: Vineet Garg
>Assignee: Vineet Garg
>Priority: Major
> Attachments: HIVE-22777.1.patch, HIVE-22777.2.patch
>
>
> *Repro*
> set hive.stats.autogather=false;
> set hive.optimize.sort.dynamic.partition.threshold=1;
> set hive.optimize.bucketingsorting = true;
> {code:sql}
> drop table if exists t1_staging;
> create table t1_staging(
>a string,
>b int,
>c int,
>d string)
> partitioned by (e  decimal(18,0))
> clustered by(a)
> into 256 buckets STORED AS TEXTFILE;
> load data local inpath '../../data/files/sortdp/00_0' overwrite into 
> table t1_staging partition (e=100);
> drop table t1_n147;
> create table t1_n147(
> a string,
> b decimal(6,0),
> c int,
> d string)
> partitioned by (e decimal(3,0))
> clustered by(a,b)
> into 10 buckets STORED AS ORC TBLPROPERTIES ('transactional'='true');
> set hive.stats.autogather=false;
> set hive.optimize.bucketingsorting = true;
> explain insert overwrite table t1_n147 partition(e) select a,b,c,d,e  from 
> t1_staging;
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22777) Sorted dynamic partition optimization doesn't work if plan require implicit cast

2020-01-27 Thread Vineet Garg (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22777?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-22777:
---
Attachment: HIVE-22777.2.patch

> Sorted dynamic partition optimization doesn't work if plan require implicit 
> cast
> 
>
> Key: HIVE-22777
> URL: https://issues.apache.org/jira/browse/HIVE-22777
> Project: Hive
>  Issue Type: Bug
>  Components: Query Planning
>Reporter: Vineet Garg
>Assignee: Vineet Garg
>Priority: Major
> Attachments: HIVE-22777.1.patch, HIVE-22777.2.patch
>
>
> *Repro*
> set hive.stats.autogather=false;
> set hive.optimize.sort.dynamic.partition.threshold=1;
> set hive.optimize.bucketingsorting = true;
> {code:sql}
> drop table if exists t1_staging;
> create table t1_staging(
>a string,
>b int,
>c int,
>d string)
> partitioned by (e  decimal(18,0))
> clustered by(a)
> into 256 buckets STORED AS TEXTFILE;
> load data local inpath '../../data/files/sortdp/00_0' overwrite into 
> table t1_staging partition (e=100);
> drop table t1_n147;
> create table t1_n147(
> a string,
> b decimal(6,0),
> c int,
> d string)
> partitioned by (e decimal(3,0))
> clustered by(a,b)
> into 10 buckets STORED AS ORC TBLPROPERTIES ('transactional'='true');
> set hive.stats.autogather=false;
> set hive.optimize.bucketingsorting = true;
> explain insert overwrite table t1_n147 partition(e) select a,b,c,d,e  from 
> t1_staging;
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22786) Agg with distinct can be optimised in HASH mode

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22786?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024850#comment-17024850
 ] 

Hive QA commented on HIVE-22786:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 10m 
27s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
18s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
52s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  5m  
0s{color} | {color:blue} ql in master has 1538 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
16s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
46s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
20s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m 
20s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
48s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  4m 
57s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
10s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
18s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 30m  1s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-20372/dev-support/hive-personality.sh
 |
| git revision | master / 6a5c0cd |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.1 |
| modules | C: ql U: ql |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-20372/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Agg with distinct can be optimised in HASH mode
> ---
>
> Key: HIVE-22786
> URL: https://issues.apache.org/jira/browse/HIVE-22786
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Reporter: Rajesh Balamohan
>Priority: Minor
> Attachments: HIVE-22786.1.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-22736) Support replication across multiple encryption zones

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22736?focusedWorklogId=378029=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-378029
 ]

ASF GitHub Bot logged work on HIVE-22736:
-

Author: ASF GitHub Bot
Created on: 28/Jan/20 03:15
Start Date: 28/Jan/20 03:15
Worklog Time Spent: 10m 
  Work Description: maheshk114 commented on pull request #883: HIVE-22736 
Support multiple encryption zones in replication
URL: https://github.com/apache/hive/pull/883#discussion_r371589321
 
 

 ##
 File path: 
standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/ReplChangeManager.java
 ##
 @@ -442,32 +451,39 @@ public static boolean isCMFileUri(Path fromPath) {
 public void run() {
   try {
 LOG.info("CMClearer started");
-
-long now = System.currentTimeMillis();
-FileSystem fs = cmroot.getFileSystem(conf);
-FileStatus[] files = fs.listStatus(cmroot);
-
-for (FileStatus file : files) {
-  long modifiedTime = file.getModificationTime();
-  if (now - modifiedTime > secRetain*1000) {
-try {
-  if 
(fs.getXAttrs(file.getPath()).containsKey(REMAIN_IN_TRASH_TAG)) {
-boolean succ = Trash.moveToAppropriateTrash(fs, 
file.getPath(), conf);
-if (succ) {
-  LOG.debug("Move " + file.toString() + " to trash");
-} else {
-  LOG.warn("Fail to move " + file.toString() + " to trash");
-}
-  } else {
-boolean succ = fs.delete(file.getPath(), false);
-if (succ) {
-  LOG.debug("Remove " + file.toString());
+for (String encryptionZone : encryptionZones) {
+  Path cmroot;
 
 Review comment:
   Do we have any existing test case for CMCleaner ?
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 378029)
Time Spent: 1h 50m  (was: 1h 40m)

> Support replication across multiple encryption zones
> 
>
> Key: HIVE-22736
> URL: https://issues.apache.org/jira/browse/HIVE-22736
> Project: Hive
>  Issue Type: Task
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch
>
>  Time Spent: 1h 50m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-22736) Support replication across multiple encryption zones

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22736?focusedWorklogId=378026=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-378026
 ]

ASF GitHub Bot logged work on HIVE-22736:
-

Author: ASF GitHub Bot
Created on: 28/Jan/20 03:15
Start Date: 28/Jan/20 03:15
Worklog Time Spent: 10m 
  Work Description: maheshk114 commented on pull request #883: HIVE-22736 
Support multiple encryption zones in replication
URL: https://github.com/apache/hive/pull/883#discussion_r371589875
 
 

 ##
 File path: 
standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java
 ##
 @@ -2836,9 +2836,14 @@ private boolean checkTableDataShouldBeDeleted(Table 
tbl, boolean deleteData) {
  *data from warehouse
  * @param shouldEnableCm If cm should be enabled
  */
-private void deleteTableData(Path tablePath, boolean ifPurge, boolean 
shouldEnableCm) throws MetaException {
+private void deleteTableData(Path tablePath, boolean ifPurge, boolean 
shouldEnableCm) {
   if (tablePath != null) {
-wh.deleteDir(tablePath, true, ifPurge, shouldEnableCm);
+try {
+  wh.deleteDir(tablePath, true, ifPurge, shouldEnableCm);
+} catch (MetaException e) {
 
 Review comment:
   shall we change it to any exception ..including run time exception ?
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 378026)
Time Spent: 1h 40m  (was: 1.5h)

> Support replication across multiple encryption zones
> 
>
> Key: HIVE-22736
> URL: https://issues.apache.org/jira/browse/HIVE-22736
> Project: Hive
>  Issue Type: Task
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch
>
>  Time Spent: 1h 40m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-22736) Support replication across multiple encryption zones

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22736?focusedWorklogId=378025=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-378025
 ]

ASF GitHub Bot logged work on HIVE-22736:
-

Author: ASF GitHub Bot
Created on: 28/Jan/20 03:15
Start Date: 28/Jan/20 03:15
Worklog Time Spent: 10m 
  Work Description: maheshk114 commented on pull request #883: HIVE-22736 
Support multiple encryption zones in replication
URL: https://github.com/apache/hive/pull/883#discussion_r371588458
 
 

 ##
 File path: 
itests/hive-unit/src/test/java/org/apache/hadoop/hive/metastore/cache/TestCachedStoreUpdateUsingEvents.java
 ##
 @@ -48,6 +48,8 @@ public void setUp() throws Exception {
 MetastoreConf.setBoolVar(conf, ConfVars.METASTORE_CACHE_CAN_USE_EVENT, 
true);
 MetastoreConf.setBoolVar(conf, ConfVars.HIVE_TXN_STATS_ENABLED, true);
 MetastoreConf.setBoolVar(conf, ConfVars.AGGREGATE_STATS_CACHE_ENABLED, 
false);
+MetastoreConf.setBoolVar(conf, ConfVars.REPLCMENABLED, true);
+MetastoreConf.setVar(conf, ConfVars.REPLCMDIR, "cmroot");
 
 Review comment:
   why we need to set this ? is it not working with default value ?
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 378025)
Time Spent: 1.5h  (was: 1h 20m)

> Support replication across multiple encryption zones
> 
>
> Key: HIVE-22736
> URL: https://issues.apache.org/jira/browse/HIVE-22736
> Project: Hive
>  Issue Type: Task
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch
>
>  Time Spent: 1.5h
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-22736) Support replication across multiple encryption zones

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22736?focusedWorklogId=378028=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-378028
 ]

ASF GitHub Bot logged work on HIVE-22736:
-

Author: ASF GitHub Bot
Created on: 28/Jan/20 03:15
Start Date: 28/Jan/20 03:15
Worklog Time Spent: 10m 
  Work Description: maheshk114 commented on pull request #883: HIVE-22736 
Support multiple encryption zones in replication
URL: https://github.com/apache/hive/pull/883#discussion_r371589038
 
 

 ##
 File path: 
standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/ReplChangeManager.java
 ##
 @@ -150,16 +152,23 @@ private ReplChangeManager(Configuration conf) throws 
MetaException {
 if (MetastoreConf.getBoolVar(conf, ConfVars.REPLCMENABLED)) {
   ReplChangeManager.enabled = true;
   ReplChangeManager.conf = conf;
-
+  cmRootDir = MetastoreConf.getVar(conf, ConfVars.REPLCMDIR);
+  encryptedCmRootDir = MetastoreConf.getVar(conf, 
ConfVars.REPLCMENCRYPTEDDIR);
   //Create default cm root
-  Path cmroot = new Path(MetastoreConf.getVar(conf, 
ConfVars.REPLCMDIR));
+  Path cmroot = new Path(cmRootDir);
+  HdfsEncryptionShim pathEncryptionShim = hadoopShims
+  .createHdfsEncryptionShim(cmroot.getFileSystem(conf), conf);
+  if (pathEncryptionShim.isPathEncrypted(cmroot)) {
+LOG.warn(ConfVars.REPLCMDIR + " should not be encrypted. To pass 
cm dir for encrypted path use "
 
 Review comment:
   It will create issue if user has a single encryption zone setup and he has 
created a directory with permission to hive as cmroot. And user has not given 
the permission to create directory on encryption zone root to hive. In that 
case we will try to create .cmroot (the default value) and will fail. Should we 
consider this case ?
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 378028)
Time Spent: 1h 50m  (was: 1h 40m)

> Support replication across multiple encryption zones
> 
>
> Key: HIVE-22736
> URL: https://issues.apache.org/jira/browse/HIVE-22736
> Project: Hive
>  Issue Type: Task
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch
>
>  Time Spent: 1h 50m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-22736) Support replication across multiple encryption zones

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22736?focusedWorklogId=378027=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-378027
 ]

ASF GitHub Bot logged work on HIVE-22736:
-

Author: ASF GitHub Bot
Created on: 28/Jan/20 03:15
Start Date: 28/Jan/20 03:15
Worklog Time Spent: 10m 
  Work Description: maheshk114 commented on pull request #883: HIVE-22736 
Support multiple encryption zones in replication
URL: https://github.com/apache/hive/pull/883#discussion_r371588545
 
 

 ##
 File path: 
itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithMiniHS2.java
 ##
 @@ -104,8 +104,11 @@
   public static void setupBeforeClass() throws Exception {
 MiniHS2.cleanupLocalDir();
 HiveConf conf = new HiveConf();
+   // conf.set(HiveConf.ConfVars.REPLCMDIR.varname, "hdfs://cmroot");
+   // conf.set(ConfVars.REPLCMENABLED.varname, "true");
 
 Review comment:
   remove the lines if not required.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 378027)
Time Spent: 1h 40m  (was: 1.5h)

> Support replication across multiple encryption zones
> 
>
> Key: HIVE-22736
> URL: https://issues.apache.org/jira/browse/HIVE-22736
> Project: Hive
>  Issue Type: Task
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, HIVE-22736.patch, 
> HIVE-22736.patch, HIVE-22736.patch
>
>  Time Spent: 1h 40m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22753) Fix gradual mem leak: Operationlog related appenders should be cleared up on errors

2020-01-27 Thread Rajesh Balamohan (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22753?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rajesh Balamohan updated HIVE-22753:

Fix Version/s: 4.0.0
   Resolution: Fixed
   Status: Resolved  (was: Patch Available)

Thanks [~ashutoshc], [~zchovan]. Committed to master.

> Fix gradual mem leak: Operationlog related appenders should be cleared up on 
> errors 
> 
>
> Key: HIVE-22753
> URL: https://issues.apache.org/jira/browse/HIVE-22753
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2
>Reporter: Rajesh Balamohan
>Assignee: Rajesh Balamohan
>Priority: Minor
> Fix For: 4.0.0
>
> Attachments: HIVE-22753.1.patch, HIVE-22753.2.patch, 
> HIVE-22753.3.patch, HIVE-22753.4.patch, image-2020-01-21-11-14-37-911.png, 
> image-2020-01-21-11-17-59-279.png, image-2020-01-21-11-18-37-294.png
>
>
> In case of exception in SQLOperation, operational log does not get cleared 
> up. This causes gradual build up of HushableRandomAccessFileAppender causing 
> HS2 to OOM after some time.
> !image-2020-01-21-11-14-37-911.png|width=431,height=267!
>  
> Allocation tree
> !image-2020-01-21-11-18-37-294.png|width=425,height=178!
>  
> Prod instance mem
> !image-2020-01-21-11-17-59-279.png|width=698,height=209!
>  
> Each HushableRandomAccessFileAppender holds internal ref to 
> RandomAccessFileAppender which holds a 256 KB bytebuffer, causing the mem 
> leak.
> Related ticket: HIVE-18820



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22786) Agg with distinct can be optimised in HASH mode

2020-01-27 Thread Rajesh Balamohan (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22786?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rajesh Balamohan updated HIVE-22786:

Attachment: HIVE-22786.1.patch

> Agg with distinct can be optimised in HASH mode
> ---
>
> Key: HIVE-22786
> URL: https://issues.apache.org/jira/browse/HIVE-22786
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Reporter: Rajesh Balamohan
>Priority: Minor
> Attachments: HIVE-22786.1.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22786) Agg with distinct can be optimised in HASH mode

2020-01-27 Thread Rajesh Balamohan (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22786?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rajesh Balamohan updated HIVE-22786:

Status: Patch Available  (was: Open)

> Agg with distinct can be optimised in HASH mode
> ---
>
> Key: HIVE-22786
> URL: https://issues.apache.org/jira/browse/HIVE-22786
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Reporter: Rajesh Balamohan
>Priority: Minor
> Attachments: HIVE-22786.1.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22779) Hive build fails with Spark 3

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22779?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024816#comment-17024816
 ] 

Hive QA commented on HIVE-22779:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12991962/HIVE-22779.001.patch

{color:red}ERROR:{color} -1 due to build exiting with an error

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/20371/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/20371/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-20371/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hiveptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ date '+%Y-%m-%d %T.%3N'
2020-01-28 01:59:12.049
+ [[ -n /usr/lib/jvm/java-8-openjdk-amd64 ]]
+ export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ export 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'MAVEN_OPTS=-Xmx1g '
+ MAVEN_OPTS='-Xmx1g '
+ cd /data/hiveptest/working/
+ tee /data/hiveptest/logs/PreCommit-HIVE-Build-20371/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ git = \s\v\n ]]
+ [[ git = \g\i\t ]]
+ [[ -z master ]]
+ [[ -d apache-github-source-source ]]
+ [[ ! -d apache-github-source-source/.git ]]
+ [[ ! -d apache-github-source-source ]]
+ date '+%Y-%m-%d %T.%3N'
2020-01-28 01:59:12.052
+ cd apache-github-source-source
+ git fetch origin
+ git reset --hard HEAD
HEAD is now at b7650eb HIVE-22366: Multiple metastore calls for same table and 
constraints during planning (Vineet Garg, reviewed by Jesus Camacho Rodriguez)
+ git clean -f -d
Removing ${project.basedir}/
Removing itests/${project.basedir}/
Removing standalone-metastore/metastore-server/src/gen/
+ git checkout master
Already on 'master'
Your branch is up-to-date with 'origin/master'.
+ git reset --hard origin/master
HEAD is now at b7650eb HIVE-22366: Multiple metastore calls for same table and 
constraints during planning (Vineet Garg, reviewed by Jesus Camacho Rodriguez)
+ git merge --ff-only origin/master
Already up-to-date.
+ date '+%Y-%m-%d %T.%3N'
2020-01-28 01:59:12.923
+ rm -rf ../yetus_PreCommit-HIVE-Build-20371
+ mkdir ../yetus_PreCommit-HIVE-Build-20371
+ git gc
+ cp -R . ../yetus_PreCommit-HIVE-Build-20371
+ mkdir /data/hiveptest/logs/PreCommit-HIVE-Build-20371/yetus
+ patchCommandPath=/data/hiveptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hiveptest/working/scratch/build.patch
+ [[ -f /data/hiveptest/working/scratch/build.patch ]]
+ chmod +x /data/hiveptest/working/scratch/smart-apply-patch.sh
+ /data/hiveptest/working/scratch/smart-apply-patch.sh 
/data/hiveptest/working/scratch/build.patch
Trying to apply the patch with -p0
error: a/ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUtils.java: does 
not exist in index
error: 
a/spark-client/src/main/java/org/apache/hive/spark/client/metrics/ShuffleWriteMetrics.java:
 does not exist in index
error: 
a/spark-client/src/main/java/org/apache/hive/spark/counter/SparkCounter.java: 
does not exist in index
Trying to apply the patch with -p1
error: patch failed: 
ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUtils.java:26
Falling back to three-way merge...
Applied patch to 
'ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUtils.java' with 
conflicts.
Going to apply patch with: git apply -p1
error: patch failed: 
ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUtils.java:26
Falling back to three-way merge...
Applied patch to 
'ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUtils.java' with 
conflicts.
U ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUtils.java
+ result=1
+ '[' 1 -ne 0 ']'
+ rm -rf yetus_PreCommit-HIVE-Build-20371
+ exit 1
'
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12991962 - PreCommit-HIVE-Build

> Hive build fails with Spark 3
> -
>
> Key: HIVE-22779
> URL: https://issues.apache.org/jira/browse/HIVE-22779
> Project: Hive
>  Issue Type: Bug
>  Components: Spark
>Affects Versions: 3.1.2
>Reporter: Igor Dvorzhak
>Priority: Major
> Attachments: HIVE-22779.001.patch
>
>
> Hive build with Spark 3 fails because of deletion of few deprecated APIs in 
> Spark 3.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22784) Boundary condition to check if there is nothing to truncate in StringExpr functions

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22784?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024814#comment-17024814
 ] 

Hive QA commented on HIVE-22784:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12991958/HIVE-22784.1.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 11 failed/errored test(s), 17953 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[orc_llap_counters1]
 (batchId=160)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[orc_llap_counters]
 (batchId=163)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[orc_ppd_schema_evol_3a]
 (batchId=161)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[schema_evol_orc_acid_part]
 (batchId=172)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[schema_evol_orc_acid_part_llap_io]
 (batchId=189)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[schema_evol_orc_acid_table]
 (batchId=175)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[schema_evol_orc_acid_table_llap_io]
 (batchId=174)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[schema_evol_orc_acidvec_part_llap_io]
 (batchId=165)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[schema_evol_orc_acidvec_table]
 (batchId=190)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[schema_evol_orc_acidvec_table_llap_io]
 (batchId=181)
org.apache.hadoop.hive.ql.exec.vector.expressions.TestVectorCastStatement.testString
 (batchId=346)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/20370/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/20370/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-20370/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 11 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12991958 - PreCommit-HIVE-Build

> Boundary condition to check if there is nothing to truncate in StringExpr 
> functions
> ---
>
> Key: HIVE-22784
> URL: https://issues.apache.org/jira/browse/HIVE-22784
> Project: Hive
>  Issue Type: Bug
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
> Attachments: HIVE-22784.1.patch
>
>
> Add a boundary condition to check if there is nothing to truncate in 
> StringExpr functions
> truncate()
> rightTrimAndTruncate()



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-22515) Support cast to decimal64 in Vectorization

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22515?focusedWorklogId=377977=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-377977
 ]

ASF GitHub Bot logged work on HIVE-22515:
-

Author: ASF GitHub Bot
Created on: 28/Jan/20 01:28
Start Date: 28/Jan/20 01:28
Worklog Time Spent: 10m 
  Work Description: ramesh0201 commented on issue #880: HIVE-22515 Support 
cast to decimal64 in Vectorization
URL: https://github.com/apache/hive/pull/880#issuecomment-579036627
 
 
   Whenever we have an expression like `long column  decimal column`, we always cast long to decimal even if the decimal 
column can be represented as decimal64. This prevents us from using Deicmal64 
class of operations that will be faster than the equivalent Decimal class of 
operations. This Jira will implements cast from long to decimal64 that can help 
in more optimizations in the decimal64 operations.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 377977)
Time Spent: 1h 10m  (was: 1h)

> Support cast to decimal64 in Vectorization
> --
>
> Key: HIVE-22515
> URL: https://issues.apache.org/jira/browse/HIVE-22515
> Project: Hive
>  Issue Type: Bug
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22515.5.patch, HIVE-22515.8.patch, 
> HIVE-22515.9.patch
>
>  Time Spent: 1h 10m
>  Remaining Estimate: 0h
>
> Support cast to decimal64 in Vectorization



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22515) Support cast to decimal64 in Vectorization

2020-01-27 Thread Ramesh Kumar Thangarajan (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22515?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024796#comment-17024796
 ] 

Ramesh Kumar Thangarajan commented on HIVE-22515:
-

Whenever we have an expression like `long column  
decimal column`, we always cast long to decimal even if the decimal column can 
be represented as decimal64. This prevents us from using Deicmal64 class of 
operations that will be faster than the equivalent Decimal class of operations. 
This Jira will implements cast from long to decimal64 that can help in more 
optimizations in the decimal64 operations.

> Support cast to decimal64 in Vectorization
> --
>
> Key: HIVE-22515
> URL: https://issues.apache.org/jira/browse/HIVE-22515
> Project: Hive
>  Issue Type: Bug
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22515.5.patch, HIVE-22515.8.patch, 
> HIVE-22515.9.patch
>
>  Time Spent: 1h
>  Remaining Estimate: 0h
>
> Support cast to decimal64 in Vectorization



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-22515) Support cast to decimal64 in Vectorization

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22515?focusedWorklogId=377961=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-377961
 ]

ASF GitHub Bot logged work on HIVE-22515:
-

Author: ASF GitHub Bot
Created on: 28/Jan/20 01:14
Start Date: 28/Jan/20 01:14
Worklog Time Spent: 10m 
  Work Description: ramesh0201 commented on pull request #880: HIVE-22515 
Support cast to decimal64 in Vectorization
URL: https://github.com/apache/hive/pull/880#discussion_r371568274
 
 

 ##
 File path: 
ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FuncLongToDecimal64.java
 ##
 @@ -0,0 +1,185 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.hadoop.hive.ql.exec.vector.expressions;
+
+import java.util.Arrays;
+
+import org.apache.hadoop.hive.ql.exec.vector.Decimal64ColumnVector;
+import org.apache.hadoop.hive.ql.exec.vector.LongColumnVector;
+import org.apache.hadoop.hive.ql.exec.vector.VectorExpressionDescriptor;
+import 
org.apache.hadoop.hive.ql.exec.vector.VectorizedExpressionsSupportDecimal64;
+import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch;
+import org.apache.hadoop.hive.ql.metadata.HiveException;
+import org.apache.hadoop.hive.serde2.typeinfo.DecimalTypeInfo;
+
+/**
+ * This is a superclass for unary long functions and expressions returning 
decimals that
+ * operate directly on the input and set the output.
+ */
+@VectorizedExpressionsSupportDecimal64()
+public abstract class FuncLongToDecimal64 extends VectorExpression {
+  private static final long serialVersionUID = 1L;
+  private final int inputColumn;
+
+  private static final long[] powerOfTenTable = {
+1L,   // 0
+10L,
+100L,
+1_000L,
+10_000L,
+100_000L,
+1_000_000L,
+10_000_000L,
+100_000_000L,   // 8
+1_000_000_000L,
+10_000_000_000L,
+100_000_000_000L,
+1_000_000_000_000L,
+10_000_000_000_000L,
+100_000_000_000_000L,
+1_000_000_000_000_000L,
+10_000_000_000_000_000L,   // 16
+100_000_000_000_000_000L,
+1_000_000_000_000_000_000L, // 18
+  };
+
+  public FuncLongToDecimal64(int inputColumn, int outputColumnNum) {
+super(outputColumnNum);
+this.inputColumn = inputColumn;
+  }
+
+  public FuncLongToDecimal64() {
+super();
+
+// Dummy final assignments.
+inputColumn = -1;
+  }
+
+  abstract protected void func(Decimal64ColumnVector outputColVector, 
LongColumnVector inputColVector, int i,
+  long scaleFactor);
+
+  @Override
+  public void evaluate(VectorizedRowBatch batch) throws HiveException {
+
+if (childExpressions != null) {
+  super.evaluateChildren(batch);
+}
+
+LongColumnVector inputColVector = (LongColumnVector) 
batch.cols[inputColumn];
+int[] sel = batch.selected;
+int n = batch.size;
+Decimal64ColumnVector outputColVector = (Decimal64ColumnVector) 
batch.cols[outputColumnNum];
+int outputScale = ((DecimalTypeInfo) outputTypeInfo).scale();
+final long scaleFactor = powerOfTenTable[outputScale];
 
 Review comment:
   Yes, before we create cast expression, we verify if the output.precision + 
output.scale <= 18(to avoid intermediate overflow) and hence implicitly 
outputscale will be less than or equal to 18.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 377961)
Time Spent: 40m  (was: 0.5h)

> Support cast to decimal64 in Vectorization
> --
>
> Key: HIVE-22515
> URL: https://issues.apache.org/jira/browse/HIVE-22515
> Project: Hive
>  Issue Type: Bug
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
>  Labels: 

[jira] [Work logged] (HIVE-22515) Support cast to decimal64 in Vectorization

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22515?focusedWorklogId=377962=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-377962
 ]

ASF GitHub Bot logged work on HIVE-22515:
-

Author: ASF GitHub Bot
Created on: 28/Jan/20 01:14
Start Date: 28/Jan/20 01:14
Worklog Time Spent: 10m 
  Work Description: ramesh0201 commented on pull request #880: HIVE-22515 
Support cast to decimal64 in Vectorization
URL: https://github.com/apache/hive/pull/880#discussion_r371568368
 
 

 ##
 File path: 
ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastLongToDecimal64.java
 ##
 @@ -0,0 +1,45 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.hadoop.hive.ql.exec.vector.expressions;
+
+import org.apache.hadoop.hive.ql.exec.vector.Decimal64ColumnVector;
+import org.apache.hadoop.hive.ql.exec.vector.LongColumnVector;
+
+/**
+ * To be used to cast long and boolean to decimal.
+ * This works for boolean too because boolean is encoded as 0
+ * for false and 1 for true.
+ */
+public class CastLongToDecimal64 extends FuncLongToDecimal64 {
 
 Review comment:
   Thank you, I have addressed this.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 377962)
Time Spent: 50m  (was: 40m)

> Support cast to decimal64 in Vectorization
> --
>
> Key: HIVE-22515
> URL: https://issues.apache.org/jira/browse/HIVE-22515
> Project: Hive
>  Issue Type: Bug
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22515.5.patch, HIVE-22515.8.patch, 
> HIVE-22515.9.patch
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>
> Support cast to decimal64 in Vectorization



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-22515) Support cast to decimal64 in Vectorization

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22515?focusedWorklogId=377963=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-377963
 ]

ASF GitHub Bot logged work on HIVE-22515:
-

Author: ASF GitHub Bot
Created on: 28/Jan/20 01:14
Start Date: 28/Jan/20 01:14
Worklog Time Spent: 10m 
  Work Description: ramesh0201 commented on pull request #880: HIVE-22515 
Support cast to decimal64 in Vectorization
URL: https://github.com/apache/hive/pull/880#discussion_r371568394
 
 

 ##
 File path: 
ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FuncLongToDecimal64.java
 ##
 @@ -0,0 +1,185 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.hadoop.hive.ql.exec.vector.expressions;
+
+import java.util.Arrays;
+
+import org.apache.hadoop.hive.ql.exec.vector.Decimal64ColumnVector;
+import org.apache.hadoop.hive.ql.exec.vector.LongColumnVector;
+import org.apache.hadoop.hive.ql.exec.vector.VectorExpressionDescriptor;
+import 
org.apache.hadoop.hive.ql.exec.vector.VectorizedExpressionsSupportDecimal64;
+import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch;
+import org.apache.hadoop.hive.ql.metadata.HiveException;
+import org.apache.hadoop.hive.serde2.typeinfo.DecimalTypeInfo;
+
+/**
+ * This is a superclass for unary long functions and expressions returning 
decimals that
+ * operate directly on the input and set the output.
+ */
+@VectorizedExpressionsSupportDecimal64()
+public abstract class FuncLongToDecimal64 extends VectorExpression {
+  private static final long serialVersionUID = 1L;
+  private final int inputColumn;
+
+  private static final long[] powerOfTenTable = {
+1L,   // 0
+10L,
+100L,
+1_000L,
+10_000L,
+100_000L,
+1_000_000L,
+10_000_000L,
+100_000_000L,   // 8
+1_000_000_000L,
+10_000_000_000L,
+100_000_000_000L,
+1_000_000_000_000L,
+10_000_000_000_000L,
+100_000_000_000_000L,
+1_000_000_000_000_000L,
+10_000_000_000_000_000L,   // 16
+100_000_000_000_000_000L,
+1_000_000_000_000_000_000L, // 18
+  };
+
+  public FuncLongToDecimal64(int inputColumn, int outputColumnNum) {
+super(outputColumnNum);
+this.inputColumn = inputColumn;
+  }
+
+  public FuncLongToDecimal64() {
+super();
+
+// Dummy final assignments.
+inputColumn = -1;
+  }
+
+  abstract protected void func(Decimal64ColumnVector outputColVector, 
LongColumnVector inputColVector, int i,
 
 Review comment:
   Thank you, I have addressed this.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 377963)
Time Spent: 1h  (was: 50m)

> Support cast to decimal64 in Vectorization
> --
>
> Key: HIVE-22515
> URL: https://issues.apache.org/jira/browse/HIVE-22515
> Project: Hive
>  Issue Type: Bug
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22515.5.patch, HIVE-22515.8.patch, 
> HIVE-22515.9.patch
>
>  Time Spent: 1h
>  Remaining Estimate: 0h
>
> Support cast to decimal64 in Vectorization



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-22515) Support cast to decimal64 in Vectorization

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22515?focusedWorklogId=377960=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-377960
 ]

ASF GitHub Bot logged work on HIVE-22515:
-

Author: ASF GitHub Bot
Created on: 28/Jan/20 01:12
Start Date: 28/Jan/20 01:12
Worklog Time Spent: 10m 
  Work Description: ramesh0201 commented on pull request #880: HIVE-22515 
Support cast to decimal64 in Vectorization
URL: https://github.com/apache/hive/pull/880#discussion_r371567891
 
 

 ##
 File path: 
ql/src/test/results/clientpositive/llap/vector_reuse_scratchcols.q.out
 ##
 @@ -369,7 +369,7 @@ STAGE PLANS:
 includeColumns: [0, 1, 2, 3, 4, 5, 6, 7, 9, 10, 11]
 dataColumns: ctinyint:tinyint, csmallint:smallint, 
cint:int, cbigint:bigint, cfloat:float, cdouble:double, cstring1:string, 
cstring2:string, ctimestamp1:timestamp, ctimestamp2:timestamp, 
cboolean1:boolean, cboolean2:boolean
 partitionColumnCount: 0
-scratchColumnTypeNames: [double, double, double, 
decimal(22,3), decimal(13,3), double, double, double, double, double, double, 
double, double]
+scratchColumnTypeNames: [double, double, double, 
decimal(22,3), decimal(13,3)/DECIMAL_64, double, double, double, double, 
double, double, double, double]
 
 Review comment:
   Yes, because we can safely store a decimal with (13,3) by scaling them up 
and storing in long vectors and tracking the scale. We do this until 
precision+scale is <=18.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 377960)
Time Spent: 0.5h  (was: 20m)

> Support cast to decimal64 in Vectorization
> --
>
> Key: HIVE-22515
> URL: https://issues.apache.org/jira/browse/HIVE-22515
> Project: Hive
>  Issue Type: Bug
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22515.5.patch, HIVE-22515.8.patch, 
> HIVE-22515.9.patch
>
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>
> Support cast to decimal64 in Vectorization



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22779) Hive build fails with Spark 3

2020-01-27 Thread Igor Dvorzhak (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22779?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Igor Dvorzhak updated HIVE-22779:
-
Attachment: HIVE-22779.001.patch
Status: Patch Available  (was: Open)

> Hive build fails with Spark 3
> -
>
> Key: HIVE-22779
> URL: https://issues.apache.org/jira/browse/HIVE-22779
> Project: Hive
>  Issue Type: Bug
>  Components: Spark
>Affects Versions: 3.1.2
>Reporter: Igor Dvorzhak
>Priority: Major
> Attachments: HIVE-22779.001.patch
>
>
> Hive build with Spark 3 fails because of deletion of few deprecated APIs in 
> Spark 3.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22779) Hive build fails with Spark 3

2020-01-27 Thread Igor Dvorzhak (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22779?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Igor Dvorzhak updated HIVE-22779:
-
Status: Open  (was: Patch Available)

> Hive build fails with Spark 3
> -
>
> Key: HIVE-22779
> URL: https://issues.apache.org/jira/browse/HIVE-22779
> Project: Hive
>  Issue Type: Bug
>  Components: Spark
>Affects Versions: 3.1.2
>Reporter: Igor Dvorzhak
>Priority: Major
> Attachments: HIVE-22779.001.patch
>
>
> Hive build with Spark 3 fails because of deletion of few deprecated APIs in 
> Spark 3.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22779) Hive build fails with Spark 3

2020-01-27 Thread Igor Dvorzhak (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22779?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Igor Dvorzhak updated HIVE-22779:
-
Attachment: (was: HIVE-22779.001.patch)

> Hive build fails with Spark 3
> -
>
> Key: HIVE-22779
> URL: https://issues.apache.org/jira/browse/HIVE-22779
> Project: Hive
>  Issue Type: Bug
>  Components: Spark
>Affects Versions: 3.1.2
>Reporter: Igor Dvorzhak
>Priority: Major
> Attachments: HIVE-22779.001.patch
>
>
> Hive build with Spark 3 fails because of deletion of few deprecated APIs in 
> Spark 3.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22784) Boundary condition to check if there is nothing to truncate in StringExpr functions

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22784?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024789#comment-17024789
 ] 

Hive QA commented on HIVE-22784:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  9m 
44s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
10s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
11s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
24s{color} | {color:blue} storage-api in master has 58 extant Findbugs 
warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
10s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
12s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m  
9s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m  
9s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
12s{color} | {color:green} storage-api: The patch generated 0 new + 9 unchanged 
- 1 fixed = 9 total (was 10) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
29s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
11s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 12m 38s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-20370/dev-support/hive-personality.sh
 |
| git revision | master / b7650eb |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.1 |
| modules | C: storage-api U: storage-api |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-20370/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Boundary condition to check if there is nothing to truncate in StringExpr 
> functions
> ---
>
> Key: HIVE-22784
> URL: https://issues.apache.org/jira/browse/HIVE-22784
> Project: Hive
>  Issue Type: Bug
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
> Attachments: HIVE-22784.1.patch
>
>
> Add a boundary condition to check if there is nothing to truncate in 
> StringExpr functions
> truncate()
> rightTrimAndTruncate()



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22771) Partition location incorrectly formed in FileOutputCommitterContainer

2020-01-27 Thread Shivam (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22771?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shivam updated HIVE-22771:
--
Status: In Progress  (was: Patch Available)

Incorporating comments to add testcase

> Partition location incorrectly formed in FileOutputCommitterContainer
> -
>
> Key: HIVE-22771
> URL: https://issues.apache.org/jira/browse/HIVE-22771
> Project: Hive
>  Issue Type: Bug
>  Components: HCatalog
>Affects Versions: 1.2.1
>Reporter: Shivam
>Assignee: Shivam
>Priority: Critical
>  Labels: pull-request-available
> Attachments: HIVE-22771.patch
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Class _HCatOutputFormat_ in package _org.apache.hive.hcatalog.mapreduce_ uses 
> function _setOutput_ to generate _idHash_ using below statement:
> *+In file org/apache/hive/hcatalog/mapreduce/HCatOutputFormat.java+*
>  *line 116: idHash = String.valueOf(Math.random());*
> The output of idHash can be similar to values like this : 7.145347157239135E-4
>  
> And, in class _FileOutputCommitterContainer_ in package 
> _org.apache.hive.hcatalog.mapreduce;_
> Uses below statement to compute final partition path:
> +*In org/apache/hive/hcatalog/mapreduce/FileOutputCommitterContainer.java*+
> *line 366: String finalLocn = jobLocation.replaceAll(Path.SEPARATOR + 
> SCRATCH_DIR_NAME + "{color:#ff}\\d\\.?
> d+"{color},"");*
> *line 367: partPath = new Path(finalLocn);*
>  
> Regex used here is incorrect, since it will only remove integers after the 
> *SCRATCH_DIR_NAME,* and hence will append  'E-4' (for the above example) in 
> the final partition location. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22753) Fix gradual mem leak: Operationlog related appenders should be cleared up on errors

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22753?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024784#comment-17024784
 ] 

Hive QA commented on HIVE-22753:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12991952/HIVE-22753.4.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 17953 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/20369/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/20369/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-20369/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12991952 - PreCommit-HIVE-Build

> Fix gradual mem leak: Operationlog related appenders should be cleared up on 
> errors 
> 
>
> Key: HIVE-22753
> URL: https://issues.apache.org/jira/browse/HIVE-22753
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2
>Reporter: Rajesh Balamohan
>Assignee: Rajesh Balamohan
>Priority: Minor
> Attachments: HIVE-22753.1.patch, HIVE-22753.2.patch, 
> HIVE-22753.3.patch, HIVE-22753.4.patch, image-2020-01-21-11-14-37-911.png, 
> image-2020-01-21-11-17-59-279.png, image-2020-01-21-11-18-37-294.png
>
>
> In case of exception in SQLOperation, operational log does not get cleared 
> up. This causes gradual build up of HushableRandomAccessFileAppender causing 
> HS2 to OOM after some time.
> !image-2020-01-21-11-14-37-911.png|width=431,height=267!
>  
> Allocation tree
> !image-2020-01-21-11-18-37-294.png|width=425,height=178!
>  
> Prod instance mem
> !image-2020-01-21-11-17-59-279.png|width=698,height=209!
>  
> Each HushableRandomAccessFileAppender holds internal ref to 
> RandomAccessFileAppender which holds a 256 KB bytebuffer, causing the mem 
> leak.
> Related ticket: HIVE-18820



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22771) Partition location incorrectly formed in FileOutputCommitterContainer

2020-01-27 Thread Mithun Radhakrishnan (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22771?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024768#comment-17024768
 ] 

Mithun Radhakrishnan commented on HIVE-22771:
-

FWIW, +1 from me as well. This is a good catch.
{quote}bq. Can we add a test for it (e.g., in TestHCatStorer)?
{quote}
This bug probably got through because while there are tests in 
{{AbstractHCatStorerTest}} for the dynamic-partitioning case, and the hybrid 
(i.e. partially dynamic) case, the static case (i.e. when the partition-ids are 
fully known _a priori_) isn't covered. There will be value in adding a test for 
Pig scripts writing to static partitions.

Thank you for working on this, [~shivam-mohan]. :]

> Partition location incorrectly formed in FileOutputCommitterContainer
> -
>
> Key: HIVE-22771
> URL: https://issues.apache.org/jira/browse/HIVE-22771
> Project: Hive
>  Issue Type: Bug
>  Components: HCatalog
>Affects Versions: 1.2.1
>Reporter: Shivam
>Assignee: Shivam
>Priority: Critical
>  Labels: pull-request-available
> Attachments: HIVE-22771.patch
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Class _HCatOutputFormat_ in package _org.apache.hive.hcatalog.mapreduce_ uses 
> function _setOutput_ to generate _idHash_ using below statement:
> *+In file org/apache/hive/hcatalog/mapreduce/HCatOutputFormat.java+*
>  *line 116: idHash = String.valueOf(Math.random());*
> The output of idHash can be similar to values like this : 7.145347157239135E-4
>  
> And, in class _FileOutputCommitterContainer_ in package 
> _org.apache.hive.hcatalog.mapreduce;_
> Uses below statement to compute final partition path:
> +*In org/apache/hive/hcatalog/mapreduce/FileOutputCommitterContainer.java*+
> *line 366: String finalLocn = jobLocation.replaceAll(Path.SEPARATOR + 
> SCRATCH_DIR_NAME + "{color:#ff}\\d\\.?
> d+"{color},"");*
> *line 367: partPath = new Path(finalLocn);*
>  
> Regex used here is incorrect, since it will only remove integers after the 
> *SCRATCH_DIR_NAME,* and hence will append  'E-4' (for the above example) in 
> the final partition location. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22753) Fix gradual mem leak: Operationlog related appenders should be cleared up on errors

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22753?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024765#comment-17024765
 ] 

Hive QA commented on HIVE-22753:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 10m 
12s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
13s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
42s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  4m 
32s{color} | {color:blue} ql in master has 1538 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
8s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
35s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
11s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m 
11s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
42s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  4m 
38s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
7s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
15s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 27m 39s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-20369/dev-support/hive-personality.sh
 |
| git revision | master / b7650eb |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.1 |
| modules | C: ql U: ql |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-20369/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Fix gradual mem leak: Operationlog related appenders should be cleared up on 
> errors 
> 
>
> Key: HIVE-22753
> URL: https://issues.apache.org/jira/browse/HIVE-22753
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2
>Reporter: Rajesh Balamohan
>Assignee: Rajesh Balamohan
>Priority: Minor
> Attachments: HIVE-22753.1.patch, HIVE-22753.2.patch, 
> HIVE-22753.3.patch, HIVE-22753.4.patch, image-2020-01-21-11-14-37-911.png, 
> image-2020-01-21-11-17-59-279.png, image-2020-01-21-11-18-37-294.png
>
>
> In case of exception in SQLOperation, operational log does not get cleared 
> up. This causes gradual build up of HushableRandomAccessFileAppender causing 
> HS2 to OOM after some time.
> !image-2020-01-21-11-14-37-911.png|width=431,height=267!
>  
> Allocation tree
> !image-2020-01-21-11-18-37-294.png|width=425,height=178!
>  
> Prod instance mem
> !image-2020-01-21-11-17-59-279.png|width=698,height=209!
>  
> Each HushableRandomAccessFileAppender holds internal ref to 
> RandomAccessFileAppender which holds a 256 KB bytebuffer, causing the mem 
> leak.
> Related ticket: HIVE-18820



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22771) Partition location incorrectly formed in FileOutputCommitterContainer

2020-01-27 Thread Jesus Camacho Rodriguez (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22771?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024759#comment-17024759
 ] 

Jesus Camacho Rodriguez commented on HIVE-22771:


Thanks [~shivam-mohan]. Fix makes sense. Can we add a test for it (e.g., in 
TestHCatStorer)?

> Partition location incorrectly formed in FileOutputCommitterContainer
> -
>
> Key: HIVE-22771
> URL: https://issues.apache.org/jira/browse/HIVE-22771
> Project: Hive
>  Issue Type: Bug
>  Components: HCatalog
>Affects Versions: 1.2.1
>Reporter: Shivam
>Assignee: Shivam
>Priority: Critical
>  Labels: pull-request-available
> Attachments: HIVE-22771.patch
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Class _HCatOutputFormat_ in package _org.apache.hive.hcatalog.mapreduce_ uses 
> function _setOutput_ to generate _idHash_ using below statement:
> *+In file org/apache/hive/hcatalog/mapreduce/HCatOutputFormat.java+*
>  *line 116: idHash = String.valueOf(Math.random());*
> The output of idHash can be similar to values like this : 7.145347157239135E-4
>  
> And, in class _FileOutputCommitterContainer_ in package 
> _org.apache.hive.hcatalog.mapreduce;_
> Uses below statement to compute final partition path:
> +*In org/apache/hive/hcatalog/mapreduce/FileOutputCommitterContainer.java*+
> *line 366: String finalLocn = jobLocation.replaceAll(Path.SEPARATOR + 
> SCRATCH_DIR_NAME + "{color:#ff}\\d\\.?
> d+"{color},"");*
> *line 367: partPath = new Path(finalLocn);*
>  
> Regex used here is incorrect, since it will only remove integers after the 
> *SCRATCH_DIR_NAME,* and hence will append  'E-4' (for the above example) in 
> the final partition location. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-22515) Support cast to decimal64 in Vectorization

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22515?focusedWorklogId=377910=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-377910
 ]

ASF GitHub Bot logged work on HIVE-22515:
-

Author: ASF GitHub Bot
Created on: 27/Jan/20 23:27
Start Date: 27/Jan/20 23:27
Worklog Time Spent: 10m 
  Work Description: jdere commented on pull request #880: HIVE-22515 
Support cast to decimal64 in Vectorization
URL: https://github.com/apache/hive/pull/880#discussion_r371533452
 
 

 ##
 File path: 
ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FuncLongToDecimal64.java
 ##
 @@ -0,0 +1,185 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.hadoop.hive.ql.exec.vector.expressions;
+
+import java.util.Arrays;
+
+import org.apache.hadoop.hive.ql.exec.vector.Decimal64ColumnVector;
+import org.apache.hadoop.hive.ql.exec.vector.LongColumnVector;
+import org.apache.hadoop.hive.ql.exec.vector.VectorExpressionDescriptor;
+import 
org.apache.hadoop.hive.ql.exec.vector.VectorizedExpressionsSupportDecimal64;
+import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch;
+import org.apache.hadoop.hive.ql.metadata.HiveException;
+import org.apache.hadoop.hive.serde2.typeinfo.DecimalTypeInfo;
+
+/**
+ * This is a superclass for unary long functions and expressions returning 
decimals that
+ * operate directly on the input and set the output.
+ */
+@VectorizedExpressionsSupportDecimal64()
+public abstract class FuncLongToDecimal64 extends VectorExpression {
+  private static final long serialVersionUID = 1L;
+  private final int inputColumn;
+
+  private static final long[] powerOfTenTable = {
+1L,   // 0
+10L,
+100L,
+1_000L,
+10_000L,
+100_000L,
+1_000_000L,
+10_000_000L,
+100_000_000L,   // 8
+1_000_000_000L,
+10_000_000_000L,
+100_000_000_000L,
+1_000_000_000_000L,
+10_000_000_000_000L,
+100_000_000_000_000L,
+1_000_000_000_000_000L,
+10_000_000_000_000_000L,   // 16
+100_000_000_000_000_000L,
+1_000_000_000_000_000_000L, // 18
+  };
+
+  public FuncLongToDecimal64(int inputColumn, int outputColumnNum) {
+super(outputColumnNum);
+this.inputColumn = inputColumn;
+  }
+
+  public FuncLongToDecimal64() {
+super();
+
+// Dummy final assignments.
+inputColumn = -1;
+  }
+
+  abstract protected void func(Decimal64ColumnVector outputColVector, 
LongColumnVector inputColVector, int i,
 
 Review comment:
   Can you use a slightly more descriptive method name here and describe what 
it's for?
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 377910)
Time Spent: 20m  (was: 10m)

> Support cast to decimal64 in Vectorization
> --
>
> Key: HIVE-22515
> URL: https://issues.apache.org/jira/browse/HIVE-22515
> Project: Hive
>  Issue Type: Bug
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22515.5.patch, HIVE-22515.8.patch, 
> HIVE-22515.9.patch
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> Support cast to decimal64 in Vectorization



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-22515) Support cast to decimal64 in Vectorization

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22515?focusedWorklogId=377911=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-377911
 ]

ASF GitHub Bot logged work on HIVE-22515:
-

Author: ASF GitHub Bot
Created on: 27/Jan/20 23:27
Start Date: 27/Jan/20 23:27
Worklog Time Spent: 10m 
  Work Description: jdere commented on pull request #880: HIVE-22515 
Support cast to decimal64 in Vectorization
URL: https://github.com/apache/hive/pull/880#discussion_r371537293
 
 

 ##
 File path: 
ql/src/test/results/clientpositive/llap/vector_reuse_scratchcols.q.out
 ##
 @@ -369,7 +369,7 @@ STAGE PLANS:
 includeColumns: [0, 1, 2, 3, 4, 5, 6, 7, 9, 10, 11]
 dataColumns: ctinyint:tinyint, csmallint:smallint, 
cint:int, cbigint:bigint, cfloat:float, cdouble:double, cstring1:string, 
cstring2:string, ctimestamp1:timestamp, ctimestamp2:timestamp, 
cboolean1:boolean, cboolean2:boolean
 partitionColumnCount: 0
-scratchColumnTypeNames: [double, double, double, 
decimal(22,3), decimal(13,3), double, double, double, double, double, double, 
double, double]
+scratchColumnTypeNames: [double, double, double, 
decimal(22,3), decimal(13,3)/DECIMAL_64, double, double, double, double, 
double, double, double, double]
 
 Review comment:
   any idea how this shows up as decimal(13,3)/DECIMAL_64 - would the 
decimal(13,3) part also be converted to decimal64?
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 377911)
Time Spent: 20m  (was: 10m)

> Support cast to decimal64 in Vectorization
> --
>
> Key: HIVE-22515
> URL: https://issues.apache.org/jira/browse/HIVE-22515
> Project: Hive
>  Issue Type: Bug
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22515.5.patch, HIVE-22515.8.patch, 
> HIVE-22515.9.patch
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> Support cast to decimal64 in Vectorization



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-22515) Support cast to decimal64 in Vectorization

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22515?focusedWorklogId=377909=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-377909
 ]

ASF GitHub Bot logged work on HIVE-22515:
-

Author: ASF GitHub Bot
Created on: 27/Jan/20 23:27
Start Date: 27/Jan/20 23:27
Worklog Time Spent: 10m 
  Work Description: jdere commented on pull request #880: HIVE-22515 
Support cast to decimal64 in Vectorization
URL: https://github.com/apache/hive/pull/880#discussion_r371534428
 
 

 ##
 File path: 
ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/CastLongToDecimal64.java
 ##
 @@ -0,0 +1,45 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.hadoop.hive.ql.exec.vector.expressions;
+
+import org.apache.hadoop.hive.ql.exec.vector.Decimal64ColumnVector;
+import org.apache.hadoop.hive.ql.exec.vector.LongColumnVector;
+
+/**
+ * To be used to cast long and boolean to decimal.
+ * This works for boolean too because boolean is encoded as 0
+ * for false and 1 for true.
+ */
+public class CastLongToDecimal64 extends FuncLongToDecimal64 {
 
 Review comment:
   This is currently the only concrete class for FuncLongToDecimal64 .. is it 
expected that there are going to be more in the future (which would be the 
point of making FuncLongToDecimal64 abstract?
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 377909)
Time Spent: 20m  (was: 10m)

> Support cast to decimal64 in Vectorization
> --
>
> Key: HIVE-22515
> URL: https://issues.apache.org/jira/browse/HIVE-22515
> Project: Hive
>  Issue Type: Bug
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22515.5.patch, HIVE-22515.8.patch, 
> HIVE-22515.9.patch
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> Support cast to decimal64 in Vectorization



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-22515) Support cast to decimal64 in Vectorization

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22515?focusedWorklogId=377908=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-377908
 ]

ASF GitHub Bot logged work on HIVE-22515:
-

Author: ASF GitHub Bot
Created on: 27/Jan/20 23:27
Start Date: 27/Jan/20 23:27
Worklog Time Spent: 10m 
  Work Description: jdere commented on pull request #880: HIVE-22515 
Support cast to decimal64 in Vectorization
URL: https://github.com/apache/hive/pull/880#discussion_r371539542
 
 

 ##
 File path: 
ql/src/java/org/apache/hadoop/hive/ql/exec/vector/expressions/FuncLongToDecimal64.java
 ##
 @@ -0,0 +1,185 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.hadoop.hive.ql.exec.vector.expressions;
+
+import java.util.Arrays;
+
+import org.apache.hadoop.hive.ql.exec.vector.Decimal64ColumnVector;
+import org.apache.hadoop.hive.ql.exec.vector.LongColumnVector;
+import org.apache.hadoop.hive.ql.exec.vector.VectorExpressionDescriptor;
+import 
org.apache.hadoop.hive.ql.exec.vector.VectorizedExpressionsSupportDecimal64;
+import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch;
+import org.apache.hadoop.hive.ql.metadata.HiveException;
+import org.apache.hadoop.hive.serde2.typeinfo.DecimalTypeInfo;
+
+/**
+ * This is a superclass for unary long functions and expressions returning 
decimals that
+ * operate directly on the input and set the output.
+ */
+@VectorizedExpressionsSupportDecimal64()
+public abstract class FuncLongToDecimal64 extends VectorExpression {
+  private static final long serialVersionUID = 1L;
+  private final int inputColumn;
+
+  private static final long[] powerOfTenTable = {
+1L,   // 0
+10L,
+100L,
+1_000L,
+10_000L,
+100_000L,
+1_000_000L,
+10_000_000L,
+100_000_000L,   // 8
+1_000_000_000L,
+10_000_000_000L,
+100_000_000_000L,
+1_000_000_000_000L,
+10_000_000_000_000L,
+100_000_000_000_000L,
+1_000_000_000_000_000L,
+10_000_000_000_000_000L,   // 16
+100_000_000_000_000_000L,
+1_000_000_000_000_000_000L, // 18
+  };
+
+  public FuncLongToDecimal64(int inputColumn, int outputColumnNum) {
+super(outputColumnNum);
+this.inputColumn = inputColumn;
+  }
+
+  public FuncLongToDecimal64() {
+super();
+
+// Dummy final assignments.
+inputColumn = -1;
+  }
+
+  abstract protected void func(Decimal64ColumnVector outputColVector, 
LongColumnVector inputColVector, int i,
+  long scaleFactor);
+
+  @Override
+  public void evaluate(VectorizedRowBatch batch) throws HiveException {
+
+if (childExpressions != null) {
+  super.evaluateChildren(batch);
+}
+
+LongColumnVector inputColVector = (LongColumnVector) 
batch.cols[inputColumn];
+int[] sel = batch.selected;
+int n = batch.size;
+Decimal64ColumnVector outputColVector = (Decimal64ColumnVector) 
batch.cols[outputColumnNum];
+int outputScale = ((DecimalTypeInfo) outputTypeInfo).scale();
+final long scaleFactor = powerOfTenTable[outputScale];
 
 Review comment:
   Is outputScale always guaranteed to be between [0, 18]?
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 377908)
Time Spent: 20m  (was: 10m)

> Support cast to decimal64 in Vectorization
> --
>
> Key: HIVE-22515
> URL: https://issues.apache.org/jira/browse/HIVE-22515
> Project: Hive
>  Issue Type: Bug
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22515.5.patch, HIVE-22515.8.patch, 
> HIVE-22515.9.patch
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>

[jira] [Updated] (HIVE-22785) Update/delete/merge statements not optimized through CBO

2020-01-27 Thread Jesus Camacho Rodriguez (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22785?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-22785:
---
Description: 
Currently, CBO is bypassed for update/delete/merge statements.

To support optimizing these statements through CBO, we need to complete three 
main tasks: 1) support for sort in Calcite planner, 2) support for SORT in AST 
converter, and 3) {{RewriteSemanticAnalyzer}} should extend {{CalcitePlanner}} 
instead of {{SemanticAnalyzer}}.

  was:
Currently, CBO is bypassed for update/delete/merge statements.

To support optimizing these statements through CBO, we need to complete three 
main tasks: 1) support for sort in CBO, 2) support for SORT in AST converter, 
and 3) {{RewriteSemanticAnalyzer}} should extend {{CalcitePlanner}} instead of 
{{SemanticAnalyzer}}.


> Update/delete/merge statements not optimized through CBO
> 
>
> Key: HIVE-22785
> URL: https://issues.apache.org/jira/browse/HIVE-22785
> Project: Hive
>  Issue Type: Improvement
>  Components: CBO
>Reporter: Jesus Camacho Rodriguez
>Priority: Critical
>
> Currently, CBO is bypassed for update/delete/merge statements.
> To support optimizing these statements through CBO, we need to complete three 
> main tasks: 1) support for sort in Calcite planner, 2) support for SORT in 
> AST converter, and 3) {{RewriteSemanticAnalyzer}} should extend 
> {{CalcitePlanner}} instead of {{SemanticAnalyzer}}.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22784) Boundary condition to check if there is nothing to truncate in StringExpr functions

2020-01-27 Thread Ramesh Kumar Thangarajan (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22784?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ramesh Kumar Thangarajan updated HIVE-22784:

Attachment: HIVE-22784.1.patch
Status: Patch Available  (was: Open)

> Boundary condition to check if there is nothing to truncate in StringExpr 
> functions
> ---
>
> Key: HIVE-22784
> URL: https://issues.apache.org/jira/browse/HIVE-22784
> Project: Hive
>  Issue Type: Bug
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
> Attachments: HIVE-22784.1.patch
>
>
> Add a boundary condition to check if there is nothing to truncate in 
> StringExpr functions
> truncate()
> rightTrimAndTruncate()



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22753) Fix gradual mem leak: Operationlog related appenders should be cleared up on errors

2020-01-27 Thread Rajesh Balamohan (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22753?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024716#comment-17024716
 ] 

Rajesh Balamohan commented on HIVE-22753:
-

Fixed the comment. Config for this would be an overkill. 

> Fix gradual mem leak: Operationlog related appenders should be cleared up on 
> errors 
> 
>
> Key: HIVE-22753
> URL: https://issues.apache.org/jira/browse/HIVE-22753
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2
>Reporter: Rajesh Balamohan
>Assignee: Rajesh Balamohan
>Priority: Minor
> Attachments: HIVE-22753.1.patch, HIVE-22753.2.patch, 
> HIVE-22753.3.patch, HIVE-22753.4.patch, image-2020-01-21-11-14-37-911.png, 
> image-2020-01-21-11-17-59-279.png, image-2020-01-21-11-18-37-294.png
>
>
> In case of exception in SQLOperation, operational log does not get cleared 
> up. This causes gradual build up of HushableRandomAccessFileAppender causing 
> HS2 to OOM after some time.
> !image-2020-01-21-11-14-37-911.png|width=431,height=267!
>  
> Allocation tree
> !image-2020-01-21-11-18-37-294.png|width=425,height=178!
>  
> Prod instance mem
> !image-2020-01-21-11-17-59-279.png|width=698,height=209!
>  
> Each HushableRandomAccessFileAppender holds internal ref to 
> RandomAccessFileAppender which holds a 256 KB bytebuffer, causing the mem 
> leak.
> Related ticket: HIVE-18820



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22753) Fix gradual mem leak: Operationlog related appenders should be cleared up on errors

2020-01-27 Thread Rajesh Balamohan (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22753?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rajesh Balamohan updated HIVE-22753:

Attachment: HIVE-22753.4.patch

> Fix gradual mem leak: Operationlog related appenders should be cleared up on 
> errors 
> 
>
> Key: HIVE-22753
> URL: https://issues.apache.org/jira/browse/HIVE-22753
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2
>Reporter: Rajesh Balamohan
>Assignee: Rajesh Balamohan
>Priority: Minor
> Attachments: HIVE-22753.1.patch, HIVE-22753.2.patch, 
> HIVE-22753.3.patch, HIVE-22753.4.patch, image-2020-01-21-11-14-37-911.png, 
> image-2020-01-21-11-17-59-279.png, image-2020-01-21-11-18-37-294.png
>
>
> In case of exception in SQLOperation, operational log does not get cleared 
> up. This causes gradual build up of HushableRandomAccessFileAppender causing 
> HS2 to OOM after some time.
> !image-2020-01-21-11-14-37-911.png|width=431,height=267!
>  
> Allocation tree
> !image-2020-01-21-11-18-37-294.png|width=425,height=178!
>  
> Prod instance mem
> !image-2020-01-21-11-17-59-279.png|width=698,height=209!
>  
> Each HushableRandomAccessFileAppender holds internal ref to 
> RandomAccessFileAppender which holds a 256 KB bytebuffer, causing the mem 
> leak.
> Related ticket: HIVE-18820



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22779) Hive build fails with Spark 3

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22779?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024708#comment-17024708
 ] 

Hive QA commented on HIVE-22779:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12991935/HIVE-22779.001.patch

{color:red}ERROR:{color} -1 due to build exiting with an error

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/20368/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/20368/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-20368/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hiveptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ date '+%Y-%m-%d %T.%3N'
2020-01-27 22:10:44.783
+ [[ -n /usr/lib/jvm/java-8-openjdk-amd64 ]]
+ export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ export 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'MAVEN_OPTS=-Xmx1g '
+ MAVEN_OPTS='-Xmx1g '
+ cd /data/hiveptest/working/
+ tee /data/hiveptest/logs/PreCommit-HIVE-Build-20368/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ git = \s\v\n ]]
+ [[ git = \g\i\t ]]
+ [[ -z master ]]
+ [[ -d apache-github-source-source ]]
+ [[ ! -d apache-github-source-source/.git ]]
+ [[ ! -d apache-github-source-source ]]
+ date '+%Y-%m-%d %T.%3N'
2020-01-27 22:10:44.787
+ cd apache-github-source-source
+ git fetch origin
+ git reset --hard HEAD
HEAD is now at b7650eb HIVE-22366: Multiple metastore calls for same table and 
constraints during planning (Vineet Garg, reviewed by Jesus Camacho Rodriguez)
+ git clean -f -d
Removing standalone-metastore/metastore-server/src/gen/
+ git checkout master
Already on 'master'
Your branch is up-to-date with 'origin/master'.
+ git reset --hard origin/master
HEAD is now at b7650eb HIVE-22366: Multiple metastore calls for same table and 
constraints during planning (Vineet Garg, reviewed by Jesus Camacho Rodriguez)
+ git merge --ff-only origin/master
Already up-to-date.
+ date '+%Y-%m-%d %T.%3N'
2020-01-27 22:10:46.008
+ rm -rf ../yetus_PreCommit-HIVE-Build-20368
+ mkdir ../yetus_PreCommit-HIVE-Build-20368
+ git gc
+ cp -R . ../yetus_PreCommit-HIVE-Build-20368
+ mkdir /data/hiveptest/logs/PreCommit-HIVE-Build-20368/yetus
+ patchCommandPath=/data/hiveptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hiveptest/working/scratch/build.patch
+ [[ -f /data/hiveptest/working/scratch/build.patch ]]
+ chmod +x /data/hiveptest/working/scratch/smart-apply-patch.sh
+ /data/hiveptest/working/scratch/smart-apply-patch.sh 
/data/hiveptest/working/scratch/build.patch
Trying to apply the patch with -p0
error: a/ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUtils.java: does 
not exist in index
error: 
a/spark-client/src/main/java/org/apache/hive/spark/client/metrics/ShuffleWriteMetrics.java:
 does not exist in index
error: 
a/spark-client/src/main/java/org/apache/hive/spark/counter/SparkCounter.java: 
does not exist in index
Trying to apply the patch with -p1
error: patch failed: 
ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUtils.java:26
Falling back to three-way merge...
Applied patch to 
'ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUtils.java' with 
conflicts.
Going to apply patch with: git apply -p1
error: patch failed: 
ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUtils.java:26
Falling back to three-way merge...
Applied patch to 
'ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUtils.java' with 
conflicts.
U ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUtils.java
+ result=1
+ '[' 1 -ne 0 ']'
+ rm -rf yetus_PreCommit-HIVE-Build-20368
+ exit 1
'
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12991935 - PreCommit-HIVE-Build

> Hive build fails with Spark 3
> -
>
> Key: HIVE-22779
> URL: https://issues.apache.org/jira/browse/HIVE-22779
> Project: Hive
>  Issue Type: Bug
>  Components: Spark
>Affects Versions: 3.1.2
>Reporter: Igor Dvorzhak
>Priority: Major
> Attachments: HIVE-22779.001.patch
>
>
> Hive build with Spark 3 fails because of deletion of few deprecated APIs in 
> Spark 3.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22770) Skip interning of MapWork fields for tez during deserialization

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22770?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024707#comment-17024707
 ] 

Hive QA commented on HIVE-22770:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12991934/HIVE-22770.03.patch

{color:red}ERROR:{color} -1 due to build exiting with an error

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/20367/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/20367/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-20367/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Tests exited with: Exception: Patch URL 
https://issues.apache.org/jira/secure/attachment/12991934/HIVE-22770.03.patch 
was found in seen patch url's cache and a test was probably run already on it. 
Aborting...
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12991934 - PreCommit-HIVE-Build

> Skip interning of MapWork fields for tez during deserialization
> ---
>
> Key: HIVE-22770
> URL: https://issues.apache.org/jira/browse/HIVE-22770
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-22770.01.patch, HIVE-22770.01.patch, 
> HIVE-22770.02.patch, HIVE-22770.02.patch, HIVE-22770.03.patch
>
>
> HIVE-19937 introduced some interning logic into mapwork deserialization 
> process, but it's only related to spark, maybe we should skip this for tez, 
> reducing the cpu pressure in tez tasks.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22780) Upgrade slf4j version to 1.7.30

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22780?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024706#comment-17024706
 ] 

Hive QA commented on HIVE-22780:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12991920/HIVE-22780.01.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 17953 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/20366/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/20366/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-20366/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12991920 - PreCommit-HIVE-Build

> Upgrade slf4j version to 1.7.30
> ---
>
> Key: HIVE-22780
> URL: https://issues.apache.org/jira/browse/HIVE-22780
> Project: Hive
>  Issue Type: Improvement
>Reporter: David Lavati
>Assignee: David Lavati
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22780.01.patch
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22780) Upgrade slf4j version to 1.7.30

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22780?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024683#comment-17024683
 ] 

Hive QA commented on HIVE-22780:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
49s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
45s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  8m 
48s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  8m  
9s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
27s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  9m 
36s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  7m 
41s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  7m 
41s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} whitespace {color} | {color:red}  0m  
0s{color} | {color:red} The patch 1 line(s) with tabs. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
7s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  7m 
21s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:red}-1{color} | {color:red} asflicense {color} | {color:red}  0m 
12s{color} | {color:red} The patch generated 3 ASF License warnings. {color} |
| {color:black}{color} | {color:black} {color} | {color:black} 52m 42s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  xml  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-20366/dev-support/hive-personality.sh
 |
| git revision | master / b7650eb |
| Default Java | 1.8.0_111 |
| whitespace | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-20366/yetus/whitespace-tabs.txt
 |
| asflicense | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-20366/yetus/patch-asflicense-problems.txt
 |
| modules | C: storage-api kafka-handler standalone-metastore/metastore-tools . 
itests/qtest itests/qtest-druid testutils/ptest2 U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-20366/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Upgrade slf4j version to 1.7.30
> ---
>
> Key: HIVE-22780
> URL: https://issues.apache.org/jira/browse/HIVE-22780
> Project: Hive
>  Issue Type: Improvement
>Reporter: David Lavati
>Assignee: David Lavati
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22780.01.patch
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (HIVE-22784) Boundary condition to check if there is nothing to truncate in StringExpr functions

2020-01-27 Thread Ramesh Kumar Thangarajan (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22784?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ramesh Kumar Thangarajan reassigned HIVE-22784:
---

Assignee: Ramesh Kumar Thangarajan

> Boundary condition to check if there is nothing to truncate in StringExpr 
> functions
> ---
>
> Key: HIVE-22784
> URL: https://issues.apache.org/jira/browse/HIVE-22784
> Project: Hive
>  Issue Type: Bug
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
>
> Add a boundary condition to check if there is nothing to truncate in 
> StringExpr functions
> truncate()
> rightTrimAndTruncate()



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22538) RS deduplication does not always enforce hive.optimize.reducededuplication.min.reducer

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22538?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024648#comment-17024648
 ] 

Hive QA commented on HIVE-22538:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12991918/HIVE-22538.8.patch

{color:red}ERROR:{color} -1 due to build exiting with an error

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/20365/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/20365/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-20365/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Tests exited with: Exception: Patch URL 
https://issues.apache.org/jira/secure/attachment/12991918/HIVE-22538.8.patch 
was found in seen patch url's cache and a test was probably run already on it. 
Aborting...
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12991918 - PreCommit-HIVE-Build

> RS deduplication does not always enforce 
> hive.optimize.reducededuplication.min.reducer
> --
>
> Key: HIVE-22538
> URL: https://issues.apache.org/jira/browse/HIVE-22538
> Project: Hive
>  Issue Type: Bug
>  Components: Physical Optimizer
>Reporter: Jesus Camacho Rodriguez
>Assignee: Krisztian Kasa
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22538.2.patch, HIVE-22538.3.patch, 
> HIVE-22538.4.patch, HIVE-22538.5.patch, HIVE-22538.6.patch, 
> HIVE-22538.6.patch, HIVE-22538.7.patch, HIVE-22538.8.patch, HIVE-22538.patch
>
>  Time Spent: 1.5h
>  Remaining Estimate: 0h
>
> For transactional tables, that property might be overriden to 1, which can 
> lead to merging final aggregation into a single stage (hence leading to 
> performance degradation). For instance, when autogather column stats is 
> enabled, this can happen for the following query:
> {code}
> set hive.support.concurrency=true;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
> EXPLAIN
> CREATE TABLE x STORED AS ORC TBLPROPERTIES('transactional'='true') AS
> SELECT * FROM SRC x CLUSTER BY x.key;
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22770) Skip interning of MapWork fields for tez during deserialization

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22770?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024625#comment-17024625
 ] 

Hive QA commented on HIVE-22770:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
44s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
6s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
36s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m 
55s{color} | {color:blue} ql in master has 1538 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
3s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
41s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
8s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m  
8s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
41s{color} | {color:red} ql: The patch generated 1 new + 162 unchanged - 0 
fixed = 163 total (was 162) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  4m  
7s{color} | {color:red} ql generated 1 new + 1538 unchanged - 0 fixed = 1539 
total (was 1538) {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
2s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
16s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 24m 51s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| FindBugs | module:ql |
|  |  Unchecked/unconfirmed cast from com.esotericsoftware.kryo.Kryo to 
org.apache.hadoop.hive.ql.exec.SerializationUtilities$KryoWithHooks in 
org.apache.hadoop.hive.ql.exec.SerializationUtilities.releaseKryo(Kryo)  At 
SerializationUtilities.java:org.apache.hadoop.hive.ql.exec.SerializationUtilities$KryoWithHooks
 in org.apache.hadoop.hive.ql.exec.SerializationUtilities.releaseKryo(Kryo)  At 
SerializationUtilities.java:[line 307] |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-20364/dev-support/hive-personality.sh
 |
| git revision | master / b7650eb |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.1 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-20364/yetus/diff-checkstyle-ql.txt
 |
| findbugs | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-20364/yetus/new-findbugs-ql.html
 |
| modules | C: ql U: ql |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-20364/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Skip interning of MapWork fields for tez during deserialization
> ---
>
> Key: HIVE-22770
> URL: https://issues.apache.org/jira/browse/HIVE-22770
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-22770.01.patch, HIVE-22770.01.patch, 
> HIVE-22770.02.patch, HIVE-22770.02.patch, HIVE-22770.03.patch
>
>
> HIVE-19937 introduced some interning logic into mapwork deserialization 
> process, but it's only related to spark, maybe we should skip this for tez, 
> reducing the cpu pressure in tez tasks.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22538) RS deduplication does not always enforce hive.optimize.reducededuplication.min.reducer

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22538?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024605#comment-17024605
 ] 

Hive QA commented on HIVE-22538:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12991918/HIVE-22538.8.patch

{color:green}SUCCESS:{color} +1 due to 2 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 17954 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestTezPerfConstraintsCliDriver.testCliDriver[cbo_query25]
 (batchId=303)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/20362/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/20362/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-20362/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12991918 - PreCommit-HIVE-Build

> RS deduplication does not always enforce 
> hive.optimize.reducededuplication.min.reducer
> --
>
> Key: HIVE-22538
> URL: https://issues.apache.org/jira/browse/HIVE-22538
> Project: Hive
>  Issue Type: Bug
>  Components: Physical Optimizer
>Reporter: Jesus Camacho Rodriguez
>Assignee: Krisztian Kasa
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22538.2.patch, HIVE-22538.3.patch, 
> HIVE-22538.4.patch, HIVE-22538.5.patch, HIVE-22538.6.patch, 
> HIVE-22538.6.patch, HIVE-22538.7.patch, HIVE-22538.8.patch, HIVE-22538.patch
>
>  Time Spent: 1.5h
>  Remaining Estimate: 0h
>
> For transactional tables, that property might be overriden to 1, which can 
> lead to merging final aggregation into a single stage (hence leading to 
> performance degradation). For instance, when autogather column stats is 
> enabled, this can happen for the following query:
> {code}
> set hive.support.concurrency=true;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
> EXPLAIN
> CREATE TABLE x STORED AS ORC TBLPROPERTIES('transactional'='true') AS
> SELECT * FROM SRC x CLUSTER BY x.key;
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (HIVE-22783) Add test for HIVE-22366

2020-01-27 Thread Vineet Garg (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22783?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg reassigned HIVE-22783:
--


> Add test for HIVE-22366
> ---
>
> Key: HIVE-22783
> URL: https://issues.apache.org/jira/browse/HIVE-22783
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Vineet Garg
>Assignee: Vineet Garg
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22366) Multiple metastore calls for same table and constraints during planning

2020-01-27 Thread Vineet Garg (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22366?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-22366:
---
Fix Version/s: 4.0.0

> Multiple metastore calls for same table and constraints during planning
> ---
>
> Key: HIVE-22366
> URL: https://issues.apache.org/jira/browse/HIVE-22366
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Planning
>Reporter: Vineet Garg
>Assignee: Vineet Garg
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0
>
> Attachments: HIVE-22366.1.patch, HIVE-22366.2.patch, 
> HIVE-22366.3.patch, HIVE-22366.4.patch, HIVE-22366.5.patch, 
> HIVE-22366.6.patch, HIVE-22366.7.patch, HIVE-22366.8.patch
>
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>
> e.g. for a simple query like
> {code:sql}
> explain select count(i) from test2 group by j limit 3;
> {code}
> planner makes the following calls:
> {code}
> 2019-10-17T22:44:49,892  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  cmd=get_table : 
> tbl=hive.default.test2
> 2019-10-17T22:44:49,908  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_primary_keys : tbl=hive.default.test2
> 2019-10-17T22:44:49,910  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_foreign_keys : parentdb=null parenttbl=null foreigndb=default 
> foreigntbl=test2
> 2019-10-17T22:44:49,911  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_unique_constraints : tbl=hive.default.test2
> 2019-10-17T22:44:49,913  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_not_null_constraints : tbl=hive.default.test2
> 2019-10-17T22:44:49,979  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_partitions : tbl=hive.default.test2
> 2019-10-17T22:44:49,997  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_aggr_stats_for: table=hive.default.test2
> 2019-10-17T22:44:50,014  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  cmd=get_table : 
> tbl=hive.default.test2
> 2019-10-17T22:44:50,023  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_primary_keys : tbl=hive.default.test2
> 2019-10-17T22:44:50,024  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_foreign_keys : parentdb=null parenttbl=null foreigndb=default 
> foreigntbl=test2
> 2019-10-17T22:44:50,025  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_unique_constraints : tbl=hive.default.test2
> 2019-10-17T22:44:50,026  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_not_null_constraints : tbl=hive.default.test2
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22366) Multiple metastore calls for same table and constraints during planning

2020-01-27 Thread Vineet Garg (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22366?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-22366:
---
Resolution: Fixed
Status: Resolved  (was: Patch Available)

Pushed to master, thanks [~jcamachorodriguez]

> Multiple metastore calls for same table and constraints during planning
> ---
>
> Key: HIVE-22366
> URL: https://issues.apache.org/jira/browse/HIVE-22366
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Planning
>Reporter: Vineet Garg
>Assignee: Vineet Garg
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22366.1.patch, HIVE-22366.2.patch, 
> HIVE-22366.3.patch, HIVE-22366.4.patch, HIVE-22366.5.patch, 
> HIVE-22366.6.patch, HIVE-22366.7.patch, HIVE-22366.8.patch
>
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>
> e.g. for a simple query like
> {code:sql}
> explain select count(i) from test2 group by j limit 3;
> {code}
> planner makes the following calls:
> {code}
> 2019-10-17T22:44:49,892  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  cmd=get_table : 
> tbl=hive.default.test2
> 2019-10-17T22:44:49,908  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_primary_keys : tbl=hive.default.test2
> 2019-10-17T22:44:49,910  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_foreign_keys : parentdb=null parenttbl=null foreigndb=default 
> foreigntbl=test2
> 2019-10-17T22:44:49,911  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_unique_constraints : tbl=hive.default.test2
> 2019-10-17T22:44:49,913  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_not_null_constraints : tbl=hive.default.test2
> 2019-10-17T22:44:49,979  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_partitions : tbl=hive.default.test2
> 2019-10-17T22:44:49,997  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_aggr_stats_for: table=hive.default.test2
> 2019-10-17T22:44:50,014  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  cmd=get_table : 
> tbl=hive.default.test2
> 2019-10-17T22:44:50,023  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_primary_keys : tbl=hive.default.test2
> 2019-10-17T22:44:50,024  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_foreign_keys : parentdb=null parenttbl=null foreigndb=default 
> foreigntbl=test2
> 2019-10-17T22:44:50,025  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_unique_constraints : tbl=hive.default.test2
> 2019-10-17T22:44:50,026  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_not_null_constraints : tbl=hive.default.test2
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22538) RS deduplication does not always enforce hive.optimize.reducededuplication.min.reducer

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22538?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024564#comment-17024564
 ] 

Hive QA commented on HIVE-22538:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
48s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
44s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
54s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
 8s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  4m 
10s{color} | {color:blue} ql in master has 1532 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
47s{color} | {color:blue} itests/hive-unit in master has 2 extant Findbugs 
warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
38s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
29s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  2m 
20s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  2m  
1s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  2m  
1s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
49s{color} | {color:green} ql: The patch generated 0 new + 499 unchanged - 4 
fixed = 499 total (was 503) {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
19s{color} | {color:green} itests/hive-unit: The patch generated 0 new + 170 
unchanged - 2 fixed = 170 total (was 172) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  5m 
14s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
25s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 32m 46s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-20362/dev-support/hive-personality.sh
 |
| git revision | master / f7eb8b6 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.1 |
| modules | C: ql itests/hive-unit U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-20362/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> RS deduplication does not always enforce 
> hive.optimize.reducededuplication.min.reducer
> --
>
> Key: HIVE-22538
> URL: https://issues.apache.org/jira/browse/HIVE-22538
> Project: Hive
>  Issue Type: Bug
>  Components: Physical Optimizer
>Reporter: Jesus Camacho Rodriguez
>Assignee: Krisztian Kasa
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22538.2.patch, HIVE-22538.3.patch, 
> HIVE-22538.4.patch, HIVE-22538.5.patch, HIVE-22538.6.patch, 
> HIVE-22538.6.patch, HIVE-22538.7.patch, HIVE-22538.8.patch, HIVE-22538.patch
>
>  Time Spent: 1.5h
>  Remaining Estimate: 0h
>
> For transactional tables, that property might be overriden to 1, which can 
> lead to merging final aggregation into a single stage (hence leading to 
> 

[jira] [Updated] (HIVE-22366) Multiple metastore calls for same table and constraints during planning

2020-01-27 Thread Vineet Garg (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22366?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-22366:
---
Attachment: HIVE-22366.8.patch

> Multiple metastore calls for same table and constraints during planning
> ---
>
> Key: HIVE-22366
> URL: https://issues.apache.org/jira/browse/HIVE-22366
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Planning
>Reporter: Vineet Garg
>Assignee: Vineet Garg
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22366.1.patch, HIVE-22366.2.patch, 
> HIVE-22366.3.patch, HIVE-22366.4.patch, HIVE-22366.5.patch, 
> HIVE-22366.6.patch, HIVE-22366.7.patch, HIVE-22366.8.patch
>
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>
> e.g. for a simple query like
> {code:sql}
> explain select count(i) from test2 group by j limit 3;
> {code}
> planner makes the following calls:
> {code}
> 2019-10-17T22:44:49,892  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  cmd=get_table : 
> tbl=hive.default.test2
> 2019-10-17T22:44:49,908  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_primary_keys : tbl=hive.default.test2
> 2019-10-17T22:44:49,910  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_foreign_keys : parentdb=null parenttbl=null foreigndb=default 
> foreigntbl=test2
> 2019-10-17T22:44:49,911  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_unique_constraints : tbl=hive.default.test2
> 2019-10-17T22:44:49,913  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_not_null_constraints : tbl=hive.default.test2
> 2019-10-17T22:44:49,979  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_partitions : tbl=hive.default.test2
> 2019-10-17T22:44:49,997  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_aggr_stats_for: table=hive.default.test2
> 2019-10-17T22:44:50,014  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  cmd=get_table : 
> tbl=hive.default.test2
> 2019-10-17T22:44:50,023  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_primary_keys : tbl=hive.default.test2
> 2019-10-17T22:44:50,024  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_foreign_keys : parentdb=null parenttbl=null foreigndb=default 
> foreigntbl=test2
> 2019-10-17T22:44:50,025  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_unique_constraints : tbl=hive.default.test2
> 2019-10-17T22:44:50,026  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_not_null_constraints : tbl=hive.default.test2
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (HIVE-22782) Consolidate metastore call to fetch constraints

2020-01-27 Thread Vineet Garg (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22782?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg reassigned HIVE-22782:
--


> Consolidate metastore call to fetch constraints
> ---
>
> Key: HIVE-22782
> URL: https://issues.apache.org/jira/browse/HIVE-22782
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Planning
>Reporter: Vineet Garg
>Assignee: Vineet Garg
>Priority: Major
>
> Currently separate calls are made to metastore to fetch constraints like Pk, 
> fk, not null etc. Since planner always retrieve these constraints we should 
> retrieve all of them in one call.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-22366) Multiple metastore calls for same table and constraints during planning

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22366?focusedWorklogId=377798=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-377798
 ]

ASF GitHub Bot logged work on HIVE-22366:
-

Author: ASF GitHub Bot
Created on: 27/Jan/20 18:38
Start Date: 27/Jan/20 18:38
Worklog Time Spent: 10m 
  Work Description: vineetgarg02 commented on pull request #888: 
HIVE-22366: Multiple metastore calls for same table and constraints d…
URL: https://github.com/apache/hive/pull/888#discussion_r371412106
 
 

 ##
 File path: ql/src/java/org/apache/hadoop/hive/ql/metadata/Table.java
 ##
 @@ -1126,71 +1149,167 @@ public Boolean isOutdatedForRewriting() {
 return outdatedForRewritingMaterializedView;
   }
 
-  /* These are only populated during optimization and describing */
+  public ColumnStatistics getColStats() {
+return tTable.isSetColStats() ? tTable.getColStats() : null;
+  }
+
+  /**
+   * Setup the table level stats as if the table is new. Used when setting up 
Table for a new
+   * table or during replication.
+   */
+  public void setStatsStateLikeNewTable() {
+if (isPartitioned()) {
+  StatsSetupConst.setStatsStateForCreateTable(getParameters(), null,
+  StatsSetupConst.FALSE);
+} else {
+  StatsSetupConst.setStatsStateForCreateTable(getParameters(),
+  MetaStoreUtils.getColumnNames(getCols()), StatsSetupConst.TRUE);
+}
+  }
+
+  /** Constraints related methods
+   *  Note that set apis are used by DESCRIBE only, although get apis return 
RELY or ENABLE
+   *  constraints DESCRIBE could set all type of constraints
+   * */
+
+  /* This only return PK with RELY */
   public PrimaryKeyInfo getPrimaryKeyInfo() {
+if(!this.isPKFetched) {
+  try {
+pki = Hive.get().getReliablePrimaryKeys(this.getDbName(), 
this.getTableName());
+this.isPKFetched = true;
+  } catch (HiveException e) {
+LOG.warn("Cannot retrieve PK info for table : " + this.getTableName()
++ " ignoring exception: " + e);
+  }
+}
 return pki;
   }
 
   public void setPrimaryKeyInfo(PrimaryKeyInfo pki) {
 this.pki = pki;
+this.isPKFetched = true;
   }
 
+  /* This only return FK with RELY */
   public ForeignKeyInfo getForeignKeyInfo() {
 
 Review comment:
   Created https://issues.apache.org/jira/browse/HIVE-22782 for follow-up
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 377798)
Time Spent: 0.5h  (was: 20m)

> Multiple metastore calls for same table and constraints during planning
> ---
>
> Key: HIVE-22366
> URL: https://issues.apache.org/jira/browse/HIVE-22366
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Planning
>Reporter: Vineet Garg
>Assignee: Vineet Garg
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22366.1.patch, HIVE-22366.2.patch, 
> HIVE-22366.3.patch, HIVE-22366.4.patch, HIVE-22366.5.patch, 
> HIVE-22366.6.patch, HIVE-22366.7.patch
>
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>
> e.g. for a simple query like
> {code:sql}
> explain select count(i) from test2 group by j limit 3;
> {code}
> planner makes the following calls:
> {code}
> 2019-10-17T22:44:49,892  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  cmd=get_table : 
> tbl=hive.default.test2
> 2019-10-17T22:44:49,908  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_primary_keys : tbl=hive.default.test2
> 2019-10-17T22:44:49,910  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_foreign_keys : parentdb=null parenttbl=null foreigndb=default 
> foreigntbl=test2
> 2019-10-17T22:44:49,911  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_unique_constraints : tbl=hive.default.test2
> 2019-10-17T22:44:49,913  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_not_null_constraints : tbl=hive.default.test2
> 2019-10-17T22:44:49,979  INFO [812910d4-bc6a-450c-aa8b-3487b4edceab main] 
> HiveMetaStore.audit: ugi=vgarg  ip=unknown-ip-addr  
> cmd=get_partitions : tbl=hive.default.test2
> 2019-10-17T22:44:49,997  INFO 

[jira] [Commented] (HIVE-22765) Quote table names in CompactionTxnHandler

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22765?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024536#comment-17024536
 ] 

Hive QA commented on HIVE-22765:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12991910/HIVE-22765.3.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 17953 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/20361/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/20361/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-20361/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12991910 - PreCommit-HIVE-Build

> Quote table names in CompactionTxnHandler
> -
>
> Key: HIVE-22765
> URL: https://issues.apache.org/jira/browse/HIVE-22765
> Project: Hive
>  Issue Type: Improvement
>Affects Versions: 4.0.0
>Reporter: Laszlo Pinter
>Assignee: Zoltan Chovan
>Priority: Major
> Attachments: HIVE-22765.2.patch, HIVE-22765.3.patch, HIVE-22765.patch
>
>
> When I run a metastore backed up by a postgres db, I get the following 
> exception: 
> {code:java}
> MetaException(message:Unable to connect to transaction database 
> org.postgresql.util.PSQLException: ERROR: relation "compaction_queue" does 
> not existMetaException(message:Unable to connect to transaction database 
> org.postgresql.util.PSQLException: ERROR: relation "compaction_queue" does 
> not exist  Position: 83 at 
> org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2468)
>  at 
> org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:2211)
>  at 
> org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:309) 
> at org.postgresql.jdbc.PgStatement.executeInternal(PgStatement.java:446) at 
> org.postgresql.jdbc.PgStatement.execute(PgStatement.java:370) at 
> org.postgresql.jdbc.PgStatement.executeWithFlags(PgStatement.java:311) at 
> org.postgresql.jdbc.PgStatement.executeCachedSql(PgStatement.java:297) at 
> org.postgresql.jdbc.PgStatement.executeWithFlags(PgStatement.java:274) at 
> org.postgresql.jdbc.PgStatement.executeQuery(PgStatement.java:225) at 
> org.apache.hive.com.zaxxer.hikari.pool.ProxyStatement.executeQuery(ProxyStatement.java:108)
>  at 
> org.apache.hive.com.zaxxer.hikari.pool.HikariProxyStatement.executeQuery(HikariProxyStatement.java)
>  at 
> org.apache.hadoop.hive.metastore.txn.CompactionTxnHandler.findNextToCompact(CompactionTxnHandler.java:157)
>  at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.find_next_compact(HiveMetaStore.java:8061)
>  at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source) at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>  at java.lang.reflect.Method.invoke(Method.java:498) at 
> org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:147)
>  at 
> org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:108)
>  at com.sun.proxy.$Proxy28.find_next_compact(Unknown Source) at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.findNextCompact(HiveMetaStoreClient.java:4046)
>  at sun.reflect.GeneratedMethodAccessor8.invoke(Unknown Source) at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>  at java.lang.reflect.Method.invoke(Method.java:498) at 
> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:208)
>  at com.sun.proxy.$Proxy29.findNextCompact(Unknown Source) at 
> org.apache.hadoop.hive.ql.txn.compactor.Worker.run(Worker.java:102){code}
> This is due to missing quotes around table names. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22779) Hive build fails with Spark 3

2020-01-27 Thread Igor Dvorzhak (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22779?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Igor Dvorzhak updated HIVE-22779:
-
  Attachment: HIVE-22779.001.patch
Target Version/s:   (was: 3.1.3)
  Status: Patch Available  (was: Open)

> Hive build fails with Spark 3
> -
>
> Key: HIVE-22779
> URL: https://issues.apache.org/jira/browse/HIVE-22779
> Project: Hive
>  Issue Type: Bug
>  Components: Spark
>Affects Versions: 3.1.2
>Reporter: Igor Dvorzhak
>Priority: Major
> Attachments: HIVE-22779.001.patch
>
>
> Hive build with Spark 3 fails because of deletion of few deprecated APIs in 
> Spark 3.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22779) Hive build fails with Spark 3

2020-01-27 Thread Igor Dvorzhak (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22779?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Igor Dvorzhak updated HIVE-22779:
-
Status: Open  (was: Patch Available)

> Hive build fails with Spark 3
> -
>
> Key: HIVE-22779
> URL: https://issues.apache.org/jira/browse/HIVE-22779
> Project: Hive
>  Issue Type: Bug
>  Components: Spark
>Affects Versions: 3.1.2
>Reporter: Igor Dvorzhak
>Priority: Major
>
> Hive build with Spark 3 fails because of deletion of few deprecated APIs in 
> Spark 3.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22779) Hive build fails with Spark 3

2020-01-27 Thread Igor Dvorzhak (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22779?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Igor Dvorzhak updated HIVE-22779:
-
Attachment: (was: HIVE-22779.patch)

> Hive build fails with Spark 3
> -
>
> Key: HIVE-22779
> URL: https://issues.apache.org/jira/browse/HIVE-22779
> Project: Hive
>  Issue Type: Bug
>  Components: Spark
>Affects Versions: 3.1.2
>Reporter: Igor Dvorzhak
>Priority: Major
>
> Hive build with Spark 3 fails because of deletion of few deprecated APIs in 
> Spark 3.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22770) Skip interning of MapWork fields for tez during deserialization

2020-01-27 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-22770?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor updated HIVE-22770:

Attachment: HIVE-22770.03.patch

> Skip interning of MapWork fields for tez during deserialization
> ---
>
> Key: HIVE-22770
> URL: https://issues.apache.org/jira/browse/HIVE-22770
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-22770.01.patch, HIVE-22770.01.patch, 
> HIVE-22770.02.patch, HIVE-22770.02.patch, HIVE-22770.03.patch
>
>
> HIVE-19937 introduced some interning logic into mapwork deserialization 
> process, but it's only related to spark, maybe we should skip this for tez, 
> reducing the cpu pressure in tez tasks.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-22746) Make TypeCheckProcFactory generic

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22746?focusedWorklogId=377740=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-377740
 ]

ASF GitHub Bot logged work on HIVE-22746:
-

Author: ASF GitHub Bot
Created on: 27/Jan/20 17:04
Start Date: 27/Jan/20 17:04
Worklog Time Spent: 10m 
  Work Description: asfgit commented on pull request #887: HIVE-22746
URL: https://github.com/apache/hive/pull/887
 
 
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 377740)
Time Spent: 1h  (was: 50m)

> Make TypeCheckProcFactory generic
> -
>
> Key: HIVE-22746
> URL: https://issues.apache.org/jira/browse/HIVE-22746
> Project: Hive
>  Issue Type: Improvement
>  Components: CBO
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0
>
> Attachments: HIVE-22746.01.patch, HIVE-22746.01.patch, 
> HIVE-22746.02.patch, HIVE-22746.patch
>
>  Time Spent: 1h
>  Remaining Estimate: 0h
>
> {{TypeCheckProcFactory}} is responsible for processing AST nodes and 
> generating ExprNode objects from them. When we generate the expressions for 
> Calcite planning, we go through a {{AST node -> ExprNode -> RexNode}} 
> transformation. We would like to avoid the overhead of going through the 
> ExprNode, and thus generate directly the RexNode from the AST.
> To do that, the first step is to make {{TypeCheckProcFactory}} generic, so it 
> can receive an expression factory and create expressions in different realms. 
> For the time being, the only factory implementation is the ExprNode factory. 
> Thus, this patch focuses mainly on refactoring {{TypeCheckProcFactory}} 
> without breaking anything that is already working.
> In a follow-up patch, we will create a {{RexNode}} factory and use it when we 
> parse the query in CalcitePlanner.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22746) Make TypeCheckProcFactory generic

2020-01-27 Thread Jesus Camacho Rodriguez (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22746?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-22746:
---
Fix Version/s: 4.0.0
   Resolution: Fixed
   Status: Resolved  (was: Patch Available)

Pushed to master, thanks [~mgergely], [~vgarg].

> Make TypeCheckProcFactory generic
> -
>
> Key: HIVE-22746
> URL: https://issues.apache.org/jira/browse/HIVE-22746
> Project: Hive
>  Issue Type: Improvement
>  Components: CBO
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0
>
> Attachments: HIVE-22746.01.patch, HIVE-22746.01.patch, 
> HIVE-22746.02.patch, HIVE-22746.patch
>
>  Time Spent: 1h
>  Remaining Estimate: 0h
>
> {{TypeCheckProcFactory}} is responsible for processing AST nodes and 
> generating ExprNode objects from them. When we generate the expressions for 
> Calcite planning, we go through a {{AST node -> ExprNode -> RexNode}} 
> transformation. We would like to avoid the overhead of going through the 
> ExprNode, and thus generate directly the RexNode from the AST.
> To do that, the first step is to make {{TypeCheckProcFactory}} generic, so it 
> can receive an expression factory and create expressions in different realms. 
> For the time being, the only factory implementation is the ExprNode factory. 
> Thus, this patch focuses mainly on refactoring {{TypeCheckProcFactory}} 
> without breaking anything that is already working.
> In a follow-up patch, we will create a {{RexNode}} factory and use it when we 
> parse the query in CalcitePlanner.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22538) RS deduplication does not always enforce hive.optimize.reducededuplication.min.reducer

2020-01-27 Thread Jesus Camacho Rodriguez (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22538?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024504#comment-17024504
 ] 

Jesus Camacho Rodriguez commented on HIVE-22538:


+1 (pending tests)

> RS deduplication does not always enforce 
> hive.optimize.reducededuplication.min.reducer
> --
>
> Key: HIVE-22538
> URL: https://issues.apache.org/jira/browse/HIVE-22538
> Project: Hive
>  Issue Type: Bug
>  Components: Physical Optimizer
>Reporter: Jesus Camacho Rodriguez
>Assignee: Krisztian Kasa
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22538.2.patch, HIVE-22538.3.patch, 
> HIVE-22538.4.patch, HIVE-22538.5.patch, HIVE-22538.6.patch, 
> HIVE-22538.6.patch, HIVE-22538.7.patch, HIVE-22538.8.patch, HIVE-22538.patch
>
>  Time Spent: 1.5h
>  Remaining Estimate: 0h
>
> For transactional tables, that property might be overriden to 1, which can 
> lead to merging final aggregation into a single stage (hence leading to 
> performance degradation). For instance, when autogather column stats is 
> enabled, this can happen for the following query:
> {code}
> set hive.support.concurrency=true;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
> EXPLAIN
> CREATE TABLE x STORED AS ORC TBLPROPERTIES('transactional'='true') AS
> SELECT * FROM SRC x CLUSTER BY x.key;
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22765) Quote table names in CompactionTxnHandler

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22765?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024503#comment-17024503
 ] 

Hive QA commented on HIVE-22765:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  9m 
28s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
26s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
17s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  1m 
14s{color} | {color:blue} standalone-metastore/metastore-server in master has 
181 extant Findbugs warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
21s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
32s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
26s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
26s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
18s{color} | {color:red} standalone-metastore/metastore-server: The patch 
generated 24 new + 55 unchanged - 25 fixed = 79 total (was 80) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
20s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
21s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 15m 31s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-20361/dev-support/hive-personality.sh
 |
| git revision | master / 8dec57c |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-20361/yetus/diff-checkstyle-standalone-metastore_metastore-server.txt
 |
| modules | C: standalone-metastore/metastore-server U: 
standalone-metastore/metastore-server |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-20361/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Quote table names in CompactionTxnHandler
> -
>
> Key: HIVE-22765
> URL: https://issues.apache.org/jira/browse/HIVE-22765
> Project: Hive
>  Issue Type: Improvement
>Affects Versions: 4.0.0
>Reporter: Laszlo Pinter
>Assignee: Zoltan Chovan
>Priority: Major
> Attachments: HIVE-22765.2.patch, HIVE-22765.3.patch, HIVE-22765.patch
>
>
> When I run a metastore backed up by a postgres db, I get the following 
> exception: 
> {code:java}
> MetaException(message:Unable to connect to transaction database 
> org.postgresql.util.PSQLException: ERROR: relation "compaction_queue" does 
> not existMetaException(message:Unable to connect to transaction database 
> org.postgresql.util.PSQLException: ERROR: relation "compaction_queue" does 
> not exist  Position: 83 at 
> org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2468)
>  at 
> org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:2211)
>  at 
> org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:309) 
> at org.postgresql.jdbc.PgStatement.executeInternal(PgStatement.java:446) at 
> org.postgresql.jdbc.PgStatement.execute(PgStatement.java:370) at 
> 

[jira] [Commented] (HIVE-22753) Fix gradual mem leak: Operationlog related appenders should be cleared up on errors

2020-01-27 Thread Zoltan Chovan (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22753?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024486#comment-17024486
 ] 

Zoltan Chovan commented on HIVE-22753:
--

[~rajesh.balamohan] the code comments state that the elements will have 5s ttl, 
but this line shows only 1s:

.expireAfterWrite(1, TimeUnit.SECONDS)

Also do you think having a config value for this would be overkill? Would more 
time be needed potentially, in case there's a heavy IO load for some reason on 
the host of HS2?

> Fix gradual mem leak: Operationlog related appenders should be cleared up on 
> errors 
> 
>
> Key: HIVE-22753
> URL: https://issues.apache.org/jira/browse/HIVE-22753
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2
>Reporter: Rajesh Balamohan
>Assignee: Rajesh Balamohan
>Priority: Minor
> Attachments: HIVE-22753.1.patch, HIVE-22753.2.patch, 
> HIVE-22753.3.patch, image-2020-01-21-11-14-37-911.png, 
> image-2020-01-21-11-17-59-279.png, image-2020-01-21-11-18-37-294.png
>
>
> In case of exception in SQLOperation, operational log does not get cleared 
> up. This causes gradual build up of HushableRandomAccessFileAppender causing 
> HS2 to OOM after some time.
> !image-2020-01-21-11-14-37-911.png|width=431,height=267!
>  
> Allocation tree
> !image-2020-01-21-11-18-37-294.png|width=425,height=178!
>  
> Prod instance mem
> !image-2020-01-21-11-17-59-279.png|width=698,height=209!
>  
> Each HushableRandomAccessFileAppender holds internal ref to 
> RandomAccessFileAppender which holds a 256 KB bytebuffer, causing the mem 
> leak.
> Related ticket: HIVE-18820



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-10362) Support Type check/conversion in dynamic partition column

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-10362?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024485#comment-17024485
 ] 

Hive QA commented on HIVE-10362:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12991909/HIVE-10362.06.patch

{color:red}ERROR:{color} -1 due to build exiting with an error

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/20360/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/20360/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-20360/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Tests exited with: Exception: Patch URL 
https://issues.apache.org/jira/secure/attachment/12991909/HIVE-10362.06.patch 
was found in seen patch url's cache and a test was probably run already on it. 
Aborting...
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12991909 - PreCommit-HIVE-Build

> Support Type check/conversion in dynamic partition column
> -
>
> Key: HIVE-10362
> URL: https://issues.apache.org/jira/browse/HIVE-10362
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Processor, Types
>Affects Versions: 1.0.0, 3.0.0, 4.0.0
>Reporter: Chaoyu Tang
>Assignee: Karen Coppage
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-10362.01.patch, HIVE-10362.02.patch, 
> HIVE-10362.03.patch, HIVE-10362.04.patch, HIVE-10362.04.patch, 
> HIVE-10362.06.patch
>
>
> There are quite a lot of issues associated with the non-noramlized or 
> type-mismatched values for partition column. Hive has many ways to introduce 
> such problematic data. HIVE-10307 mainly provides the support to type 
> check/convert/normalize the partition column value in static partition 
> specification. This JIRA tries to deal with the partition column type  in 
> dynamic partition insert. Currently any data can be inserted as a partition 
> column value as long as it is quoted as a string. For example,
> create table dynparttypechecknum (key int, value string) partitioned by (part 
> int);
> insert into dynparttypechecknum partition (part) select key, value, '1' 
> from src limit 1;
> show partitions dynparttypechecknum;
> --
> part=1
> The partition column value is non-normalized int 1. It causes some 
> unnecessary problems such as integer partition column JDO filter pushdown 
> (see HIVE-6052) and others like HIVE-10210.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-10362) Support Type check/conversion in dynamic partition column

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-10362?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024484#comment-17024484
 ] 

Hive QA commented on HIVE-10362:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12991909/HIVE-10362.06.patch

{color:red}ERROR:{color} -1 due to build exiting with an error

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/20359/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/20359/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-20359/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Tests exited with: Exception: Patch URL 
https://issues.apache.org/jira/secure/attachment/12991909/HIVE-10362.06.patch 
was found in seen patch url's cache and a test was probably run already on it. 
Aborting...
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12991909 - PreCommit-HIVE-Build

> Support Type check/conversion in dynamic partition column
> -
>
> Key: HIVE-10362
> URL: https://issues.apache.org/jira/browse/HIVE-10362
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Processor, Types
>Affects Versions: 1.0.0, 3.0.0, 4.0.0
>Reporter: Chaoyu Tang
>Assignee: Karen Coppage
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-10362.01.patch, HIVE-10362.02.patch, 
> HIVE-10362.03.patch, HIVE-10362.04.patch, HIVE-10362.04.patch, 
> HIVE-10362.06.patch
>
>
> There are quite a lot of issues associated with the non-noramlized or 
> type-mismatched values for partition column. Hive has many ways to introduce 
> such problematic data. HIVE-10307 mainly provides the support to type 
> check/convert/normalize the partition column value in static partition 
> specification. This JIRA tries to deal with the partition column type  in 
> dynamic partition insert. Currently any data can be inserted as a partition 
> column value as long as it is quoted as a string. For example,
> create table dynparttypechecknum (key int, value string) partitioned by (part 
> int);
> insert into dynparttypechecknum partition (part) select key, value, '1' 
> from src limit 1;
> show partitions dynparttypechecknum;
> --
> part=1
> The partition column value is non-normalized int 1. It causes some 
> unnecessary problems such as integer partition column JDO filter pushdown 
> (see HIVE-6052) and others like HIVE-10210.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22775) Use the qt:authorizer option in qtests

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22775?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024483#comment-17024483
 ] 

Hive QA commented on HIVE-22775:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12991904/HIVE-22775.02.patch

{color:red}ERROR:{color} -1 due to build exiting with an error

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/20358/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/20358/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-20358/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hiveptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ date '+%Y-%m-%d %T.%3N'
2020-01-27 16:34:55.003
+ [[ -n /usr/lib/jvm/java-8-openjdk-amd64 ]]
+ export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ export 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'MAVEN_OPTS=-Xmx1g '
+ MAVEN_OPTS='-Xmx1g '
+ cd /data/hiveptest/working/
+ tee /data/hiveptest/logs/PreCommit-HIVE-Build-20358/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ git = \s\v\n ]]
+ [[ git = \g\i\t ]]
+ [[ -z master ]]
+ [[ -d apache-github-source-source ]]
+ [[ ! -d apache-github-source-source/.git ]]
+ [[ ! -d apache-github-source-source ]]
+ date '+%Y-%m-%d %T.%3N'
2020-01-27 16:34:55.005
+ cd apache-github-source-source
+ git fetch origin
>From https://github.com/apache/hive
   7bb1d1e..8dec57c  master -> origin/master
+ git reset --hard HEAD
HEAD is now at 7bb1d1e HIVE-22518: SQLStdHiveAuthorizerFactoryForTest doesn't 
work correctly for llap tests (Zoltan Haindrich reviewed by Miklos Gergely)
+ git clean -f -d
Removing standalone-metastore/metastore-server/src/gen/
+ git checkout master
Already on 'master'
Your branch is behind 'origin/master' by 3 commits, and can be fast-forwarded.
  (use "git pull" to update your local branch)
+ git reset --hard origin/master
HEAD is now at 8dec57c HIVE-22679: Replace Base64 in metastore-common Package 
(David Mollitor via Naveen Gangam)
+ git merge --ff-only origin/master
Already up-to-date.
+ date '+%Y-%m-%d %T.%3N'
2020-01-27 16:34:57.416
+ rm -rf ../yetus_PreCommit-HIVE-Build-20358
+ mkdir ../yetus_PreCommit-HIVE-Build-20358
+ git gc
+ cp -R . ../yetus_PreCommit-HIVE-Build-20358
+ mkdir /data/hiveptest/logs/PreCommit-HIVE-Build-20358/yetus
+ patchCommandPath=/data/hiveptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hiveptest/working/scratch/build.patch
+ [[ -f /data/hiveptest/working/scratch/build.patch ]]
+ chmod +x /data/hiveptest/working/scratch/smart-apply-patch.sh
+ /data/hiveptest/working/scratch/smart-apply-patch.sh 
/data/hiveptest/working/scratch/build.patch
Trying to apply the patch with -p0
Going to apply patch with: git apply -p0
/data/hiveptest/working/scratch/build.patch:25: trailing whitespace.

warning: 1 line adds whitespace errors.
+ [[ maven == \m\a\v\e\n ]]
+ rm -rf /data/hiveptest/working/maven/org/apache/hive
+ mvn -B clean install -DskipTests -T 4 -q 
-Dmaven.repo.local=/data/hiveptest/working/maven
protoc-jar: executing: [/tmp/protoc1330264086802028845.exe, --version]
libprotoc 2.5.0
protoc-jar: executing: [/tmp/protoc1330264086802028845.exe, 
-I/data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-common/src/main/protobuf/org/apache/hadoop/hive/metastore,
 
--java_out=/data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-common/target/generated-sources,
 
/data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-common/src/main/protobuf/org/apache/hadoop/hive/metastore/metastore.proto]
ANTLR Parser Generator  Version 3.5.2
[ERROR] Failed to execute goal 
org.apache.maven.plugins:maven-remote-resources-plugin:1.5:process 
(process-resource-bundles) on project hive-shims-0.23: Execution 
process-resource-bundles of goal 
org.apache.maven.plugins:maven-remote-resources-plugin:1.5:process failed. 
ConcurrentModificationException -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please 
read the following articles:
[ERROR] [Help 1] 
http://cwiki.apache.org/confluence/display/MAVEN/PluginExecutionException
[ERROR] 
[ERROR] After correcting the problems, you can resume the 

[jira] [Commented] (HIVE-10362) Support Type check/conversion in dynamic partition column

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-10362?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024478#comment-17024478
 ] 

Hive QA commented on HIVE-10362:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12991909/HIVE-10362.06.patch

{color:green}SUCCESS:{color} +1 due to 2 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 23 failed/errored test(s), 17947 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[autoColumnStats_6] 
(batchId=75)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[dynpart_sort_optimization_acid2]
 (batchId=36)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[infer_bucket_sort_num_buckets]
 (batchId=69)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[orc_merge10] (batchId=73)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[orc_merge1] (batchId=23)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[orc_merge2] (batchId=103)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[orc_merge_diff_fs] 
(batchId=1)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[smb_join_partition_key] 
(batchId=15)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[orc_merge10] 
(batchId=161)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[orc_merge1] 
(batchId=158)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[orc_merge2] 
(batchId=163)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[orc_merge_diff_fs]
 (batchId=158)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[rcfile_merge2] 
(batchId=162)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[auto_sortmerge_join_16]
 (batchId=180)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[dynpart_sort_optimization_acid]
 (batchId=177)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[llap_smb] 
(batchId=191)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[tez_dml] 
(batchId=173)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[auto_sortmerge_join_16]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[infer_bucket_sort_num_buckets]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[orc_merge1]
 (batchId=197)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[orc_merge2]
 (batchId=200)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[orc_merge_diff_fs]
 (batchId=197)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[auto_sortmerge_join_16]
 (batchId=141)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/20357/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/20357/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-20357/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 23 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12991909 - PreCommit-HIVE-Build

> Support Type check/conversion in dynamic partition column
> -
>
> Key: HIVE-10362
> URL: https://issues.apache.org/jira/browse/HIVE-10362
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Processor, Types
>Affects Versions: 1.0.0, 3.0.0, 4.0.0
>Reporter: Chaoyu Tang
>Assignee: Karen Coppage
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-10362.01.patch, HIVE-10362.02.patch, 
> HIVE-10362.03.patch, HIVE-10362.04.patch, HIVE-10362.04.patch, 
> HIVE-10362.06.patch
>
>
> There are quite a lot of issues associated with the non-noramlized or 
> type-mismatched values for partition column. Hive has many ways to introduce 
> such problematic data. HIVE-10307 mainly provides the support to type 
> check/convert/normalize the partition column value in static partition 
> specification. This JIRA tries to deal with the partition column type  in 
> dynamic partition insert. Currently any data can be inserted as a partition 
> column value as long as it is quoted as a string. For example,
> create table dynparttypechecknum (key int, value string) partitioned by (part 
> int);
> insert into dynparttypechecknum partition (part) select key, value, '1' 
> from src limit 1;
> show partitions dynparttypechecknum;
> --
> part=1
> The partition column value is non-normalized int 1. It causes some 
> unnecessary problems such as integer 

[jira] [Comment Edited] (HIVE-22771) Partition location incorrectly formed in FileOutputCommitterContainer

2020-01-27 Thread Shivam (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22771?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17023549#comment-17023549
 ] 

Shivam edited comment on HIVE-22771 at 1/27/20 4:32 PM:


[~mithun] [~jcamachorodriguez] [~kgyrtkirk] Can anyone please review this?

Thanks.


was (Author: shivam-mohan):
[~mithun] [~jcamachorodriguez] Can you please review this?

Thanks.

> Partition location incorrectly formed in FileOutputCommitterContainer
> -
>
> Key: HIVE-22771
> URL: https://issues.apache.org/jira/browse/HIVE-22771
> Project: Hive
>  Issue Type: Bug
>  Components: HCatalog
>Affects Versions: 1.2.1
>Reporter: Shivam
>Assignee: Shivam
>Priority: Critical
>  Labels: pull-request-available
> Attachments: HIVE-22771.patch
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Class _HCatOutputFormat_ in package _org.apache.hive.hcatalog.mapreduce_ uses 
> function _setOutput_ to generate _idHash_ using below statement:
> *+In file org/apache/hive/hcatalog/mapreduce/HCatOutputFormat.java+*
>  *line 116: idHash = String.valueOf(Math.random());*
> The output of idHash can be similar to values like this : 7.145347157239135E-4
>  
> And, in class _FileOutputCommitterContainer_ in package 
> _org.apache.hive.hcatalog.mapreduce;_
> Uses below statement to compute final partition path:
> +*In org/apache/hive/hcatalog/mapreduce/FileOutputCommitterContainer.java*+
> *line 366: String finalLocn = jobLocation.replaceAll(Path.SEPARATOR + 
> SCRATCH_DIR_NAME + "{color:#ff}\\d\\.?
> d+"{color},"");*
> *line 367: partPath = new Path(finalLocn);*
>  
> Regex used here is incorrect, since it will only remove integers after the 
> *SCRATCH_DIR_NAME,* and hence will append  'E-4' (for the above example) in 
> the final partition location. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22753) Fix gradual mem leak: Operationlog related appenders should be cleared up on errors

2020-01-27 Thread Ashutosh Chauhan (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22753?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024471#comment-17024471
 ] 

Ashutosh Chauhan commented on HIVE-22753:
-

+1

> Fix gradual mem leak: Operationlog related appenders should be cleared up on 
> errors 
> 
>
> Key: HIVE-22753
> URL: https://issues.apache.org/jira/browse/HIVE-22753
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2
>Reporter: Rajesh Balamohan
>Assignee: Rajesh Balamohan
>Priority: Minor
> Attachments: HIVE-22753.1.patch, HIVE-22753.2.patch, 
> HIVE-22753.3.patch, image-2020-01-21-11-14-37-911.png, 
> image-2020-01-21-11-17-59-279.png, image-2020-01-21-11-18-37-294.png
>
>
> In case of exception in SQLOperation, operational log does not get cleared 
> up. This causes gradual build up of HushableRandomAccessFileAppender causing 
> HS2 to OOM after some time.
> !image-2020-01-21-11-14-37-911.png|width=431,height=267!
>  
> Allocation tree
> !image-2020-01-21-11-18-37-294.png|width=425,height=178!
>  
> Prod instance mem
> !image-2020-01-21-11-17-59-279.png|width=698,height=209!
>  
> Each HushableRandomAccessFileAppender holds internal ref to 
> RandomAccessFileAppender which holds a 256 KB bytebuffer, causing the mem 
> leak.
> Related ticket: HIVE-18820



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22770) Skip interning of MapWork fields for tez during deserialization

2020-01-27 Thread Ashutosh Chauhan (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22770?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024468#comment-17024468
 ] 

Ashutosh Chauhan commented on HIVE-22770:
-

   Configuration conf = ((KryoWithHooks) kryo).getConf();
+  if (conf != null && 
conf.get(HiveConf.ConfVars.HIVE_EXECUTION_ENGINE.varname, 
"").equalsIgnoreCase("spark")) {
+mapWork.internFields();
+  }
+

Lets remove above code block. With that removed we won't need {{new 
Configuration}} which is an expensive call. That means we will never intern 
fields.

> Skip interning of MapWork fields for tez during deserialization
> ---
>
> Key: HIVE-22770
> URL: https://issues.apache.org/jira/browse/HIVE-22770
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-22770.01.patch, HIVE-22770.01.patch, 
> HIVE-22770.02.patch, HIVE-22770.02.patch
>
>
> HIVE-19937 introduced some interning logic into mapwork deserialization 
> process, but it's only related to spark, maybe we should skip this for tez, 
> reducing the cpu pressure in tez tasks.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22769) Incorrect query results and query failure during split generation for compressed text files

2020-01-27 Thread Chiran Ravani (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22769?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024455#comment-17024455
 ] 

Chiran Ravani commented on HIVE-22769:
--

[~kuczoram] Can we skip get split operation if the files are text and 
compressed?

> Incorrect query results and query failure during split generation for 
> compressed text files
> ---
>
> Key: HIVE-22769
> URL: https://issues.apache.org/jira/browse/HIVE-22769
> Project: Hive
>  Issue Type: Bug
>  Components: File Formats
>Affects Versions: 3.0.0, 3.1.0
>Reporter: Chiran Ravani
>Assignee: Marta Kuczora
>Priority: Critical
> Attachments: testcase1.csv.bz2, testcase2.csv.bz2
>
>
> Hive Query produces incorrect results when data is in text format and 
> compressed and for certain data the query fails during split generation.
> This behavior is seen when skip.header.line.count and skip.footer.line.count 
> are set for table.
> Case 1: Select count/aggregate query produces Incorrect row counts/displays 
> all rows (when hive.fetch.task.conversion=none)
> Steps to reproduce:
> 1. Create table as below
> {code}
> CREATE EXTERNAL TABLE `testcase1`(id int, name string) ROW FORMAT SERDE 
> 'org.apache.hadoop.hive.serde2.OpenCSVSerde' LOCATION '/user/hive/testcase1' 
> TBLPROPERTIES ("skip.header.line.count"="1", "skip.footer.line.count"="1");
> {code}
> 2. Upload attached testcase1.csv.bz2 file to /user/hive/testcase1
> 3. Run count(*) on table.
> {code}
> > select * from testcase1;
> INFO  : Compiling 
> command(queryId=hive_20200124053854_454b03c1-d4c5-4dba-a2c2-91c09f4b670f): 
> select * from testcase1
> INFO  : Semantic Analysis Completed (retrial = false)
> INFO  : Returning Hive schema: 
> Schema(fieldSchemas:[FieldSchema(name:testcase1.id, type:string, 
> comment:null), FieldSchema(name:testcase1.name, type:string, comment:null)], 
> properties:null)
> INFO  : Completed compiling 
> command(queryId=hive_20200124053854_454b03c1-d4c5-4dba-a2c2-91c09f4b670f); 
> Time taken: 0.07 seconds
> INFO  : Executing 
> command(queryId=hive_20200124053854_454b03c1-d4c5-4dba-a2c2-91c09f4b670f): 
> select * from testcase1
> INFO  : Completed executing 
> command(queryId=hive_20200124053854_454b03c1-d4c5-4dba-a2c2-91c09f4b670f); 
> Time taken: 0.007 seconds
> INFO  : OK
> +---+-+
> | testcase1.id  | testcase1.name  |
> +---+-+
> | 2 | 2019-12-31  |
> +---+-+
> 1 row selected (0.111 seconds)
> > select count(*) from testcase1
> INFO  : Compiling 
> command(queryId=hive_20200124053645_a7d699b7-c7e1-4d92-8d99-666b0a010ba7): 
> select count(*) from testcase1
> INFO  : Semantic Analysis Completed (retrial = false)
> INFO  : Returning Hive schema: Schema(fieldSchemas:[FieldSchema(name:_c0, 
> type:bigint, comment:null)], properties:null)
> INFO  : Completed compiling 
> command(queryId=hive_20200124053645_a7d699b7-c7e1-4d92-8d99-666b0a010ba7); 
> Time taken: 0.073 seconds
> INFO  : Executing 
> command(queryId=hive_20200124053645_a7d699b7-c7e1-4d92-8d99-666b0a010ba7): 
> select count(*) from testcase1
> INFO  : Query ID = hive_20200124053645_a7d699b7-c7e1-4d92-8d99-666b0a010ba7
> INFO  : Total jobs = 1
> INFO  : Launching Job 1 out of 1
> INFO  : Starting task [Stage-1:MAPRED] in serial mode
> INFO  : Subscribed to counters: [] for queryId: 
> hive_20200124053645_a7d699b7-c7e1-4d92-8d99-666b0a010ba7
> INFO  : Session is already open
> INFO  : Dag name: select count(*) from testcase1 (Stage-1)
> INFO  : Status: Running (Executing on YARN cluster with App id 
> application_1579811438512_0046)
> .
> .
> .
> INFO  : Completed executing 
> command(queryId=hive_20200124053645_a7d699b7-c7e1-4d92-8d99-666b0a010ba7); 
> Time taken: 4.228 seconds
> INFO  : OK
> +--+
> | _c0  |
> +--+
> | 3|
> +--+
> 1 row selected (4.335 seconds)
> {code}
> Case 2: Select count/aggregate query fails with java.lang.ClassCastException: 
> java.io.PushbackInputStream cannot be cast to org.apache.hadoop.fs.Seekable
> The issue is only seen when there is a space in a field (eg:- "3,2019-12-31 
> 01" second column has a space)
> Steps to reproduce:
> 1. Create table as below
> {code}
> CREATE EXTERNAL TABLE `testcase2`(id int, name string) ROW FORMAT SERDE 
> 'org.apache.hadoop.hive.serde2.OpenCSVSerde' LOCATION '/user/hive/testcase2' 
> TBLPROPERTIES ("skip.header.line.count"="1", "skip.footer.line.count"="1");
> {code}
> 2. Upload attached testcase2.csv.bz2 file to /user/hive/testcase2
> 3. Run count(*) on table.
> {code}
> 0: > select * from testcase2;
> INFO  : Compiling 
> command(queryId=hive_20200124053159_5d8ce56a-183d-4359-a147-bd470d82e134): 
> select * from testcase2
> INFO  : Semantic Analysis Completed 

[jira] [Updated] (HIVE-22780) Upgrade slf4j version to 1.7.30

2020-01-27 Thread David Lavati (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22780?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Lavati updated HIVE-22780:

Attachment: HIVE-22780.01.patch
Status: Patch Available  (was: Open)

> Upgrade slf4j version to 1.7.30
> ---
>
> Key: HIVE-22780
> URL: https://issues.apache.org/jira/browse/HIVE-22780
> Project: Hive
>  Issue Type: Improvement
>Reporter: David Lavati
>Assignee: David Lavati
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22780.01.patch
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-22780) Upgrade slf4j version to 1.7.30

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22780?focusedWorklogId=377708=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-377708
 ]

ASF GitHub Bot logged work on HIVE-22780:
-

Author: ASF GitHub Bot
Created on: 27/Jan/20 15:39
Start Date: 27/Jan/20 15:39
Worklog Time Spent: 10m 
  Work Description: dlavati commented on pull request #890: HIVE-22780: 
Upgrade slf4j to 1.7.30
URL: https://github.com/apache/hive/pull/890
 
 
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 377708)
Remaining Estimate: 0h
Time Spent: 10m

> Upgrade slf4j version to 1.7.30
> ---
>
> Key: HIVE-22780
> URL: https://issues.apache.org/jira/browse/HIVE-22780
> Project: Hive
>  Issue Type: Improvement
>Reporter: David Lavati
>Assignee: David Lavati
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22780) Upgrade slf4j version to 1.7.30

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22780?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated HIVE-22780:
--
Labels: pull-request-available  (was: )

> Upgrade slf4j version to 1.7.30
> ---
>
> Key: HIVE-22780
> URL: https://issues.apache.org/jira/browse/HIVE-22780
> Project: Hive
>  Issue Type: Improvement
>Reporter: David Lavati
>Assignee: David Lavati
>Priority: Major
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-10362) Support Type check/conversion in dynamic partition column

2020-01-27 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-10362?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024436#comment-17024436
 ] 

Hive QA commented on HIVE-10362:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
16s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
3s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
39s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m 
49s{color} | {color:blue} ql in master has 1533 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
57s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
27s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
1s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m  
1s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
40s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  3m 
53s{color} | {color:red} ql generated 2 new + 1533 unchanged - 0 fixed = 1535 
total (was 1533) {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
57s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
13s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 23m 23s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| FindBugs | module:ql |
|  |  Redundant nullcheck of destinationTable which is known to be null in 
org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genFileSinkPlan(String, QB, 
Operator)  Redundant null check at SemanticAnalyzer.java:is known to be null in 
org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genFileSinkPlan(String, QB, 
Operator)  Redundant null check at SemanticAnalyzer.java:[line 7680] |
|  |  Nullcheck of destinationTable at line 7335 of value previously 
dereferenced in 
org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genFileSinkPlan(String, QB, 
Operator)  At SemanticAnalyzer.java:7335 of value previously dereferenced in 
org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genFileSinkPlan(String, QB, 
Operator)  At SemanticAnalyzer.java:[line 7335] |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-20357/dev-support/hive-personality.sh
 |
| git revision | master / 7bb1d1e |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.1 |
| findbugs | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-20357/yetus/new-findbugs-ql.html
 |
| modules | C: ql U: ql |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-20357/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Support Type check/conversion in dynamic partition column
> -
>
> Key: HIVE-10362
> URL: https://issues.apache.org/jira/browse/HIVE-10362
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Processor, Types
>Affects Versions: 1.0.0, 3.0.0, 4.0.0
>Reporter: Chaoyu Tang
>Assignee: Karen Coppage
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-10362.01.patch, HIVE-10362.02.patch, 
> HIVE-10362.03.patch, HIVE-10362.04.patch, HIVE-10362.04.patch, 
> HIVE-10362.06.patch
>
>
> There are quite a lot of issues associated with the non-noramlized or 
> type-mismatched values for 

[jira] [Comment Edited] (HIVE-22767) Beeline doesn't parse semicolons in comments properly

2020-01-27 Thread Zoltan Haindrich (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22767?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024433#comment-17024433
 ] 

Zoltan Haindrich edited comment on HIVE-22767 at 1/27/20 3:36 PM:
--

because there were no real code changes...waiting for another ptest will not 
change anything...I've run the new test manually.
pushed to master. Thank you [~zmatyus]!


was (Author: kgyrtkirk):
pushed to master. Thank you [~zmatyus]

> Beeline doesn't parse semicolons in comments properly
> -
>
> Key: HIVE-22767
> URL: https://issues.apache.org/jira/browse/HIVE-22767
> Project: Hive
>  Issue Type: Bug
>  Components: Beeline
>Reporter: Zoltan Matyus
>Assignee: Zoltan Matyus
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-22767.01.patch, HIVE-22767.02.patch, 
> HIVE-22767.03.patch
>
>
> HIVE-12646 fixed the handling of semicolons in quoted strings, but leaves the 
> problem of semicolons in comments. E.g. with beeline connected to any 
> database...
> this works: {code:sql}select 1; select /*   */ 2; select /*   */ 3;{code}
> this doesn't work: {code:sql}select 1; select /* ; */ 2; select /* ; */ 
> 3;{code}
> This has been fixed and reintroduced before (possibly multiple times). 
> Ideally, there should be a single utility method somewhere to separate 
> comments, strings and commands -- with the proper testing in place (q files).
> However, I'm trying to make this fix back-portable, so a light touch is 
> needed. I'm focusing on beeline for now, and only writing (very thorough) 
> unit tests, as I cannot exclude any new q files from TestCliDriver (which 
> would break, since it's using a different parsing method).
> 
> P.S. excerpt of the error message:
> {noformat}
> 0: jdbc:hive2://...> select 1; select /* ; */ 2; select /* ; */ 3;
> INFO  : Compiling command(queryId=...): select 1
> INFO  : Semantic Analysis Completed (retrial = false)
> INFO  : Returning Hive schema: Schema(fieldSchemas:[FieldSchema(name:_c0, 
> type:int, comment:null)], properties:null)
> INFO  : Completed compiling command(queryId=...); Time taken: 0.38 seconds
> INFO  : Executing command(queryId=...): select 1
> INFO  : Completed executing command(queryId=...); Time taken: 0.004 seconds
> INFO  : OK
> +--+
> | _c0  |
> +--+
> | 1|
> +--+
> 1 row selected (2.007 seconds)
> INFO  : Compiling command(queryId=...): select /*
> ERROR : FAILED: ParseException line 1:9 cannot recognize input near '' 
> '' '' in select clause
> org.apache.hadoop.hive.ql.parse.ParseException: line 1:9 cannot recognize 
> input near '' '' '' in select clause
>   at 
> org.apache.hadoop.hive.ql.parse.ParseDriver.parse(ParseDriver.java:233)
>   at org.apache.hadoop.hive.ql.parse.ParseUtils.parse(ParseUtils.java:79)
>   at org.apache.hadoop.hive.ql.parse.ParseUtils.parse(ParseUtils.java:72)
>   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:598)
>   at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1505)
>   at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:1452)
>   at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:1447)
>   at 
> org.apache.hadoop.hive.ql.reexec.ReExecDriver.compileAndRespond(ReExecDriver.java:126)
>   at ...
> {noformat}
> 
> Similarly, the following query also fails:
> {code:sql}select /* ' */ 1; select /* ' */ 2;{code}
> I suspect line comments are also not handled properly but I cannot reproduce 
> this in interactive beeline...



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22767) Beeline doesn't parse semicolons in comments properly

2020-01-27 Thread Zoltan Haindrich (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22767?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltan Haindrich updated HIVE-22767:

Fix Version/s: 4.0.0
   Resolution: Fixed
   Status: Resolved  (was: Patch Available)

pushed to master. Thank you [~zmatyus]

> Beeline doesn't parse semicolons in comments properly
> -
>
> Key: HIVE-22767
> URL: https://issues.apache.org/jira/browse/HIVE-22767
> Project: Hive
>  Issue Type: Bug
>  Components: Beeline
>Reporter: Zoltan Matyus
>Assignee: Zoltan Matyus
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-22767.01.patch, HIVE-22767.02.patch, 
> HIVE-22767.03.patch
>
>
> HIVE-12646 fixed the handling of semicolons in quoted strings, but leaves the 
> problem of semicolons in comments. E.g. with beeline connected to any 
> database...
> this works: {code:sql}select 1; select /*   */ 2; select /*   */ 3;{code}
> this doesn't work: {code:sql}select 1; select /* ; */ 2; select /* ; */ 
> 3;{code}
> This has been fixed and reintroduced before (possibly multiple times). 
> Ideally, there should be a single utility method somewhere to separate 
> comments, strings and commands -- with the proper testing in place (q files).
> However, I'm trying to make this fix back-portable, so a light touch is 
> needed. I'm focusing on beeline for now, and only writing (very thorough) 
> unit tests, as I cannot exclude any new q files from TestCliDriver (which 
> would break, since it's using a different parsing method).
> 
> P.S. excerpt of the error message:
> {noformat}
> 0: jdbc:hive2://...> select 1; select /* ; */ 2; select /* ; */ 3;
> INFO  : Compiling command(queryId=...): select 1
> INFO  : Semantic Analysis Completed (retrial = false)
> INFO  : Returning Hive schema: Schema(fieldSchemas:[FieldSchema(name:_c0, 
> type:int, comment:null)], properties:null)
> INFO  : Completed compiling command(queryId=...); Time taken: 0.38 seconds
> INFO  : Executing command(queryId=...): select 1
> INFO  : Completed executing command(queryId=...); Time taken: 0.004 seconds
> INFO  : OK
> +--+
> | _c0  |
> +--+
> | 1|
> +--+
> 1 row selected (2.007 seconds)
> INFO  : Compiling command(queryId=...): select /*
> ERROR : FAILED: ParseException line 1:9 cannot recognize input near '' 
> '' '' in select clause
> org.apache.hadoop.hive.ql.parse.ParseException: line 1:9 cannot recognize 
> input near '' '' '' in select clause
>   at 
> org.apache.hadoop.hive.ql.parse.ParseDriver.parse(ParseDriver.java:233)
>   at org.apache.hadoop.hive.ql.parse.ParseUtils.parse(ParseUtils.java:79)
>   at org.apache.hadoop.hive.ql.parse.ParseUtils.parse(ParseUtils.java:72)
>   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:598)
>   at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1505)
>   at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:1452)
>   at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:1447)
>   at 
> org.apache.hadoop.hive.ql.reexec.ReExecDriver.compileAndRespond(ReExecDriver.java:126)
>   at ...
> {noformat}
> 
> Similarly, the following query also fails:
> {code:sql}select /* ' */ 1; select /* ' */ 2;{code}
> I suspect line comments are also not handled properly but I cannot reproduce 
> this in interactive beeline...



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22774) Usability improvements of scheduled queries

2020-01-27 Thread Zoltan Haindrich (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22774?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltan Haindrich updated HIVE-22774:

Fix Version/s: 4.0.0
   Resolution: Fixed
   Status: Resolved  (was: Patch Available)

pushed to master thank you Jesus for reviewing the changes!

> Usability improvements of scheduled queries
> ---
>
> Key: HIVE-22774
> URL: https://issues.apache.org/jira/browse/HIVE-22774
> Project: Hive
>  Issue Type: Improvement
>  Components: Standalone Metastore
>Reporter: Zoltan Haindrich
>Assignee: Zoltan Haindrich
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-22774.01.patch, HIVE-22774.02.patch, 
> HIVE-22774.03.patch
>
>
> * add feature toggle switch to HMS side rpc call sites - make it possible to 
> disable it
> * rename query state ERRORED state to FAILED
> * ProgressReporter thread should not enter a busy wait loop in case the 
> thread is interrupted



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22679) Replace Base64 in metastore-common Package

2020-01-27 Thread Zoltan Haindrich (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22679?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltan Haindrich updated HIVE-22679:

Fix Version/s: 4.0.0
   Resolution: Fixed
   Status: Resolved  (was: Patch Available)

pushed to master. Thank you David!

> Replace Base64 in metastore-common Package
> --
>
> Key: HIVE-22679
> URL: https://issues.apache.org/jira/browse/HIVE-22679
> Project: Hive
>  Issue Type: Sub-task
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Minor
> Fix For: 4.0.0
>
> Attachments: HIVE-22679.1.patch, HIVE-22679.1.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-22767) Beeline doesn't parse semicolons in comments properly

2020-01-27 Thread Zoltan Haindrich (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22767?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltan Haindrich updated HIVE-22767:

Summary: Beeline doesn't parse semicolons in comments properly  (was: 
beeline doesn't parse semicolons in comments properly)

> Beeline doesn't parse semicolons in comments properly
> -
>
> Key: HIVE-22767
> URL: https://issues.apache.org/jira/browse/HIVE-22767
> Project: Hive
>  Issue Type: Bug
>  Components: Beeline
>Reporter: Zoltan Matyus
>Assignee: Zoltan Matyus
>Priority: Major
> Attachments: HIVE-22767.01.patch, HIVE-22767.02.patch, 
> HIVE-22767.03.patch
>
>
> HIVE-12646 fixed the handling of semicolons in quoted strings, but leaves the 
> problem of semicolons in comments. E.g. with beeline connected to any 
> database...
> this works: {code:sql}select 1; select /*   */ 2; select /*   */ 3;{code}
> this doesn't work: {code:sql}select 1; select /* ; */ 2; select /* ; */ 
> 3;{code}
> This has been fixed and reintroduced before (possibly multiple times). 
> Ideally, there should be a single utility method somewhere to separate 
> comments, strings and commands -- with the proper testing in place (q files).
> However, I'm trying to make this fix back-portable, so a light touch is 
> needed. I'm focusing on beeline for now, and only writing (very thorough) 
> unit tests, as I cannot exclude any new q files from TestCliDriver (which 
> would break, since it's using a different parsing method).
> 
> P.S. excerpt of the error message:
> {noformat}
> 0: jdbc:hive2://...> select 1; select /* ; */ 2; select /* ; */ 3;
> INFO  : Compiling command(queryId=...): select 1
> INFO  : Semantic Analysis Completed (retrial = false)
> INFO  : Returning Hive schema: Schema(fieldSchemas:[FieldSchema(name:_c0, 
> type:int, comment:null)], properties:null)
> INFO  : Completed compiling command(queryId=...); Time taken: 0.38 seconds
> INFO  : Executing command(queryId=...): select 1
> INFO  : Completed executing command(queryId=...); Time taken: 0.004 seconds
> INFO  : OK
> +--+
> | _c0  |
> +--+
> | 1|
> +--+
> 1 row selected (2.007 seconds)
> INFO  : Compiling command(queryId=...): select /*
> ERROR : FAILED: ParseException line 1:9 cannot recognize input near '' 
> '' '' in select clause
> org.apache.hadoop.hive.ql.parse.ParseException: line 1:9 cannot recognize 
> input near '' '' '' in select clause
>   at 
> org.apache.hadoop.hive.ql.parse.ParseDriver.parse(ParseDriver.java:233)
>   at org.apache.hadoop.hive.ql.parse.ParseUtils.parse(ParseUtils.java:79)
>   at org.apache.hadoop.hive.ql.parse.ParseUtils.parse(ParseUtils.java:72)
>   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:598)
>   at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1505)
>   at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:1452)
>   at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:1447)
>   at 
> org.apache.hadoop.hive.ql.reexec.ReExecDriver.compileAndRespond(ReExecDriver.java:126)
>   at ...
> {noformat}
> 
> Similarly, the following query also fails:
> {code:sql}select /* ' */ 1; select /* ' */ 2;{code}
> I suspect line comments are also not handled properly but I cannot reproduce 
> this in interactive beeline...



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-22538) RS deduplication does not always enforce hive.optimize.reducededuplication.min.reducer

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22538?focusedWorklogId=377703=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-377703
 ]

ASF GitHub Bot logged work on HIVE-22538:
-

Author: ASF GitHub Bot
Created on: 27/Jan/20 15:31
Start Date: 27/Jan/20 15:31
Worklog Time Spent: 10m 
  Work Description: kasakrisz commented on pull request #877: HIVE-22538: 
RS deduplication does not always enforce 
hive.optimize.reducededuplication.min.reducer
URL: https://github.com/apache/hive/pull/877#discussion_r371308488
 
 

 ##
 File path: 
ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction_3.q.out
 ##
 @@ -814,30 +791,20 @@ STAGE PLANS:
   Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
   Group By Operator
 aggregations: compute_stats(a, 'hll'), compute_stats(b, 
'hll')
-minReductionHashAggr: 0.99
-mode: hash
+mode: complete
 
 Review comment:
   Enabling parallelism when inserting eliminates this change.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 377703)
Time Spent: 1.5h  (was: 1h 20m)

> RS deduplication does not always enforce 
> hive.optimize.reducededuplication.min.reducer
> --
>
> Key: HIVE-22538
> URL: https://issues.apache.org/jira/browse/HIVE-22538
> Project: Hive
>  Issue Type: Bug
>  Components: Physical Optimizer
>Reporter: Jesus Camacho Rodriguez
>Assignee: Krisztian Kasa
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22538.2.patch, HIVE-22538.3.patch, 
> HIVE-22538.4.patch, HIVE-22538.5.patch, HIVE-22538.6.patch, 
> HIVE-22538.6.patch, HIVE-22538.7.patch, HIVE-22538.8.patch, HIVE-22538.patch
>
>  Time Spent: 1.5h
>  Remaining Estimate: 0h
>
> For transactional tables, that property might be overriden to 1, which can 
> lead to merging final aggregation into a single stage (hence leading to 
> performance degradation). For instance, when autogather column stats is 
> enabled, this can happen for the following query:
> {code}
> set hive.support.concurrency=true;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
> EXPLAIN
> CREATE TABLE x STORED AS ORC TBLPROPERTIES('transactional'='true') AS
> SELECT * FROM SRC x CLUSTER BY x.key;
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-22538) RS deduplication does not always enforce hive.optimize.reducededuplication.min.reducer

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22538?focusedWorklogId=377702=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-377702
 ]

ASF GitHub Bot logged work on HIVE-22538:
-

Author: ASF GitHub Bot
Created on: 27/Jan/20 15:30
Start Date: 27/Jan/20 15:30
Worklog Time Spent: 10m 
  Work Description: kasakrisz commented on pull request #877: HIVE-22538: 
RS deduplication does not always enforce 
hive.optimize.reducededuplication.min.reducer
URL: https://github.com/apache/hive/pull/877#discussion_r371308303
 
 

 ##
 File path: 
ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction_3.q.out
 ##
 @@ -261,31 +260,21 @@ STAGE PLANS:
   Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE 
Column stats: NONE
   Group By Operator
 aggregations: compute_stats(a, 'hll'), compute_stats(b, 
'hll')
-minReductionHashAggr: 0.99
-mode: hash
+mode: complete
 
 Review comment:
   Enabling parallelism when inserting eliminates this change.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 377702)
Time Spent: 1h 20m  (was: 1h 10m)

> RS deduplication does not always enforce 
> hive.optimize.reducededuplication.min.reducer
> --
>
> Key: HIVE-22538
> URL: https://issues.apache.org/jira/browse/HIVE-22538
> Project: Hive
>  Issue Type: Bug
>  Components: Physical Optimizer
>Reporter: Jesus Camacho Rodriguez
>Assignee: Krisztian Kasa
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22538.2.patch, HIVE-22538.3.patch, 
> HIVE-22538.4.patch, HIVE-22538.5.patch, HIVE-22538.6.patch, 
> HIVE-22538.6.patch, HIVE-22538.7.patch, HIVE-22538.8.patch, HIVE-22538.patch
>
>  Time Spent: 1h 20m
>  Remaining Estimate: 0h
>
> For transactional tables, that property might be overriden to 1, which can 
> lead to merging final aggregation into a single stage (hence leading to 
> performance degradation). For instance, when autogather column stats is 
> enabled, this can happen for the following query:
> {code}
> set hive.support.concurrency=true;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
> EXPLAIN
> CREATE TABLE x STORED AS ORC TBLPROPERTIES('transactional'='true') AS
> SELECT * FROM SRC x CLUSTER BY x.key;
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-22538) RS deduplication does not always enforce hive.optimize.reducededuplication.min.reducer

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22538?focusedWorklogId=377700=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-377700
 ]

ASF GitHub Bot logged work on HIVE-22538:
-

Author: ASF GitHub Bot
Created on: 27/Jan/20 15:29
Start Date: 27/Jan/20 15:29
Worklog Time Spent: 10m 
  Work Description: kasakrisz commented on pull request #877: HIVE-22538: 
RS deduplication does not always enforce 
hive.optimize.reducededuplication.min.reducer
URL: https://github.com/apache/hive/pull/877#discussion_r371307699
 
 

 ##
 File path: ql/src/test/results/clientpositive/autoColumnStats_4.q.out
 ##
 @@ -128,16 +127,20 @@ STAGE PLANS:
 Statistics: Num rows: 10 Data size: 1728 Basic stats: COMPLETE 
Column stats: COMPLETE
 Group By Operator
   aggregations: compute_stats(a, 'hll'), compute_stats(b, 'hll')
-  minReductionHashAggr: 0.99
-  mode: hash
+  mode: complete
 
 Review comment:
   Enabling parallelism when inserting eliminates this change.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 377700)
Time Spent: 1h 10m  (was: 1h)

> RS deduplication does not always enforce 
> hive.optimize.reducededuplication.min.reducer
> --
>
> Key: HIVE-22538
> URL: https://issues.apache.org/jira/browse/HIVE-22538
> Project: Hive
>  Issue Type: Bug
>  Components: Physical Optimizer
>Reporter: Jesus Camacho Rodriguez
>Assignee: Krisztian Kasa
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22538.2.patch, HIVE-22538.3.patch, 
> HIVE-22538.4.patch, HIVE-22538.5.patch, HIVE-22538.6.patch, 
> HIVE-22538.6.patch, HIVE-22538.7.patch, HIVE-22538.8.patch, HIVE-22538.patch
>
>  Time Spent: 1h 10m
>  Remaining Estimate: 0h
>
> For transactional tables, that property might be overriden to 1, which can 
> lead to merging final aggregation into a single stage (hence leading to 
> performance degradation). For instance, when autogather column stats is 
> enabled, this can happen for the following query:
> {code}
> set hive.support.concurrency=true;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
> EXPLAIN
> CREATE TABLE x STORED AS ORC TBLPROPERTIES('transactional'='true') AS
> SELECT * FROM SRC x CLUSTER BY x.key;
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-22538) RS deduplication does not always enforce hive.optimize.reducededuplication.min.reducer

2020-01-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22538?focusedWorklogId=377698=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-377698
 ]

ASF GitHub Bot logged work on HIVE-22538:
-

Author: ASF GitHub Bot
Created on: 27/Jan/20 15:28
Start Date: 27/Jan/20 15:28
Worklog Time Spent: 10m 
  Work Description: kasakrisz commented on pull request #877: HIVE-22538: 
RS deduplication does not always enforce 
hive.optimize.reducededuplication.min.reducer
URL: https://github.com/apache/hive/pull/877#discussion_r371306847
 
 

 ##
 File path: ql/src/test/results/clientpositive/acid_table_directories_test.q.out
 ##
 @@ -154,6 +154,7 @@ POSTHOOK: Input: default@acidparttbl@p=200
 ### ACID BASE DIR ###
 ### ACID BASE DIR ###
 ### ACID BASE DIR ###
+### ACID BASE DIR ###
 
 Review comment:
   Enabling parallelism when inserting eliminates this change.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 377698)
Time Spent: 1h  (was: 50m)

> RS deduplication does not always enforce 
> hive.optimize.reducededuplication.min.reducer
> --
>
> Key: HIVE-22538
> URL: https://issues.apache.org/jira/browse/HIVE-22538
> Project: Hive
>  Issue Type: Bug
>  Components: Physical Optimizer
>Reporter: Jesus Camacho Rodriguez
>Assignee: Krisztian Kasa
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22538.2.patch, HIVE-22538.3.patch, 
> HIVE-22538.4.patch, HIVE-22538.5.patch, HIVE-22538.6.patch, 
> HIVE-22538.6.patch, HIVE-22538.7.patch, HIVE-22538.8.patch, HIVE-22538.patch
>
>  Time Spent: 1h
>  Remaining Estimate: 0h
>
> For transactional tables, that property might be overriden to 1, which can 
> lead to merging final aggregation into a single stage (hence leading to 
> performance degradation). For instance, when autogather column stats is 
> enabled, this can happen for the following query:
> {code}
> set hive.support.concurrency=true;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
> EXPLAIN
> CREATE TABLE x STORED AS ORC TBLPROPERTIES('transactional'='true') AS
> SELECT * FROM SRC x CLUSTER BY x.key;
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22680) Replace Base64 in druid-handler Package

2020-01-27 Thread Zoltan Haindrich (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22680?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17024428#comment-17024428
 ] 

Zoltan Haindrich commented on HIVE-22680:
-

+1

> Replace Base64 in druid-handler Package
> ---
>
> Key: HIVE-22680
> URL: https://issues.apache.org/jira/browse/HIVE-22680
> Project: Hive
>  Issue Type: Sub-task
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Minor
> Attachments: HIVE-22680.1.patch, HIVE-22680.1.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


  1   2   >