[jira] [Commented] (HIVE-23082) PK/FK stat rescale doesn't work in some cases

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23082?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073385#comment-17073385
 ] 

Hive QA commented on HIVE-23082:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
30s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
10s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
2s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
41s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m 
42s{color} | {color:blue} ql in master has 1528 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
56s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
28s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
23s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
0s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m  
0s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
43s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
45s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
53s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 25m 46s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-21380/dev-support/hive-personality.sh
 |
| git revision | master / 709235c |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.1 |
| modules | C: ql itests U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21380/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> PK/FK stat rescale doesn't work in some cases
> -
>
> Key: HIVE-23082
> URL: https://issues.apache.org/jira/browse/HIVE-23082
> Project: Hive
>  Issue Type: Bug
>  Components: Statistics
>Reporter: Zoltan Haindrich
>Assignee: Zoltan Haindrich
>Priority: Major
> Attachments: HIVE-23082.01.patch, HIVE-23082.02.patch, 
> HIVE-23082.03.patch
>
>
> As a result in Joins may retain the original estimate; see MAPJOIN_33 in this 
> plan ; which retained the estimate of SEL_32
> {code}
> ++
> |  Explain   |
> ++
> | Plan optimized by CBO. |
> ||
> | Vertex dependency in root stage|
> | Map 1 <- Map 2 (BROADCAST_EDGE)|
> ||
> | Stage-0|
> |   Fetch Operator   |
> | limit:12   |
> | Stage-1|
> |   Map 1 vectorized |
> |   File Output Operator [FS_36] 

[jira] [Updated] (HIVE-23004) Support Decimal64 operations across multiple vertices

2020-04-01 Thread Ramesh Kumar Thangarajan (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23004?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ramesh Kumar Thangarajan updated HIVE-23004:

Attachment: HIVE-23004.17.patch
Status: Patch Available  (was: Open)

> Support Decimal64 operations across multiple vertices
> -
>
> Key: HIVE-23004
> URL: https://issues.apache.org/jira/browse/HIVE-23004
> Project: Hive
>  Issue Type: Bug
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
> Attachments: HIVE-23004.1.patch, HIVE-23004.10.patch, 
> HIVE-23004.11.patch, HIVE-23004.12.patch, HIVE-23004.13.patch, 
> HIVE-23004.14.patch, HIVE-23004.15.patch, HIVE-23004.16.patch, 
> HIVE-23004.17.patch, HIVE-23004.2.patch, HIVE-23004.4.patch, 
> HIVE-23004.6.patch, HIVE-23004.7.patch, HIVE-23004.8.patch, HIVE-23004.9.patch
>
>
> Support Decimal64 operations across multiple vertices



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23004) Support Decimal64 operations across multiple vertices

2020-04-01 Thread Ramesh Kumar Thangarajan (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23004?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ramesh Kumar Thangarajan updated HIVE-23004:

Status: Open  (was: Patch Available)

> Support Decimal64 operations across multiple vertices
> -
>
> Key: HIVE-23004
> URL: https://issues.apache.org/jira/browse/HIVE-23004
> Project: Hive
>  Issue Type: Bug
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
> Attachments: HIVE-23004.1.patch, HIVE-23004.10.patch, 
> HIVE-23004.11.patch, HIVE-23004.12.patch, HIVE-23004.13.patch, 
> HIVE-23004.14.patch, HIVE-23004.15.patch, HIVE-23004.16.patch, 
> HIVE-23004.17.patch, HIVE-23004.2.patch, HIVE-23004.4.patch, 
> HIVE-23004.6.patch, HIVE-23004.7.patch, HIVE-23004.8.patch, HIVE-23004.9.patch
>
>
> Support Decimal64 operations across multiple vertices



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23030) Enable sketch union-s to be rolled up

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23030?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073377#comment-17073377
 ] 

Hive QA commented on HIVE-23030:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12998404/HIVE-23030.06.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 18163 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.metastore.TestMetastoreHousekeepingLeaderEmptyConfig.testHouseKeepingThreadExistence
 (batchId=252)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/21379/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/21379/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-21379/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12998404 - PreCommit-HIVE-Build

> Enable sketch union-s to be rolled up
> -
>
> Key: HIVE-23030
> URL: https://issues.apache.org/jira/browse/HIVE-23030
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Zoltan Haindrich
>Assignee: Zoltan Haindrich
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23030.01.patch, HIVE-23030.02.patch, 
> HIVE-23030.03.patch, HIVE-23030.04.patch, HIVE-23030.04.patch, 
> HIVE-23030.05.patch, HIVE-23030.06.patch, HIVE-23030.06.patch
>
>  Time Spent: 5h
>  Remaining Estimate: 0h
>
> Enabling rolling up sketch aggregates could enable the matching of 
> materialized views created for higher dimensions to be applied for lower 
> dimension cases.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-22959) Extend storage-api to expose FilterContext

2020-04-01 Thread Jesus Camacho Rodriguez (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-22959?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073359#comment-17073359
 ] 

Jesus Camacho Rodriguez commented on HIVE-22959:


[~omalley], did you have the chance to look into this? {{storage-api}} release 
is on hold until this is solved. Thanks

> Extend storage-api to expose FilterContext
> --
>
> Key: HIVE-22959
> URL: https://issues.apache.org/jira/browse/HIVE-22959
> Project: Hive
>  Issue Type: Sub-task
>  Components: storage-api
>Reporter: Panagiotis Garefalakis
>Assignee: Panagiotis Garefalakis
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0, storage-2.7.2
>
> Attachments: HIVE-22959.1.patch, HIVE-22959.2.patch, 
> HIVE-22959.3.patch, HIVE-22959.4.patch
>
>  Time Spent: 1h 20m
>  Remaining Estimate: 0h
>
> To enable row-level filtering at the ORC level ORC-577, or as an extension 
> ProDecode MapJoin HIVE-22731 we need a common context class that will hold 
> all the needed information for the filter.
> I propose this class to be part of the storage-api – similar to 
> VectorizedRowBatch class and hold the information below:
>  * A boolean variable showing if the filter is enabled
>  * A int array storing the row Ids that are actually selected (passing the 
> filter)
>  * An int variable storing the the number or rows that passed the filter
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23128) SHOW CREATE TABLE Creates Incorrect Syntax When Database Specified

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23128?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor updated HIVE-23128:
--
Summary: SHOW CREATE TABLE Creates Incorrect Syntax When Database Specified 
 (was: SHOW CREATE TABLE Creates Invalid Syntax When Database Specified)

> SHOW CREATE TABLE Creates Incorrect Syntax When Database Specified
> --
>
> Key: HIVE-23128
> URL: https://issues.apache.org/jira/browse/HIVE-23128
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.4.0, 3.1.2
>Reporter: David Mollitor
>Priority: Major
>
> {code:sql}
> show create table `sample_07`;
> show create table `default`.`sample_07`;
> show create table `default.sample_07`;
> {code}
> {code:none|title=Results}
> CREATE TABLE `sample_07`(...)
> CREATE TABLE `default.sample_07`(...)
> CREATE TABLE `default.sample_07`(...);
> {code}
> All three queries complete in Hive 2.x (not sure about 3.x) but the results 
> are consistent in both versions.  The first result is correct because it does 
> not include the database name, however, the subsequent two results are 
> invalid: each field must be quoted individually.
> {quote}
> If any components of a multiple-part name require quoting, quote them 
> individually rather than quoting the name as a whole. For example, write 
> `my-table`.`my-column`, not `my-table.my-column`.
> * https://dev.mysql.com/doc/refman/8.0/en/identifier-qualifiers.html
> {quote}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23030) Enable sketch union-s to be rolled up

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23030?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073358#comment-17073358
 ] 

Hive QA commented on HIVE-23030:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
52s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
45s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
59s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
43s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m 
36s{color} | {color:blue} ql in master has 1528 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
51s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
28s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
23s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
58s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
58s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
43s{color} | {color:red} ql: The patch generated 4 new + 174 unchanged - 0 
fixed = 178 total (was 174) {color} |
| {color:red}-1{color} | {color:red} whitespace {color} | {color:red}  0m  
0s{color} | {color:red} The patch has 1 line(s) that end in whitespace. Use git 
apply --whitespace=fix <>. Refer https://git-scm.com/docs/git-apply 
{color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
51s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
51s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 24m 45s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-21379/dev-support/hive-personality.sh
 |
| git revision | master / 709235c |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.1 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21379/yetus/diff-checkstyle-ql.txt
 |
| whitespace | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21379/yetus/whitespace-eol.txt
 |
| modules | C: ql itests U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21379/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Enable sketch union-s to be rolled up
> -
>
> Key: HIVE-23030
> URL: https://issues.apache.org/jira/browse/HIVE-23030
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Zoltan Haindrich
>Assignee: Zoltan Haindrich
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23030.01.patch, HIVE-23030.02.patch, 
> HIVE-23030.03.patch, HIVE-23030.04.patch, HIVE-23030.04.patch, 
> HIVE-23030.05.patch, HIVE-23030.06.patch, HIVE-23030.06.patch
>
>  Time Spent: 5h
>  Remaining Estimate: 0h
>
> Enabling rolling up sketch aggregates could enable the matching of 
> materialized views created for higher dimensions to be applied for lower 
> dimension cases.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Resolved] (HIVE-10775) Frequent calls to printStackTrace() obscuring legitimate problems

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-10775?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor resolved HIVE-10775.
---
Resolution: Duplicate

> Frequent calls to printStackTrace() obscuring legitimate problems
> -
>
> Key: HIVE-10775
> URL: https://issues.apache.org/jira/browse/HIVE-10775
> Project: Hive
>  Issue Type: Improvement
>  Components: Metastore, Query Processor
>Reporter: Andrew Cowie
>Assignee: Andrew Cowie
>Priority: Minor
> Attachments: HIVE-10775.1.patch
>
>
> When running test suites built on top of libraries that build on top of ... 
> that use Hive, the signal to noise ratio with exceptions flying past is 
> appalling. Most of this is down to calls to printStackTrace() embedded in 
> this library. HIVE-7697 showed someone cleaning that up and replacing with 
> logging the exception instead. That seems wise (logging can be redirected by 
> the calling test suite).
> So, if you don't object, I'll hunt down the calls to printStackTrace() and 
> replace them with LOG.warn() instead. I'm about half way through the patch 
> now.
> AfC



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23004) Support Decimal64 operations across multiple vertices

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23004?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073342#comment-17073342
 ] 

Hive QA commented on HIVE-23004:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12998395/HIVE-23004.16.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 18163 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[vectorization_17] 
(batchId=158)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/21378/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/21378/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-21378/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12998395 - PreCommit-HIVE-Build

> Support Decimal64 operations across multiple vertices
> -
>
> Key: HIVE-23004
> URL: https://issues.apache.org/jira/browse/HIVE-23004
> Project: Hive
>  Issue Type: Bug
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
> Attachments: HIVE-23004.1.patch, HIVE-23004.10.patch, 
> HIVE-23004.11.patch, HIVE-23004.12.patch, HIVE-23004.13.patch, 
> HIVE-23004.14.patch, HIVE-23004.15.patch, HIVE-23004.16.patch, 
> HIVE-23004.2.patch, HIVE-23004.4.patch, HIVE-23004.6.patch, 
> HIVE-23004.7.patch, HIVE-23004.8.patch, HIVE-23004.9.patch
>
>
> Support Decimal64 operations across multiple vertices



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-21492) VectorizedParquetRecordReader can't to read parquet file generated using thrift/custom tool

2020-04-01 Thread Ganesha Shreedhara (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21492?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ganesha Shreedhara updated HIVE-21492:
--
Status: Patch Available  (was: In Progress)

> VectorizedParquetRecordReader can't to read parquet file generated using 
> thrift/custom tool
> ---
>
> Key: HIVE-21492
> URL: https://issues.apache.org/jira/browse/HIVE-21492
> Project: Hive
>  Issue Type: Bug
>Reporter: Ganesha Shreedhara
>Assignee: Ganesha Shreedhara
>Priority: Major
> Attachments: HIVE-21492.2.patch, HIVE-21492.3.patch, HIVE-21492.patch
>
>
> Taking an example of a parquet table having array of integers as below. 
> {code:java}
> CREATE EXTERNAL TABLE ( list_of_ints` array)
> STORED AS PARQUET 
> LOCATION '{location}';
> {code}
> Parquet file generated using hive will have schema for Type as below:
> {code:java}
> group list_of_ints (LIST) { repeated group bag { optional int32 array;\n};\n} 
> {code}
> Parquet file generated using thrift or any custom tool (using 
> org.apache.parquet.io.api.RecordConsumer)
> may have schema for Type as below:
> {code:java}
> required group list_of_ints (LIST) { repeated int32 list_of_tuple} {code}
> VectorizedParquetRecordReader handles only parquet file generated using hive. 
> It throws the following exception when parquet file generated using thrift is 
> read because of the changes done as part of HIVE-18553 .
> {code:java}
> Caused by: java.lang.ClassCastException: repeated int32 list_of_ints_tuple is 
> not a group
>  at org.apache.parquet.schema.Type.asGroupType(Type.java:207)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.getElementType(VectorizedParquetRecordReader.java:479)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.buildVectorizedParquetReader(VectorizedParquetRecordReader.java:532)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.checkEndOfRowGroup(VectorizedParquetRecordReader.java:440)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.nextBatch(VectorizedParquetRecordReader.java:401)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:353)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:92)
>  at 
> org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.doNext(HiveContextAwareRecordReader.java:365){code}
>  
>  I have done a small change to handle the case where the child type of group 
> type can be PrimitiveType.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work started] (HIVE-21492) VectorizedParquetRecordReader can't to read parquet file generated using thrift/custom tool

2020-04-01 Thread Ganesha Shreedhara (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21492?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HIVE-21492 started by Ganesha Shreedhara.
-
> VectorizedParquetRecordReader can't to read parquet file generated using 
> thrift/custom tool
> ---
>
> Key: HIVE-21492
> URL: https://issues.apache.org/jira/browse/HIVE-21492
> Project: Hive
>  Issue Type: Bug
>Reporter: Ganesha Shreedhara
>Assignee: Ganesha Shreedhara
>Priority: Major
> Attachments: HIVE-21492.2.patch, HIVE-21492.3.patch, HIVE-21492.patch
>
>
> Taking an example of a parquet table having array of integers as below. 
> {code:java}
> CREATE EXTERNAL TABLE ( list_of_ints` array)
> STORED AS PARQUET 
> LOCATION '{location}';
> {code}
> Parquet file generated using hive will have schema for Type as below:
> {code:java}
> group list_of_ints (LIST) { repeated group bag { optional int32 array;\n};\n} 
> {code}
> Parquet file generated using thrift or any custom tool (using 
> org.apache.parquet.io.api.RecordConsumer)
> may have schema for Type as below:
> {code:java}
> required group list_of_ints (LIST) { repeated int32 list_of_tuple} {code}
> VectorizedParquetRecordReader handles only parquet file generated using hive. 
> It throws the following exception when parquet file generated using thrift is 
> read because of the changes done as part of HIVE-18553 .
> {code:java}
> Caused by: java.lang.ClassCastException: repeated int32 list_of_ints_tuple is 
> not a group
>  at org.apache.parquet.schema.Type.asGroupType(Type.java:207)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.getElementType(VectorizedParquetRecordReader.java:479)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.buildVectorizedParquetReader(VectorizedParquetRecordReader.java:532)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.checkEndOfRowGroup(VectorizedParquetRecordReader.java:440)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.nextBatch(VectorizedParquetRecordReader.java:401)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:353)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:92)
>  at 
> org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.doNext(HiveContextAwareRecordReader.java:365){code}
>  
>  I have done a small change to handle the case where the child type of group 
> type can be PrimitiveType.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-21492) VectorizedParquetRecordReader can't to read parquet file generated using thrift/custom tool

2020-04-01 Thread Ganesha Shreedhara (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21492?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ganesha Shreedhara updated HIVE-21492:
--
Status: Open  (was: Patch Available)

> VectorizedParquetRecordReader can't to read parquet file generated using 
> thrift/custom tool
> ---
>
> Key: HIVE-21492
> URL: https://issues.apache.org/jira/browse/HIVE-21492
> Project: Hive
>  Issue Type: Bug
>Reporter: Ganesha Shreedhara
>Assignee: Ganesha Shreedhara
>Priority: Major
> Attachments: HIVE-21492.2.patch, HIVE-21492.3.patch, HIVE-21492.patch
>
>
> Taking an example of a parquet table having array of integers as below. 
> {code:java}
> CREATE EXTERNAL TABLE ( list_of_ints` array)
> STORED AS PARQUET 
> LOCATION '{location}';
> {code}
> Parquet file generated using hive will have schema for Type as below:
> {code:java}
> group list_of_ints (LIST) { repeated group bag { optional int32 array;\n};\n} 
> {code}
> Parquet file generated using thrift or any custom tool (using 
> org.apache.parquet.io.api.RecordConsumer)
> may have schema for Type as below:
> {code:java}
> required group list_of_ints (LIST) { repeated int32 list_of_tuple} {code}
> VectorizedParquetRecordReader handles only parquet file generated using hive. 
> It throws the following exception when parquet file generated using thrift is 
> read because of the changes done as part of HIVE-18553 .
> {code:java}
> Caused by: java.lang.ClassCastException: repeated int32 list_of_ints_tuple is 
> not a group
>  at org.apache.parquet.schema.Type.asGroupType(Type.java:207)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.getElementType(VectorizedParquetRecordReader.java:479)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.buildVectorizedParquetReader(VectorizedParquetRecordReader.java:532)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.checkEndOfRowGroup(VectorizedParquetRecordReader.java:440)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.nextBatch(VectorizedParquetRecordReader.java:401)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:353)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:92)
>  at 
> org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.doNext(HiveContextAwareRecordReader.java:365){code}
>  
>  I have done a small change to handle the case where the child type of group 
> type can be PrimitiveType.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-21492) VectorizedParquetRecordReader can't to read parquet file generated using thrift/custom tool

2020-04-01 Thread Ganesha Shreedhara (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-21492?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073324#comment-17073324
 ] 

Ganesha Shreedhara commented on HIVE-21492:
---

Test failures are unrelated and mostly because the metstore server was down.
{code:java}
Could not connect to meta store using any of the URIs provided. Most recent 
failure: org.apache.thrift.transport.TTransportException: 
java.net.ConnectException: Connection refused at 
org.apache.thrift.transport.TSocket.open(TSocket.java:226) at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:686){code}
Can we rerun the tests?

> VectorizedParquetRecordReader can't to read parquet file generated using 
> thrift/custom tool
> ---
>
> Key: HIVE-21492
> URL: https://issues.apache.org/jira/browse/HIVE-21492
> Project: Hive
>  Issue Type: Bug
>Reporter: Ganesha Shreedhara
>Assignee: Ganesha Shreedhara
>Priority: Major
> Attachments: HIVE-21492.2.patch, HIVE-21492.3.patch, HIVE-21492.patch
>
>
> Taking an example of a parquet table having array of integers as below. 
> {code:java}
> CREATE EXTERNAL TABLE ( list_of_ints` array)
> STORED AS PARQUET 
> LOCATION '{location}';
> {code}
> Parquet file generated using hive will have schema for Type as below:
> {code:java}
> group list_of_ints (LIST) { repeated group bag { optional int32 array;\n};\n} 
> {code}
> Parquet file generated using thrift or any custom tool (using 
> org.apache.parquet.io.api.RecordConsumer)
> may have schema for Type as below:
> {code:java}
> required group list_of_ints (LIST) { repeated int32 list_of_tuple} {code}
> VectorizedParquetRecordReader handles only parquet file generated using hive. 
> It throws the following exception when parquet file generated using thrift is 
> read because of the changes done as part of HIVE-18553 .
> {code:java}
> Caused by: java.lang.ClassCastException: repeated int32 list_of_ints_tuple is 
> not a group
>  at org.apache.parquet.schema.Type.asGroupType(Type.java:207)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.getElementType(VectorizedParquetRecordReader.java:479)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.buildVectorizedParquetReader(VectorizedParquetRecordReader.java:532)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.checkEndOfRowGroup(VectorizedParquetRecordReader.java:440)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.nextBatch(VectorizedParquetRecordReader.java:401)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:353)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:92)
>  at 
> org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.doNext(HiveContextAwareRecordReader.java:365){code}
>  
>  I have done a small change to handle the case where the child type of group 
> type can be PrimitiveType.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23004) Support Decimal64 operations across multiple vertices

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23004?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073320#comment-17073320
 ] 

Hive QA commented on HIVE-23004:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
51s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
41s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
14s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
 3s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
41s{color} | {color:blue} serde in master has 197 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m 
38s{color} | {color:blue} ql in master has 1528 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
6s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
29s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
43s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
17s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m 
17s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
19s{color} | {color:red} serde: The patch generated 7 new + 713 unchanged - 1 
fixed = 720 total (was 714) {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
43s{color} | {color:red} ql: The patch generated 3 new + 410 unchanged - 1 
fixed = 413 total (was 411) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  0m 
47s{color} | {color:red} serde generated 4 new + 194 unchanged - 3 fixed = 198 
total (was 197) {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
49s{color} | {color:green} ql generated 0 new + 1527 unchanged - 1 fixed = 1527 
total (was 1528) {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
10s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 28m 30s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| FindBugs | module:serde |
|  |  new 
org.apache.hadoop.hive.serde2.binarysortable.fast.BinarySortableDeserializeRead(TypeInfo[],
 DataTypePhysicalVariation[], boolean, boolean[], byte[], byte[]) may expose 
internal representation by storing an externally mutable object into 
BinarySortableDeserializeRead.columnNotNullMarker  At 
BinarySortableDeserializeRead.java:byte[], byte[]) may expose internal 
representation by storing an externally mutable object into 
BinarySortableDeserializeRead.columnNotNullMarker  At 
BinarySortableDeserializeRead.java:[line 151] |
|  |  new 
org.apache.hadoop.hive.serde2.binarysortable.fast.BinarySortableDeserializeRead(TypeInfo[],
 DataTypePhysicalVariation[], boolean, boolean[], byte[], byte[]) may expose 
internal representation by storing an externally mutable object into 
BinarySortableDeserializeRead.columnNullMarker  At 
BinarySortableDeserializeRead.java:byte[], byte[]) may expose internal 
representation by storing an externally mutable object into 
BinarySortableDeserializeRead.columnNullMarker  At 
BinarySortableDeserializeRead.java:[line 150] |
|  |  new 
org.apache.hadoop.hive.serde2.binarysortable.fast.BinarySortableDeserializeRead(TypeInfo[],
 DataTypePhysicalVariation[], boolean, boolean[], byte[], byte[]) may expose 
internal representation by storing an externally mutable object into 
BinarySortableDeserializeRead.columnSortOrderIsDesc  At 

[jira] [Commented] (HIVE-21492) VectorizedParquetRecordReader can't to read parquet file generated using thrift/custom tool

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-21492?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073305#comment-17073305
 ] 

Hive QA commented on HIVE-21492:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12998415/HIVE-21492.3.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 68 failed/errored test(s), 18162 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.metastore.TestMetastoreHousekeepingLeaderEmptyConfig.testHouseKeepingThreadExistence
 (batchId=252)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.alterTableBogusCatalog[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.createTableInBogusCatalog[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.dropTableBogusCatalog[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.getAllTablesInBogusCatalog[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.getMaterializedViewsInBogusCatalog[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.getTableInBogusCatalog[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.getTableObjectsByNameBogusCatalog[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.moveTablesBetweenCatalogsOnAlter[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.tablesInOtherCatalogs[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.testAlterTableAlreadyExists[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.testAlterTableCascade[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.testAlterTableChangeCols[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.testAlterTableChangingDatabase[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.testAlterTableEmptyTableNameInNew[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.testAlterTableExternalTableChangeLocation[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.testAlterTableExternalTable[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.testAlterTableInvalidStorageDescriptorAddPartitionColumns[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.testAlterTableInvalidStorageDescriptorAlterPartitionColumnName[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.testAlterTableInvalidStorageDescriptorInvalidColumnType[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.testAlterTableInvalidStorageDescriptorNullCols[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.testAlterTableInvalidStorageDescriptorNullColumnType[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.testAlterTableInvalidStorageDescriptorNullLocation[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.testAlterTableInvalidStorageDescriptorNullSerdeInfo[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.testAlterTableInvalidStorageDescriptorRemovePartitionColumn[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.testAlterTableInvalidTableNameInNew[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.testAlterTableNoSuchDatabase[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.testAlterTableNoSuchTableInThisDatabase[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.testAlterTableNoSuchTable[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.testAlterTableNullDatabaseInNew[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.testAlterTableNullDatabase[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.testAlterTableNullNewTable[Remote]
 (batchId=230)
org.apache.hadoop.hive.metastore.client.TestTablesCreateDropAlterTruncate.testAlterTableNullStorageDescriptorInNew[Remote]
 (batchId=230)

[jira] [Commented] (HIVE-21492) VectorizedParquetRecordReader can't to read parquet file generated using thrift/custom tool

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-21492?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073278#comment-17073278
 ] 

Hive QA commented on HIVE-21492:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  9m 
36s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
1s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
39s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m 
41s{color} | {color:blue} ql in master has 1528 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
52s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
22s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
3s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m  
3s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
40s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
57s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
57s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
15s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 24m 32s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-21377/dev-support/hive-personality.sh
 |
| git revision | master / 709235c |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.1 |
| modules | C: ql U: ql |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21377/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> VectorizedParquetRecordReader can't to read parquet file generated using 
> thrift/custom tool
> ---
>
> Key: HIVE-21492
> URL: https://issues.apache.org/jira/browse/HIVE-21492
> Project: Hive
>  Issue Type: Bug
>Reporter: Ganesha Shreedhara
>Assignee: Ganesha Shreedhara
>Priority: Major
> Attachments: HIVE-21492.2.patch, HIVE-21492.3.patch, HIVE-21492.patch
>
>
> Taking an example of a parquet table having array of integers as below. 
> {code:java}
> CREATE EXTERNAL TABLE ( list_of_ints` array)
> STORED AS PARQUET 
> LOCATION '{location}';
> {code}
> Parquet file generated using hive will have schema for Type as below:
> {code:java}
> group list_of_ints (LIST) { repeated group bag { optional int32 array;\n};\n} 
> {code}
> Parquet file generated using thrift or any custom tool (using 
> org.apache.parquet.io.api.RecordConsumer)
> may have schema for Type as below:
> {code:java}
> required group list_of_ints (LIST) { repeated int32 list_of_tuple} {code}
> VectorizedParquetRecordReader handles only parquet file generated using hive. 
> It throws the following exception when parquet file generated using thrift is 
> read because of the changes done as part of HIVE-18553 .
> {code:java}
> Caused by: java.lang.ClassCastException: repeated int32 list_of_ints_tuple is 
> not a group
>  at org.apache.parquet.schema.Type.asGroupType(Type.java:207)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.getElementType(VectorizedParquetRecordReader.java:479)
>  at 
> 

[jira] [Commented] (HIVE-21492) VectorizedParquetRecordReader can't to read parquet file generated using thrift/custom tool

2020-04-01 Thread Ferdinand Xu (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-21492?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073277#comment-17073277
 ] 

Ferdinand Xu commented on HIVE-21492:
-

LGTM +1 pending on the test

> VectorizedParquetRecordReader can't to read parquet file generated using 
> thrift/custom tool
> ---
>
> Key: HIVE-21492
> URL: https://issues.apache.org/jira/browse/HIVE-21492
> Project: Hive
>  Issue Type: Bug
>Reporter: Ganesha Shreedhara
>Assignee: Ganesha Shreedhara
>Priority: Major
> Attachments: HIVE-21492.2.patch, HIVE-21492.3.patch, HIVE-21492.patch
>
>
> Taking an example of a parquet table having array of integers as below. 
> {code:java}
> CREATE EXTERNAL TABLE ( list_of_ints` array)
> STORED AS PARQUET 
> LOCATION '{location}';
> {code}
> Parquet file generated using hive will have schema for Type as below:
> {code:java}
> group list_of_ints (LIST) { repeated group bag { optional int32 array;\n};\n} 
> {code}
> Parquet file generated using thrift or any custom tool (using 
> org.apache.parquet.io.api.RecordConsumer)
> may have schema for Type as below:
> {code:java}
> required group list_of_ints (LIST) { repeated int32 list_of_tuple} {code}
> VectorizedParquetRecordReader handles only parquet file generated using hive. 
> It throws the following exception when parquet file generated using thrift is 
> read because of the changes done as part of HIVE-18553 .
> {code:java}
> Caused by: java.lang.ClassCastException: repeated int32 list_of_ints_tuple is 
> not a group
>  at org.apache.parquet.schema.Type.asGroupType(Type.java:207)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.getElementType(VectorizedParquetRecordReader.java:479)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.buildVectorizedParquetReader(VectorizedParquetRecordReader.java:532)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.checkEndOfRowGroup(VectorizedParquetRecordReader.java:440)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.nextBatch(VectorizedParquetRecordReader.java:401)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:353)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:92)
>  at 
> org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.doNext(HiveContextAwareRecordReader.java:365){code}
>  
>  I have done a small change to handle the case where the child type of group 
> type can be PrimitiveType.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23118) Option for exposing compile time counters as tez counters

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23118?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073270#comment-17073270
 ] 

Hive QA commented on HIVE-23118:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12998390/HIVE-23118.1.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 18163 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[infer_bucket_sort_reducers_power_two]
 (batchId=15)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[tez_compile_counters]
 (batchId=186)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/21376/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/21376/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-21376/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12998390 - PreCommit-HIVE-Build

> Option for exposing compile time counters as tez counters
> -
>
> Key: HIVE-23118
> URL: https://issues.apache.org/jira/browse/HIVE-23118
> Project: Hive
>  Issue Type: Improvement
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
>Priority: Minor
>  Labels: pull-request-available
> Attachments: HIVE-23118.1.patch
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> TezCounters currently are runtime only. Some compile time information from 
> optimizer can be exposed as counters which can then be used by workload 
> management to make runtime decisions. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23126) Add TXNS table to sysdb

2020-04-01 Thread Ramesh Kumar Thangarajan (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23126?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ramesh Kumar Thangarajan updated HIVE-23126:

Attachment: HIVE-23126.2.patch
Status: Patch Available  (was: Open)

> Add TXNS table to sysdb
> ---
>
> Key: HIVE-23126
> URL: https://issues.apache.org/jira/browse/HIVE-23126
> Project: Hive
>  Issue Type: Task
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
> Attachments: HIVE-23126.1.patch, HIVE-23126.2.patch
>
>
> Add TXNS table to sysdb



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23126) Add TXNS table to sysdb

2020-04-01 Thread Ramesh Kumar Thangarajan (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23126?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ramesh Kumar Thangarajan updated HIVE-23126:

Status: Open  (was: Patch Available)

> Add TXNS table to sysdb
> ---
>
> Key: HIVE-23126
> URL: https://issues.apache.org/jira/browse/HIVE-23126
> Project: Hive
>  Issue Type: Task
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
> Attachments: HIVE-23126.1.patch
>
>
> Add TXNS table to sysdb



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23127) Replace listPartitionsByExpr with GetPartitionsWithSpecs in Partition pruner

2020-04-01 Thread Vineet Garg (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23127?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-23127:
---
Status: Patch Available  (was: Open)

> Replace listPartitionsByExpr with GetPartitionsWithSpecs in Partition pruner
> 
>
> Key: HIVE-23127
> URL: https://issues.apache.org/jira/browse/HIVE-23127
> Project: Hive
>  Issue Type: Task
>  Components: HiveServer2
>Reporter: Vineet Garg
>Assignee: Vineet Garg
>Priority: Major
> Attachments: HIVE-23127.1.patch
>
>
> GetPartitionsWithSpecs reduces data transfer by deduplicating storage 
> descriptor



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23127) Replace listPartitionsByExpr with GetPartitionsWithSpecs in Partition pruner

2020-04-01 Thread Vineet Garg (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23127?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-23127:
---
Attachment: HIVE-23127.1.patch

> Replace listPartitionsByExpr with GetPartitionsWithSpecs in Partition pruner
> 
>
> Key: HIVE-23127
> URL: https://issues.apache.org/jira/browse/HIVE-23127
> Project: Hive
>  Issue Type: Task
>  Components: HiveServer2
>Reporter: Vineet Garg
>Assignee: Vineet Garg
>Priority: Major
> Attachments: HIVE-23127.1.patch
>
>
> GetPartitionsWithSpecs reduces data transfer by deduplicating storage 
> descriptor



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (HIVE-23127) Replace listPartitionsByExpr with GetPartitionsWithSpecs in Partition pruner

2020-04-01 Thread Vineet Garg (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23127?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg reassigned HIVE-23127:
--


> Replace listPartitionsByExpr with GetPartitionsWithSpecs in Partition pruner
> 
>
> Key: HIVE-23127
> URL: https://issues.apache.org/jira/browse/HIVE-23127
> Project: Hive
>  Issue Type: Task
>  Components: HiveServer2
>Reporter: Vineet Garg
>Assignee: Vineet Garg
>Priority: Major
>
> GetPartitionsWithSpecs reduces data transfer by deduplicating storage 
> descriptor



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23118) Option for exposing compile time counters as tez counters

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23118?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073248#comment-17073248
 ] 

Hive QA commented on HIVE-23118:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
50s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
51s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
0s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
40s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m 
44s{color} | {color:blue} ql in master has 1528 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
52s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
29s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
25s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
0s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m  
0s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
42s{color} | {color:red} ql: The patch generated 4 new + 88 unchanged - 0 fixed 
= 92 total (was 88) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 1s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
50s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
54s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 25m  6s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-21376/dev-support/hive-personality.sh
 |
| git revision | master / 709235c |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.1 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21376/yetus/diff-checkstyle-ql.txt
 |
| modules | C: ql itests U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21376/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Option for exposing compile time counters as tez counters
> -
>
> Key: HIVE-23118
> URL: https://issues.apache.org/jira/browse/HIVE-23118
> Project: Hive
>  Issue Type: Improvement
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
>Priority: Minor
>  Labels: pull-request-available
> Attachments: HIVE-23118.1.patch
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> TezCounters currently are runtime only. Some compile time information from 
> optimizer can be exposed as counters which can then be used by workload 
> management to make runtime decisions. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-21328) Call To Hadoop Text getBytes() Without Call to getLength()

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21328?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor updated HIVE-21328:
--
Status: Patch Available  (was: Open)

> Call To Hadoop Text getBytes() Without Call to getLength()
> --
>
> Key: HIVE-21328
> URL: https://issues.apache.org/jira/browse/HIVE-21328
> Project: Hive
>  Issue Type: Bug
>  Components: Query Planning
>Affects Versions: 4.0.0, 3.2.0
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Major
> Attachments: HIVE-21328.1.patch
>
>
> I'm not sure if there is actually a bug, but this looks highly suspect:
> {code:java}
>   public Object set(final Object o, final Text text) {
> return new BytesWritable(text == null ? null : text.getBytes());
>   }
> {code}
> https://github.com/apache/hive/blob/master/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/serde/primitive/ParquetStringInspector.java#L104-L106
> There are two components to a Text object.  There are the internal bytes and 
> the length of the bytes.  The two are independent.  I.e., a quick "reset" on 
> the Text object simply sets the internal length counter to zero.  This code 
> is potentially looking at obsolete data that it shouldn't be seeing because 
> it is not considering the length of the Text.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23117) Review of HiveStatement Class

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23117?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073230#comment-17073230
 ] 

Hive QA commented on HIVE-23117:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12998383/HIVE-23117.1.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 18162 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/21375/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/21375/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-21375/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12998383 - PreCommit-HIVE-Build

> Review of HiveStatement Class
> -
>
> Key: HIVE-23117
> URL: https://issues.apache.org/jira/browse/HIVE-23117
> Project: Hive
>  Issue Type: Improvement
>  Components: JDBC
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Minor
> Attachments: HIVE-23117.1.patch
>
>
> * Remove unused instance variable(s)
>  * Remove non-JavaDoc comments
>  * Make inPlaceUpdateStream Optional (and remove NO-OP class) (inconsistent 
> behavior with 'null' values)
>  * {{getQueryTimeout()}} returns incorrect value
>  * Unify and improve Exception messages
>  * Checkstyle fixes



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23096) Review Code Path for getResults

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23096?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor updated HIVE-23096:
--
Fix Version/s: 4.0.0
   Resolution: Fixed
   Status: Resolved  (was: Patch Available)

Pushed to master.  Thanks [~ngangam] for the review!

> Review Code Path for getResults
> ---
>
> Key: HIVE-23096
> URL: https://issues.apache.org/jira/browse/HIVE-23096
> Project: Hive
>  Issue Type: Sub-task
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Minor
> Fix For: 4.0.0
>
> Attachments: HIVE-23096.1.patch
>
>
> * Log when a buffer returned to a client is completely full.  This can be 
> helpful for identifying misconfigured clients that have a fetchSize that is 
> too small
> * See below code snippet
> * Rows are loaded into an ArrayList called "convey."  This ArrayList is 
> instantiated with the default size (10) and is expanded on-demand as it is 
> filled.  I changed it so that this ArrayList's size is set once, on the first 
> call to fetch, so that it doesn't have to "grow" and waste cycles and memory 
> with GC thrashing for that.
> * Some minor cleanup
> {code:java|Driver.java}
>  try {
> ss = Utilities.readColumn(driverContext.getResStream(), bos);
> if (bos.getLength() > 0) {
>   // use JDK StandardCharsets
>   row = new String(bos.getData(), 0, bos.getLength(), "UTF-8");
> } else if (ss == Utilities.StreamStatus.TERMINATED) {
>   // Do not create a new string for this.  Just use a hard-coded 
> empty string.  JDK: "Note that use of this constructor is unnecessary since 
> Strings are immutable."
>   row = new String();
> }
> ...
> }
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23117) Review of HiveStatement Class

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23117?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073198#comment-17073198
 ] 

Hive QA commented on HIVE-23117:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  9m 
32s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
17s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
12s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
30s{color} | {color:blue} jdbc in master has 16 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
13s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
40s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
18s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
18s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
11s{color} | {color:green} jdbc: The patch generated 0 new + 6 unchanged - 1 
fixed = 6 total (was 7) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
36s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
13s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
15s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 13m 24s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-21375/dev-support/hive-personality.sh
 |
| git revision | master / 709235c |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.1 |
| modules | C: jdbc U: jdbc |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21375/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Review of HiveStatement Class
> -
>
> Key: HIVE-23117
> URL: https://issues.apache.org/jira/browse/HIVE-23117
> Project: Hive
>  Issue Type: Improvement
>  Components: JDBC
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Minor
> Attachments: HIVE-23117.1.patch
>
>
> * Remove unused instance variable(s)
>  * Remove non-JavaDoc comments
>  * Make inPlaceUpdateStream Optional (and remove NO-OP class) (inconsistent 
> behavior with 'null' values)
>  * {{getQueryTimeout()}} returns incorrect value
>  * Unify and improve Exception messages
>  * Checkstyle fixes



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23106) Cleanup CalcitePlanner genOPTree exception handling

2020-04-01 Thread Jesus Camacho Rodriguez (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23106?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073191#comment-17073191
 ] 

Jesus Camacho Rodriguez commented on HIVE-23106:


+1

> Cleanup CalcitePlanner genOPTree exception handling
> ---
>
> Key: HIVE-23106
> URL: https://issues.apache.org/jira/browse/HIVE-23106
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2
>Reporter: John Sherman
>Assignee: John Sherman
>Priority: Minor
> Attachments: HIVE-23106.1.patch, HIVE-23106.2.patch
>
>
> The logic where genOPTree handles exceptions during CBO planning is a bit 
> twisty and could use some cleanup and comments.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23106) Cleanup CalcitePlanner genOPTree exception handling

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23106?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073189#comment-17073189
 ] 

Hive QA commented on HIVE-23106:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12998377/HIVE-23106.2.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 18162 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/21374/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/21374/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-21374/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12998377 - PreCommit-HIVE-Build

> Cleanup CalcitePlanner genOPTree exception handling
> ---
>
> Key: HIVE-23106
> URL: https://issues.apache.org/jira/browse/HIVE-23106
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2
>Reporter: John Sherman
>Assignee: John Sherman
>Priority: Minor
> Attachments: HIVE-23106.1.patch, HIVE-23106.2.patch
>
>
> The logic where genOPTree handles exceptions during CBO planning is a bit 
> twisty and could use some cleanup and comments.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23106) Cleanup CalcitePlanner genOPTree exception handling

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23106?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073159#comment-17073159
 ] 

Hive QA commented on HIVE-23106:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  9m 
26s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
1s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
40s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m 
43s{color} | {color:blue} ql in master has 1528 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
55s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
28s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
2s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m  
2s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
42s{color} | {color:green} ql: The patch generated 0 new + 115 unchanged - 7 
fixed = 115 total (was 122) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  4m  
0s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
56s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
15s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 24m 39s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-21374/dev-support/hive-personality.sh
 |
| git revision | master / 709235c |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.1 |
| modules | C: ql U: ql |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21374/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Cleanup CalcitePlanner genOPTree exception handling
> ---
>
> Key: HIVE-23106
> URL: https://issues.apache.org/jira/browse/HIVE-23106
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2
>Reporter: John Sherman
>Assignee: John Sherman
>Priority: Minor
> Attachments: HIVE-23106.1.patch, HIVE-23106.2.patch
>
>
> The logic where genOPTree handles exceptions during CBO planning is a bit 
> twisty and could use some cleanup and comments.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23126) Add TXNS table to sysdb

2020-04-01 Thread Ramesh Kumar Thangarajan (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23126?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ramesh Kumar Thangarajan updated HIVE-23126:

Attachment: HIVE-23126.1.patch
Status: Patch Available  (was: Open)

> Add TXNS table to sysdb
> ---
>
> Key: HIVE-23126
> URL: https://issues.apache.org/jira/browse/HIVE-23126
> Project: Hive
>  Issue Type: Task
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
> Attachments: HIVE-23126.1.patch
>
>
> Add TXNS table to sysdb



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (HIVE-23126) Add TXNS table to sysdb

2020-04-01 Thread Ramesh Kumar Thangarajan (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23126?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ramesh Kumar Thangarajan reassigned HIVE-23126:
---


> Add TXNS table to sysdb
> ---
>
> Key: HIVE-23126
> URL: https://issues.apache.org/jira/browse/HIVE-23126
> Project: Hive
>  Issue Type: Task
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
>
> Add TXNS table to sysdb



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23094) Implement Explain CBO of Update and Delete statements

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23094?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073141#comment-17073141
 ] 

Hive QA commented on HIVE-23094:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12998369/HIVE-23094.1.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 18162 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/21373/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/21373/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-21373/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12998369 - PreCommit-HIVE-Build

> Implement Explain CBO of Update and Delete statements
> -
>
> Key: HIVE-23094
> URL: https://issues.apache.org/jira/browse/HIVE-23094
> Project: Hive
>  Issue Type: Improvement
>  Components: CBO
>Reporter: Krisztian Kasa
>Assignee: Krisztian Kasa
>Priority: Major
> Attachments: HIVE-23094.1.patch
>
>
> {code}
> set hive.support.concurrency=true;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
> create table acidtlb(a int, b int) clustered by (a) into 2 buckets stored as 
> orc TBLPROPERTIES ('transactional'='true');
> explain cbo
> update acidtlb set b=777;
> {code}
> doesn't print CBO plan.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-21328) Call To Hadoop Text getBytes() Without Call to getLength()

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21328?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor updated HIVE-21328:
--
Attachment: HIVE-21328.1.patch

> Call To Hadoop Text getBytes() Without Call to getLength()
> --
>
> Key: HIVE-21328
> URL: https://issues.apache.org/jira/browse/HIVE-21328
> Project: Hive
>  Issue Type: Bug
>  Components: Query Planning
>Affects Versions: 4.0.0, 3.2.0
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Major
> Attachments: HIVE-21328.1.patch
>
>
> I'm not sure if there is actually a bug, but this looks highly suspect:
> {code:java}
>   public Object set(final Object o, final Text text) {
> return new BytesWritable(text == null ? null : text.getBytes());
>   }
> {code}
> https://github.com/apache/hive/blob/master/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/serde/primitive/ParquetStringInspector.java#L104-L106
> There are two components to a Text object.  There are the internal bytes and 
> the length of the bytes.  The two are independent.  I.e., a quick "reset" on 
> the Text object simply sets the internal length counter to zero.  This code 
> is potentially looking at obsolete data that it shouldn't be seeing because 
> it is not considering the length of the Text.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (HIVE-21328) Call To Hadoop Text getBytes() Without Call to getLength()

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21328?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor reassigned HIVE-21328:
-

Assignee: David Mollitor

> Call To Hadoop Text getBytes() Without Call to getLength()
> --
>
> Key: HIVE-21328
> URL: https://issues.apache.org/jira/browse/HIVE-21328
> Project: Hive
>  Issue Type: Bug
>  Components: Query Planning
>Affects Versions: 4.0.0, 3.2.0
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Major
> Attachments: HIVE-21328.1.patch
>
>
> I'm not sure if there is actually a bug, but this looks highly suspect:
> {code:java}
>   public Object set(final Object o, final Text text) {
> return new BytesWritable(text == null ? null : text.getBytes());
>   }
> {code}
> https://github.com/apache/hive/blob/master/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/serde/primitive/ParquetStringInspector.java#L104-L106
> There are two components to a Text object.  There are the internal bytes and 
> the length of the bytes.  The two are independent.  I.e., a quick "reset" on 
> the Text object simply sets the internal length counter to zero.  This code 
> is potentially looking at obsolete data that it shouldn't be seeing because 
> it is not considering the length of the Text.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23067) Use batch DB calls in TxnHandler for commitTxn and abortTxns

2020-04-01 Thread Marton Bod (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23067?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Marton Bod updated HIVE-23067:
--
Attachment: (was: HIVE-23067.3.patch)

> Use batch DB calls in TxnHandler for commitTxn and abortTxns
> 
>
> Key: HIVE-23067
> URL: https://issues.apache.org/jira/browse/HIVE-23067
> Project: Hive
>  Issue Type: Improvement
>Reporter: Marton Bod
>Assignee: Marton Bod
>Priority: Major
> Attachments: HIVE-23067.1.patch, HIVE-23067.1.patch, 
> HIVE-23067.1.patch, HIVE-23067.1.patch, HIVE-23067.2.patch, 
> HIVE-23067.2.patch, HIVE-23067.4.patch
>
>
> To reduce the number of database calls and network roundtrips, we could use 
> more batching in TxnHandler, where currently in many places we call insert 
> commands in loops sequentially.
> Some examples:
>  * openTxns (TXNS, REPL_TXN_MAP)
>  * commitTxn (COMPLETED_TXN_COMPONENTS)
>  * replTableWriteIdState (TXN_TO_WRITE_ID)
>  * allocateTableWriteIds (TXN_TO_WRITE_ID)
>  * 
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23067) Use batch DB calls in TxnHandler for commitTxn and abortTxns

2020-04-01 Thread Marton Bod (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23067?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Marton Bod updated HIVE-23067:
--
Attachment: HIVE-23067.4.patch

> Use batch DB calls in TxnHandler for commitTxn and abortTxns
> 
>
> Key: HIVE-23067
> URL: https://issues.apache.org/jira/browse/HIVE-23067
> Project: Hive
>  Issue Type: Improvement
>Reporter: Marton Bod
>Assignee: Marton Bod
>Priority: Major
> Attachments: HIVE-23067.1.patch, HIVE-23067.1.patch, 
> HIVE-23067.1.patch, HIVE-23067.1.patch, HIVE-23067.2.patch, 
> HIVE-23067.2.patch, HIVE-23067.4.patch
>
>
> To reduce the number of database calls and network roundtrips, we could use 
> more batching in TxnHandler, where currently in many places we call insert 
> commands in loops sequentially.
> Some examples:
>  * openTxns (TXNS, REPL_TXN_MAP)
>  * commitTxn (COMPLETED_TXN_COMPONENTS)
>  * replTableWriteIdState (TXN_TO_WRITE_ID)
>  * allocateTableWriteIds (TXN_TO_WRITE_ID)
>  * 
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23067) Use batch DB calls in TxnHandler for commitTxn and abortTxns

2020-04-01 Thread Marton Bod (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23067?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Marton Bod updated HIVE-23067:
--
Attachment: HIVE-23067.3.patch

> Use batch DB calls in TxnHandler for commitTxn and abortTxns
> 
>
> Key: HIVE-23067
> URL: https://issues.apache.org/jira/browse/HIVE-23067
> Project: Hive
>  Issue Type: Improvement
>Reporter: Marton Bod
>Assignee: Marton Bod
>Priority: Major
> Attachments: HIVE-23067.1.patch, HIVE-23067.1.patch, 
> HIVE-23067.1.patch, HIVE-23067.1.patch, HIVE-23067.2.patch, 
> HIVE-23067.2.patch, HIVE-23067.3.patch
>
>
> To reduce the number of database calls and network roundtrips, we could use 
> more batching in TxnHandler, where currently in many places we call insert 
> commands in loops sequentially.
> Some examples:
>  * openTxns (TXNS, REPL_TXN_MAP)
>  * commitTxn (COMPLETED_TXN_COMPONENTS)
>  * replTableWriteIdState (TXN_TO_WRITE_ID)
>  * allocateTableWriteIds (TXN_TO_WRITE_ID)
>  * 
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23094) Implement Explain CBO of Update and Delete statements

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23094?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073085#comment-17073085
 ] 

Hive QA commented on HIVE-23094:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  9m 
13s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
1s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
39s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m 
45s{color} | {color:blue} ql in master has 1528 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
53s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
28s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
59s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
59s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
41s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
53s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
54s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 24m 13s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-21373/dev-support/hive-personality.sh
 |
| git revision | master / 709235c |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.1 |
| modules | C: ql U: ql |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21373/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Implement Explain CBO of Update and Delete statements
> -
>
> Key: HIVE-23094
> URL: https://issues.apache.org/jira/browse/HIVE-23094
> Project: Hive
>  Issue Type: Improvement
>  Components: CBO
>Reporter: Krisztian Kasa
>Assignee: Krisztian Kasa
>Priority: Major
> Attachments: HIVE-23094.1.patch
>
>
> {code}
> set hive.support.concurrency=true;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
> create table acidtlb(a int, b int) clustered by (a) into 2 buckets stored as 
> orc TBLPROPERTIES ('transactional'='true');
> explain cbo
> update acidtlb set b=777;
> {code}
> doesn't print CBO plan.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23039) Checkpointing for repl dump bootstrap phase

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23039?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073058#comment-17073058
 ] 

Hive QA commented on HIVE-23039:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12998454/HIVE-23039.02.patch

{color:green}SUCCESS:{color} +1 due to 3 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 8 failed/errored test(s), 18168 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.ql.parse.TestReplWithJsonMessageFormat.testCheckPointing 
(batchId=260)
org.apache.hadoop.hive.ql.parse.TestReplWithJsonMessageFormat.testCheckPointingInDumpFailure
 (batchId=260)
org.apache.hadoop.hive.ql.parse.TestReplicationOnHDFSEncryptedZones.targetAndSourceHaveDifferentEncryptionZoneKeys
 (batchId=283)
org.apache.hadoop.hive.ql.parse.TestReplicationScenarios.testCheckPointing 
(batchId=270)
org.apache.hadoop.hive.ql.parse.TestReplicationScenarios.testCheckPointingInDumpFailure
 (batchId=270)
org.apache.hadoop.hive.ql.parse.TestReplicationScenariosExternalTables.externalTableIncrementalReplication
 (batchId=268)
org.apache.hadoop.hive.ql.parse.TestReplicationScenariosExternalTables.externalTableReplicationWithCustomPaths
 (batchId=268)
org.apache.hadoop.hive.ql.parse.TestReplicationScenariosExternalTables.externalTableWithPartitions
 (batchId=268)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/21372/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/21372/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-21372/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 8 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12998454 - PreCommit-HIVE-Build

> Checkpointing for repl dump bootstrap phase
> ---
>
> Key: HIVE-23039
> URL: https://issues.apache.org/jira/browse/HIVE-23039
> Project: Hive
>  Issue Type: Bug
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23039.01.patch, HIVE-23039.02.patch, 
> HIVE-23039.03.patch
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23039) Checkpointing for repl dump bootstrap phase

2020-04-01 Thread Aasha Medhi (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23039?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aasha Medhi updated HIVE-23039:
---
Attachment: HIVE-23039.03.patch
Status: Patch Available  (was: In Progress)

> Checkpointing for repl dump bootstrap phase
> ---
>
> Key: HIVE-23039
> URL: https://issues.apache.org/jira/browse/HIVE-23039
> Project: Hive
>  Issue Type: Bug
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23039.01.patch, HIVE-23039.02.patch, 
> HIVE-23039.03.patch
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23039) Checkpointing for repl dump bootstrap phase

2020-04-01 Thread Aasha Medhi (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23039?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aasha Medhi updated HIVE-23039:
---
Status: In Progress  (was: Patch Available)

> Checkpointing for repl dump bootstrap phase
> ---
>
> Key: HIVE-23039
> URL: https://issues.apache.org/jira/browse/HIVE-23039
> Project: Hive
>  Issue Type: Bug
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23039.01.patch, HIVE-23039.02.patch
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Resolved] (HIVE-18897) Hive is Double-Logging Invalid UDF Error

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-18897?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor resolved HIVE-18897.
---
Resolution: Not A Problem

Based on the current master branch, it looks like the logging was fixed to 
remove the ERROR log in the {{Driver}} class.

> Hive is Double-Logging Invalid UDF Error
> 
>
> Key: HIVE-18897
> URL: https://issues.apache.org/jira/browse/HIVE-18897
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2
>Affects Versions: 3.0.0, 2.4.0
>Reporter: David Mollitor
>Assignee: Ramesh Kumar Thangarajan
>Priority: Minor
>  Labels: newbie, noob
>
> It logs the "invalid" function twice.  Once at ERROR level and once at WARN 
> level.  Please change so that the error is logged once, at the WARN level.  
> The stack trace seems overkill here as well for such a trivial error... 
> usually a user typo or the function needs to be registered.
> {code:java}
> 2018-03-05 07:50:44,473  ERROR org.apache.hadoop.hive.ql.Driver: 
> [HiveServer2-Handler-Pool: Thread-43]: FAILED: SemanticException [Error 
> 10011]: Line 1:7 Invalid function 'aes_encrypt'
> org.apache.hadoop.hive.ql.parse.SemanticException: Line 1:7 Invalid function 
> 'aes_encrypt'
>   at 
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$DefaultExprProcessor.getXpathOrFuncExprNodeDesc(TypeCheckProcFactory.java:836)
>   at 
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$DefaultExprProcessor.process(TypeCheckProcFactory.java:1176)
>   at 
> org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:90)
>   at 
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatchAndReturn(DefaultGraphWalker.java:94)
>   at 
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:78)
>   at 
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:132)
>   at 
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:109)
>   at 
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory.genExprNode(TypeCheckProcFactory.java:193)
>   at 
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory.genExprNode(TypeCheckProcFactory.java:146)
>   at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genAllExprNodeDesc(SemanticAnalyzer.java:10422)
>   at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:10378)
>   at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genSelectPlan(SemanticAnalyzer.java:3771)
>   at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genSelectPlan(SemanticAnalyzer.java:3550)
>   at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPostGroupByBodyPlan(SemanticAnalyzer.java:8830)
>   at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:8785)
>   at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:9652)
>   at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:9545)
>   at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genOPTree(SemanticAnalyzer.java:10018)
>   at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:10029)
>   at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:9909)
>   at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:223)
>   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:488)
>   at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1274)
>   at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:1261)
>   at 
> org.apache.hive.service.cli.operation.SQLOperation.prepare(SQLOperation.java:143)
>   at 
> org.apache.hive.service.cli.operation.SQLOperation.runInternal(SQLOperation.java:215)
>   at 
> org.apache.hive.service.cli.operation.Operation.run(Operation.java:337)
>   at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:425)
>   at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync(HiveSessionImpl.java:402)
>   at 
> org.apache.hive.service.cli.CLIService.executeStatementAsync(CLIService.java:258)
>   at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:500)
>   at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1313)
>   at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1298)
>   at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>   at 

[jira] [Resolved] (HIVE-18956) AvroSerDe Race Condition

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-18956?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor resolved HIVE-18956.
---
Resolution: Won't Fix

> AvroSerDe Race Condition
> 
>
> Key: HIVE-18956
> URL: https://issues.apache.org/jira/browse/HIVE-18956
> Project: Hive
>  Issue Type: Bug
>  Components: Serializers/Deserializers
>Affects Versions: 3.0.0, 2.3.2
>Reporter: David Mollitor
>Priority: Trivial
>
> {code}
>   @Override
>   public Writable serialize(Object o, ObjectInspector objectInspector) throws 
> SerDeException {
> if(badSchema) {
>   throw new BadSchemaException();
> }
> return getSerializer().serialize(o, objectInspector, columnNames, 
> columnTypes, schema);
>   }
>   @Override
>   public Object deserialize(Writable writable) throws SerDeException {
> if(badSchema) {
>   throw new BadSchemaException();
> }
> return getDeserializer().deserialize(columnNames, columnTypes, writable, 
> schema);
>   }
> ...
>   private AvroDeserializer getDeserializer() {
> if(avroDeserializer == null) {
>   avroDeserializer = new AvroDeserializer();
> }
> return avroDeserializer;
>   }
>   private AvroSerializer getSerializer() {
> if(avroSerializer == null) {
>   avroSerializer = new AvroSerializer();
> }
> return avroSerializer;
>   }
> {code}
> {{getDeserializer}} and {{getSerializer}} methods are not thread safe, so 
> neither are {{deserialize}} and {{serialize}} methods.  It probably didn't 
> matter with MapReduce, but now that we have Spark/Tez, it may be an issue.
> You could visualize a scenario where three threads all enter 
> {{getSerializer}} and all see that {{avroSerializer}} is _null_ and create 
> three instances, then they would fight to assign the new object to the 
> {{avroSerializer}} variable.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23039) Checkpointing for repl dump bootstrap phase

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23039?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073001#comment-17073001
 ] 

Hive QA commented on HIVE-23039:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
46s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
 7s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
44s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
 8s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m 
44s{color} | {color:blue} ql in master has 1528 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
40s{color} | {color:blue} itests/hive-unit in master has 2 extant Findbugs 
warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
22s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
29s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  2m 
 8s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
47s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m 
47s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
40s{color} | {color:red} ql: The patch generated 9 new + 59 unchanged - 0 fixed 
= 68 total (was 59) {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
23s{color} | {color:red} itests/hive-unit: The patch generated 2 new + 688 
unchanged - 0 fixed = 690 total (was 688) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  4m 
41s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
24s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 31m  9s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-21372/dev-support/hive-personality.sh
 |
| git revision | master / 709235c |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.1 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21372/yetus/diff-checkstyle-ql.txt
 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21372/yetus/diff-checkstyle-itests_hive-unit.txt
 |
| modules | C: ql itests/hive-unit U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21372/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Checkpointing for repl dump bootstrap phase
> ---
>
> Key: HIVE-23039
> URL: https://issues.apache.org/jira/browse/HIVE-23039
> Project: Hive
>  Issue Type: Bug
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23039.01.patch, HIVE-23039.02.patch
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23114) Insert overwrite with dynamic partitioning is not working correctly with direct insert

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23114?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072955#comment-17072955
 ] 

Hive QA commented on HIVE-23114:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12998356/HIVE-23114.1.patch

{color:green}SUCCESS:{color} +1 due to 2 test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 18166 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/21371/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/21371/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-21371/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12998356 - PreCommit-HIVE-Build

> Insert overwrite with dynamic partitioning is not working correctly with 
> direct insert
> --
>
> Key: HIVE-23114
> URL: https://issues.apache.org/jira/browse/HIVE-23114
> Project: Hive
>  Issue Type: Bug
>Reporter: Marta Kuczora
>Assignee: Marta Kuczora
>Priority: Major
> Attachments: HIVE-23114.1.patch
>
>
> This is a follow-up Jira for the 
> [conversation|https://issues.apache.org/jira/browse/HIVE-21164?focusedCommentId=17059280=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-17059280]
>  in HIVE-21164
>  Doing an insert overwrite from a multi-insert statement with dynamic 
> partitioning will give wrong results for ACID tables when 
> 'hive.acid.direct.insert.enabled' is true or for insert-only tables.
> Reproduction:
> {noformat}
> set hive.acid.direct.insert.enabled=true;
> set hive.support.concurrency=true;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
> set hive.vectorized.execution.enabled=false;
> set hive.stats.autogather=false;
> create external table multiinsert_test_text (a int, b int, c int) stored as 
> textfile;
> insert into multiinsert_test_text values (, 11, ), (, 22, ), 
> (, 33, ), (, 44, NULL), (, 55, NULL);
> create table multiinsert_test_acid (a int, b int) partitioned by (c int) 
> stored as orc tblproperties('transactional'='true');
> create table multiinsert_test_mm (a int, b int) partitioned by (c int) stored 
> as orc tblproperties('transactional'='true', 
> 'transactional_properties'='insert_only');
> from multiinsert_test_text a
> insert overwrite table multiinsert_test_acid partition (c)
> select
>  a.a,
>  a.b,
>  a.c
>  where a.c is not null
> insert overwrite table multiinsert_test_acid partition (c)
> select
>  a.a,
>  a.b,
>  a.c
> where a.c is null;
> select * from multiinsert_test_acid;
> from multiinsert_test_text a
> insert overwrite table multiinsert_test_mm partition (c)
> select
>  a.a,
>  a.b,
>  a.c
>  where a.c is not null
> insert overwrite table multiinsert_test_mm partition (c)
> select
>  a.a,
>  a.b,
>  a.c
> where a.c is null;
> select * from multiinsert_test_mm;
> {noformat}
> The result of these steps can be different, it depends on the execution order 
> of the FileSinkOperators of the insert overwrite statements. It can happen 
> that an error occurs due to manifest file collision, it can happen that no 
> error occurs but the result will be incorrect.
>  Running the same insert query with an external table of with and ACID table 
> with 'hive.acid.direct.insert.enabled=false' will give the follwing result:
> {noformat}
> 11  
> 22  
> 33  
> 44  NULL
> 55  NULL
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23114) Insert overwrite with dynamic partitioning is not working correctly with direct insert

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23114?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072883#comment-17072883
 ] 

Hive QA commented on HIVE-23114:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
42s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
57s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
0s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
44s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m 
45s{color} | {color:blue} ql in master has 1528 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
54s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
29s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
26s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
2s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m  
2s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
44s{color} | {color:red} ql: The patch generated 1 new + 314 unchanged - 0 
fixed = 315 total (was 314) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  3m 
54s{color} | {color:red} ql generated 1 new + 1528 unchanged - 0 fixed = 1529 
total (was 1528) {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
55s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 25m 25s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| FindBugs | module:ql |
|  |  The field 
org.apache.hadoop.hive.ql.exec.FileSinkOperator.dynamicPartitionSpecs is 
transient but isn't set by deserialization  In FileSinkOperator.java:but isn't 
set by deserialization  In FileSinkOperator.java |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-21371/dev-support/hive-personality.sh
 |
| git revision | master / 709235c |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.1 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21371/yetus/diff-checkstyle-ql.txt
 |
| findbugs | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21371/yetus/new-findbugs-ql.html
 |
| modules | C: ql itests U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21371/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Insert overwrite with dynamic partitioning is not working correctly with 
> direct insert
> --
>
> Key: HIVE-23114
> URL: https://issues.apache.org/jira/browse/HIVE-23114
> Project: Hive
>  Issue Type: Bug
>Reporter: Marta Kuczora
>Assignee: Marta Kuczora
>Priority: Major
> Attachments: HIVE-23114.1.patch
>
>
> This is a follow-up Jira for the 
> [conversation|https://issues.apache.org/jira/browse/HIVE-21164?focusedCommentId=17059280=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-17059280]
>  in HIVE-21164
>  Doing an insert overwrite from a multi-insert statement with dynamic 
> partitioning will give wrong results for ACID tables 

[jira] [Commented] (HIVE-23109) Query-based compaction omits database

2020-04-01 Thread Jira


[ 
https://issues.apache.org/jira/browse/HIVE-23109?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072869#comment-17072869
 ] 

László Pintér commented on HIVE-23109:
--

+1

> Query-based compaction omits database
> -
>
> Key: HIVE-23109
> URL: https://issues.apache.org/jira/browse/HIVE-23109
> Project: Hive
>  Issue Type: Bug
>Reporter: Karen Coppage
>Assignee: Karen Coppage
>Priority: Major
> Attachments: HIVE-23109.01.patch, HIVE-23109.02.patch
>
>
> E.g. MM major compaction query looks like:
> {code:java}
> insert into tmp_table select * from src_table;
> {code}
> it should be
> {code:java}
> insert into tmp_table select * from src_db.src_table;
> {code}
> Therefore compaction fails if db of source table isn't default.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23109) Query-based compaction omits database

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23109?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072850#comment-17072850
 ] 

Hive QA commented on HIVE-23109:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12998353/HIVE-23109.02.patch

{color:green}SUCCESS:{color} +1 due to 3 test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 18166 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/21370/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/21370/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-21370/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12998353 - PreCommit-HIVE-Build

> Query-based compaction omits database
> -
>
> Key: HIVE-23109
> URL: https://issues.apache.org/jira/browse/HIVE-23109
> Project: Hive
>  Issue Type: Bug
>Reporter: Karen Coppage
>Assignee: Karen Coppage
>Priority: Major
> Attachments: HIVE-23109.01.patch, HIVE-23109.02.patch
>
>
> E.g. MM major compaction query looks like:
> {code:java}
> insert into tmp_table select * from src_table;
> {code}
> it should be
> {code:java}
> insert into tmp_table select * from src_db.src_table;
> {code}
> Therefore compaction fails if db of source table isn't default.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23122) LLAP: TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Jira


[ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072821#comment-17072821
 ] 

László Bodor commented on HIVE-23122:
-

cc: [~rajesh.balamohan]

> LLAP: TaskExecutorService should log details about task eviction decision 
> details
> -
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-23122.02.patch
>
>
> TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
> another. Under the hood, the queue uses a configurable 
> [comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
>  The currently available comparators typically use the following properties 
> of a task(wrapper):
> getWithinDagPriority: related to vertex
> currentAttemptStartTime
> firstAttemptStartTime
> knownPending: remaining upstream tasks
> The problem is, when an eviction happens, the INFO level message doesn't 
> provide any insight about the decision, only attempts ids like below:
> {code}
> attempt_1585248378306_0010_72_02_96_8 evicted from wait queue in favor of 
> attempt_1585248378306_0003_175_02_79_175 because of lower priority
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23098) Allow Operation assertState to Accept a Collection

2020-04-01 Thread David Mollitor (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23098?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072814#comment-17072814
 ] 

David Mollitor commented on HIVE-23098:
---

[~ngangam] Here is another trivial review if you have any spare cycles (lol).

> Allow Operation assertState to Accept a Collection
> --
>
> Key: HIVE-23098
> URL: https://issues.apache.org/jira/browse/HIVE-23098
> Project: Hive
>  Issue Type: Improvement
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Minor
> Attachments: HIVE-23098.1.patch, HIVE-23098.2.patch, 
> HIVE-23098.2.patch, HIVE-23098.2.patch, HIVE-23098.3.patch
>
>
> {code:java|title=Operation.java}
>   protected final void assertState(List states) throws 
> HiveSQLException {
> if (!states.contains(state)) {
>   throw new HiveSQLException("Expected states: " + states.toString() + ", 
> but found "
>   + this.state);
> }
> this.lastAccessTime = System.currentTimeMillis();
>   }
> /*/
> public void someMethod() {
> assertState(new 
> ArrayList(Arrays.asList(OperationState.FINISHED)));
> }
> {code}
> By allowing {{assertState}} to accept a {{Collection}}, one can save an 
> allocation and simplify the code:
> {code:java}
> assertState(Collections.singleton(OperationState.FINISHED));
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Resolved] (HIVE-22484) Remove Calls to printStackTrace

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22484?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor resolved HIVE-22484.
---
Fix Version/s: 4.0.0
   Resolution: Fixed

> Remove Calls to printStackTrace
> ---
>
> Key: HIVE-22484
> URL: https://issues.apache.org/jira/browse/HIVE-22484
> Project: Hive
>  Issue Type: Improvement
>Affects Versions: 3.2.0
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Major
> Fix For: 4.0.0
>
>
> In many cases, the call to {{printStackTrace}} bypasses the logging 
> framework, in other cases, the error stack trace is printed and the exception 
> is re-thrown (log-and-throw is a bad pattern), and then there are some other 
> edge cases.
> Remove this call and replace with calls to the logging framework or remove 
> altogether if exception is wrapped and re-thrown.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23124) Review of SQLOperation Class

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23124?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor updated HIVE-23124:
--
Description: 
* Use ConcurrentHashMap instead of synchronized methods to improve 
multi-threaded access
 * Use JDK 8 facilities where applicable
 * General cleanup
 * Better log messages and Exception messages
 * Use {{switch}} statement instead of if/else blocks
 * Checkstyle fixes

> Review of SQLOperation Class
> 
>
> Key: HIVE-23124
> URL: https://issues.apache.org/jira/browse/HIVE-23124
> Project: Hive
>  Issue Type: Improvement
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Major
> Attachments: HIVE-23124.1.patch
>
>
> * Use ConcurrentHashMap instead of synchronized methods to improve 
> multi-threaded access
>  * Use JDK 8 facilities where applicable
>  * General cleanup
>  * Better log messages and Exception messages
>  * Use {{switch}} statement instead of if/else blocks
>  * Checkstyle fixes



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23124) Review of SQLOperation Class

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23124?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor updated HIVE-23124:
--
Attachment: HIVE-23124.1.patch

> Review of SQLOperation Class
> 
>
> Key: HIVE-23124
> URL: https://issues.apache.org/jira/browse/HIVE-23124
> Project: Hive
>  Issue Type: Improvement
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Major
> Attachments: HIVE-23124.1.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23089) Add constraint checks to CBO plan

2020-04-01 Thread Krisztian Kasa (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23089?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Krisztian Kasa updated HIVE-23089:
--
Attachment: HIVE-23089.1.patch

> Add constraint checks to CBO plan
> -
>
> Key: HIVE-23089
> URL: https://issues.apache.org/jira/browse/HIVE-23089
> Project: Hive
>  Issue Type: Improvement
>  Components: CBO
>Reporter: Krisztian Kasa
>Assignee: Krisztian Kasa
>Priority: Major
> Attachments: HIVE-23089.1.patch
>
>
> {code}
> create table acid_uami(i int,
>  de decimal(5,2) constraint nn1 not null enforced,
>  vc varchar(128) constraint nn2 not null enforced) clustered 
> by (i) into 2 buckets stored as orc TBLPROPERTIES ('transactional'='true');
> explain
> update acid_uami set de=null where i=1;
> {code}
> Non-CBO path:
> {code:java}
> Map Operator Tree:
> TableScan
> alias: acid_uami
> filterExpr: ((i = 1) and enforce_constraint(vc is not null)) 
> (type: boolean)
> Statistics: Num rows: 1 Data size: 216 Basic stats: COMPLETE 
> Column stats: NONE
> Filter Operator
>   predicate: ((i = 1) and enforce_constraint(vc is not null)) 
> (type: boolean)
> {code}
> CBO path:
> {code:java}
> Map Reduce
>   Map Operator Tree:
>   TableScan
> alias: acid_uami
> filterExpr: (i = 1) (type: boolean)
> Statistics: Num rows: 1 Data size: 216 Basic stats: COMPLETE 
> Column stats: NONE
> Filter Operator
>   predicate: (i = 1) (type: boolean)
> ...
>   Reduce Operator Tree:
> ...
>  Filter Operator
> predicate: enforce_constraint((null is not null and _col3 is not 
> null)) (type: boolean)
> {code}
> In CBO path the enforce_constraint function is added to the plan when CBO 
> plan is already generated and optimized.
> {code}
> HiveSortExchange(distribution=[any], collation=[[0]])
>   HiveProject(row__id=[$5], i=[CAST(1):INTEGER], _o__c2=[null:NULL], vc=[$2])
> HiveFilter(condition=[=($0, 1)])
>   HiveTableScan(table=[[default, acid_uami]], table:alias=[acid_uami])
> {code} 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23089) Add constraint checks to CBO plan

2020-04-01 Thread Krisztian Kasa (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23089?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Krisztian Kasa updated HIVE-23089:
--
Status: Patch Available  (was: Open)

> Add constraint checks to CBO plan
> -
>
> Key: HIVE-23089
> URL: https://issues.apache.org/jira/browse/HIVE-23089
> Project: Hive
>  Issue Type: Improvement
>  Components: CBO
>Reporter: Krisztian Kasa
>Assignee: Krisztian Kasa
>Priority: Major
> Attachments: HIVE-23089.1.patch
>
>
> {code}
> create table acid_uami(i int,
>  de decimal(5,2) constraint nn1 not null enforced,
>  vc varchar(128) constraint nn2 not null enforced) clustered 
> by (i) into 2 buckets stored as orc TBLPROPERTIES ('transactional'='true');
> explain
> update acid_uami set de=null where i=1;
> {code}
> Non-CBO path:
> {code:java}
> Map Operator Tree:
> TableScan
> alias: acid_uami
> filterExpr: ((i = 1) and enforce_constraint(vc is not null)) 
> (type: boolean)
> Statistics: Num rows: 1 Data size: 216 Basic stats: COMPLETE 
> Column stats: NONE
> Filter Operator
>   predicate: ((i = 1) and enforce_constraint(vc is not null)) 
> (type: boolean)
> {code}
> CBO path:
> {code:java}
> Map Reduce
>   Map Operator Tree:
>   TableScan
> alias: acid_uami
> filterExpr: (i = 1) (type: boolean)
> Statistics: Num rows: 1 Data size: 216 Basic stats: COMPLETE 
> Column stats: NONE
> Filter Operator
>   predicate: (i = 1) (type: boolean)
> ...
>   Reduce Operator Tree:
> ...
>  Filter Operator
> predicate: enforce_constraint((null is not null and _col3 is not 
> null)) (type: boolean)
> {code}
> In CBO path the enforce_constraint function is added to the plan when CBO 
> plan is already generated and optimized.
> {code}
> HiveSortExchange(distribution=[any], collation=[[0]])
>   HiveProject(row__id=[$5], i=[CAST(1):INTEGER], _o__c2=[null:NULL], vc=[$2])
> HiveFilter(condition=[=($0, 1)])
>   HiveTableScan(table=[[default, acid_uami]], table:alias=[acid_uami])
> {code} 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (HIVE-23124) Review of SQLOperation

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23124?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor reassigned HIVE-23124:
-

Assignee: David Mollitor

> Review of SQLOperation 
> ---
>
> Key: HIVE-23124
> URL: https://issues.apache.org/jira/browse/HIVE-23124
> Project: Hive
>  Issue Type: Improvement
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Major
> Attachments: HIVE-23124.1.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23124) Review of SQLOperation Class

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23124?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor updated HIVE-23124:
--
Summary: Review of SQLOperation Class  (was: Review of SQLOperation )

> Review of SQLOperation Class
> 
>
> Key: HIVE-23124
> URL: https://issues.apache.org/jira/browse/HIVE-23124
> Project: Hive
>  Issue Type: Improvement
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Major
> Attachments: HIVE-23124.1.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23124) Review of SQLOperation Class

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23124?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor updated HIVE-23124:
--
Status: Patch Available  (was: Open)

> Review of SQLOperation Class
> 
>
> Key: HIVE-23124
> URL: https://issues.apache.org/jira/browse/HIVE-23124
> Project: Hive
>  Issue Type: Improvement
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Major
> Attachments: HIVE-23124.1.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23109) Query-based compaction omits database

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23109?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072809#comment-17072809
 ] 

Hive QA commented on HIVE-23109:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
39s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
49s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
46s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
57s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m 
43s{color} | {color:blue} ql in master has 1528 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
43s{color} | {color:blue} itests/hive-unit in master has 2 extant Findbugs 
warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
21s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
27s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  2m 
 9s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
43s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m 
43s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
40s{color} | {color:green} ql: The patch generated 0 new + 0 unchanged - 1 
fixed = 0 total (was 1) {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
17s{color} | {color:green} The patch hive-unit passed checkstyle {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  4m 
31s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
20s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
15s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 30m  9s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-21370/dev-support/hive-personality.sh
 |
| git revision | master / 484d182 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.1 |
| modules | C: ql itests/hive-unit U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21370/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Query-based compaction omits database
> -
>
> Key: HIVE-23109
> URL: https://issues.apache.org/jira/browse/HIVE-23109
> Project: Hive
>  Issue Type: Bug
>Reporter: Karen Coppage
>Assignee: Karen Coppage
>Priority: Major
> Attachments: HIVE-23109.01.patch, HIVE-23109.02.patch
>
>
> E.g. MM major compaction query looks like:
> {code:java}
> insert into tmp_table select * from src_table;
> {code}
> it should be
> {code:java}
> insert into tmp_table select * from src_db.src_table;
> {code}
> Therefore compaction fails if db of source table isn't default.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23122) LLAP: TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Panagiotis Garefalakis (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072785#comment-17072785
 ] 

Panagiotis Garefalakis commented on HIVE-23122:
---

Thanks [~abstractdog]

+1 tests pending

> LLAP: TaskExecutorService should log details about task eviction decision 
> details
> -
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-23122.02.patch
>
>
> TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
> another. Under the hood, the queue uses a configurable 
> [comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
>  The currently available comparators typically use the following properties 
> of a task(wrapper):
> getWithinDagPriority: related to vertex
> currentAttemptStartTime
> firstAttemptStartTime
> knownPending: remaining upstream tasks
> The problem is, when an eviction happens, the INFO level message doesn't 
> provide any insight about the decision, only attempts ids like below:
> {code}
> attempt_1585248378306_0010_72_02_96_8 evicted from wait queue in favor of 
> attempt_1585248378306_0003_175_02_79_175 because of lower priority
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23122) LLAP: TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor updated HIVE-23122:

Summary: LLAP: TaskExecutorService should log details about task eviction 
decision details  (was: LLAP TaskExecutorService should log details about task 
eviction decision details)

> LLAP: TaskExecutorService should log details about task eviction decision 
> details
> -
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-23122.02.patch
>
>
> TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
> another. Under the hood, the queue uses a configurable 
> [comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
>  The currently available comparators typically use the following properties 
> of a task(wrapper):
> getWithinDagPriority: related to vertex
> currentAttemptStartTime
> firstAttemptStartTime
> knownPending: remaining upstream tasks
> The problem is, when an eviction happens, the INFO level message doesn't 
> provide any insight about the decision, only attempts ids like below:
> {code}
> attempt_1585248378306_0010_72_02_96_8 evicted from wait queue in favor of 
> attempt_1585248378306_0003_175_02_79_175 because of lower priority
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23122) LLAP TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Jira


[ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072783#comment-17072783
 ] 

László Bodor commented on HIVE-23122:
-

thanks for the comments [~pgaref], included canFinishForPriority
removing 01.patch for saving ptest time

> LLAP TaskExecutorService should log details about task eviction decision 
> details
> 
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-23122.02.patch
>
>
> TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
> another. Under the hood, the queue uses a configurable 
> [comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
>  The currently available comparators typically use the following properties 
> of a task(wrapper):
> getWithinDagPriority: related to vertex
> currentAttemptStartTime
> firstAttemptStartTime
> knownPending: remaining upstream tasks
> The problem is, when an eviction happens, the INFO level message doesn't 
> provide any insight about the decision, only attempts ids like below:
> {code}
> attempt_1585248378306_0010_72_02_96_8 evicted from wait queue in favor of 
> attempt_1585248378306_0003_175_02_79_175 because of lower priority
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23122) LLAP TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor updated HIVE-23122:

Attachment: (was: HIVE-23122.01.patch)

> LLAP TaskExecutorService should log details about task eviction decision 
> details
> 
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-23122.02.patch
>
>
> TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
> another. Under the hood, the queue uses a configurable 
> [comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
>  The currently available comparators typically use the following properties 
> of a task(wrapper):
> getWithinDagPriority: related to vertex
> currentAttemptStartTime
> firstAttemptStartTime
> knownPending: remaining upstream tasks
> The problem is, when an eviction happens, the INFO level message doesn't 
> provide any insight about the decision, only attempts ids like below:
> {code}
> attempt_1585248378306_0010_72_02_96_8 evicted from wait queue in favor of 
> attempt_1585248378306_0003_175_02_79_175 because of lower priority
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23122) LLAP TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor updated HIVE-23122:

Attachment: HIVE-23122.02.patch

> LLAP TaskExecutorService should log details about task eviction decision 
> details
> 
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-23122.02.patch
>
>
> TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
> another. Under the hood, the queue uses a configurable 
> [comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
>  The currently available comparators typically use the following properties 
> of a task(wrapper):
> getWithinDagPriority: related to vertex
> currentAttemptStartTime
> firstAttemptStartTime
> knownPending: remaining upstream tasks
> The problem is, when an eviction happens, the INFO level message doesn't 
> provide any insight about the decision, only attempts ids like below:
> {code}
> attempt_1585248378306_0010_72_02_96_8 evicted from wait queue in favor of 
> attempt_1585248378306_0003_175_02_79_175 because of lower priority
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23107) Remove MIN_HISTORY_LEVEL table

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23107?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072781#comment-17072781
 ] 

Hive QA commented on HIVE-23107:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12998338/HIVE-23107.02.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 160 failed/errored test(s), 18162 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.metastore.txn.TestCompactionTxnHandler.testFindNextToClean
 (batchId=318)
org.apache.hadoop.hive.metastore.txn.TestCompactionTxnHandler.testMarkCleaned 
(batchId=318)
org.apache.hadoop.hive.metastore.txn.TestCompactionTxnHandler.testMarkCleanedCleansTxnsAndTxnComponents
 (batchId=318)
org.apache.hadoop.hive.metastore.txn.TestCompactionTxnHandler.testMarkCompacted 
(batchId=318)
org.apache.hadoop.hive.ql.TestAcidOnTez.testInsertWithRemoveUnion (batchId=253)
org.apache.hadoop.hive.ql.TestAcidOnTez.testMapJoinOnMR (batchId=253)
org.apache.hadoop.hive.ql.TestAcidOnTez.testMapJoinOnTez (batchId=253)
org.apache.hadoop.hive.ql.TestAcidOnTez.testMergeJoinOnMR (batchId=253)
org.apache.hadoop.hive.ql.TestAcidOnTez.testMergeJoinOnTez (batchId=253)
org.apache.hadoop.hive.ql.TestTxnCommands.testNonAcidToAcidConversion01 
(batchId=363)
org.apache.hadoop.hive.ql.TestTxnCommands.testVersioning (batchId=363)
org.apache.hadoop.hive.ql.TestTxnCommands2.testACIDwithSchemaEvolutionAndCompaction
 (batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testCleanerForTxnToWriteId 
(batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testCompactWithDelete (batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testEmptyInTblproperties 
(batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testInitiatorWithMultipleFailedCompactions
 (batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testInsertOverwrite1 (batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testInsertOverwrite2 (batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testNonAcidToAcidConversion02 
(batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testNonAcidToAcidConversion1 
(batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testNonAcidToAcidConversion2 
(batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testNonAcidToAcidConversion3 
(batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testOriginalFileReaderWhenNonAcidConvertedToAcid
 (batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testSchemaEvolutionCompaction 
(batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.writeBetweenWorkerAndCleaner 
(batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testACIDwithSchemaEvolutionAndCompaction
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testCleanerForTxnToWriteId
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testCompactWithDelete
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testEmptyInTblproperties
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testInitiatorWithMultipleFailedCompactions
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testInsertOverwrite1
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testInsertOverwrite2
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testNonAcidToAcidConversion02
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testNonAcidToAcidConversion1
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testNonAcidToAcidConversion2
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testNonAcidToAcidConversion3
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testOriginalFileReaderWhenNonAcidConvertedToAcid
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testSchemaEvolutionCompaction
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.writeBetweenWorkerAndCleaner
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands3.testAcidMetaColumsDecode 
(batchId=359)
org.apache.hadoop.hive.ql.TestTxnCommands3.testCleaner2 (batchId=359)
org.apache.hadoop.hive.ql.TestTxnCommands3.testCompactionAbort (batchId=359)
org.apache.hadoop.hive.ql.TestTxnCommands3.testDeleteEventPruningOff 
(batchId=359)
org.apache.hadoop.hive.ql.TestTxnCommands3.testDeleteEventPruningOn 
(batchId=359)
org.apache.hadoop.hive.ql.TestTxnCommands3.testNotEnoughToCompactNeedsCleaning 
(batchId=359)
org.apache.hadoop.hive.ql.TestTxnCommandsForMmTable.testInsertOverwriteForPartitionedMmTable
 (batchId=318)

[jira] [Updated] (HIVE-23123) Export/Import of views and materialized views

2020-04-01 Thread Miklos Gergely (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23123?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Miklos Gergely updated HIVE-23123:
--
Description: 
According to 
[https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport] 
import and export can be done by using the
{code:java}
export table ...
import table ... 
{code}
commands. The document doesn't mention views or materialized views at all, and 
in fact we don't support commands like
{code:java}
export view ...
import view ...
export materialized view ...
import materialized view ... 
{code}
they can not be parsed at all. The word table is often used though in a broader 
sense, when it means all table like entities, including views and materialized 
views. For example the various Table classes may represent any of these as well.

If I try to export a view with the export table ... command, it goes fine. A 
_metadata file will be created, but no data directory, which is what we'd 
expect. If I try to import it back, an exception is thrown due to the lack of 
the data dir:
{code:java}
java.lang.AssertionError: null==getPath() for exim_view
 at org.apache.hadoop.hive.ql.metadata.Hive.loadTable(Hive.java:3088)
 at org.apache.hadoop.hive.ql.exec.MoveTask.execute(MoveTask.java:419)
 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:213)
 at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:105)
 at org.apache.hadoop.hive.ql.Executor.launchTask(Executor.java:364)
 at org.apache.hadoop.hive.ql.Executor.launchTasks(Executor.java:335)
 at org.apache.hadoop.hive.ql.Executor.runTasks(Executor.java:246)
 at org.apache.hadoop.hive.ql.Executor.execute(Executor.java:109)
 at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:722)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:491)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:485) 
{code}
Still the view gets imported successfully, as data movement wasn't even 
necessary.

If we try to export a materialized view which is transactional, then this 
exception occurs:
{code:java}
org.apache.hadoop.hive.ql.parse.SemanticException: 
org.apache.hadoop.hive.ql.metadata.InvalidTableException: Table not found 
exim_materialized_view_da21d41a_9fe4_4446_9c72_d251496abf9d
 at 
org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyzeAcidExport(AcidExportSemanticAnalyzer.java:163)
 at 
org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyze(AcidExportSemanticAnalyzer.java:71)
 at 
org.apache.hadoop.hive.ql.parse.RewriteSemanticAnalyzer.analyzeInternal(RewriteSemanticAnalyzer.java:72)
 at 
org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:289)
 at org.apache.hadoop.hive.ql.Compiler.analyze(Compiler.java:220)
 at org.apache.hadoop.hive.ql.Compiler.compile(Compiler.java:104)
 at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:183)
 at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:601)
 at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:547)
 at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:541) 
{code}
So the export process can not handle it, as the temporary table is not getting 
created.

 

The import command handling have a lot of codes dedicated to importing views 
and materialized views, which suggests that we support the importing (and thus 
also suggests implicitly that we support the exporting) of views and 
materialiezed views.

 

So the conclusion is that we have to decide if we support exporting/importing 
of views and materialized views.

If we decide not to support them then:
 - export process should throw an exception if a view or materialized view is 
the subject
 - the codes specific to view imports should be removed

If we decide to support them, then:
 - the commands mentioned above above should be introduced
 - exception should be thrown if not the proper command used (e.g. export view 
on a table)
 - the exceptions mentioned above should be fixed

I prefer not to support them, I don't think we should support the exporting / 
importing of views. The point of exporting / importing is the transfer of data, 
not DDL, it causes more issues than it solves. Our current documentation also 
suggests that it is only supported for tables.

  was:
According to 
[https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport] 
import and export can be done by using the
{code:java}
export table ...
import table ... 
{code}
commands. The document doesn't mention views or materialized views at all, and 
in fact we don't support commands like
{code:java}
export view ...
import view ...
export materialized view ...
import materialized view ... 
{code}
they can not be parsed at all. The word table is often used though in a broader 
sense, when it means all table like entities, including views and materialized 
views. For example the various Table classes may represent any of 

[jira] [Commented] (HIVE-23123) Export/Import of views and materialized views

2020-04-01 Thread Miklos Gergely (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23123?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072779#comment-17072779
 ] 

Miklos Gergely commented on HIVE-23123:
---

[~jcamachorodriguez] please let me know what you think.

> Export/Import of views and materialized views
> -
>
> Key: HIVE-23123
> URL: https://issues.apache.org/jira/browse/HIVE-23123
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Miklos Gergely
>Assignee: Miklos Gergely
>Priority: Major
>
> According to 
> [https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport]
>  import and export can be done by using the
>  
> {code:java}
> export table ...
> import table ... 
> {code}
> commands. The document doesn't mention views or materialized views at all, 
> and in fact we don't support commands like
>  
> {code:java}
> export view ...
> import view ...
> export materialized view ...
> import materialized view ... 
> {code}
> they can not be parsed at all. The word table is often used though in a 
> broader sense, when it means all table like entities, including views and 
> materialized views. For example the various Table classes may represent any 
> of these as well.
> If I try to export a view with the export table ... command, it goes fine. A 
> _metadata file will be created, but no data directory, which is what we'd 
> expect. If I try to import it back, an exception is thrown due to the lack of 
> the data dir:
>  
> {code:java}
> java.lang.AssertionError: null==getPath() for exim_view
>  at org.apache.hadoop.hive.ql.metadata.Hive.loadTable(Hive.java:3088)
>  at org.apache.hadoop.hive.ql.exec.MoveTask.execute(MoveTask.java:419)
>  at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:213)
>  at 
> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:105)
>  at org.apache.hadoop.hive.ql.Executor.launchTask(Executor.java:364)
>  at org.apache.hadoop.hive.ql.Executor.launchTasks(Executor.java:335)
>  at org.apache.hadoop.hive.ql.Executor.runTasks(Executor.java:246)
>  at org.apache.hadoop.hive.ql.Executor.execute(Executor.java:109)
>  at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:722)
>  at org.apache.hadoop.hive.ql.Driver.run(Driver.java:491)
>  at org.apache.hadoop.hive.ql.Driver.run(Driver.java:485) 
> {code}
> Still the view gets imported successfully, as data movement wasn't even 
> necessary.
> If we try to export a materialized view which is transactional, then this 
> exception occurs:
>  
> {code:java}
> org.apache.hadoop.hive.ql.parse.SemanticException: 
> org.apache.hadoop.hive.ql.metadata.InvalidTableException: Table not found 
> exim_materialized_view_da21d41a_9fe4_4446_9c72_d251496abf9d
>  at 
> org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyzeAcidExport(AcidExportSemanticAnalyzer.java:163)
>  at 
> org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyze(AcidExportSemanticAnalyzer.java:71)
>  at 
> org.apache.hadoop.hive.ql.parse.RewriteSemanticAnalyzer.analyzeInternal(RewriteSemanticAnalyzer.java:72)
>  at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:289)
>  at org.apache.hadoop.hive.ql.Compiler.analyze(Compiler.java:220)
>  at org.apache.hadoop.hive.ql.Compiler.compile(Compiler.java:104)
>  at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:183)
>  at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:601)
>  at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:547)
>  at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:541) 
> {code}
> So the export process can not handle it, as the temporary table is not 
> getting created.
>  
> The import command handling have a lot of codes dedicated to importing views 
> and materialized views, which suggests that we support the importing (and 
> thus also suggests implicitly that we support the exporting) of views and 
> materialiezed views.
>  
> So the conclusion is that we have to decide if we support exporting/importing 
> of views and materialized views.
> If we decide not to support them then:
>  - export process should throw an exception if a view or materialized view is 
> the subject
>  - the codes specific to view imports should be removed
> If we decide to support them, then:
>  - the commands mentioned above above should be introduced
>  - exception should be thrown if not the proper command used (e.g. export 
> view on a table)
>  - the exceptions mentioned above should be fixed
> I prefer not to support them, I don't think we should support the exporting / 
> importing of views. The point of exporting / importing is the transfer of 
> data, not DDL, it causes more issues than it solves. Our current 
> documentation also suggests that it is only supported for tables.
>  



[jira] [Updated] (HIVE-23123) Export/Import of views and materialized views

2020-04-01 Thread Miklos Gergely (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23123?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Miklos Gergely updated HIVE-23123:
--
Description: 
According to 
[https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport] 
import and export can be done by using the
{code:java}
export table ...
import table ... 
{code}
commands. The document doesn't mention views or materialized views at all, and 
in fact we don't support commands like
{code:java}
export view ...
import view ...
export materialized view ...
import materialized view ... 
{code}
they can not be parsed at all. The word table is often used though in a broader 
sense, when it means all table like entities, including views and materialized 
views. For example the various Table classes may represent any of these as well.

If I try to export a view with the export table ... command, it goes fine. A 
_metadata file will be created, but no data directory, which is what we'd 
expect. If I try to import it back, an exception is thrown due to the lack of 
the data dir:
{code:java}
java.lang.AssertionError: null==getPath() for exim_view
 at org.apache.hadoop.hive.ql.metadata.Hive.loadTable(Hive.java:3088)
 at org.apache.hadoop.hive.ql.exec.MoveTask.execute(MoveTask.java:419)
 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:213)
 at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:105)
 at org.apache.hadoop.hive.ql.Executor.launchTask(Executor.java:364)
 at org.apache.hadoop.hive.ql.Executor.launchTasks(Executor.java:335)
 at org.apache.hadoop.hive.ql.Executor.runTasks(Executor.java:246)
 at org.apache.hadoop.hive.ql.Executor.execute(Executor.java:109)
 at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:722)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:491)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:485) 
{code}
Still the view gets imported successfully, as data movement wasn't even 
necessary.

If we try to export a materialized view which is transactional, then this 
exception occurs:
{code:java}
org.apache.hadoop.hive.ql.parse.SemanticException: 
org.apache.hadoop.hive.ql.metadata.InvalidTableException: Table not found 
exim_materialized_view_da21d41a_9fe4_4446_9c72_d251496abf9d
 at 
org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyzeAcidExport(AcidExportSemanticAnalyzer.java:163)
 at 
org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyze(AcidExportSemanticAnalyzer.java:71)
 at 
org.apache.hadoop.hive.ql.parse.RewriteSemanticAnalyzer.analyzeInternal(RewriteSemanticAnalyzer.java:72)
 at 
org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:289)
 at org.apache.hadoop.hive.ql.Compiler.analyze(Compiler.java:220)
 at org.apache.hadoop.hive.ql.Compiler.compile(Compiler.java:104)
 at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:183)
 at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:601)
 at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:547)
 at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:541) 
{code}
So the export process can not handle it, as the temporary table is not getting 
created.

 

The import command handling have a lot of codes dedicated to importing views 
and materialized views, which suggests that we support the importing (and thus 
also suggests implicitly that we support the exporting) of views and 
materialiezed views.

 

So the conclusion is that we have to decide if we support exporting/importing 
of views and materialized views.

If we decide not to support them then:
 - export process should throw an exception if a view or materialized view is 
the subject
 - the codes specific to view imports should be removed

If we decide to support them, then:
 - the commands mentioned above above should be introduced
 - exception should be thrown if not the proper command used (e.g. export view 
on a table)
 - the exceptions mentioned above should be fixed

I prefer not to support them, I don't think we should support the exporting / 
importing of views. The point of exporting / importing is the transfer of data, 
not DDL, it causes more issues than it solves. Our current documentation also 
suggests that it is only supported for tables.

 

  was:
According to 
[https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport] 
import and export can be done by using the

 
{code:java}
export table ...
import table ... 
{code}
commands. The document doesn't mention views or materialized views at all, and 
in fact we don't support commands like

 
{code:java}
export view ...
import view ...
export materialized view ...
import materialized view ... 
{code}
they can not be parsed at all. The word table is often used though in a broader 
sense, when it means all table like entities, including views and materialized 
views. For example the various Table classes may represent 

[jira] [Updated] (HIVE-23123) Export/Import of views and materialized views

2020-04-01 Thread Miklos Gergely (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23123?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Miklos Gergely updated HIVE-23123:
--
Description: 
According to 
[https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport] 
import and export can be done by using the

 
{code:java}
export table ...
import table ... 
{code}
commands. The document doesn't mention views or materialized views at all, and 
in fact we don't support commands like

 
{code:java}
export view ...
import view ...
export materialized view ...
import materialized view ... 
{code}
they can not be parsed at all. The word table is often used though in a broader 
sense, when it means all table like entities, including views and materialized 
views. For example the various Table classes may represent any of these as well.

If I try to export a view with the export table ... command, it goes fine. A 
_metadata file will be created, but no data directory, which is what we'd 
expect. If I try to import it back, an exception is thrown due to the lack of 
the data dir:

 
{code:java}
java.lang.AssertionError: null==getPath() for exim_view
 at org.apache.hadoop.hive.ql.metadata.Hive.loadTable(Hive.java:3088)
 at org.apache.hadoop.hive.ql.exec.MoveTask.execute(MoveTask.java:419)
 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:213)
 at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:105)
 at org.apache.hadoop.hive.ql.Executor.launchTask(Executor.java:364)
 at org.apache.hadoop.hive.ql.Executor.launchTasks(Executor.java:335)
 at org.apache.hadoop.hive.ql.Executor.runTasks(Executor.java:246)
 at org.apache.hadoop.hive.ql.Executor.execute(Executor.java:109)
 at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:722)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:491)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:485) 
{code}
Still the view gets imported successfully, as data movement wasn't even 
necessary.

If we try to export a materialized view which is transactional, then this 
exception occurs:

 
{code:java}
org.apache.hadoop.hive.ql.parse.SemanticException: 
org.apache.hadoop.hive.ql.metadata.InvalidTableException: Table not found 
exim_materialized_view_da21d41a_9fe4_4446_9c72_d251496abf9d
 at 
org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyzeAcidExport(AcidExportSemanticAnalyzer.java:163)
 at 
org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyze(AcidExportSemanticAnalyzer.java:71)
 at 
org.apache.hadoop.hive.ql.parse.RewriteSemanticAnalyzer.analyzeInternal(RewriteSemanticAnalyzer.java:72)
 at 
org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:289)
 at org.apache.hadoop.hive.ql.Compiler.analyze(Compiler.java:220)
 at org.apache.hadoop.hive.ql.Compiler.compile(Compiler.java:104)
 at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:183)
 at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:601)
 at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:547)
 at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:541) 
{code}
So the export process can not handle it, as the temporary table is not getting 
created.

 

The import command handling have a lot of codes dedicated to importing views 
and materialized views, which suggests that we support the importing (and thus 
also suggests implicitly that we support the exporting) of views and 
materialiezed views.

 

So the conclusion is that we have to decide if we support exporting/importing 
of views and materialized views.

If we decide not to support them then:
 - export process should throw an exception if a view or materialized view is 
the subject
 - the codes specific to view imports should be removed

If we decide to support them, then:
 - the commands mentioned above above should be introduced
 - exception should be thrown if not the proper command used (e.g. export view 
on a table)
 - the exceptions mentioned above should be fixed

I prefer not to support them, I don't think we should support the exporting / 
importing of views. The point of exporting / importing is the transfer of data, 
not DDL, it causes more issues than it solves. Our current documentation also 
suggests that it is only supported for tables.

 

  was:
According to 
[https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport] 
import and export can be done by using the

 
{code:java}
export table ...
import table ... 
{code}
commands. The document doesn't mention views or materialized views at all, and 
in fact we don't support commands like

 
{code:java}
export view ...
import view ...
export materialized view ...
import materialized view ... 
{code}
they can not be parsed at all. The word table is often used though in a broader 
sense, when it means all table like entities, including views and materialized 
views. For example the various Table classes 

[jira] [Updated] (HIVE-23123) Export/Import of views and materialized views

2020-04-01 Thread Miklos Gergely (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23123?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Miklos Gergely updated HIVE-23123:
--
Description: 
According to 
[https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport] 
import and export can be done by using the

 
{code:java}
export table ...
import table ... 
{code}
commands. The document doesn't mention views or materialized views at all, and 
in fact we don't support commands like

 
{code:java}
export view ...
import view ...
export materialized view ...
import materialized view ... 
{code}
they can not be parsed at all. The word table is often used though in a broader 
sense, when it means all table like entities, including views and materialized 
views. For example the various Table classes may represent any of these as well.

If I try to export a view with the export table ... command, it goes fine. A 
_metadata file will be created, but no data directory, which is what we'd 
expect. If I try to import it back, an exception is thrown due to the lack of 
the data dir:

 
{code:java}
java.lang.AssertionError: null==getPath() for exim_view
 at org.apache.hadoop.hive.ql.metadata.Hive.loadTable(Hive.java:3088)
 at org.apache.hadoop.hive.ql.exec.MoveTask.execute(MoveTask.java:419)
 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:213)
 at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:105)
 at org.apache.hadoop.hive.ql.Executor.launchTask(Executor.java:364)
 at org.apache.hadoop.hive.ql.Executor.launchTasks(Executor.java:335)
 at org.apache.hadoop.hive.ql.Executor.runTasks(Executor.java:246)
 at org.apache.hadoop.hive.ql.Executor.execute(Executor.java:109)
 at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:722)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:491)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:485) 
{code}
Still the view gets imported successfully, as data movement wasn't even 
necessary.

If we try to export a materialized view which is transactional, then this 
exception occurs:

 
{code:java}
org.apache.hadoop.hive.ql.parse.SemanticException: 
org.apache.hadoop.hive.ql.metadata.InvalidTableException: Table not found 
exim_materialized_view_da21d41a_9fe4_4446_9c72_d251496abf9d
 at 
org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyzeAcidExport(AcidExportSemanticAnalyzer.java:163)
 at 
org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyze(AcidExportSemanticAnalyzer.java:71)
 at 
org.apache.hadoop.hive.ql.parse.RewriteSemanticAnalyzer.analyzeInternal(RewriteSemanticAnalyzer.java:72)
 at 
org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:289)
 at org.apache.hadoop.hive.ql.Compiler.analyze(Compiler.java:220)
 at org.apache.hadoop.hive.ql.Compiler.compile(Compiler.java:104)
 at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:183)
 at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:601)
 at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:547)
 at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:541) 
{code}
So the export process can not handle it, as the temporary table is not getting 
created.

 

The import command handling have a lot of codes dedicated to importing views 
and materialized views, which suggests that we support the importing (and thus 
also suggests implicitly that we support the exporting) of views and 
materialiezed views.

 

So the conclusion is that we have to decide if we support exporting/importing 
of views and materialized views.
 #  If we decide not to support them then:
 # 

 - export process should throw an exception if a view or materialized view is 
the subject
 - the codes specific to view imports should be removed

 # If we decide to support them, then:

 - the commands mentioned above above should be introduced
 - exception should be thrown if not the proper command used (e.g. export view 
on a table)
 - the exceptions mentioned above should be fixed

I prefer #1, I don't think we should support the exporting / importing of 
views. The point of exporting / importing is the transfer of data, not DDL, it 
causes more issues than it solves. Our current documentation also suggests that 
it is only supported for tables.

 

  was:
According to 
https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport 
import and export can be done by using the

 
{code:java}
export table ...
import table ... 
{code}
commands. The document doesn't mention views or materialized views at all, and 
in fact we don't support commands like

 
{code:java}
export view ...
import view ...
export materialized view ...
import materialized view ... 
{code}
they can not be parsed at all. The word table is often used though in a broader 
sense, when it means all table like entities, including views and materialized 
views. For example the various Table classes may 

[jira] [Updated] (HIVE-23123) Export/Import of views and materialized views

2020-04-01 Thread Miklos Gergely (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23123?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Miklos Gergely updated HIVE-23123:
--
Description: 
According to 
[https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport] 
import and export can be done by using the

 
{code:java}
export table ...
import table ... 
{code}
commands. The document doesn't mention views or materialized views at all, and 
in fact we don't support commands like

 
{code:java}
export view ...
import view ...
export materialized view ...
import materialized view ... 
{code}
they can not be parsed at all. The word table is often used though in a broader 
sense, when it means all table like entities, including views and materialized 
views. For example the various Table classes may represent any of these as well.

If I try to export a view with the export table ... command, it goes fine. A 
_metadata file will be created, but no data directory, which is what we'd 
expect. If I try to import it back, an exception is thrown due to the lack of 
the data dir:

 
{code:java}
java.lang.AssertionError: null==getPath() for exim_view
 at org.apache.hadoop.hive.ql.metadata.Hive.loadTable(Hive.java:3088)
 at org.apache.hadoop.hive.ql.exec.MoveTask.execute(MoveTask.java:419)
 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:213)
 at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:105)
 at org.apache.hadoop.hive.ql.Executor.launchTask(Executor.java:364)
 at org.apache.hadoop.hive.ql.Executor.launchTasks(Executor.java:335)
 at org.apache.hadoop.hive.ql.Executor.runTasks(Executor.java:246)
 at org.apache.hadoop.hive.ql.Executor.execute(Executor.java:109)
 at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:722)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:491)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:485) 
{code}
Still the view gets imported successfully, as data movement wasn't even 
necessary.

If we try to export a materialized view which is transactional, then this 
exception occurs:

 
{code:java}
org.apache.hadoop.hive.ql.parse.SemanticException: 
org.apache.hadoop.hive.ql.metadata.InvalidTableException: Table not found 
exim_materialized_view_da21d41a_9fe4_4446_9c72_d251496abf9d
 at 
org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyzeAcidExport(AcidExportSemanticAnalyzer.java:163)
 at 
org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyze(AcidExportSemanticAnalyzer.java:71)
 at 
org.apache.hadoop.hive.ql.parse.RewriteSemanticAnalyzer.analyzeInternal(RewriteSemanticAnalyzer.java:72)
 at 
org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:289)
 at org.apache.hadoop.hive.ql.Compiler.analyze(Compiler.java:220)
 at org.apache.hadoop.hive.ql.Compiler.compile(Compiler.java:104)
 at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:183)
 at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:601)
 at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:547)
 at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:541) 
{code}
So the export process can not handle it, as the temporary table is not getting 
created.

 

The import command handling have a lot of codes dedicated to importing views 
and materialized views, which suggests that we support the importing (and thus 
also suggests implicitly that we support the exporting) of views and 
materialiezed views.

 

So the conclusion is that we have to decide if we support exporting/importing 
of views and materialized views.

If we decide not to support them then:
 - export process should throw an exception if a view or materialized view is 
the subject
 - the codes specific to view imports should be removed

If we decide to support them, then:
 - the commands mentioned above above should be introduced
 - exception should be thrown if not the proper command used (e.g. export view 
on a table)
 - the exceptions mentioned above should be fixed

I prefer #1, I don't think we should support the exporting / importing of 
views. The point of exporting / importing is the transfer of data, not DDL, it 
causes more issues than it solves. Our current documentation also suggests that 
it is only supported for tables.

 

  was:
According to 
[https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport] 
import and export can be done by using the

 
{code:java}
export table ...
import table ... 
{code}
commands. The document doesn't mention views or materialized views at all, and 
in fact we don't support commands like

 
{code:java}
export view ...
import view ...
export materialized view ...
import materialized view ... 
{code}
they can not be parsed at all. The word table is often used though in a broader 
sense, when it means all table like entities, including views and materialized 
views. For example the various Table classes may represent any 

[jira] [Assigned] (HIVE-23123) Export/Import of views and materialized views

2020-04-01 Thread Miklos Gergely (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23123?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Miklos Gergely reassigned HIVE-23123:
-


> Export/Import of views and materialized views
> -
>
> Key: HIVE-23123
> URL: https://issues.apache.org/jira/browse/HIVE-23123
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Miklos Gergely
>Assignee: Miklos Gergely
>Priority: Major
>
> According to 
> https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport 
> import and export can be done by using the
>  
> {code:java}
> export table ...
> import table ... 
> {code}
> commands. The document doesn't mention views or materialized views at all, 
> and in fact we don't support commands like
>  
> {code:java}
> export view ...
> import view ...
> export materialized view ...
> import materialized view ... 
> {code}
> they can not be parsed at all. The word table is often used though in a 
> broader sense, when it means all table like entities, including views and 
> materialized views. For example the various Table classes may represent any 
> of these as well.
> If I try to export a view with the export table ... command, it goes fine. A 
> _metadata file will be created, but no data directory, which is what we'd 
> expect. If I try to import it back, an exception is thrown due to the lack of 
> the data dir:
>  
> {code:java}
> java.lang.AssertionError: null==getPath() for exim_view
>  at org.apache.hadoop.hive.ql.metadata.Hive.loadTable(Hive.java:3088)
>  at org.apache.hadoop.hive.ql.exec.MoveTask.execute(MoveTask.java:419)
>  at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:213)
>  at 
> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:105)
>  at org.apache.hadoop.hive.ql.Executor.launchTask(Executor.java:364)
>  at org.apache.hadoop.hive.ql.Executor.launchTasks(Executor.java:335)
>  at org.apache.hadoop.hive.ql.Executor.runTasks(Executor.java:246)
>  at org.apache.hadoop.hive.ql.Executor.execute(Executor.java:109)
>  at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:722)
>  at org.apache.hadoop.hive.ql.Driver.run(Driver.java:491)
>  at org.apache.hadoop.hive.ql.Driver.run(Driver.java:485) 
> {code}
> Still the view gets imported successfully, as data movement wasn't even 
> necessary.
> If we try to export a materialized view which is transactional, then this 
> exception occurs:
>  
> {code:java}
> org.apache.hadoop.hive.ql.parse.SemanticException: 
> org.apache.hadoop.hive.ql.metadata.InvalidTableException: Table not found 
> exim_materialized_view_da21d41a_9fe4_4446_9c72_d251496abf9d
>  at 
> org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyzeAcidExport(AcidExportSemanticAnalyzer.java:163)
>  at 
> org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyze(AcidExportSemanticAnalyzer.java:71)
>  at 
> org.apache.hadoop.hive.ql.parse.RewriteSemanticAnalyzer.analyzeInternal(RewriteSemanticAnalyzer.java:72)
>  at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:289)
>  at org.apache.hadoop.hive.ql.Compiler.analyze(Compiler.java:220)
>  at org.apache.hadoop.hive.ql.Compiler.compile(Compiler.java:104)
>  at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:183)
>  at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:601)
>  at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:547)
>  at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:541) 
> {code}
> So the export process can not handle it, as the temporary table is not 
> getting created.
>  
> The import command handling have a lot of codes dedicated to importing views 
> and materialized views, which suggests that we support the importing (and 
> thus also suggests implicitly that we support the exporting) of views and 
> materialiezed views.
>  
> So the conclusion is that we have to decide if we support exporting/importing 
> of views and materialized views.
>  #  If we decide not to support them then:
>  - export process should throw an exception if a view or materialized view is 
> the subject
>  - the codes specific to view imports should be removed
>  # If we decide to support them, then:
>  - the commands mentioned above above should be introduced
>  - exception should be thrown if not the proper command used (e.g. export 
> view on a table)
>  - the exceptions mentioned above should be fixed
> I prefer #1, I don't think we should support the exporting / importing of 
> views. The point of exporting / importing is the transfer of data, not DDL, 
> it causes more issues than it solves. Our current documentation also suggests 
> that it is only supported for tables.
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23079) Remove Calls to printStackTrace in Module hive-serde

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23079?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor updated HIVE-23079:
--
Fix Version/s: 4.0.0
   Resolution: Fixed
   Status: Resolved  (was: Patch Available)

Pushed to master.  Thanks [~pvary] for the review!

> Remove Calls to printStackTrace in Module hive-serde
> 
>
> Key: HIVE-23079
> URL: https://issues.apache.org/jira/browse/HIVE-23079
> Project: Hive
>  Issue Type: Sub-task
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-23079.1.patch, HIVE-23079.1.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23113) Clean Up HiveCallableStatement

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23113?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor updated HIVE-23113:
--
Attachment: HIVE-23113.1.patch

> Clean Up HiveCallableStatement
> --
>
> Key: HIVE-23113
> URL: https://issues.apache.org/jira/browse/HIVE-23113
> Project: Hive
>  Issue Type: Improvement
>  Components: JDBC
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Trivial
> Attachments: HIVE-23113.1.patch, HIVE-23113.1.patch
>
>
> * Add a useful class comment
>  * Remove all non-javadoc comments
>  * Remove 'TODO' tags
>  * Add {{@override}} tags
>  * Checkstyle formatting



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23107) Remove MIN_HISTORY_LEVEL table

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23107?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072736#comment-17072736
 ] 

Hive QA commented on HIVE-23107:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
43s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
45s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
27s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
 2s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  1m  
9s{color} | {color:blue} standalone-metastore/metastore-server in master has 
190 extant Findbugs warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m 
40s{color} | {color:blue} ql in master has 1529 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
12s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
27s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
57s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
26s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m 
26s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
23s{color} | {color:red} standalone-metastore/metastore-server: The patch 
generated 4 new + 669 unchanged - 2 fixed = 673 total (was 671) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
22s{color} | {color:green} standalone-metastore/metastore-server generated 0 
new + 187 unchanged - 3 fixed = 187 total (was 190) {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
46s{color} | {color:green} ql in the patch passed. {color} |
| {color:red}-1{color} | {color:red} javadoc {color} | {color:red}  0m 
20s{color} | {color:red} standalone-metastore_metastore-server generated 1 new 
+ 24 unchanged - 0 fixed = 25 total (was 24) {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 30m 10s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-21369/dev-support/hive-personality.sh
 |
| git revision | master / 94f90bc |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21369/yetus/diff-checkstyle-standalone-metastore_metastore-server.txt
 |
| javadoc | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21369/yetus/diff-javadoc-javadoc-standalone-metastore_metastore-server.txt
 |
| modules | C: standalone-metastore/metastore-server ql U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21369/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Remove MIN_HISTORY_LEVEL table
> --
>
> Key: HIVE-23107
> URL: https://issues.apache.org/jira/browse/HIVE-23107
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Reporter: László Pintér
>Assignee: László Pintér
>Priority: Major
> Attachments: HIVE-23107.01.patch, HIVE-23107.02.patch
>
>
> MIN_HISTORY_LEVEL table is used in two places:
>  * Cleaner uses it to decide if the files can 

[jira] [Updated] (HIVE-23039) Checkpointing for repl dump bootstrap phase

2020-04-01 Thread Aasha Medhi (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23039?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aasha Medhi updated HIVE-23039:
---
Status: In Progress  (was: Patch Available)

> Checkpointing for repl dump bootstrap phase
> ---
>
> Key: HIVE-23039
> URL: https://issues.apache.org/jira/browse/HIVE-23039
> Project: Hive
>  Issue Type: Bug
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23039.01.patch, HIVE-23039.02.patch
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23039) Checkpointing for repl dump bootstrap phase

2020-04-01 Thread Aasha Medhi (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23039?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aasha Medhi updated HIVE-23039:
---
Attachment: HIVE-23039.02.patch
Status: Patch Available  (was: In Progress)

> Checkpointing for repl dump bootstrap phase
> ---
>
> Key: HIVE-23039
> URL: https://issues.apache.org/jira/browse/HIVE-23039
> Project: Hive
>  Issue Type: Bug
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23039.01.patch, HIVE-23039.02.patch
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23122) LLAP TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Panagiotis Garefalakis (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072716#comment-17072716
 ] 

Panagiotis Garefalakis commented on HIVE-23122:
---

Thanks for the patch [~abstractdog] – pretty useful – would it make sense to 
add the canFinishForPriority property used by the BaseComparator as well?

> LLAP TaskExecutorService should log details about task eviction decision 
> details
> 
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-23122.01.patch
>
>
> TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
> another. Under the hood, the queue uses a configurable 
> [comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
>  The currently available comparators typically use the following properties 
> of a task(wrapper):
> getWithinDagPriority: related to vertex
> currentAttemptStartTime
> firstAttemptStartTime
> knownPending: remaining upstream tasks
> The problem is, when an eviction happens, the INFO level message doesn't 
> provide any insight about the decision, only attempts ids like below:
> {code}
> attempt_1585248378306_0010_72_02_96_8 evicted from wait queue in favor of 
> attempt_1585248378306_0003_175_02_79_175 because of lower priority
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23113) Clean Up HiveCallableStatement

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23113?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072699#comment-17072699
 ] 

Hive QA commented on HIVE-23113:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12998347/HIVE-23113.1.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 18162 tests 
executed
*Failed tests:*
{noformat}
org.apache.hive.beeline.TestBeeLineWithArgs.testRowsAffected (batchId=286)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/21368/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/21368/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-21368/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12998347 - PreCommit-HIVE-Build

> Clean Up HiveCallableStatement
> --
>
> Key: HIVE-23113
> URL: https://issues.apache.org/jira/browse/HIVE-23113
> Project: Hive
>  Issue Type: Improvement
>  Components: JDBC
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Trivial
> Attachments: HIVE-23113.1.patch
>
>
> * Add a useful class comment
>  * Remove all non-javadoc comments
>  * Remove 'TODO' tags
>  * Add {{@override}} tags
>  * Checkstyle formatting



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23122) LLAP TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor updated HIVE-23122:

Description: 
TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
another. Under the hood, the queue uses a configurable 
[comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
 The currently available comparators typically use the following properties of 
a task(wrapper):
getWithinDagPriority: related to vertex
currentAttemptStartTime
firstAttemptStartTime
knownPending: remaining upstream tasks

The problem is, when an eviction happens, the INFO level message doesn't 
provide any insight about the decision, only attempts ids like below:
{code}
attempt_1585248378306_0010_72_02_96_8 evicted from wait queue in favor of 
attempt_1585248378306_0003_175_02_79_175 because of lower priority
{code}



  was:
TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
another. Under the hood, the queue uses a configurable 
[comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
 The currently available comparators typically use the following properties of 
a task:
getWithinDagPriority: related to vertex



> LLAP TaskExecutorService should log details about task eviction decision 
> details
> 
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-23122.01.patch
>
>
> TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
> another. Under the hood, the queue uses a configurable 
> [comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
>  The currently available comparators typically use the following properties 
> of a task(wrapper):
> getWithinDagPriority: related to vertex
> currentAttemptStartTime
> firstAttemptStartTime
> knownPending: remaining upstream tasks
> The problem is, when an eviction happens, the INFO level message doesn't 
> provide any insight about the decision, only attempts ids like below:
> {code}
> attempt_1585248378306_0010_72_02_96_8 evicted from wait queue in favor of 
> attempt_1585248378306_0003_175_02_79_175 because of lower priority
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23122) LLAP TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor updated HIVE-23122:

Description: 
TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
another. Under the hood, the queue uses a configurable 
[comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
 The currently available comparators typically use the following properties of 
a task:
getWithinDagPriority: related to vertex


> LLAP TaskExecutorService should log details about task eviction decision 
> details
> 
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-23122.01.patch
>
>
> TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
> another. Under the hood, the queue uses a configurable 
> [comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
>  The currently available comparators typically use the following properties 
> of a task:
> getWithinDagPriority: related to vertex



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23122) LLAP TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor updated HIVE-23122:

Status: Patch Available  (was: Open)

> LLAP TaskExecutorService should log details about task eviction decision 
> details
> 
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-23122.01.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23122) LLAP TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor updated HIVE-23122:

Attachment: HIVE-23122.01.patch

> LLAP TaskExecutorService should log details about task eviction decision 
> details
> 
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-23122.01.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23122) LLAP TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor updated HIVE-23122:

Summary: LLAP TaskExecutorService should log details about task eviction 
decision details  (was: LLAP TaskSchedulerService should log details about task 
eviction decision details)

> LLAP TaskExecutorService should log details about task eviction decision 
> details
> 
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (HIVE-23122) LLAP TaskSchedulerService should log details about task eviction decision details

2020-04-01 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor reassigned HIVE-23122:
---

Assignee: László Bodor

> LLAP TaskSchedulerService should log details about task eviction decision 
> details
> -
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-21304) Show Bucketing version for ReduceSinkOp in explain extended plan

2020-04-01 Thread Zoltan Haindrich (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21304?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltan Haindrich updated HIVE-21304:

Attachment: HIVE-21304.20.patch

> Show Bucketing version for ReduceSinkOp in explain extended plan
> 
>
> Key: HIVE-21304
> URL: https://issues.apache.org/jira/browse/HIVE-21304
> Project: Hive
>  Issue Type: Bug
>Reporter: Deepak Jaiswal
>Assignee: Zoltan Haindrich
>Priority: Major
> Attachments: HIVE-21304.01.patch, HIVE-21304.02.patch, 
> HIVE-21304.03.patch, HIVE-21304.04.patch, HIVE-21304.05.patch, 
> HIVE-21304.06.patch, HIVE-21304.07.patch, HIVE-21304.08.patch, 
> HIVE-21304.09.patch, HIVE-21304.10.patch, HIVE-21304.11.patch, 
> HIVE-21304.12.patch, HIVE-21304.13.patch, HIVE-21304.14.patch, 
> HIVE-21304.15.patch, HIVE-21304.16.patch, HIVE-21304.17.patch, 
> HIVE-21304.18.patch, HIVE-21304.19.patch, HIVE-21304.20.patch
>
>
> Show Bucketing version for ReduceSinkOp in explain extended plan.
> This helps identify what hashing algorithm is being used by by ReduceSinkOp.
>  
> cc [~vgarg]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (HIVE-23121) Re-examine TestWarehouseExternalDir to see if it uses HMS translation.

2020-04-01 Thread Naveen Gangam (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23121?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Naveen Gangam reassigned HIVE-23121:



> Re-examine TestWarehouseExternalDir to see if it uses HMS translation.
> --
>
> Key: HIVE-23121
> URL: https://issues.apache.org/jira/browse/HIVE-23121
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Naveen Gangam
>Assignee: Naveen Gangam
>Priority: Major
>
> TestWarehouseExternalDir currently passes with just one change related to 
> HIVE-22995. But that change was assuming it was using HMS Translation to 
> convert non-acid managed table to external. 
> Ensure that it still does.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23113) Clean Up HiveCallableStatement

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23113?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072649#comment-17072649
 ] 

Hive QA commented on HIVE-23113:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
59s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
18s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
12s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
29s{color} | {color:blue} jdbc in master has 16 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
13s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
38s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
18s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
18s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
11s{color} | {color:green} jdbc: The patch generated 0 new + 0 unchanged - 1 
fixed = 0 total (was 1) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
35s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
13s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 12m 55s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-21368/dev-support/hive-personality.sh
 |
| git revision | master / 94f90bc |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.1 |
| modules | C: jdbc U: jdbc |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21368/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Clean Up HiveCallableStatement
> --
>
> Key: HIVE-23113
> URL: https://issues.apache.org/jira/browse/HIVE-23113
> Project: Hive
>  Issue Type: Improvement
>  Components: JDBC
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Trivial
> Attachments: HIVE-23113.1.patch
>
>
> * Add a useful class comment
>  * Remove all non-javadoc comments
>  * Remove 'TODO' tags
>  * Add {{@override}} tags
>  * Checkstyle formatting



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-23020) Avoid using _files for replication data copy during incremental run

2020-04-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23020?focusedWorklogId=413976=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-413976
 ]

ASF GitHub Bot logged work on HIVE-23020:
-

Author: ASF GitHub Bot
Created on: 01/Apr/20 10:51
Start Date: 01/Apr/20 10:51
Worklog Time Spent: 10m 
  Work Description: aasha commented on pull request #963: HIVE-23020 : 
Avoid using _files for replication data copy during incr…
URL: https://github.com/apache/hive/pull/963#discussion_r401526351
 
 

 ##
 File path: 
ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/AddPartitionHandler.java
 ##
 @@ -108,10 +108,8 @@ public void handle(Context withinContext) throws 
Exception {
 Iterable files = partitionFilesIter.next().getFiles();
 if (files != null) {
   // encoded filename/checksum of files, write into _files
-  try (BufferedWriter fileListWriter = writer(withinContext, qlPtn)) {
-for (String file : files) {
-  writeFileEntry(qlMdTable.getDbName(), qlMdTable, file, 
fileListWriter, withinContext);
-}
+  for (String file : files) {
 
 Review comment:
   tests for this
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 413976)
Time Spent: 0.5h  (was: 20m)

> Avoid using _files for replication data copy during incremental run
> ---
>
> Key: HIVE-23020
> URL: https://issues.apache.org/jira/browse/HIVE-23020
> Project: Hive
>  Issue Type: Task
>Reporter: PRAVIN KUMAR SINHA
>Assignee: PRAVIN KUMAR SINHA
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23020.01.patch, HIVE-23020.02.patch
>
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-23020) Avoid using _files for replication data copy during incremental run

2020-04-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23020?focusedWorklogId=413974=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-413974
 ]

ASF GitHub Bot logged work on HIVE-23020:
-

Author: ASF GitHub Bot
Created on: 01/Apr/20 10:50
Start Date: 01/Apr/20 10:50
Worklog Time Spent: 10m 
  Work Description: aasha commented on pull request #963: HIVE-23020 : 
Avoid using _files for replication data copy during incr…
URL: https://github.com/apache/hive/pull/963#discussion_r401522879
 
 

 ##
 File path: 
ql/src/java/org/apache/hadoop/hive/ql/parse/ImportSemanticAnalyzer.java
 ##
 @@ -505,9 +506,8 @@ private static ImportTableDesc 
getBaseCreateTableDescFromTable(String dbName,
 
 Task copyTask = null;
 if (replicationSpec.isInReplicationScope()) {
-  boolean isImport = 
ReplicationSpec.Type.IMPORT.equals(replicationSpec.getReplSpecType());
   copyTask = ReplCopyTask.getLoadCopyTask(replicationSpec, dataPath, 
destPath, x.getConf(),
-  isAutoPurge, needRecycle, copyToMigratedTxnTable, !isImport);
+  isAutoPurge, needRecycle, copyToMigratedTxnTable, false);
 
 Review comment:
   Why is this set to false? Do we even need this boolean if everyone is 
setting it to false?
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 413974)
Time Spent: 20m  (was: 10m)

> Avoid using _files for replication data copy during incremental run
> ---
>
> Key: HIVE-23020
> URL: https://issues.apache.org/jira/browse/HIVE-23020
> Project: Hive
>  Issue Type: Task
>Reporter: PRAVIN KUMAR SINHA
>Assignee: PRAVIN KUMAR SINHA
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23020.01.patch, HIVE-23020.02.patch
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-21304) Show Bucketing version for ReduceSinkOp in explain extended plan

2020-04-01 Thread Zoltan Haindrich (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21304?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltan Haindrich updated HIVE-21304:

Attachment: HIVE-21304.19.patch

> Show Bucketing version for ReduceSinkOp in explain extended plan
> 
>
> Key: HIVE-21304
> URL: https://issues.apache.org/jira/browse/HIVE-21304
> Project: Hive
>  Issue Type: Bug
>Reporter: Deepak Jaiswal
>Assignee: Zoltan Haindrich
>Priority: Major
> Attachments: HIVE-21304.01.patch, HIVE-21304.02.patch, 
> HIVE-21304.03.patch, HIVE-21304.04.patch, HIVE-21304.05.patch, 
> HIVE-21304.06.patch, HIVE-21304.07.patch, HIVE-21304.08.patch, 
> HIVE-21304.09.patch, HIVE-21304.10.patch, HIVE-21304.11.patch, 
> HIVE-21304.12.patch, HIVE-21304.13.patch, HIVE-21304.14.patch, 
> HIVE-21304.15.patch, HIVE-21304.16.patch, HIVE-21304.17.patch, 
> HIVE-21304.18.patch, HIVE-21304.19.patch
>
>
> Show Bucketing version for ReduceSinkOp in explain extended plan.
> This helps identify what hashing algorithm is being used by by ReduceSinkOp.
>  
> cc [~vgarg]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23076) Add batching for openTxn

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23076?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072629#comment-17072629
 ] 

Hive QA commented on HIVE-23076:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12998336/HIVE-23076.9.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 18163 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/21367/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/21367/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-21367/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12998336 - PreCommit-HIVE-Build

> Add batching for openTxn
> 
>
> Key: HIVE-23076
> URL: https://issues.apache.org/jira/browse/HIVE-23076
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Reporter: Peter Vary
>Assignee: Peter Vary
>Priority: Major
> Attachments: HIVE-23076.2.patch, HIVE-23076.3.patch, 
> HIVE-23076.4.patch, HIVE-23076.5.patch, HIVE-23076.6.patch, 
> HIVE-23076.7.patch, HIVE-23076.8.patch, HIVE-23076.9.patch, HIVE-23076.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-23039) Checkpointing for repl dump bootstrap phase

2020-04-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23039?focusedWorklogId=413958=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-413958
 ]

ASF GitHub Bot logged work on HIVE-23039:
-

Author: ASF GitHub Bot
Created on: 01/Apr/20 10:23
Start Date: 01/Apr/20 10:23
Worklog Time Spent: 10m 
  Work Description: anishek commented on pull request #965: HIVE-23039 
Checkpointing for repl dump bootstrap phase
URL: https://github.com/apache/hive/pull/965#discussion_r401496558
 
 

 ##
 File path: ql/src/java/org/apache/hadoop/hive/ql/exec/repl/util/ReplUtils.java
 ##
 @@ -296,4 +300,17 @@ public static boolean includeAcidTableInDump(HiveConf 
conf) {
   public static boolean tableIncludedInReplScope(ReplScope replScope, String 
tableName) {
 return ((replScope == null) || 
replScope.tableIncludedInReplScope(tableName));
   }
+
+  public static boolean dataCopyCompleted(Path toPath, HiveConf conf) throws 
IOException {
+FileSystem dstFs = null;
+dstFs = toPath.getFileSystem(conf);
+if (dstFs.exists(new Path(toPath, ReplUtils.COPY_ACKNOWLEDGEMENT))) {
 
 Review comment:
   do return of condition directly rather than using if clause
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 413958)
Time Spent: 0.5h  (was: 20m)

> Checkpointing for repl dump bootstrap phase
> ---
>
> Key: HIVE-23039
> URL: https://issues.apache.org/jira/browse/HIVE-23039
> Project: Hive
>  Issue Type: Bug
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23039.01.patch
>
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-23039) Checkpointing for repl dump bootstrap phase

2020-04-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23039?focusedWorklogId=413959=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-413959
 ]

ASF GitHub Bot logged work on HIVE-23039:
-

Author: ASF GitHub Bot
Created on: 01/Apr/20 10:23
Start Date: 01/Apr/20 10:23
Worklog Time Spent: 10m 
  Work Description: anishek commented on pull request #965: HIVE-23039 
Checkpointing for repl dump bootstrap phase
URL: https://github.com/apache/hive/pull/965#discussion_r401491670
 
 

 ##
 File path: ql/src/java/org/apache/hadoop/hive/ql/plan/ReplCopyWork.java
 ##
 @@ -120,4 +122,12 @@ public boolean isNeedCheckDuplicateCopy() {
   public void setCheckDuplicateCopy(boolean flag) {
 checkDuplicateCopy = flag;
   }
+
+  public boolean isCheckpointEnabled() {
+return checkpointEnabled;
+  }
+
+  public void setCheckpointEnabled(boolean checkpointEnabled) {
 
 Review comment:
   do initialization based on constructor rather than setters ?
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 413959)

> Checkpointing for repl dump bootstrap phase
> ---
>
> Key: HIVE-23039
> URL: https://issues.apache.org/jira/browse/HIVE-23039
> Project: Hive
>  Issue Type: Bug
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23039.01.patch
>
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-23039) Checkpointing for repl dump bootstrap phase

2020-04-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23039?focusedWorklogId=413960=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-413960
 ]

ASF GitHub Bot logged work on HIVE-23039:
-

Author: ASF GitHub Bot
Created on: 01/Apr/20 10:23
Start Date: 01/Apr/20 10:23
Worklog Time Spent: 10m 
  Work Description: anishek commented on pull request #965: HIVE-23039 
Checkpointing for repl dump bootstrap phase
URL: https://github.com/apache/hive/pull/965#discussion_r401496301
 
 

 ##
 File path: ql/src/java/org/apache/hadoop/hive/ql/exec/repl/util/ReplUtils.java
 ##
 @@ -98,6 +100,8 @@
   public static final String DUMP_ACKNOWLEDGEMENT = "_finished_dump";
   //Acknowledgement for repl load complete
   public static final String LOAD_ACKNOWLEDGEMENT = "_finished_load";
+  //Acknowledgement for data copy complete. Used for checkpointing
+  public static final String COPY_ACKNOWLEDGEMENT = "_finished_copy";
 
 Review comment:
   looks like we need a constants Class or a Enum class, the utils one seem to 
have a lot of magic variables.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 413960)
Time Spent: 40m  (was: 0.5h)

> Checkpointing for repl dump bootstrap phase
> ---
>
> Key: HIVE-23039
> URL: https://issues.apache.org/jira/browse/HIVE-23039
> Project: Hive
>  Issue Type: Bug
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23039.01.patch
>
>  Time Spent: 40m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-23039) Checkpointing for repl dump bootstrap phase

2020-04-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23039?focusedWorklogId=413961=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-413961
 ]

ASF GitHub Bot logged work on HIVE-23039:
-

Author: ASF GitHub Bot
Created on: 01/Apr/20 10:23
Start Date: 01/Apr/20 10:23
Worklog Time Spent: 10m 
  Work Description: anishek commented on pull request #965: HIVE-23039 
Checkpointing for repl dump bootstrap phase
URL: https://github.com/apache/hive/pull/965#discussion_r401496956
 
 

 ##
 File path: ql/src/java/org/apache/hadoop/hive/ql/exec/repl/util/ReplUtils.java
 ##
 @@ -296,4 +300,17 @@ public static boolean includeAcidTableInDump(HiveConf 
conf) {
   public static boolean tableIncludedInReplScope(ReplScope replScope, String 
tableName) {
 return ((replScope == null) || 
replScope.tableIncludedInReplScope(tableName));
   }
+
+  public static boolean dataCopyCompleted(Path toPath, HiveConf conf) throws 
IOException {
+FileSystem dstFs = null;
+dstFs = toPath.getFileSystem(conf);
+if (dstFs.exists(new Path(toPath, ReplUtils.COPY_ACKNOWLEDGEMENT))) {
+  return true;
+}
+return false;
+  }
+
+  public static void setDataCopyComplete(Path toPath, HiveConf conf) throws 
SemanticException {
 
 Review comment:
   rename method to ackCopy() ?
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 413961)
Time Spent: 50m  (was: 40m)

> Checkpointing for repl dump bootstrap phase
> ---
>
> Key: HIVE-23039
> URL: https://issues.apache.org/jira/browse/HIVE-23039
> Project: Hive
>  Issue Type: Bug
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23039.01.patch
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-23039) Checkpointing for repl dump bootstrap phase

2020-04-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23039?focusedWorklogId=413957=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-413957
 ]

ASF GitHub Bot logged work on HIVE-23039:
-

Author: ASF GitHub Bot
Created on: 01/Apr/20 10:23
Start Date: 01/Apr/20 10:23
Worklog Time Spent: 10m 
  Work Description: anishek commented on pull request #965: HIVE-23039 
Checkpointing for repl dump bootstrap phase
URL: https://github.com/apache/hive/pull/965#discussion_r401486847
 
 

 ##
 File path: 
itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosExternalTables.java
 ##
 @@ -906,6 +908,131 @@ public void replicationWithTableNameContainsKeywords() 
throws Throwable {
 .verifyReplTargetProperty(replicatedDbName);
   }
 
+  @Test
+  public void testCheckPointing() throws Throwable {
+List withClauseOptions = externalTableBasePathWithClause();
+WarehouseInstance.Tuple bootstrapDump = primary.run("use " + primaryDbName)
+.run("CREATE TABLE t1(a string) STORED AS TEXTFILE")
+.run("CREATE EXTERNAL TABLE t2(a string) STORED AS TEXTFILE")
+.run("insert into t1 values (1)")
+.run("insert into t1 values (2)")
+.run("insert into t2 values (11)")
+.run("insert into t2 values (21)")
+.dump(primaryDbName, withClauseOptions);
+
+// verify that the external table info is written correctly for bootstrap
+assertExternalFileInfo(Arrays.asList("t2"), bootstrapDump.dumpLocation, 
primaryDbName);
 
 Review comment:
   might be better to use another location here completely, sometimes since we 
are writing to same dump location and also the db directory is there you might 
have files in locations that are slight off and you wont realize it till you 
run actual prod scenario
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 413957)
Time Spent: 20m  (was: 10m)

> Checkpointing for repl dump bootstrap phase
> ---
>
> Key: HIVE-23039
> URL: https://issues.apache.org/jira/browse/HIVE-23039
> Project: Hive
>  Issue Type: Bug
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23039.01.patch
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-5312) Let HiveServer2 run simultaneously in HTTP (over thrift) and Binary (normal thrift transport) mode

2020-04-01 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-5312?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ádám Szita updated HIVE-5312:
-
Attachment: HIVE-5312.3.patch

> Let HiveServer2 run simultaneously in HTTP (over thrift) and Binary (normal 
> thrift transport) mode 
> ---
>
> Key: HIVE-5312
> URL: https://issues.apache.org/jira/browse/HIVE-5312
> Project: Hive
>  Issue Type: Sub-task
>  Components: HiveServer2
>Reporter: Vaibhav Gumashta
>Assignee: Ádám Szita
>Priority: Major
> Attachments: HIVE-5132.branch-2.3.patch, HIVE-5312.1.patch, 
> HIVE-5312.2.patch, HIVE-5312.3.patch
>
>
> [HIVE-4763|https://issues.apache.org/jira/browse/HIVE-4763] adds support for 
> HTTP transport over thrift. With that, HS2 can be configured to run either 
> using using HTTP or using normal thrift binary transport. Ideally HS2 should 
> be supporting both modes simultaneously and the client should be able to 
> specify the mode used in serving the request.  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


  1   2   >