[jira] [Commented] (HIVE-14388) Add number of rows inserted message after insert command in Beeline

2018-05-10 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-14388?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471507#comment-16471507
 ] 

Hive QA commented on HIVE-14388:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
30s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
14s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  2m 
47s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
35s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
29s{color} | {color:blue} beeline in master has 69 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
36s{color} | {color:blue} itests/hive-unit in master has 2 extant Findbugs 
warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
32s{color} | {color:blue} jdbc in master has 17 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  4m  
5s{color} | {color:blue} ql in master has 2321 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
39s{color} | {color:blue} service in master has 50 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  2m  
8s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
10s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  3m 
44s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  2m 
49s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  2m 
49s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
 6s{color} | {color:green} The patch service-rpc passed checkstyle {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
10s{color} | {color:green} The patch beeline passed checkstyle {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
17s{color} | {color:green} The patch hive-unit passed checkstyle {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
11s{color} | {color:green} The patch jdbc passed checkstyle {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
39s{color} | {color:green} ql: The patch generated 0 new + 217 unchanged - 13 
fixed = 217 total (was 230) {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
12s{color} | {color:green} The patch service passed checkstyle {color} |
| {color:red}-1{color} | {color:red} whitespace {color} | {color:red}  0m  
0s{color} | {color:red} The patch has 2 line(s) that end in whitespace. Use git 
apply --whitespace=fix <>. Refer https://git-scm.com/docs/git-apply 
{color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  7m  
0s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  2m 
10s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:red}-1{color} | {color:red} asflicense {color} | {color:red}  0m 
14s{color} | {color:red} The patch generated 1 ASF License warnings. {color} |
| {color:black}{color} | {color:black} {color} | {color:black} 39m 34s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-10818/dev-support/hive-personality.sh
 |
| git revision | master / 6c364c6 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| whitespace | 

[jira] [Commented] (HIVE-19417) Modify metastore to have/access persistent tables for stats

2018-05-10 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19417?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471486#comment-16471486
 ] 

Hive QA commented on HIVE-19417:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12922684/HIVE-19417.01.patch

{color:red}ERROR:{color} -1 due to build exiting with an error

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/10815/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/10815/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-10815/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hiveptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ date '+%Y-%m-%d %T.%3N'
2018-05-11 04:38:25.196
+ [[ -n /usr/lib/jvm/java-8-openjdk-amd64 ]]
+ export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ export 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'MAVEN_OPTS=-Xmx1g '
+ MAVEN_OPTS='-Xmx1g '
+ cd /data/hiveptest/working/
+ tee /data/hiveptest/logs/PreCommit-HIVE-Build-10815/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ git = \s\v\n ]]
+ [[ git = \g\i\t ]]
+ [[ -z master ]]
+ [[ -d apache-github-source-source ]]
+ [[ ! -d apache-github-source-source/.git ]]
+ [[ ! -d apache-github-source-source ]]
+ date '+%Y-%m-%d %T.%3N'
2018-05-11 04:38:25.199
+ cd apache-github-source-source
+ git fetch origin
+ git reset --hard HEAD
HEAD is now at 6c364c6 HIVE-19306: Arrow batch serializer (Teddy Choi, reviewed 
by Matt McCline and Eric Wohlstadter (non-binding))
+ git clean -f -d
+ git checkout master
Already on 'master'
Your branch is up-to-date with 'origin/master'.
+ git reset --hard origin/master
HEAD is now at 6c364c6 HIVE-19306: Arrow batch serializer (Teddy Choi, reviewed 
by Matt McCline and Eric Wohlstadter (non-binding))
+ git merge --ff-only origin/master
Already up-to-date.
+ date '+%Y-%m-%d %T.%3N'
2018-05-11 04:38:28.550
+ rm -rf ../yetus_PreCommit-HIVE-Build-10815
+ mkdir ../yetus_PreCommit-HIVE-Build-10815
+ git gc
+ cp -R . ../yetus_PreCommit-HIVE-Build-10815
+ mkdir /data/hiveptest/logs/PreCommit-HIVE-Build-10815/yetus
+ patchCommandPath=/data/hiveptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hiveptest/working/scratch/build.patch
+ [[ -f /data/hiveptest/working/scratch/build.patch ]]
+ chmod +x /data/hiveptest/working/scratch/smart-apply-patch.sh
+ /data/hiveptest/working/scratch/smart-apply-patch.sh 
/data/hiveptest/working/scratch/build.patch
error: patch failed: 
standalone-metastore/src/main/sql/derby/upgrade-2.3.0-to-3.0.0.derby.sql:254
Falling back to three-way merge...
Applied patch to 
'standalone-metastore/src/main/sql/derby/upgrade-2.3.0-to-3.0.0.derby.sql' with 
conflicts.
error: patch failed: 
standalone-metastore/src/main/sql/mssql/upgrade-2.3.0-to-3.0.0.mssql.sql:321
Falling back to three-way merge...
Applied patch to 
'standalone-metastore/src/main/sql/mssql/upgrade-2.3.0-to-3.0.0.mssql.sql' with 
conflicts.
error: patch failed: 
standalone-metastore/src/main/sql/mysql/upgrade-2.3.0-to-3.0.0.mysql.sql:292
Falling back to three-way merge...
Applied patch to 
'standalone-metastore/src/main/sql/mysql/upgrade-2.3.0-to-3.0.0.mysql.sql' with 
conflicts.
error: patch failed: 
standalone-metastore/src/main/sql/oracle/upgrade-2.3.0-to-3.0.0.oracle.sql:310
Falling back to three-way merge...
Applied patch to 
'standalone-metastore/src/main/sql/oracle/upgrade-2.3.0-to-3.0.0.oracle.sql' 
with conflicts.
error: patch failed: 
standalone-metastore/src/main/sql/postgres/upgrade-2.3.0-to-3.0.0.postgres.sql:329
Falling back to three-way merge...
Applied patch to 
'standalone-metastore/src/main/sql/postgres/upgrade-2.3.0-to-3.0.0.postgres.sql'
 with conflicts.
Going to apply patch with: git apply -p0
/data/hiveptest/working/scratch/build.patch:1317: trailing whitespace.
tmpMap.put(_Fields.WRITE_ID, new 
org.apache.thrift.meta_data.FieldMetaData("writeId", 
org.apache.thrift.TFieldRequirementType.OPTIONAL, 
/data/hiveptest/working/scratch/build.patch:1319: trailing whitespace.
tmpMap.put(_Fields.VALID_WRITE_ID_LIST, new 
org.apache.thrift.meta_data.FieldMetaData("validWriteIdList", 
org.apache.thrift.TFieldRequirementType.OPTIONAL, 
/data/hiveptest/working/scratch/build.patch:1551: trailing whitespace.
} else { 
/data/hiveptest/working/scratch/build.patch:1559: trailing whitespace.
} else { 
/data/hiveptest/working/scratch/build.patch:1695: trailing whitespace.
   

[jira] [Commented] (HIVE-18117) Create TestCliDriver for HDFS EC

2018-05-10 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-18117?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471484#comment-16471484
 ] 

Hive QA commented on HIVE-18117:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12922690/HIVE-18117.2.patch

{color:green}SUCCESS:{color} +1 due to 5 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 43 failed/errored test(s), 13578 tests 
executed
*Failed tests:*
{noformat}
TestDbNotificationListener - did not produce a TEST-*.xml file (likely timed 
out) (batchId=247)
TestHCatHiveCompatibility - did not produce a TEST-*.xml file (likely timed 
out) (batchId=247)
TestNegativeCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=95)


[jira] [Comment Edited] (HIVE-19384) Vectorization: IfExprTimestamp* do not handle NULLs correctly

2018-05-10 Thread Matt McCline (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19384?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16470483#comment-16470483
 ] 

Matt McCline edited comment on HIVE-19384 at 5/11/18 4:11 AM:
--

#10816 ?


was (Author: mmccline):
#10824

> Vectorization: IfExprTimestamp* do not handle NULLs correctly
> -
>
> Key: HIVE-19384
> URL: https://issues.apache.org/jira/browse/HIVE-19384
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Matt McCline
>Assignee: Matt McCline
>Priority: Blocker
> Fix For: 3.0.0, 3.1.0
>
> Attachments: HIVE-19384.01.patch, HIVE-19384.02.patch, 
> HIVE-19384.04.patch
>
>
> HIVE-18622: "Vectorization: IF Statements, Comparisons, and more do not 
> handle NULLs correctly" didn't quite fix the IfExprTimestamp* classes 
> right
> {noformat}
> // Carefully handle NULLs...
> outputColVector.noNulls = false;{noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-18117) Create TestCliDriver for HDFS EC

2018-05-10 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-18117?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471454#comment-16471454
 ] 

Hive QA commented on HIVE-18117:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
42s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
 0s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  2m 
25s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
27s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
18s{color} | {color:blue} shims/common in master has 6 extant Findbugs 
warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
20s{color} | {color:blue} shims/0.23 in master has 7 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
42s{color} | {color:blue} itests/util in master has 55 extant Findbugs 
warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  4m  
9s{color} | {color:blue} ql in master has 2321 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
26s{color} | {color:blue} testutils/ptest2 in master has 24 extant Findbugs 
warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  2m  
1s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m  
8s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:red}-1{color} | {color:red} mvninstall {color} | {color:red}  0m 
28s{color} | {color:red} qtest in the patch failed. {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  2m 
26s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  2m 
26s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m  
9s{color} | {color:red} shims/common: The patch generated 5 new + 95 unchanged 
- 0 fixed = 100 total (was 95) {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m  
9s{color} | {color:red} shims/0.23: The patch generated 7 new + 69 unchanged - 
0 fixed = 76 total (was 69) {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
13s{color} | {color:red} itests/qtest: The patch generated 2 new + 0 unchanged 
- 0 fixed = 2 total (was 0) {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
13s{color} | {color:red} itests/util: The patch generated 26 new + 100 
unchanged - 0 fixed = 126 total (was 100) {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
38s{color} | {color:red} ql: The patch generated 46 new + 106 unchanged - 0 
fixed = 152 total (was 106) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 1s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  0m 
29s{color} | {color:red} shims/0.23 generated 1 new + 7 unchanged - 0 fixed = 8 
total (was 7) {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
55s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:red}-1{color} | {color:red} asflicense {color} | {color:red}  0m 
13s{color} | {color:red} The patch generated 4 ASF License warnings. {color} |
| {color:black}{color} | {color:black} {color} | {color:black} 36m 12s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| FindBugs | module:shims/0.23 |
|  |  Should org.apache.hadoop.hive.shims.Hadoop23Shims$HdfsErasureCodingShim 
be a _static_ inner class?  At Hadoop23Shims.java:inner class?  At 
Hadoop23Shims.java:[lines 1523-1606] |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 

[jira] [Commented] (HIVE-19306) Arrow batch serializer

2018-05-10 Thread Teddy Choi (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19306?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471444#comment-16471444
 ] 

Teddy Choi commented on HIVE-19306:
---

[~mmccline], thanks for committing!

> Arrow batch serializer
> --
>
> Key: HIVE-19306
> URL: https://issues.apache.org/jira/browse/HIVE-19306
> Project: Hive
>  Issue Type: Task
>  Components: Serializers/Deserializers
>Reporter: Eric Wohlstadter
>Assignee: Teddy Choi
>Priority: Major
> Fix For: 3.1.0
>
> Attachments: HIVE-19306.2.patch, HIVE-19306.3.patch, 
> HIVE-19306.4.patch
>
>
> Leverage the ThriftJDBCBinarySerDe code path that already exists in 
> SemanticAnalyzer/FileSinkOperator to create a serializer that batches rows 
> into Arrow vector batches.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19433) HiveJoinPushTransitivePredicatesRule hangs

2018-05-10 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19433?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471425#comment-16471425
 ] 

Hive QA commented on HIVE-19433:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12922692/HIVE-19433.2.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 41 failed/errored test(s), 13576 tests 
executed
*Failed tests:*
{noformat}
TestDbNotificationListener - did not produce a TEST-*.xml file (likely timed 
out) (batchId=247)
TestHCatHiveCompatibility - did not produce a TEST-*.xml file (likely timed 
out) (batchId=247)
TestNegativeCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=95)


[jira] [Commented] (HIVE-19433) HiveJoinPushTransitivePredicatesRule hangs

2018-05-10 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19433?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471383#comment-16471383
 ] 

Hive QA commented on HIVE-19433:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  6m 
57s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
0s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
38s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  4m  
0s{color} | {color:blue} ql in master has 2321 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
53s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
21s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
1s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m  
1s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
37s{color} | {color:red} ql: The patch generated 1 new + 8 unchanged - 0 fixed 
= 9 total (was 8) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  4m  
4s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
54s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:red}-1{color} | {color:red} asflicense {color} | {color:red}  0m 
13s{color} | {color:red} The patch generated 1 ASF License warnings. {color} |
| {color:black}{color} | {color:black} {color} | {color:black} 22m  3s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-10813/dev-support/hive-personality.sh
 |
| git revision | master / 6c364c6 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-10813/yetus/diff-checkstyle-ql.txt
 |
| asflicense | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-10813/yetus/patch-asflicense-problems.txt
 |
| modules | C: ql U: ql |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-10813/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> HiveJoinPushTransitivePredicatesRule hangs
> --
>
> Key: HIVE-19433
> URL: https://issues.apache.org/jira/browse/HIVE-19433
> Project: Hive
>  Issue Type: Bug
>  Components: Logical Optimizer
>Reporter: Vineet Garg
>Assignee: Vineet Garg
>Priority: Major
> Attachments: HIVE-19433.1.patch, HIVE-19433.2.patch
>
>
> *Reproducer*
> {code:sql}
> CREATE TABLE `table1`(
>`idp_warehouse_id` bigint,
>`idp_audit_id` bigint,
>`idp_effective_date` date,
>`idp_end_date` date,
>`idp_delete_date` date,
>`pruid` varchar(32),
>`prid` bigint,
>`prtimesheetid` bigint,
>`prassignmentid` bigint,
>`prchargecodeid` bigint,
>`prtypecodeid` bigint,
>`prsequence` bigint,
>`prmodby` varchar(96),
>`prmodtime` timestamp,
>`prrmexported` bigint,
>`prrmckdel` bigint,
>`slice_status` int,
>`role_id` bigint,
>`user_lov1` varchar(30),
>`user_lov2` varchar(30),
>`incident_id` bigint,
>`incident_investment_id` bigint,
>`odf_ss_actuals` bigint,
>`practsum` decimal(38,20));
> CREATE TABLE `table2`(
>`idp_warehouse_id` bigint,
>`idp_audit_id` bigint,
>`idp_effective_date` date,
>`idp_end_date` date,
>

[jira] [Assigned] (HIVE-19498) Vectorization: CAST expressions produce wrong results

2018-05-10 Thread Matt McCline (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19498?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matt McCline reassigned HIVE-19498:
---


> Vectorization: CAST expressions produce wrong results
> -
>
> Key: HIVE-19498
> URL: https://issues.apache.org/jira/browse/HIVE-19498
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Matt McCline
>Assignee: Matt McCline
>Priority: Critical
> Fix For: 3.1.0
>
>
> DATE --> BOOLEAN
> DOUBLE --> DECIMAL
> STRING|CHAR|VARCHAR --> DECIMAL
> TIMESTAMP --> LONG



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19453) Extend Load Data statement to take Input file format and Serde as parameters

2018-05-10 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19453?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471369#comment-16471369
 ] 

Hive QA commented on HIVE-19453:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12922702/HIVE-19453.01-branch-3.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 30 failed/errored test(s), 13373 tests 
executed
*Failed tests:*
{noformat}
TestBeeLineDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=253)
TestDbNotificationListener - did not produce a TEST-*.xml file (likely timed 
out) (batchId=247)
TestDummy - did not produce a TEST-*.xml file (likely timed out) (batchId=253)
TestHCatHiveCompatibility - did not produce a TEST-*.xml file (likely timed 
out) (batchId=247)
TestMiniDruidCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=253)
TestMiniDruidKafkaCliDriver - did not produce a TEST-*.xml file (likely timed 
out) (batchId=253)
TestNegativeCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=95)


[jira] [Commented] (HIVE-19399) Down cast from int to tinyint generating incorrect value for vectorization

2018-05-10 Thread Bharathkrishna Guruvayoor Murali (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19399?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471364#comment-16471364
 ] 

Bharathkrishna Guruvayoor Murali commented on HIVE-19399:
-

I am interested in working on this Jira as it is a chance for me to get 
familiarized with vectorization related stuff!
If nobody is planning to work on it soon, I can take a look at this :)

> Down cast from int to tinyint generating incorrect value for vectorization
> --
>
> Key: HIVE-19399
> URL: https://issues.apache.org/jira/browse/HIVE-19399
> Project: Hive
>  Issue Type: Bug
>  Components: Vectorization
>Affects Versions: 3.1.0
>Reporter: Haifeng Chen
>Priority: Major
>
>  The following sql scripts generating different result for vectorization 
> disabled and enabled (both for ORC and for parquet).
>    drop table test_schema;
>    create table test_schema (f int) stored as parquet;
>    insert into test_schema values ('9');
>    select cast(f as tinyint) + 1 from test_schema;
> For non-vectorization, the result is -96 while for vectorization mode, it is 
> 10
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19497) SessionHiveMetaStoreClient.getTable should respect default catalog

2018-05-10 Thread Dongjoon Hyun (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19497?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471365#comment-16471365
 ] 

Dongjoon Hyun commented on HIVE-19497:
--

Based on the previous commit, could you review this, [~alangates] and [~thejas]?

> SessionHiveMetaStoreClient.getTable should respect default catalog
> --
>
> Key: HIVE-19497
> URL: https://issues.apache.org/jira/browse/HIVE-19497
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Affects Versions: 3.0.0
>Reporter: Dongjoon Hyun
>Priority: Major
> Attachments: 
> 0001-HIVE-19497-SessionHiveMetaStoreClient.getTable-shoul.patch
>
>
> Currently, SessionHiveMetaStoreClient always use `DEFAULT_CATALOG_NAME` 
> during looking up the underlying tables.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-19016) Vectorization and Parquet: When vectorized, parquet_nested_complex.q produces RuntimeException: Unsupported type used

2018-05-10 Thread Haifeng Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19016?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haifeng Chen reassigned HIVE-19016:
---

Assignee: Haifeng Chen  (was: Vihang Karajgaonkar)

> Vectorization and Parquet: When vectorized, parquet_nested_complex.q produces 
> RuntimeException: Unsupported type used
> -
>
> Key: HIVE-19016
> URL: https://issues.apache.org/jira/browse/HIVE-19016
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 3.0.0
>Reporter: Matt McCline
>Assignee: Haifeng Chen
>Priority: Critical
>
> Adding "SET hive.vectorized.execution.enabled=true;" to 
> parquet_nested_complex.q triggers this call stack:
> {noformat}
> Caused by: java.lang.RuntimeException: Unsupported type used in 
> list:array
>   at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.checkListColumnSupport(VectorizedParquetRecordReader.java:589)
>  ~[hive-exec-3.0.0-SNAPSHOT.jar:3.0.0-SNAPSHOT]
>   at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.buildVectorizedParquetReader(VectorizedParquetRecordReader.java:525)
>  ~[hive-exec-3.0.0-SNAPSHOT.jar:3.0.0-SNAPSHOT]
>   at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.checkEndOfRowGroup(VectorizedParquetRecordReader.java:440)
>  ~[hive-exec-3.0.0-SNAPSHOT.jar:3.0.0-SNAPSHOT]
>   at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.nextBatch(VectorizedParquetRecordReader.java:401)
>  ~[hive-exec-3.0.0-SNAPSHOT.jar:3.0.0-SNAPSHOT]
>   at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:353)
>  ~[hive-exec-3.0.0-SNAPSHOT.jar:3.0.0-SNAPSHOT]
>   at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:92)
>  ~[hive-exec-3.0.0-SNAPSHOT.jar:3.0.0-SNAPSHOT]
>   at 
> org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.doNext(HiveContextAwareRecordReader.java:360)
>  ~[hive-exec-3.0.0-SNAPSHOT.jar:3.0.0-SNAPSHOT]
> {noformat}
> FYI: [~vihangk1]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-19015) Vectorization and Parquet: When vectorized, parquet_map_of_arrays_of_ints.q gets a ClassCastException

2018-05-10 Thread Haifeng Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19015?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haifeng Chen reassigned HIVE-19015:
---

Assignee: Haifeng Chen  (was: Vihang Karajgaonkar)

> Vectorization and Parquet: When vectorized, parquet_map_of_arrays_of_ints.q 
> gets a ClassCastException
> -
>
> Key: HIVE-19015
> URL: https://issues.apache.org/jira/browse/HIVE-19015
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 3.0.0
>Reporter: Matt McCline
>Assignee: Haifeng Chen
>Priority: Critical
>
> Adding "SET hive.vectorized.execution.enabled=true;"  to 
> parquet_map_of_arrays_of_ints.q triggers this call stack:
> {noformat}
> Caused by: java.lang.ClassCastException: 
> org.apache.hadoop.hive.serde2.typeinfo.ListTypeInfo cannot be cast to 
> org.apache.hadoop.hive.serde2.typeinfo.PrimitiveTypeInfo
>   at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedListColumnReader.readBatch(VectorizedListColumnReader.java:67)
>  ~[hive-exec-3.0.0-SNAPSHOT.jar:3.0.0-SNAPSHOT]
>   at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedMapColumnReader.readBatch(VectorizedMapColumnReader.java:57)
>  ~[hive-exec-3.0.0-SNAPSHOT.jar:3.0.0-SNAPSHOT]
>   at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.nextBatch(VectorizedParquetRecordReader.java:410)
>  ~[hive-exec-3.0.0-SNAPSHOT.jar:3.0.0-SNAPSHOT]
>   at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:353)
>  ~[hive-exec-3.0.0-SNAPSHOT.jar:3.0.0-SNAPSHOT]
>   at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:92)
>  ~[hive-exec-3.0.0-SNAPSHOT.jar:3.0.0-SNAPSHOT]
>   at 
> org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.doNext(HiveContextAwareRecordReader.java:360)
>  ~[hive-exec-3.0.0-SNAPSHOT.jar:3.0.0-SNAPSHOT]
> {noformat}
> FYI: [~vihangk1]
> Adding parquet_map_of_maps.q, too.  Stack trace seems related.
> {noformat}
> Caused by: java.lang.ClassCastException: optional group value (MAP) {
>   repeated group key_value {
> optional binary key (UTF8);
> required int32 value;
>   }
> } is not primitive
>   at org.apache.parquet.schema.Type.asPrimitiveType(Type.java:213) 
> ~[parquet-hadoop-bundle-1.9.0.jar:1.9.0]
>   at 
> org.apache.hadoop.hive.ql.io.parquet.vector.BaseVectorizedColumnReader.(BaseVectorizedColumnReader.java:130)
>  ~[hive-exec-3.0.0-SNAPSHOT.jar:3.0.0-SNAPSHOT]
>   at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedListColumnReader.(VectorizedListColumnReader.java:52)
>  ~[hive-exec-3.0.0-SNAPSHOT.jar:3.0.0-SNAPSHOT]
>   at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.buildVectorizedParquetReader(VectorizedParquetRecordReader.java:568)
>  ~[hive-exec-3.0.0-SNAPSHOT.jar:3.0.0-SNAPSHOT]
>   at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.checkEndOfRowGroup(VectorizedParquetRecordReader.java:440)
>  ~[hive-exec-3.0.0-SNAPSHOT.jar:3.0.0-SNAPSHOT]
>   at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.nextBatch(VectorizedParquetRecordReader.java:401)
>  ~[hive-exec-3.0.0-SNAPSHOT.jar:3.0.0-SNAPSHOT]
>   at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:353)
>  ~[hive-exec-3.0.0-SNAPSHOT.jar:3.0.0-SNAPSHOT]
>   at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:92)
>  ~[hive-exec-3.0.0-SNAPSHOT.jar:3.0.0-SNAPSHOT]
>   at 
> org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.doNext(HiveContextAwareRecordReader.java:360)
>  ~[hive-exec-3.0.0-SNAPSHOT.jar:3.0.0-SNAPSHOT]
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19497) SessionHiveMetaStoreClient.getTable should respect default catalog

2018-05-10 Thread Dongjoon Hyun (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19497?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dongjoon Hyun updated HIVE-19497:
-
Attachment: 0001-HIVE-19497-SessionHiveMetaStoreClient.getTable-shoul.patch

> SessionHiveMetaStoreClient.getTable should respect default catalog
> --
>
> Key: HIVE-19497
> URL: https://issues.apache.org/jira/browse/HIVE-19497
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Affects Versions: 3.0.0
>Reporter: Dongjoon Hyun
>Priority: Major
> Attachments: 
> 0001-HIVE-19497-SessionHiveMetaStoreClient.getTable-shoul.patch
>
>
> Currently, SessionHiveMetaStoreClient always use `DEFAULT_CATALOG_NAME` 
> during looking up the underlying tables.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19493) VectorUDFDateDiffColCol copySelected does not handle nulls correctly

2018-05-10 Thread Vihang Karajgaonkar (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19493?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471354#comment-16471354
 ] 

Vihang Karajgaonkar commented on HIVE-19493:


Thanks [~mmccline] for the review. Updated the diff with the suggested change. 
Actually I am having hard time to understand the complete logic of when noNulls 
needs to be set to {{true}}. For example,

in the below code snippet from {{copySelected}} method
{code:java}
// Handle repeating case
if (input.isRepeating) {
  if (input.noNulls || !input.isNull[0]) {
String string = new String(input.vector[0], input.start[0], 
input.length[0]);
try {
  date.setTime(formatter.parse(string).getTime());
  output.vector[0] = DateWritable.dateToDays(date);
  output.isNull[0] = false;
} catch (ParseException e) {
  output.isNull[0] = true;
  output.noNulls = false;
}
  } else {
output.isNull[0] = true;
output.noNulls = false;
  }
  output.isRepeating = true;
  return;
}
{code}
Can you please help me understand the following?
 * Why do we *not* set output.notNulls = true in the {{if (input.noNulls || 
!input.isNull[0])}} code block when we know that {{input.isRepeating == true}}? 
Shouldn't we be setting it to true and reset output.isNull[] array to all false 
values? My guess is that we are not doing this for performance reasons since it 
doesn't make sense to reset the whole {{output.isNull[]}} when we know that we 
only need to look for the first element. Just want to confirm if this 
understanding is correct (may be even add a comment so that its easy to 
remember next time)
 * As a corollary of the above statement it means both the following conditions 
are valid but they represent the same state of the columnVector.
 {{vector.isRepeating == true && vector.noNulls == true}} --> vector has a 
non-null repeating value
 {{vector.isRepeating == true && vector.noNulls == false && vector.isNull[0] == 
false}} --> vector has non-null repeating value. Is this understanding correct?
 * In case of {{input.isRepeating == false && selected == true}} case when 
{{input.noNulls == true && output.noNull == false}} why don't we set 
{{output.noNull = true}}. Is it because we think that when {{selected == true}} 
there may be less number of rows to be updated and hence its unnecessary work 
to reset {{output.isNull[]}} which we have to do everytime when set 
{{output.noNulls = true}}?

I think the bug which was fixed in HIVE-18622 manifests when the 
{{output.noNulls}} is flipped from {{false to true}} and when the {{isNull}} 
array has some entries which are {{true}}. So may be we create an expression 
such that the columnVector is reused and this flag gets flipped. Just adding a 
row of nulls may not exercise the code which has the bug. Let me see if I can 
create an expression to test this particular issue.

> VectorUDFDateDiffColCol copySelected does not handle nulls correctly
> 
>
> Key: HIVE-19493
> URL: https://issues.apache.org/jira/browse/HIVE-19493
> Project: Hive
>  Issue Type: Bug
>  Components: Vectorization
>Reporter: Vihang Karajgaonkar
>Assignee: Vihang Karajgaonkar
>Priority: Major
> Attachments: HIVE-19493.01.patch, HIVE-19493.02.patch
>
>
> The {{copySelected}} method in {{VectorUDFDateDiffColCol}} class was missed 
> during HIVE-18622



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19493) VectorUDFDateDiffColCol copySelected does not handle nulls correctly

2018-05-10 Thread Vihang Karajgaonkar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19493?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vihang Karajgaonkar updated HIVE-19493:
---
Attachment: HIVE-19493.02.patch

> VectorUDFDateDiffColCol copySelected does not handle nulls correctly
> 
>
> Key: HIVE-19493
> URL: https://issues.apache.org/jira/browse/HIVE-19493
> Project: Hive
>  Issue Type: Bug
>  Components: Vectorization
>Reporter: Vihang Karajgaonkar
>Assignee: Vihang Karajgaonkar
>Priority: Major
> Attachments: HIVE-19493.01.patch, HIVE-19493.02.patch
>
>
> The {{copySelected}} method in {{VectorUDFDateDiffColCol}} class was missed 
> during HIVE-18622



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19453) Extend Load Data statement to take Input file format and Serde as parameters

2018-05-10 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19453?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471308#comment-16471308
 ] 

Hive QA commented on HIVE-19453:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:red}-1{color} | {color:red} patch {color} | {color:red}  0m  8s{color} 
| {color:red} 
/data/hiveptest/logs/PreCommit-HIVE-Build-10812/patches/PreCommit-HIVE-Build-10812.patch
 does not apply to master. Rebase required? Wrong Branch? See 
http://cwiki.apache.org/confluence/display/Hive/HowToContribute for help. 
{color} |
\\
\\
|| Subsystem || Report/Notes ||
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-10812/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Extend Load Data statement to take Input file format and Serde as parameters
> 
>
> Key: HIVE-19453
> URL: https://issues.apache.org/jira/browse/HIVE-19453
> Project: Hive
>  Issue Type: Task
>Reporter: Deepak Jaiswal
>Assignee: Deepak Jaiswal
>Priority: Major
> Attachments: HIVE-19453.01-branch-3.patch, HIVE-19453.1.patch
>
>
> Extend the load data statement to take the inputformat of the source files 
> and the serde to interpret it as parameter. For eg,
>  
> load data local inpath 
> '../../data/files/load_data_job/partitions/load_data_2_partitions.txt' INTO 
> TABLE srcbucket_mapjoin
> INPUTFORMAT 'org.apache.hadoop.mapred.TextInputFormat'
> SERDE 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe';



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19259) Create view on tables having union all fail with "Table not found"

2018-05-10 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19259?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471300#comment-16471300
 ] 

Hive QA commented on HIVE-19259:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12922652/HIVE-19259.01.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 46 failed/errored test(s), 13577 tests 
executed
*Failed tests:*
{noformat}
TestDbNotificationListener - did not produce a TEST-*.xml file (likely timed 
out) (batchId=247)
TestHCatHiveCompatibility - did not produce a TEST-*.xml file (likely timed 
out) (batchId=247)
TestNegativeCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=95)


[jira] [Commented] (HIVE-14388) Add number of rows inserted message after insert command in Beeline

2018-05-10 Thread Sahil Takiar (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-14388?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471286#comment-16471286
 ] 

Sahil Takiar commented on HIVE-14388:
-

+1 LGTM

> Add number of rows inserted message after insert command in Beeline
> ---
>
> Key: HIVE-14388
> URL: https://issues.apache.org/jira/browse/HIVE-14388
> Project: Hive
>  Issue Type: Improvement
>  Components: Beeline
>Reporter: Vihang Karajgaonkar
>Assignee: Bharathkrishna Guruvayoor Murali
>Priority: Minor
> Attachments: HIVE-14388-WIP.patch, HIVE-14388.02.patch, 
> HIVE-14388.03.patch, HIVE-14388.05.patch, HIVE-14388.06.patch, 
> HIVE-14388.07.patch, HIVE-14388.08.patch, HIVE-14388.09.patch
>
>
> Currently, when you run insert command on beeline, it returns a message 
> saying "No rows affected .."
> A better and more intuitive msg would be "xxx rows inserted (26.068 seconds)"



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19493) VectorUDFDateDiffColCol copySelected does not handle nulls correctly

2018-05-10 Thread Matt McCline (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19493?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471281#comment-16471281
 ] 

Matt McCline commented on HIVE-19493:
-

[~vihangk1] thank you for catching this. I've started writing random data 
VectorExpression unit tests (e.g. HIVE-19384 for IF statements). Given this 
miss, I after I finish CAST statements I'm going to do date/timestamp area next.

The one method propagateNullsColCol in VectorUDFDateDiffColCol makes me a 
little nervous.

I tried to add a comment to the review board but it hung...

{noformat}
output.isNull[0] = false;
{noformat}

is missing is the isRepeated and input.noNulls || !input.isNull[0] block.

The isRepeating block can be forced by inserting single rows in the Q file into 
ORC (and Parquet?) and that causes ORC to return a VectorizedRowBlock with 
isRepeating set.

I'm not sure how to stimulate the other paths other than making sure there are 
more NULLs in the data...

So, some Q file changes and/or UT changes would be great.

> VectorUDFDateDiffColCol copySelected does not handle nulls correctly
> 
>
> Key: HIVE-19493
> URL: https://issues.apache.org/jira/browse/HIVE-19493
> Project: Hive
>  Issue Type: Bug
>  Components: Vectorization
>Reporter: Vihang Karajgaonkar
>Assignee: Vihang Karajgaonkar
>Priority: Major
> Attachments: HIVE-19493.01.patch
>
>
> The {{copySelected}} method in {{VectorUDFDateDiffColCol}} class was missed 
> during HIVE-18622



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-18774) ACID: Use the _copy_N files copyNumber as the implicit statement-id

2018-05-10 Thread Eugene Koifman (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-18774?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471275#comment-16471275
 ] 

Eugene Koifman commented on HIVE-18774:
---

suppose you have 2 Load Data statements in 1 txn
1st one creates delta_7_7_0/ with _0 and _0_copy_1
2nd one creates delta_7_7_1/ with _0 and _0_copy_1

so you can't simply take N from copy_N and use it as stmt id

> ACID: Use the _copy_N files copyNumber as the implicit statement-id
> ---
>
> Key: HIVE-18774
> URL: https://issues.apache.org/jira/browse/HIVE-18774
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
> Environment: if this is not done in 3.0 it cannot be done at all
>Reporter: Gopal V
>Assignee: Eugene Koifman
>Priority: Blocker
>
> When upgrading flat ORC files to ACID, use the _copy_N numbering as a 
> statement-id to avoid having to align the row numbering between _copy_1 and 
> _copy_2 files.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19259) Create view on tables having union all fail with "Table not found"

2018-05-10 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19259?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471260#comment-16471260
 ] 

Hive QA commented on HIVE-19259:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
28s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
2s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
42s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  4m  
1s{color} | {color:blue} ql in master has 2321 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
54s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
21s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
58s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
58s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
40s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} whitespace {color} | {color:red}  0m  
0s{color} | {color:red} The patch has 28 line(s) that end in whitespace. Use 
git apply --whitespace=fix <>. Refer 
https://git-scm.com/docs/git-apply {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  4m 
13s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
56s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:red}-1{color} | {color:red} asflicense {color} | {color:red}  0m 
15s{color} | {color:red} The patch generated 1 ASF License warnings. {color} |
| {color:black}{color} | {color:black} {color} | {color:black} 23m  9s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-10810/dev-support/hive-personality.sh
 |
| git revision | master / 6c364c6 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| whitespace | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-10810/yetus/whitespace-eol.txt
 |
| asflicense | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-10810/yetus/patch-asflicense-problems.txt
 |
| modules | C: ql U: ql |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-10810/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Create view on tables having union all fail with "Table not found"
> --
>
> Key: HIVE-19259
> URL: https://issues.apache.org/jira/browse/HIVE-19259
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Affects Versions: 1.2.1
> Environment: hive-1.2.1
>  
>Reporter: Rajkumar Singh
>Assignee: Jesus Camacho Rodriguez
>Priority: Major
> Attachments: HIVE-19259.01.patch, HIVE-19259.patch
>
>
> create view on table with union work well while "union all" failed with table 
> not found, here are the reproduce steps.
> {code}
> _hive> create table foo(id int);_
> _OK_
> _Time taken: 0.401 seconds_
> _hive> create table bar(id int);_
> _OK_
>  
> _// view on table union_
> _hive> create view unionview as with tmp_1 as ( select * from foo ), tmp_2 as 
> (select * from bar ) select * from tmp_1 union  select * from tmp_2;_ 
> _OK_
> _Time taken: 0.517 seconds_
> _hive> select * from unionview;_
> _OK_
> _Time taken: 5.805 seconds_
>  
>  
> _// view on union all_ 
> _hive> create view unionallview as with tmp_1 as ( select * from foo ), tmp_2 
> as (select * from bar ) select * from tmp_1 union all  select * from tmp_2;_ 
> _OK_
> _Time taken: 1.535 seconds_
> _hive> select * from unionallview;_
> _FAILED: SemanticException Line 1:134 Table not 

[jira] [Updated] (HIVE-19370) Issue: ADD Months function on timestamp datatype fields in hive

2018-05-10 Thread Bharathkrishna Guruvayoor Murali (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19370?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bharathkrishna Guruvayoor Murali updated HIVE-19370:

Status: Patch Available  (was: Open)

> Issue: ADD Months function on timestamp datatype fields in hive
> ---
>
> Key: HIVE-19370
> URL: https://issues.apache.org/jira/browse/HIVE-19370
> Project: Hive
>  Issue Type: Bug
>Reporter: Amit Chauhan
>Assignee: Bharathkrishna Guruvayoor Murali
>Priority: Major
> Attachments: HIVE-19370.01.patch
>
>
> *Issue:*
> while using ADD_Months function on a timestamp datatype column the output 
> omits the time part[HH:MM:SS] part from output.
> which should not be the case.
> *query:* EMAIL_FAILURE_DTMZ is of datatype timestamp in hive.
> hive> select CUSTOMER_ID,EMAIL_FAILURE_DTMZ,ADD_MONTHS (EMAIL_FAILURE_DTMZ , 
> 1) from TABLE1 where CUSTOMER_ID=125674937;
> OK
> 125674937   2015-12-09 12:25:53 2016-01-09
> *hiver version :*
> hive> !hive --version;
>  Hive 1.2.1000.2.5.6.0-40
>  
> can you please help if somehow I can get below as output:
>  
> 125674937   2015-12-09 12:25:53   2016-01-09 12:25:53



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19370) Issue: ADD Months function on timestamp datatype fields in hive

2018-05-10 Thread Bharathkrishna Guruvayoor Murali (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19370?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bharathkrishna Guruvayoor Murali updated HIVE-19370:

Attachment: HIVE-19370.01.patch

> Issue: ADD Months function on timestamp datatype fields in hive
> ---
>
> Key: HIVE-19370
> URL: https://issues.apache.org/jira/browse/HIVE-19370
> Project: Hive
>  Issue Type: Bug
>Reporter: Amit Chauhan
>Assignee: Bharathkrishna Guruvayoor Murali
>Priority: Major
> Attachments: HIVE-19370.01.patch
>
>
> *Issue:*
> while using ADD_Months function on a timestamp datatype column the output 
> omits the time part[HH:MM:SS] part from output.
> which should not be the case.
> *query:* EMAIL_FAILURE_DTMZ is of datatype timestamp in hive.
> hive> select CUSTOMER_ID,EMAIL_FAILURE_DTMZ,ADD_MONTHS (EMAIL_FAILURE_DTMZ , 
> 1) from TABLE1 where CUSTOMER_ID=125674937;
> OK
> 125674937   2015-12-09 12:25:53 2016-01-09
> *hiver version :*
> hive> !hive --version;
>  Hive 1.2.1000.2.5.6.0-40
>  
> can you please help if somehow I can get below as output:
>  
> 125674937   2015-12-09 12:25:53   2016-01-09 12:25:53



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19159) TestMTQueries.testMTQueries1 failure

2018-05-10 Thread Laszlo Bodor (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19159?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471245#comment-16471245
 ] 

Laszlo Bodor commented on HIVE-19159:
-

[~vgarg]: It has passed on master. The same patch is available for branch-3.

> TestMTQueries.testMTQueries1 failure
> 
>
> Key: HIVE-19159
> URL: https://issues.apache.org/jira/browse/HIVE-19159
> Project: Hive
>  Issue Type: Sub-task
>  Components: Test
>Reporter: Vineet Garg
>Assignee: Laszlo Bodor
>Priority: Blocker
> Fix For: 3.0.0
>
> Attachments: HIVE-19159.01.patch, HIVE-19159.02-branch-3.patch, 
> HIVE-19159.02.patch, HIVE-19159.03-branch-3.patch, HIVE-19159.03.patch
>
>
> I have confirmed that HIVE-18051 caused this failure



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19307) Support ArrowOutputStream in LlapOutputFormatService

2018-05-10 Thread Eric Wohlstadter (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19307?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eric Wohlstadter updated HIVE-19307:

Attachment: HIVE-19307.5.patch

> Support ArrowOutputStream in LlapOutputFormatService
> 
>
> Key: HIVE-19307
> URL: https://issues.apache.org/jira/browse/HIVE-19307
> Project: Hive
>  Issue Type: Task
>  Components: llap
>Reporter: Eric Wohlstadter
>Assignee: Eric Wohlstadter
>Priority: Major
> Attachments: HIVE-19307.3.patch, HIVE-19307.4.patch, 
> HIVE-19307.5.patch
>
>
> Support pushing arrow batches through 
> org.apache.arrow.vector.ipc.ArrowOutputStream in LllapOutputFormatService.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19307) Support ArrowOutputStream in LlapOutputFormatService

2018-05-10 Thread Eric Wohlstadter (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19307?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eric Wohlstadter updated HIVE-19307:

Status: Patch Available  (was: In Progress)

> Support ArrowOutputStream in LlapOutputFormatService
> 
>
> Key: HIVE-19307
> URL: https://issues.apache.org/jira/browse/HIVE-19307
> Project: Hive
>  Issue Type: Task
>  Components: llap
>Reporter: Eric Wohlstadter
>Assignee: Eric Wohlstadter
>Priority: Major
> Attachments: HIVE-19307.3.patch, HIVE-19307.4.patch, 
> HIVE-19307.5.patch
>
>
> Support pushing arrow batches through 
> org.apache.arrow.vector.ipc.ArrowOutputStream in LllapOutputFormatService.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19166) TestMiniLlapLocalCliDriver sysdb failure

2018-05-10 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471234#comment-16471234
 ] 

Hive QA commented on HIVE-19166:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12922663/HIVE-19166.05.patch

{color:red}ERROR:{color} -1 due to build exiting with an error

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/10809/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/10809/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-10809/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hiveptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ date '+%Y-%m-%d %T.%3N'
2018-05-10 22:26:12.092
+ [[ -n /usr/lib/jvm/java-8-openjdk-amd64 ]]
+ export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ export 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'MAVEN_OPTS=-Xmx1g '
+ MAVEN_OPTS='-Xmx1g '
+ cd /data/hiveptest/working/
+ tee /data/hiveptest/logs/PreCommit-HIVE-Build-10809/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ git = \s\v\n ]]
+ [[ git = \g\i\t ]]
+ [[ -z master ]]
+ [[ -d apache-github-source-source ]]
+ [[ ! -d apache-github-source-source/.git ]]
+ [[ ! -d apache-github-source-source ]]
+ date '+%Y-%m-%d %T.%3N'
2018-05-10 22:26:12.095
+ cd apache-github-source-source
+ git fetch origin
>From https://github.com/apache/hive
   2d8e48c..6c364c6  master -> origin/master
+ git reset --hard HEAD
HEAD is now at 2d8e48c HIVE-19472: HiveStreamingConnection swallows exception 
on partition creation (Prasanth Jayachandran reviewed by Sergey Shelukhin)
+ git clean -f -d
+ git checkout master
Already on 'master'
Your branch is behind 'origin/master' by 2 commits, and can be fast-forwarded.
  (use "git pull" to update your local branch)
+ git reset --hard origin/master
HEAD is now at 6c364c6 HIVE-19306: Arrow batch serializer (Teddy Choi, reviewed 
by Matt McCline and Eric Wohlstadter (non-binding))
+ git merge --ff-only origin/master
Already up-to-date.
+ date '+%Y-%m-%d %T.%3N'
2018-05-10 22:26:17.298
+ rm -rf ../yetus_PreCommit-HIVE-Build-10809
+ mkdir ../yetus_PreCommit-HIVE-Build-10809
+ git gc
+ cp -R . ../yetus_PreCommit-HIVE-Build-10809
+ mkdir /data/hiveptest/logs/PreCommit-HIVE-Build-10809/yetus
+ patchCommandPath=/data/hiveptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hiveptest/working/scratch/build.patch
+ [[ -f /data/hiveptest/working/scratch/build.patch ]]
+ chmod +x /data/hiveptest/working/scratch/smart-apply-patch.sh
+ /data/hiveptest/working/scratch/smart-apply-patch.sh 
/data/hiveptest/working/scratch/build.patch
error: patch failed: ql/src/test/results/clientpositive/llap/sysdb.q.out:273
Falling back to three-way merge...
Applied patch to 'ql/src/test/results/clientpositive/llap/sysdb.q.out' with 
conflicts.
Going to apply patch with: git apply -p0
error: patch failed: ql/src/test/results/clientpositive/llap/sysdb.q.out:273
Falling back to three-way merge...
Applied patch to 'ql/src/test/results/clientpositive/llap/sysdb.q.out' with 
conflicts.
U ql/src/test/results/clientpositive/llap/sysdb.q.out
+ exit 1
'
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12922663 - PreCommit-HIVE-Build

> TestMiniLlapLocalCliDriver sysdb failure
> 
>
> Key: HIVE-19166
> URL: https://issues.apache.org/jira/browse/HIVE-19166
> Project: Hive
>  Issue Type: Sub-task
>  Components: Test
>Reporter: Vineet Garg
>Assignee: Zoltan Haindrich
>Priority: Major
> Attachments: HIVE-19166.04.patch, HIVE-19166.05.patch, 
> HIVE-19166.1.patch, HIVE-19166.2.patch, HIVE-19166.3.patch
>
>
> Broken by HIVE-18715



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19159) TestMTQueries.testMTQueries1 failure

2018-05-10 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19159?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471228#comment-16471228
 ] 

Hive QA commented on HIVE-19159:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12922650/HIVE-19159.03.patch

{color:green}SUCCESS:{color} +1 due to 2 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 42 failed/errored test(s), 13546 tests 
executed
*Failed tests:*
{noformat}
TestDbNotificationListener - did not produce a TEST-*.xml file (likely timed 
out) (batchId=247)
TestHCatHiveCompatibility - did not produce a TEST-*.xml file (likely timed 
out) (batchId=247)
TestNegativeCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=95)


[jira] [Updated] (HIVE-19440) Make StorageBasedAuthorizer work with information schema

2018-05-10 Thread Daniel Dai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19440?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daniel Dai updated HIVE-19440:
--
Status: Patch Available  (was: Open)

> Make StorageBasedAuthorizer work with information schema
> 
>
> Key: HIVE-19440
> URL: https://issues.apache.org/jira/browse/HIVE-19440
> Project: Hive
>  Issue Type: Improvement
>Reporter: Daniel Dai
>Assignee: Daniel Dai
>Priority: Blocker
> Fix For: 3.0.0
>
> Attachments: HIVE-19440.1.patch
>
>
> With HIVE-19161, Hive information schema works with external authorizer (such 
> as ranger). However, we also need to make StorageBasedAuthorizer 
> synchronization work as it is also widely use.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19440) Make StorageBasedAuthorizer work with information schema

2018-05-10 Thread Daniel Dai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19440?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daniel Dai updated HIVE-19440:
--
Attachment: HIVE-19440.1.patch

> Make StorageBasedAuthorizer work with information schema
> 
>
> Key: HIVE-19440
> URL: https://issues.apache.org/jira/browse/HIVE-19440
> Project: Hive
>  Issue Type: Improvement
>Reporter: Daniel Dai
>Assignee: Daniel Dai
>Priority: Blocker
> Fix For: 3.0.0
>
> Attachments: HIVE-19440.1.patch
>
>
> With HIVE-19161, Hive information schema works with external authorizer (such 
> as ranger). However, we also need to make StorageBasedAuthorizer 
> synchronization work as it is also widely use.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19381) Function replication in cloud fail when download resource from AWS

2018-05-10 Thread Daniel Dai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19381?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daniel Dai updated HIVE-19381:
--
Attachment: HIVE-19381.3.patch

> Function replication in cloud fail when download resource from AWS
> --
>
> Key: HIVE-19381
> URL: https://issues.apache.org/jira/browse/HIVE-19381
> Project: Hive
>  Issue Type: Bug
>  Components: repl
>Reporter: Daniel Dai
>Assignee: Daniel Dai
>Priority: Major
> Fix For: 3.0.0, 3.1.0
>
> Attachments: HIVE-19381.1.patch, HIVE-19381.2.patch, 
> HIVE-19381.3.patch
>
>
> Another case replication shall use the config in with clause.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19494) Accept shade prefix during reflective instantiation of output format

2018-05-10 Thread Prasanth Jayachandran (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19494?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471216#comment-16471216
 ] 

Prasanth Jayachandran commented on HIVE-19494:
--

Only streaming ingest encountered this recently, but technically this can 
happen for the case where multiple hive versions are involved. Spark compiles 
against 1.2.1 but streaming ingest API requires hive 3.0.0 and in its codepath, 
so only this instantiation in hive streaming API fails as of now (in my limited 
testing). I left the config generic enough for future in case if someone runs 
into the same issue they can reuse the hive config. 

> Accept shade prefix during reflective instantiation of output format
> 
>
> Key: HIVE-19494
> URL: https://issues.apache.org/jira/browse/HIVE-19494
> Project: Hive
>  Issue Type: Sub-task
>  Components: Streaming
>Affects Versions: 3.0.0, 3.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
>Priority: Major
> Attachments: HIVE-19494.1.patch
>
>
> Hive Streaming API jars are sometimes shaded with a different prefix when 
> used in environments where another version of hive already exists (spark for 
> example). In most cases, shading is done with rename of classes with some 
> prefix. If an uber/assembly jar is generated with renamed prefix, Hive 
> Streaming API will not work as Hive Streaming API will reflectively 
> instantiate outputformat class using FQCN string provided by metastore table 
> storage descriptor object. 
> For example: 
> RecordWriter will create instance of OutputFormat using string 
> "org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat". When a shaded jar with 
> renamed class references are used, this class will not be found by the 
> classloader. 
> We can optionally accept a shade prefix from user via config which will be 
> tried (as fallback) when ClassNotFoundException is thrown.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19494) Accept shade prefix during reflective instantiation of output format

2018-05-10 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19494?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471200#comment-16471200
 ] 

Sergey Shelukhin commented on HIVE-19494:
-

The name and description seem to indicate this is a global setting for 
everywhere Hive does this, however it's only used in streaming. Should it be 
scoped to streaming w.r.t. name and desc? 
Otherwise +1

> Accept shade prefix during reflective instantiation of output format
> 
>
> Key: HIVE-19494
> URL: https://issues.apache.org/jira/browse/HIVE-19494
> Project: Hive
>  Issue Type: Sub-task
>  Components: Streaming
>Affects Versions: 3.0.0, 3.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
>Priority: Major
> Attachments: HIVE-19494.1.patch
>
>
> Hive Streaming API jars are sometimes shaded with a different prefix when 
> used in environments where another version of hive already exists (spark for 
> example). In most cases, shading is done with rename of classes with some 
> prefix. If an uber/assembly jar is generated with renamed prefix, Hive 
> Streaming API will not work as Hive Streaming API will reflectively 
> instantiate outputformat class using FQCN string provided by metastore table 
> storage descriptor object. 
> For example: 
> RecordWriter will create instance of OutputFormat using string 
> "org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat". When a shaded jar with 
> renamed class references are used, this class will not be found by the 
> classloader. 
> We can optionally accept a shade prefix from user via config which will be 
> tried (as fallback) when ClassNotFoundException is thrown.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19493) VectorUDFDateDiffColCol copySelected does not handle nulls correctly

2018-05-10 Thread Vihang Karajgaonkar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19493?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vihang Karajgaonkar updated HIVE-19493:
---
Status: Patch Available  (was: Open)

> VectorUDFDateDiffColCol copySelected does not handle nulls correctly
> 
>
> Key: HIVE-19493
> URL: https://issues.apache.org/jira/browse/HIVE-19493
> Project: Hive
>  Issue Type: Bug
>  Components: Vectorization
>Reporter: Vihang Karajgaonkar
>Assignee: Vihang Karajgaonkar
>Priority: Major
> Attachments: HIVE-19493.01.patch
>
>
> The {{copySelected}} method in {{VectorUDFDateDiffColCol}} class was missed 
> during HIVE-18622



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19493) VectorUDFDateDiffColCol copySelected does not handle nulls correctly

2018-05-10 Thread Vihang Karajgaonkar (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19493?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471198#comment-16471198
 ] 

Vihang Karajgaonkar commented on HIVE-19493:


adding review board link. [~mmccline] Can you please review?

> VectorUDFDateDiffColCol copySelected does not handle nulls correctly
> 
>
> Key: HIVE-19493
> URL: https://issues.apache.org/jira/browse/HIVE-19493
> Project: Hive
>  Issue Type: Bug
>  Components: Vectorization
>Reporter: Vihang Karajgaonkar
>Assignee: Vihang Karajgaonkar
>Priority: Major
> Attachments: HIVE-19493.01.patch
>
>
> The {{copySelected}} method in {{VectorUDFDateDiffColCol}} class was missed 
> during HIVE-18622



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19493) VectorUDFDateDiffColCol copySelected does not handle nulls correctly

2018-05-10 Thread Vihang Karajgaonkar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19493?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vihang Karajgaonkar updated HIVE-19493:
---
Attachment: HIVE-19493.01.patch

> VectorUDFDateDiffColCol copySelected does not handle nulls correctly
> 
>
> Key: HIVE-19493
> URL: https://issues.apache.org/jira/browse/HIVE-19493
> Project: Hive
>  Issue Type: Bug
>  Components: Vectorization
>Reporter: Vihang Karajgaonkar
>Assignee: Vihang Karajgaonkar
>Priority: Major
> Attachments: HIVE-19493.01.patch
>
>
> The {{copySelected}} method in {{VectorUDFDateDiffColCol}} class was missed 
> during HIVE-18622



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19494) Accept shade prefix during reflective instantiation of output format

2018-05-10 Thread Prasanth Jayachandran (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19494?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471192#comment-16471192
 ] 

Prasanth Jayachandran commented on HIVE-19494:
--

[~sershe] can you please take a look? small patch that accepts a prefix for 
classloading. 

> Accept shade prefix during reflective instantiation of output format
> 
>
> Key: HIVE-19494
> URL: https://issues.apache.org/jira/browse/HIVE-19494
> Project: Hive
>  Issue Type: Sub-task
>  Components: Streaming
>Affects Versions: 3.0.0, 3.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
>Priority: Major
> Attachments: HIVE-19494.1.patch
>
>
> Hive Streaming API jars are sometimes shaded with a different prefix when 
> used in environments where another version of hive already exists (spark for 
> example). In most cases, shading is done with rename of classes with some 
> prefix. If an uber/assembly jar is generated with renamed prefix, Hive 
> Streaming API will not work as Hive Streaming API will reflectively 
> instantiate outputformat class using FQCN string provided by metastore table 
> storage descriptor object. 
> For example: 
> RecordWriter will create instance of OutputFormat using string 
> "org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat". When a shaded jar with 
> renamed class references are used, this class will not be found by the 
> classloader. 
> We can optionally accept a shade prefix from user via config which will be 
> tried (as fallback) when ClassNotFoundException is thrown.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19494) Accept shade prefix during reflective instantiation of output format

2018-05-10 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19494?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran updated HIVE-19494:
-
Attachment: HIVE-19494.1.patch

> Accept shade prefix during reflective instantiation of output format
> 
>
> Key: HIVE-19494
> URL: https://issues.apache.org/jira/browse/HIVE-19494
> Project: Hive
>  Issue Type: Sub-task
>  Components: Streaming
>Affects Versions: 3.0.0, 3.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
>Priority: Major
> Attachments: HIVE-19494.1.patch
>
>
> Hive Streaming API jars are sometimes shaded with a different prefix when 
> used in environments where another version of hive already exists (spark for 
> example). In most cases, shading is done with rename of classes with some 
> prefix. If an uber/assembly jar is generated with renamed prefix, Hive 
> Streaming API will not work as Hive Streaming API will reflectively 
> instantiate outputformat class using FQCN string provided by metastore table 
> storage descriptor object. 
> For example: 
> RecordWriter will create instance of OutputFormat using string 
> "org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat". When a shaded jar with 
> renamed class references are used, this class will not be found by the 
> classloader. 
> We can optionally accept a shade prefix from user via config which will be 
> tried (as fallback) when ClassNotFoundException is thrown.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19494) Accept shade prefix during reflective instantiation of output format

2018-05-10 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19494?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran updated HIVE-19494:
-
Status: Patch Available  (was: Open)

> Accept shade prefix during reflective instantiation of output format
> 
>
> Key: HIVE-19494
> URL: https://issues.apache.org/jira/browse/HIVE-19494
> Project: Hive
>  Issue Type: Sub-task
>  Components: Streaming
>Affects Versions: 3.0.0, 3.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
>Priority: Major
> Attachments: HIVE-19494.1.patch
>
>
> Hive Streaming API jars are sometimes shaded with a different prefix when 
> used in environments where another version of hive already exists (spark for 
> example). In most cases, shading is done with rename of classes with some 
> prefix. If an uber/assembly jar is generated with renamed prefix, Hive 
> Streaming API will not work as Hive Streaming API will reflectively 
> instantiate outputformat class using FQCN string provided by metastore table 
> storage descriptor object. 
> For example: 
> RecordWriter will create instance of OutputFormat using string 
> "org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat". When a shaded jar with 
> renamed class references are used, this class will not be found by the 
> classloader. 
> We can optionally accept a shade prefix from user via config which will be 
> tried (as fallback) when ClassNotFoundException is thrown.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19464) Upgrade Parquet to 1.10.0

2018-05-10 Thread Prasanth Jayachandran (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19464?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471189#comment-16471189
 ] 

Prasanth Jayachandran commented on HIVE-19464:
--

+1

> Upgrade Parquet to 1.10.0
> -
>
> Key: HIVE-19464
> URL: https://issues.apache.org/jira/browse/HIVE-19464
> Project: Hive
>  Issue Type: Improvement
>  Components: File Formats
>Affects Versions: 3.0.0
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
>Priority: Major
> Attachments: HIVE-19464.01.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19496) Check untar folder

2018-05-10 Thread Sahil Takiar (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19496?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471186#comment-16471186
 ] 

Sahil Takiar commented on HIVE-19496:
-

+1 LGTM

> Check untar folder
> --
>
> Key: HIVE-19496
> URL: https://issues.apache.org/jira/browse/HIVE-19496
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Aihua Xu
>Assignee: Aihua Xu
>Priority: Major
> Attachments: HIVE-19496.1.patch
>
>
> We need to check if the file is under untar folder.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19496) Check untar folder

2018-05-10 Thread Aihua Xu (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19496?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471185#comment-16471185
 ] 

Aihua Xu commented on HIVE-19496:
-

[~stakiar_impala_496e] Can you review the change?

> Check untar folder
> --
>
> Key: HIVE-19496
> URL: https://issues.apache.org/jira/browse/HIVE-19496
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Aihua Xu
>Assignee: Aihua Xu
>Priority: Major
> Attachments: HIVE-19496.1.patch
>
>
> We need to check if the file is under untar folder.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19496) Check untar folder

2018-05-10 Thread Aihua Xu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19496?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aihua Xu updated HIVE-19496:

Status: Patch Available  (was: Open)

> Check untar folder
> --
>
> Key: HIVE-19496
> URL: https://issues.apache.org/jira/browse/HIVE-19496
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Aihua Xu
>Assignee: Aihua Xu
>Priority: Major
> Attachments: HIVE-19496.1.patch
>
>
> We need to check if the file is under untar folder.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19496) Check untar folder

2018-05-10 Thread Aihua Xu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19496?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aihua Xu updated HIVE-19496:

Attachment: HIVE-19496.1.patch

> Check untar folder
> --
>
> Key: HIVE-19496
> URL: https://issues.apache.org/jira/browse/HIVE-19496
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Aihua Xu
>Assignee: Aihua Xu
>Priority: Major
> Attachments: HIVE-19496.1.patch
>
>
> We need to check if the file is under untar folder.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19496) Check untar folder

2018-05-10 Thread Aihua Xu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19496?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aihua Xu updated HIVE-19496:

Description: We need to check if the file is under untar folder.  (was: We 
need to check untar folder.)

> Check untar folder
> --
>
> Key: HIVE-19496
> URL: https://issues.apache.org/jira/browse/HIVE-19496
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Aihua Xu
>Assignee: Aihua Xu
>Priority: Major
>
> We need to check if the file is under untar folder.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-19496) Check untar folder

2018-05-10 Thread Aihua Xu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19496?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aihua Xu reassigned HIVE-19496:
---


> Check untar folder
> --
>
> Key: HIVE-19496
> URL: https://issues.apache.org/jira/browse/HIVE-19496
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Aihua Xu
>Assignee: Aihua Xu
>Priority: Major
>
> We need to check untar folder.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-19495) Arrow SerDe itest failure

2018-05-10 Thread Eric Wohlstadter (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19495?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eric Wohlstadter reassigned HIVE-19495:
---


> Arrow SerDe itest failure
> -
>
> Key: HIVE-19495
> URL: https://issues.apache.org/jira/browse/HIVE-19495
> Project: Hive
>  Issue Type: Sub-task
>  Components: Serializers/Deserializers
>Reporter: Eric Wohlstadter
>Assignee: Teddy Choi
>Priority: Major
> Fix For: 3.1.0
>
>
> "You tried to write a Bit type when you are using a ValueWriter of type 
> NullableMapWriter."



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19306) Arrow batch serializer

2018-05-10 Thread Matt McCline (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19306?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matt McCline updated HIVE-19306:

Fix Version/s: 3.1.0

> Arrow batch serializer
> --
>
> Key: HIVE-19306
> URL: https://issues.apache.org/jira/browse/HIVE-19306
> Project: Hive
>  Issue Type: Task
>  Components: Serializers/Deserializers
>Reporter: Eric Wohlstadter
>Assignee: Teddy Choi
>Priority: Major
> Fix For: 3.1.0
>
> Attachments: HIVE-19306.2.patch, HIVE-19306.3.patch, 
> HIVE-19306.4.patch
>
>
> Leverage the ThriftJDBCBinarySerDe code path that already exists in 
> SemanticAnalyzer/FileSinkOperator to create a serializer that batches rows 
> into Arrow vector batches.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19306) Arrow batch serializer

2018-05-10 Thread Matt McCline (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19306?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471166#comment-16471166
 ] 

Matt McCline commented on HIVE-19306:
-

Committed to master.

> Arrow batch serializer
> --
>
> Key: HIVE-19306
> URL: https://issues.apache.org/jira/browse/HIVE-19306
> Project: Hive
>  Issue Type: Task
>  Components: Serializers/Deserializers
>Reporter: Eric Wohlstadter
>Assignee: Teddy Choi
>Priority: Major
> Fix For: 3.1.0
>
> Attachments: HIVE-19306.2.patch, HIVE-19306.3.patch, 
> HIVE-19306.4.patch
>
>
> Leverage the ThriftJDBCBinarySerDe code path that already exists in 
> SemanticAnalyzer/FileSinkOperator to create a serializer that batches rows 
> into Arrow vector batches.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19306) Arrow batch serializer

2018-05-10 Thread Matt McCline (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19306?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matt McCline updated HIVE-19306:

Resolution: Fixed
Status: Resolved  (was: Patch Available)

> Arrow batch serializer
> --
>
> Key: HIVE-19306
> URL: https://issues.apache.org/jira/browse/HIVE-19306
> Project: Hive
>  Issue Type: Task
>  Components: Serializers/Deserializers
>Reporter: Eric Wohlstadter
>Assignee: Teddy Choi
>Priority: Major
> Fix For: 3.1.0
>
> Attachments: HIVE-19306.2.patch, HIVE-19306.3.patch, 
> HIVE-19306.4.patch
>
>
> Leverage the ThriftJDBCBinarySerDe code path that already exists in 
> SemanticAnalyzer/FileSinkOperator to create a serializer that batches rows 
> into Arrow vector batches.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (HIVE-17547) MoveTask for Acid tables race condition

2018-05-10 Thread Eugene Koifman (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17547?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eugene Koifman resolved HIVE-17547.
---
Resolution: Not A Problem

since commit() of the txn happens after all tasks are complete, no one will 
read contents of this delta until it has all the files

> MoveTask for Acid tables race condition
> ---
>
> Key: HIVE-17547
> URL: https://issues.apache.org/jira/browse/HIVE-17547
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Affects Versions: 1.0.0
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
>Priority: Major
>
> Consider Hive.moveAcidFiles()
> it starts out with something like
> {noformat}
>   └── -ext-1
> │   └── 00_0
> │   ├── _orc_acid_version
> │   └── delta_019_019
> │   └── bucket_0
> │   └── 00_1
> │   ├── _orc_acid_version
> │   └── delta_019_019
> │   └── bucket_1
> {noformat}
> for a write to a bucketed table.
> The "move" handles each 00_N separately.  The first on creates 
> delta_019_019 under the table/partition dir, the others just add 
> bucket_N there.
> That means there is a small window where someone may "ls 
> table/part/delta_019_019" and not see all the buckets.
> Once Acid writes directly to the final location (a la MM tables) this issue 
> resolves automatically since txn 19 is uncommitted until everything is 
> written.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19306) Arrow batch serializer

2018-05-10 Thread Matt McCline (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19306?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471165#comment-16471165
 ] 

Matt McCline commented on HIVE-19306:
-

+1 LGTM

> Arrow batch serializer
> --
>
> Key: HIVE-19306
> URL: https://issues.apache.org/jira/browse/HIVE-19306
> Project: Hive
>  Issue Type: Task
>  Components: Serializers/Deserializers
>Reporter: Eric Wohlstadter
>Assignee: Teddy Choi
>Priority: Major
> Attachments: HIVE-19306.2.patch, HIVE-19306.3.patch, 
> HIVE-19306.4.patch
>
>
> Leverage the ThriftJDBCBinarySerDe code path that already exists in 
> SemanticAnalyzer/FileSinkOperator to create a serializer that batches rows 
> into Arrow vector batches.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19490) Locking on Insert into for non native and managed tables.

2018-05-10 Thread Jesus Camacho Rodriguez (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19490?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471161#comment-16471161
 ] 

Jesus Camacho Rodriguez commented on HIVE-19490:


Since this may be dependent on the handler itself (and the system that it 
relies on), we can also add a method to the storage handler interface to 
indicate whether 'insert into' needs shared or exclusive write lock.

> Locking on Insert into for non native and managed tables.
> -
>
> Key: HIVE-19490
> URL: https://issues.apache.org/jira/browse/HIVE-19490
> Project: Hive
>  Issue Type: Improvement
>  Components: Druid integration
>Reporter: slim bouguerra
>Assignee: slim bouguerra
>Priority: Major
>  Labels: druid, locking
> Attachments: HIVE-19490.patch
>
>
> Current state of the art: 
> Managed non native table like Druid Tables, will need to get a Lock on Insert 
> into or insert Over write. The nature of this lock is set to Exclusive by 
> default for any non native table.
> This implies that Inserts into Druid table will Lock any read query as well 
> during the execution of the insert into. IMO this lock (on insert into) is  
> not needed since the insert statement is appending data and the state of 
> loading it is managed partially by Hive Storage handler hook and part of it 
> by Druid. 
> What i am proposing is to relax the lock level to shared for all non native 
> tables on insert into operations and keep it as Exclusive Write for insert 
> Overwrite for now.
>  
> Any feedback is welcome.
> cc [~ekoifman] / [~ashutoshc] / [~jdere] / [~hagleitn]
> Also am not sure what is the best way to unit test this currently am using 
> debugger to check if locks are what i except, please let me know if there is 
> a better way to do this. 
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19159) TestMTQueries.testMTQueries1 failure

2018-05-10 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19159?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471132#comment-16471132
 ] 

Hive QA commented on HIVE-19159:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
55s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  6m 
 8s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
2s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
28s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
41s{color} | {color:blue} itests/util in master has 55 extant Findbugs 
warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
35s{color} | {color:blue} itests/hive-unit in master has 2 extant Findbugs 
warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
40s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
26s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
 7s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
0s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m  
0s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
13s{color} | {color:green} itests/util: The patch generated 0 new + 70 
unchanged - 2 fixed = 70 total (was 72) {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
15s{color} | {color:green} The patch hive-unit passed checkstyle {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
32s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
39s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:red}-1{color} | {color:red} asflicense {color} | {color:red}  0m 
13s{color} | {color:red} The patch generated 1 ASF License warnings. {color} |
| {color:black}{color} | {color:black} {color} | {color:black} 17m 28s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-10808/dev-support/hive-personality.sh
 |
| git revision | master / 2d8e48c |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| asflicense | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-10808/yetus/patch-asflicense-problems.txt
 |
| modules | C: itests/util itests/hive-unit U: itests |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-10808/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> TestMTQueries.testMTQueries1 failure
> 
>
> Key: HIVE-19159
> URL: https://issues.apache.org/jira/browse/HIVE-19159
> Project: Hive
>  Issue Type: Sub-task
>  Components: Test
>Reporter: Vineet Garg
>Assignee: Laszlo Bodor
>Priority: Blocker
> Fix For: 3.0.0
>
> Attachments: HIVE-19159.01.patch, HIVE-19159.02-branch-3.patch, 
> HIVE-19159.02.patch, HIVE-19159.03-branch-3.patch, HIVE-19159.03.patch
>
>
> I have confirmed that HIVE-18051 caused this failure



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19466) Update constraint violation error message

2018-05-10 Thread Vineet Garg (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19466?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-19466:
---
Status: Patch Available  (was: Open)

> Update constraint violation error message
> -
>
> Key: HIVE-19466
> URL: https://issues.apache.org/jira/browse/HIVE-19466
> Project: Hive
>  Issue Type: Improvement
>Reporter: Vineet Garg
>Assignee: Vineet Garg
>Priority: Major
> Attachments: HIVE-19466.01-branch-3.patch, HIVE-19466.1.patch
>
>
> Currently for both CHECK and NOT NULL constraint violation hive throws {{NOT 
> NULL Constraint violated}}.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19466) Update constraint violation error message

2018-05-10 Thread Vineet Garg (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19466?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-19466:
---
Status: Open  (was: Patch Available)

> Update constraint violation error message
> -
>
> Key: HIVE-19466
> URL: https://issues.apache.org/jira/browse/HIVE-19466
> Project: Hive
>  Issue Type: Improvement
>Reporter: Vineet Garg
>Assignee: Vineet Garg
>Priority: Major
> Attachments: HIVE-19466.01-branch-3.patch, HIVE-19466.1.patch
>
>
> Currently for both CHECK and NOT NULL constraint violation hive throws {{NOT 
> NULL Constraint violated}}.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19421) Upgrade version of Jetty to 9.3.20.v20170531

2018-05-10 Thread Janaki Lahorani (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19421?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Janaki Lahorani updated HIVE-19421:
---
Attachment: HIVE-19421.3.patch

> Upgrade version of Jetty to 9.3.20.v20170531
> 
>
> Key: HIVE-19421
> URL: https://issues.apache.org/jira/browse/HIVE-19421
> Project: Hive
>  Issue Type: Improvement
>Reporter: Janaki Lahorani
>Assignee: Janaki Lahorani
>Priority: Major
> Attachments: HIVE-19421.1.patch, HIVE-19421.2.patch, 
> HIVE-19421.3.patch, HIVE-19421.3.patch
>
>
> Move Jetty up to 9.3.20.v20170531



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19466) Update constraint violation error message

2018-05-10 Thread Vineet Garg (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19466?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-19466:
---
Attachment: HIVE-19466.01-branch-3.patch

> Update constraint violation error message
> -
>
> Key: HIVE-19466
> URL: https://issues.apache.org/jira/browse/HIVE-19466
> Project: Hive
>  Issue Type: Improvement
>Reporter: Vineet Garg
>Assignee: Vineet Garg
>Priority: Major
> Attachments: HIVE-19466.01-branch-3.patch, HIVE-19466.1.patch
>
>
> Currently for both CHECK and NOT NULL constraint violation hive throws {{NOT 
> NULL Constraint violated}}.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19476) Fix failures in TestReplicationScenariosAcidTables, TestReplicationOnHDFSEncryptedZones and TestCopyUtils

2018-05-10 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19476?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471092#comment-16471092
 ] 

Hive QA commented on HIVE-19476:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12922637/HIVE-19476.01.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 39 failed/errored test(s), 13546 tests 
executed
*Failed tests:*
{noformat}
TestDbNotificationListener - did not produce a TEST-*.xml file (likely timed 
out) (batchId=247)
TestHCatHiveCompatibility - did not produce a TEST-*.xml file (likely timed 
out) (batchId=247)
TestNegativeCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=95)


[jira] [Updated] (HIVE-19490) Locking on Insert into for non native and managed tables.

2018-05-10 Thread slim bouguerra (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19490?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

slim bouguerra updated HIVE-19490:
--
Description: 
Current state of the art: 

Managed non native table like Druid Tables, will need to get a Lock on Insert 
into or insert Over write. The nature of this lock is set to Exclusive by 
default for any non native table.

This implies that Inserts into Druid table will Lock any read query as well 
during the execution of the insert into. IMO this lock (on insert into) is  not 
needed since the insert statement is appending data and the state of loading it 
is managed partially by Hive Storage handler hook and part of it by Druid. 

What i am proposing is to relax the lock level to shared for all non native 
tables on insert into operations and keep it as Exclusive Write for insert 
Overwrite for now.

 

Any feedback is welcome.

cc [~ekoifman] / [~ashutoshc] / [~jdere] / [~hagleitn]

Also am not sure what is the best way to unit test this currently am using 
debugger to check if locks are what i except, please let me know if there is a 
better way to do this. 

 

  was:
Current state of the art: 

Managed non native table like Druid Tables, will need to get a Lock on Insert 
into or insert Over write. The nature of this lock is set to Exclusive by 
default for any non native table.

This implies that Inserts into Druid table will Lock any read query as well 
during the execution of the insert into. IMO this lock (on insert into) is  not 
needed since the insert statement is appending data and the state of loading it 
is managed partially by Hive Storage handler hook and part of it by Druid. 

What i am proposing is to relax the lock level to shared for all non native 
tables on insert into operations and keep it as Exclusive Write for insert 
Overwrite for now.

 

Any feedback is welcome.

cc [~ekoifman] / [~ashutoshc] / [~jdere] / [~hagleitn]

Also am not sure what is the best way to unit test this currently am using 
debugger to check of locks are what i except, please let me know if there is a 
better way to do this. 

 


> Locking on Insert into for non native and managed tables.
> -
>
> Key: HIVE-19490
> URL: https://issues.apache.org/jira/browse/HIVE-19490
> Project: Hive
>  Issue Type: Improvement
>  Components: Druid integration
>Reporter: slim bouguerra
>Assignee: slim bouguerra
>Priority: Major
>  Labels: druid, locking
> Attachments: HIVE-19490.patch
>
>
> Current state of the art: 
> Managed non native table like Druid Tables, will need to get a Lock on Insert 
> into or insert Over write. The nature of this lock is set to Exclusive by 
> default for any non native table.
> This implies that Inserts into Druid table will Lock any read query as well 
> during the execution of the insert into. IMO this lock (on insert into) is  
> not needed since the insert statement is appending data and the state of 
> loading it is managed partially by Hive Storage handler hook and part of it 
> by Druid. 
> What i am proposing is to relax the lock level to shared for all non native 
> tables on insert into operations and keep it as Exclusive Write for insert 
> Overwrite for now.
>  
> Any feedback is welcome.
> cc [~ekoifman] / [~ashutoshc] / [~jdere] / [~hagleitn]
> Also am not sure what is the best way to unit test this currently am using 
> debugger to check if locks are what i except, please let me know if there is 
> a better way to do this. 
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19250) Schema column definitions inconsistencies in MySQL

2018-05-10 Thread Naveen Gangam (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19250?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471057#comment-16471057
 ] 

Naveen Gangam commented on HIVE-19250:
--

Actually they ran .. there has been only 1 version of the patch. There were 
some failures but I cannot analyze them as the build is too old. So a re-run 
would help.

> Schema column definitions inconsistencies in MySQL
> --
>
> Key: HIVE-19250
> URL: https://issues.apache.org/jira/browse/HIVE-19250
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 2.1.0
>Reporter: Naveen Gangam
>Assignee: Naveen Gangam
>Priority: Minor
> Attachments: HIVE-19250.patch, HIVE-19250.patch
>
>
> There are some inconsistencies in column definitions in MySQL between a 
> schema that was upgraded to 2.1 (from an older release) vs installing the 
> 2.1.0 schema directly.
> >   `CQ_TBLPROPERTIES` varchar(2048) DEFAULT NULL,
> 117d117
> <   `CQ_TBLPROPERTIES` varchar(2048) DEFAULT NULL,
> 135a136
> >   `CC_TBLPROPERTIES` varchar(2048) DEFAULT NULL,
> 143d143
> <   `CC_TBLPROPERTIES` varchar(2048) DEFAULT NULL,
> 156c156
> <   `CTC_TXNID` bigint(20) DEFAULT NULL,
> ---
> >   `CTC_TXNID` bigint(20) NOT NULL,
> 158c158
> <   `CTC_TABLE` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT 
> NULL,
> ---
> >   `CTC_TABLE` varchar(256) DEFAULT NULL,
> 476c476
> <   `TBL_NAME` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT 
> NULL,
> ---
> >   `TBL_NAME` varchar(256) DEFAULT NULL,
> 664c664
> <   KEY `PCS_STATS_IDX` 
> (`DB_NAME`,`TABLE_NAME`,`COLUMN_NAME`,`PARTITION_NAME`),
> ---
> >   KEY `PCS_STATS_IDX` 
> > (`DB_NAME`,`TABLE_NAME`,`COLUMN_NAME`,`PARTITION_NAME`) USING BTREE,
> 768c768
> <   `PARAM_VALUE` mediumtext,
> ---
> >   `PARAM_VALUE` mediumtext CHARACTER SET latin1 COLLATE latin1_bin,
> 814c814
> <   `PARAM_VALUE` mediumtext,
> ---
> >   `PARAM_VALUE` mediumtext CHARACTER SET latin1 COLLATE latin1_bin,
> 934c934
> <   `PARAM_VALUE` mediumtext,
> ---
> >   `PARAM_VALUE` mediumtext CHARACTER SET latin1 COLLATE latin1_bin,
> 1066d1065
> <   `TXN_HEARTBEAT_COUNT` int(11) DEFAULT NULL,
> 1067a1067
> >   `TXN_HEARTBEAT_COUNT` int(11) DEFAULT NULL,
> 1080c1080
> <   `TC_TXNID` bigint(20) DEFAULT NULL,
> ---
> >   `TC_TXNID` bigint(20) NOT NULL,
> 1082c1082
> <   `TC_TABLE` varchar(128) DEFAULT NULL,
> ---
> >   `TC_TABLE` varchar(128) NOT NULL,
> 1084c1084
> <   `TC_OPERATION_TYPE` char(1) DEFAULT NULL,
> ---
> >   `TC_OPERATION_TYPE` char(1) NOT NULL,



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19494) Accept shade prefix during reflective instantiation of output format

2018-05-10 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19494?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran updated HIVE-19494:
-
Description: 
Hive Streaming API jars are sometimes shaded with a different prefix when used 
in environments where another version of hive already exists (spark for 
example). In most cases, shading is done with rename of classes with some 
prefix. If an uber/assembly jar is generated with renamed prefix, Hive 
Streaming API will not work as Hive Streaming API will reflectively instantiate 
outputformat class using FQCN string provided by metastore table storage 
descriptor object. 
For example: 
RecordWriter will create instance of OutputFormat using string 
"org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat". When a shaded jar with 
renamed class references are used, this class will not be found by the 
classloader. 

We can optionally accept a shade prefix from user via config which will be 
tried (as fallback) when ClassNotFoundException is thrown.

  was:
Hive Streaming API jars are sometimes shaded with a different prefix when used 
in environments where another version of hive already exists (spark for 
example). In most cases, shading is done with rename of classes with some 
prefix. If an uber/assembly jar is generated with renamed prefix, Hive 
Streaming API will not work as Hive Streaming API will reflectively instantiate 
outputformat class using FQCN string provided by metastore table storage 
descriptor object. 
For example: 
RecordWriter will create instance of OutputFormat using string 
"org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat". When a shaded jar with 
renamed class references are used, this class will not be found by the 
classloader. 

We can optionally accept a shade prefix from user which will be tried (as 
fallback) when ClassNotFoundException is thrown.


> Accept shade prefix during reflective instantiation of output format
> 
>
> Key: HIVE-19494
> URL: https://issues.apache.org/jira/browse/HIVE-19494
> Project: Hive
>  Issue Type: Sub-task
>  Components: Streaming
>Affects Versions: 3.0.0, 3.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
>Priority: Major
>
> Hive Streaming API jars are sometimes shaded with a different prefix when 
> used in environments where another version of hive already exists (spark for 
> example). In most cases, shading is done with rename of classes with some 
> prefix. If an uber/assembly jar is generated with renamed prefix, Hive 
> Streaming API will not work as Hive Streaming API will reflectively 
> instantiate outputformat class using FQCN string provided by metastore table 
> storage descriptor object. 
> For example: 
> RecordWriter will create instance of OutputFormat using string 
> "org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat". When a shaded jar with 
> renamed class references are used, this class will not be found by the 
> classloader. 
> We can optionally accept a shade prefix from user via config which will be 
> tried (as fallback) when ClassNotFoundException is thrown.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19250) Schema column definitions inconsistencies in MySQL

2018-05-10 Thread Naveen Gangam (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19250?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Naveen Gangam updated HIVE-19250:
-
Status: Open  (was: Patch Available)

pre-commits did not run. Retrying


> Schema column definitions inconsistencies in MySQL
> --
>
> Key: HIVE-19250
> URL: https://issues.apache.org/jira/browse/HIVE-19250
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 2.1.0
>Reporter: Naveen Gangam
>Assignee: Naveen Gangam
>Priority: Minor
> Attachments: HIVE-19250.patch, HIVE-19250.patch
>
>
> There are some inconsistencies in column definitions in MySQL between a 
> schema that was upgraded to 2.1 (from an older release) vs installing the 
> 2.1.0 schema directly.
> >   `CQ_TBLPROPERTIES` varchar(2048) DEFAULT NULL,
> 117d117
> <   `CQ_TBLPROPERTIES` varchar(2048) DEFAULT NULL,
> 135a136
> >   `CC_TBLPROPERTIES` varchar(2048) DEFAULT NULL,
> 143d143
> <   `CC_TBLPROPERTIES` varchar(2048) DEFAULT NULL,
> 156c156
> <   `CTC_TXNID` bigint(20) DEFAULT NULL,
> ---
> >   `CTC_TXNID` bigint(20) NOT NULL,
> 158c158
> <   `CTC_TABLE` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT 
> NULL,
> ---
> >   `CTC_TABLE` varchar(256) DEFAULT NULL,
> 476c476
> <   `TBL_NAME` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT 
> NULL,
> ---
> >   `TBL_NAME` varchar(256) DEFAULT NULL,
> 664c664
> <   KEY `PCS_STATS_IDX` 
> (`DB_NAME`,`TABLE_NAME`,`COLUMN_NAME`,`PARTITION_NAME`),
> ---
> >   KEY `PCS_STATS_IDX` 
> > (`DB_NAME`,`TABLE_NAME`,`COLUMN_NAME`,`PARTITION_NAME`) USING BTREE,
> 768c768
> <   `PARAM_VALUE` mediumtext,
> ---
> >   `PARAM_VALUE` mediumtext CHARACTER SET latin1 COLLATE latin1_bin,
> 814c814
> <   `PARAM_VALUE` mediumtext,
> ---
> >   `PARAM_VALUE` mediumtext CHARACTER SET latin1 COLLATE latin1_bin,
> 934c934
> <   `PARAM_VALUE` mediumtext,
> ---
> >   `PARAM_VALUE` mediumtext CHARACTER SET latin1 COLLATE latin1_bin,
> 1066d1065
> <   `TXN_HEARTBEAT_COUNT` int(11) DEFAULT NULL,
> 1067a1067
> >   `TXN_HEARTBEAT_COUNT` int(11) DEFAULT NULL,
> 1080c1080
> <   `TC_TXNID` bigint(20) DEFAULT NULL,
> ---
> >   `TC_TXNID` bigint(20) NOT NULL,
> 1082c1082
> <   `TC_TABLE` varchar(128) DEFAULT NULL,
> ---
> >   `TC_TABLE` varchar(128) NOT NULL,
> 1084c1084
> <   `TC_OPERATION_TYPE` char(1) DEFAULT NULL,
> ---
> >   `TC_OPERATION_TYPE` char(1) NOT NULL,



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19250) Schema column definitions inconsistencies in MySQL

2018-05-10 Thread Naveen Gangam (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19250?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Naveen Gangam updated HIVE-19250:
-
Attachment: HIVE-19250.patch

> Schema column definitions inconsistencies in MySQL
> --
>
> Key: HIVE-19250
> URL: https://issues.apache.org/jira/browse/HIVE-19250
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 2.1.0
>Reporter: Naveen Gangam
>Assignee: Naveen Gangam
>Priority: Minor
> Attachments: HIVE-19250.patch, HIVE-19250.patch
>
>
> There are some inconsistencies in column definitions in MySQL between a 
> schema that was upgraded to 2.1 (from an older release) vs installing the 
> 2.1.0 schema directly.
> >   `CQ_TBLPROPERTIES` varchar(2048) DEFAULT NULL,
> 117d117
> <   `CQ_TBLPROPERTIES` varchar(2048) DEFAULT NULL,
> 135a136
> >   `CC_TBLPROPERTIES` varchar(2048) DEFAULT NULL,
> 143d143
> <   `CC_TBLPROPERTIES` varchar(2048) DEFAULT NULL,
> 156c156
> <   `CTC_TXNID` bigint(20) DEFAULT NULL,
> ---
> >   `CTC_TXNID` bigint(20) NOT NULL,
> 158c158
> <   `CTC_TABLE` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT 
> NULL,
> ---
> >   `CTC_TABLE` varchar(256) DEFAULT NULL,
> 476c476
> <   `TBL_NAME` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT 
> NULL,
> ---
> >   `TBL_NAME` varchar(256) DEFAULT NULL,
> 664c664
> <   KEY `PCS_STATS_IDX` 
> (`DB_NAME`,`TABLE_NAME`,`COLUMN_NAME`,`PARTITION_NAME`),
> ---
> >   KEY `PCS_STATS_IDX` 
> > (`DB_NAME`,`TABLE_NAME`,`COLUMN_NAME`,`PARTITION_NAME`) USING BTREE,
> 768c768
> <   `PARAM_VALUE` mediumtext,
> ---
> >   `PARAM_VALUE` mediumtext CHARACTER SET latin1 COLLATE latin1_bin,
> 814c814
> <   `PARAM_VALUE` mediumtext,
> ---
> >   `PARAM_VALUE` mediumtext CHARACTER SET latin1 COLLATE latin1_bin,
> 934c934
> <   `PARAM_VALUE` mediumtext,
> ---
> >   `PARAM_VALUE` mediumtext CHARACTER SET latin1 COLLATE latin1_bin,
> 1066d1065
> <   `TXN_HEARTBEAT_COUNT` int(11) DEFAULT NULL,
> 1067a1067
> >   `TXN_HEARTBEAT_COUNT` int(11) DEFAULT NULL,
> 1080c1080
> <   `TC_TXNID` bigint(20) DEFAULT NULL,
> ---
> >   `TC_TXNID` bigint(20) NOT NULL,
> 1082c1082
> <   `TC_TABLE` varchar(128) DEFAULT NULL,
> ---
> >   `TC_TABLE` varchar(128) NOT NULL,
> 1084c1084
> <   `TC_OPERATION_TYPE` char(1) DEFAULT NULL,
> ---
> >   `TC_OPERATION_TYPE` char(1) NOT NULL,



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19250) Schema column definitions inconsistencies in MySQL

2018-05-10 Thread Naveen Gangam (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19250?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Naveen Gangam updated HIVE-19250:
-
Status: Patch Available  (was: Open)

> Schema column definitions inconsistencies in MySQL
> --
>
> Key: HIVE-19250
> URL: https://issues.apache.org/jira/browse/HIVE-19250
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 2.1.0
>Reporter: Naveen Gangam
>Assignee: Naveen Gangam
>Priority: Minor
> Attachments: HIVE-19250.patch, HIVE-19250.patch
>
>
> There are some inconsistencies in column definitions in MySQL between a 
> schema that was upgraded to 2.1 (from an older release) vs installing the 
> 2.1.0 schema directly.
> >   `CQ_TBLPROPERTIES` varchar(2048) DEFAULT NULL,
> 117d117
> <   `CQ_TBLPROPERTIES` varchar(2048) DEFAULT NULL,
> 135a136
> >   `CC_TBLPROPERTIES` varchar(2048) DEFAULT NULL,
> 143d143
> <   `CC_TBLPROPERTIES` varchar(2048) DEFAULT NULL,
> 156c156
> <   `CTC_TXNID` bigint(20) DEFAULT NULL,
> ---
> >   `CTC_TXNID` bigint(20) NOT NULL,
> 158c158
> <   `CTC_TABLE` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT 
> NULL,
> ---
> >   `CTC_TABLE` varchar(256) DEFAULT NULL,
> 476c476
> <   `TBL_NAME` varchar(256) CHARACTER SET latin1 COLLATE latin1_bin DEFAULT 
> NULL,
> ---
> >   `TBL_NAME` varchar(256) DEFAULT NULL,
> 664c664
> <   KEY `PCS_STATS_IDX` 
> (`DB_NAME`,`TABLE_NAME`,`COLUMN_NAME`,`PARTITION_NAME`),
> ---
> >   KEY `PCS_STATS_IDX` 
> > (`DB_NAME`,`TABLE_NAME`,`COLUMN_NAME`,`PARTITION_NAME`) USING BTREE,
> 768c768
> <   `PARAM_VALUE` mediumtext,
> ---
> >   `PARAM_VALUE` mediumtext CHARACTER SET latin1 COLLATE latin1_bin,
> 814c814
> <   `PARAM_VALUE` mediumtext,
> ---
> >   `PARAM_VALUE` mediumtext CHARACTER SET latin1 COLLATE latin1_bin,
> 934c934
> <   `PARAM_VALUE` mediumtext,
> ---
> >   `PARAM_VALUE` mediumtext CHARACTER SET latin1 COLLATE latin1_bin,
> 1066d1065
> <   `TXN_HEARTBEAT_COUNT` int(11) DEFAULT NULL,
> 1067a1067
> >   `TXN_HEARTBEAT_COUNT` int(11) DEFAULT NULL,
> 1080c1080
> <   `TC_TXNID` bigint(20) DEFAULT NULL,
> ---
> >   `TC_TXNID` bigint(20) NOT NULL,
> 1082c1082
> <   `TC_TABLE` varchar(128) DEFAULT NULL,
> ---
> >   `TC_TABLE` varchar(128) NOT NULL,
> 1084c1084
> <   `TC_OPERATION_TYPE` char(1) DEFAULT NULL,
> ---
> >   `TC_OPERATION_TYPE` char(1) NOT NULL,



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19494) Accept shade prefix during reflective instantiation of output format

2018-05-10 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19494?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran updated HIVE-19494:
-
Description: 
Hive Streaming API jars are sometimes shaded with a different prefix when used 
in environments where another version of hive already exists (spark for 
example). In most cases, shading is done with rename of classes with some 
prefix. If an uber/assembly jar is generated with renamed prefix, Hive 
Streaming API will not work as Hive Streaming API will reflectively instantiate 
outputformat class using FQCN string provided by metastore table storage 
descriptor object. 
For example: 
RecordWriter will create instance of OutputFormat using string 
"org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat". When a shaded jar with 
renamed class references are used, this class will not be found by the 
classloader. 

We can optionally accept a shade prefix from user which will be tried (as 
fallback) when ClassNotFoundException is thrown.

  was:
Hive Streaming API jars are sometimes shaded with a different prefix when used 
in environments where another version of hive already exists (spark for 
example). In most cases, shading is done with rename of classes with some 
prefix. If an uber/assembly jar is generated with renamed prefix, Hive 
Streaming API will not work as Hive Streaming API will reflectively instantiate 
outputformat class using FQCN string provided by metastore table storage 
descriptor object. 
For example: 
RecordWriter will create instance of OutputFormat using string 
"org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat". When a shaded jar with 
renames class references are used, this class will not be found by the 
classloader. 

We can optionally accept a shade prefix from user which will be tried (as 
fallback) when ClassNotFoundException is thrown.


> Accept shade prefix during reflective instantiation of output format
> 
>
> Key: HIVE-19494
> URL: https://issues.apache.org/jira/browse/HIVE-19494
> Project: Hive
>  Issue Type: Sub-task
>  Components: Streaming
>Affects Versions: 3.0.0, 3.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
>Priority: Major
>
> Hive Streaming API jars are sometimes shaded with a different prefix when 
> used in environments where another version of hive already exists (spark for 
> example). In most cases, shading is done with rename of classes with some 
> prefix. If an uber/assembly jar is generated with renamed prefix, Hive 
> Streaming API will not work as Hive Streaming API will reflectively 
> instantiate outputformat class using FQCN string provided by metastore table 
> storage descriptor object. 
> For example: 
> RecordWriter will create instance of OutputFormat using string 
> "org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat". When a shaded jar with 
> renamed class references are used, this class will not be found by the 
> classloader. 
> We can optionally accept a shade prefix from user which will be tried (as 
> fallback) when ClassNotFoundException is thrown.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19464) Upgrade Parquet to 1.10.0

2018-05-10 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19464?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-19464:
---
Attachment: (was: HIVE-19464.patch)

> Upgrade Parquet to 1.10.0
> -
>
> Key: HIVE-19464
> URL: https://issues.apache.org/jira/browse/HIVE-19464
> Project: Hive
>  Issue Type: Improvement
>  Components: File Formats
>Affects Versions: 3.0.0
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
>Priority: Major
> Attachments: HIVE-19464.01.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19464) Upgrade Parquet to 1.10.0

2018-05-10 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19464?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-19464:
---
Attachment: HIVE-19464.01.patch

> Upgrade Parquet to 1.10.0
> -
>
> Key: HIVE-19464
> URL: https://issues.apache.org/jira/browse/HIVE-19464
> Project: Hive
>  Issue Type: Improvement
>  Components: File Formats
>Affects Versions: 3.0.0
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
>Priority: Major
> Attachments: HIVE-19464.01.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19465) Upgrade ORC to 1.5.0

2018-05-10 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19465?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-19465:
---
Attachment: HIVE-19465.01.patch

> Upgrade ORC to 1.5.0
> 
>
> Key: HIVE-19465
> URL: https://issues.apache.org/jira/browse/HIVE-19465
> Project: Hive
>  Issue Type: Improvement
>  Components: File Formats
>Affects Versions: 3.0.0
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
>Priority: Major
> Attachments: HIVE-19465.01.patch, HIVE-19465.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19465) Upgrade ORC to 1.5.0

2018-05-10 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19465?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-19465:
---
Attachment: (was: HIVE-19465.01.patch)

> Upgrade ORC to 1.5.0
> 
>
> Key: HIVE-19465
> URL: https://issues.apache.org/jira/browse/HIVE-19465
> Project: Hive
>  Issue Type: Improvement
>  Components: File Formats
>Affects Versions: 3.0.0
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
>Priority: Major
> Attachments: HIVE-19465.01.patch, HIVE-19465.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19465) Upgrade ORC to 1.5.0

2018-05-10 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19465?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-19465:
---
Attachment: HIVE-19465.01.patch

> Upgrade ORC to 1.5.0
> 
>
> Key: HIVE-19465
> URL: https://issues.apache.org/jira/browse/HIVE-19465
> Project: Hive
>  Issue Type: Improvement
>  Components: File Formats
>Affects Versions: 3.0.0
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
>Priority: Major
> Attachments: HIVE-19465.01.patch, HIVE-19465.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (HIVE-17463) ORC: include orc-shims in hive-exec.jar

2018-05-10 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17463?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez resolved HIVE-17463.

   Resolution: Fixed
Fix Version/s: 3.0.0

Thanks [~gopalv]!

> ORC: include orc-shims in hive-exec.jar
> ---
>
> Key: HIVE-17463
> URL: https://issues.apache.org/jira/browse/HIVE-17463
> Project: Hive
>  Issue Type: Bug
>  Components: ORC
>Affects Versions: 3.0.0
>Reporter: Gopal V
>Assignee: Gopal V
>Priority: Minor
> Fix For: 3.0.0
>
> Attachments: HIVE-17463.1.patch
>
>
> ORC-234 added a new shims module - this needs to be part of hive-exec shading 
> to use ORC-1.5.x branch in Hive.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19465) Upgrade ORC to 1.5.0

2018-05-10 Thread Gopal V (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19465?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471038#comment-16471038
 ] 

Gopal V commented on HIVE-19465:


My nightly runs on LLAP have finished without any issue after rebuilding 
against orc-1.5 branch.

LGTM - +1 tests pending

> Upgrade ORC to 1.5.0
> 
>
> Key: HIVE-19465
> URL: https://issues.apache.org/jira/browse/HIVE-19465
> Project: Hive
>  Issue Type: Improvement
>  Components: File Formats
>Affects Versions: 3.0.0
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
>Priority: Major
> Attachments: HIVE-19465.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-17463) ORC: include orc-shims in hive-exec.jar

2018-05-10 Thread Gopal V (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17463?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471036#comment-16471036
 ] 

Gopal V commented on HIVE-17463:


Yes, +1

> ORC: include orc-shims in hive-exec.jar
> ---
>
> Key: HIVE-17463
> URL: https://issues.apache.org/jira/browse/HIVE-17463
> Project: Hive
>  Issue Type: Bug
>  Components: ORC
>Affects Versions: 3.0.0
>Reporter: Gopal V
>Assignee: Gopal V
>Priority: Minor
> Attachments: HIVE-17463.1.patch
>
>
> ORC-234 added a new shims module - this needs to be part of hive-exec shading 
> to use ORC-1.5.x branch in Hive.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-17463) ORC: include orc-shims in hive-exec.jar

2018-05-10 Thread Jesus Camacho Rodriguez (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17463?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471033#comment-16471033
 ] 

Jesus Camacho Rodriguez commented on HIVE-17463:


[~gopalv], [~owen.omalley], I will incorporate this into HIVE-19465?

> ORC: include orc-shims in hive-exec.jar
> ---
>
> Key: HIVE-17463
> URL: https://issues.apache.org/jira/browse/HIVE-17463
> Project: Hive
>  Issue Type: Bug
>  Components: ORC
>Affects Versions: 3.0.0
>Reporter: Gopal V
>Assignee: Gopal V
>Priority: Minor
> Attachments: HIVE-17463.1.patch
>
>
> ORC-234 added a new shims module - this needs to be part of hive-exec shading 
> to use ORC-1.5.x branch in Hive.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19490) Locking on Insert into for non native and managed tables.

2018-05-10 Thread Eugene Koifman (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19490?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471031#comment-16471031
 ] 

Eugene Koifman commented on HIVE-19490:
---

TestDbTxnManager2 has many tests to verify locks

I think using shared lock for Druid insert makes sense but I'm not sure if that 
is appropriate for all non-native tables.

> Locking on Insert into for non native and managed tables.
> -
>
> Key: HIVE-19490
> URL: https://issues.apache.org/jira/browse/HIVE-19490
> Project: Hive
>  Issue Type: Improvement
>  Components: Druid integration
>Reporter: slim bouguerra
>Assignee: slim bouguerra
>Priority: Major
>  Labels: druid, locking
> Attachments: HIVE-19490.patch
>
>
> Current state of the art: 
> Managed non native table like Druid Tables, will need to get a Lock on Insert 
> into or insert Over write. The nature of this lock is set to Exclusive by 
> default for any non native table.
> This implies that Inserts into Druid table will Lock any read query as well 
> during the execution of the insert into. IMO this lock (on insert into) is  
> not needed since the insert statement is appending data and the state of 
> loading it is managed partially by Hive Storage handler hook and part of it 
> by Druid. 
> What i am proposing is to relax the lock level to shared for all non native 
> tables on insert into operations and keep it as Exclusive Write for insert 
> Overwrite for now.
>  
> Any feedback is welcome.
> cc [~ekoifman] / [~ashutoshc] / [~jdere] / [~hagleitn]
> Also am not sure what is the best way to unit test this currently am using 
> debugger to check of locks are what i except, please let me know if there is 
> a better way to do this. 
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-19494) Accept shade prefix during reflective instantiation of output format

2018-05-10 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19494?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran reassigned HIVE-19494:



> Accept shade prefix during reflective instantiation of output format
> 
>
> Key: HIVE-19494
> URL: https://issues.apache.org/jira/browse/HIVE-19494
> Project: Hive
>  Issue Type: Sub-task
>  Components: Streaming
>Affects Versions: 3.0.0, 3.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
>Priority: Major
>
> Hive Streaming API jars are sometimes shaded with a different prefix when 
> used in environments where another version of hive already exists (spark for 
> example). In most cases, shading is done with rename of classes with some 
> prefix. If an uber/assembly jar is generated with renamed prefix, Hive 
> Streaming API will not work as Hive Streaming API will reflectively 
> instantiate outputformat class using FQCN string provided by metastore table 
> storage descriptor object. 
> For example: 
> RecordWriter will create instance of OutputFormat using string 
> "org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat". When a shaded jar with 
> renames class references are used, this class will not be found by the 
> classloader. 
> We can optionally accept a shade prefix from user which will be tried (as 
> fallback) when ClassNotFoundException is thrown.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19476) Fix failures in TestReplicationScenariosAcidTables, TestReplicationOnHDFSEncryptedZones and TestCopyUtils

2018-05-10 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19476?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16471021#comment-16471021
 ] 

Hive QA commented on HIVE-19476:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
12s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  6m 
42s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
43s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
 2s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
36s{color} | {color:blue} itests/hive-unit in master has 2 extant Findbugs 
warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  4m  
8s{color} | {color:blue} ql in master has 2321 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
13s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m  
8s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  2m 
 2s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
39s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m 
39s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
52s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  4m 
52s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
16s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:red}-1{color} | {color:red} asflicense {color} | {color:red}  0m 
14s{color} | {color:red} The patch generated 1 ASF License warnings. {color} |
| {color:black}{color} | {color:black} {color} | {color:black} 27m 12s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-10807/dev-support/hive-personality.sh
 |
| git revision | master / 2d8e48c |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| asflicense | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-10807/yetus/patch-asflicense-problems.txt
 |
| modules | C: itests/hive-unit ql U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-10807/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Fix failures in TestReplicationScenariosAcidTables, 
> TestReplicationOnHDFSEncryptedZones and TestCopyUtils
> -
>
> Key: HIVE-19476
> URL: https://issues.apache.org/jira/browse/HIVE-19476
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2, repl
>Affects Versions: 3.0.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>Priority: Major
>  Labels: Tests, pull-request-available
> Fix For: 3.1.0
>
> Attachments: HIVE-19476.01.patch
>
>
> TestReplicationScenariosAcidTables broken due to HIVE-17657.
> cc [~sershe]
> TestReplicationOnHDFSEncryptedZones and TestCopyUtils are failing due to 
> auto-stats gathered which thrown exception irrelevant to theses tests.
>  
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-18862) qfiles: prepare .q files for using datasets

2018-05-10 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-18862?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16470981#comment-16470981
 ] 

Hive QA commented on HIVE-18862:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12922622/HIVE-18862.10-branch-3.patch

{color:green}SUCCESS:{color} +1 due to 2062 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 47 failed/errored test(s), 13373 tests 
executed
*Failed tests:*
{noformat}
TestBeeLineDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=253)
TestDbNotificationListener - did not produce a TEST-*.xml file (likely timed 
out) (batchId=247)
TestDummy - did not produce a TEST-*.xml file (likely timed out) (batchId=253)
TestHCatHiveCompatibility - did not produce a TEST-*.xml file (likely timed 
out) (batchId=247)
TestMiniDruidCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=253)
TestMiniDruidKafkaCliDriver - did not produce a TEST-*.xml file (likely timed 
out) (batchId=253)
TestNegativeCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=95)


[jira] [Updated] (HIVE-19258) add originals support to MM tables (and make the conversion a metadata only operation)

2018-05-10 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19258?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-19258:

Attachment: HIVE-19258.05.patch

> add originals support to MM tables (and make the conversion a metadata only 
> operation)
> --
>
> Key: HIVE-19258
> URL: https://issues.apache.org/jira/browse/HIVE-19258
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
>Priority: Major
> Attachments: HIVE-19258.01.patch, HIVE-19258.02.patch, 
> HIVE-19258.03.patch, HIVE-19258.04.patch, HIVE-19258.05.patch, 
> HIVE-19258.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19258) add originals support to MM tables (and make the conversion a metadata only operation)

2018-05-10 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19258?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16470970#comment-16470970
 ] 

Sergey Shelukhin commented on HIVE-19258:
-

Fixed two more issues in BHIF, NPE and the same final path being added twice 
for non MM case

> add originals support to MM tables (and make the conversion a metadata only 
> operation)
> --
>
> Key: HIVE-19258
> URL: https://issues.apache.org/jira/browse/HIVE-19258
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
>Priority: Major
> Attachments: HIVE-19258.01.patch, HIVE-19258.02.patch, 
> HIVE-19258.03.patch, HIVE-19258.04.patch, HIVE-19258.05.patch, 
> HIVE-19258.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19493) VectorUDFDateDiffColCol copySelected does not handle nulls correctly

2018-05-10 Thread Vihang Karajgaonkar (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19493?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16470964#comment-16470964
 ] 

Vihang Karajgaonkar commented on HIVE-19493:


cc: [~mmccline]

> VectorUDFDateDiffColCol copySelected does not handle nulls correctly
> 
>
> Key: HIVE-19493
> URL: https://issues.apache.org/jira/browse/HIVE-19493
> Project: Hive
>  Issue Type: Bug
>  Components: Vectorization
>Reporter: Vihang Karajgaonkar
>Assignee: Vihang Karajgaonkar
>Priority: Major
>
> The {{copySelected}} method in {{VectorUDFDateDiffColCol}} class was missed 
> during HIVE-18622



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-19493) VectorUDFDateDiffColCol copySelected does not handle nulls correctly

2018-05-10 Thread Vihang Karajgaonkar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19493?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vihang Karajgaonkar reassigned HIVE-19493:
--


> VectorUDFDateDiffColCol copySelected does not handle nulls correctly
> 
>
> Key: HIVE-19493
> URL: https://issues.apache.org/jira/browse/HIVE-19493
> Project: Hive
>  Issue Type: Bug
>  Components: Vectorization
>Reporter: Vihang Karajgaonkar
>Assignee: Vihang Karajgaonkar
>Priority: Major
>
> The {{copySelected}} method in {{VectorUDFDateDiffColCol}} class was missed 
> during HIVE-18622



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19472) HiveStreamingConnection swallows exception on partition creation

2018-05-10 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19472?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran updated HIVE-19472:
-
   Resolution: Fixed
Fix Version/s: 3.1.0
   3.0.0
   Status: Resolved  (was: Patch Available)

Test failures are unrelated. Committed to branch-3 and master. 

> HiveStreamingConnection swallows exception on partition creation
> 
>
> Key: HIVE-19472
> URL: https://issues.apache.org/jira/browse/HIVE-19472
> Project: Hive
>  Issue Type: Sub-task
>  Components: Streaming
>Affects Versions: 3.0.0, 3.1.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
>Priority: Major
> Fix For: 3.0.0, 3.1.0
>
> Attachments: HIVE-19472.1.patch
>
>
> HiveStreamingConnection swallows exception on partition creation



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19041) Thrift deserialization of Partition objects should intern fields

2018-05-10 Thread Misha Dmitriev (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19041?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16470954#comment-16470954
 ] 

Misha Dmitriev commented on HIVE-19041:
---

Thank you for looking into the details, [~vihangk1] I've checked the code and I 
agree with you - these strings are really short-lived. Furthemore, the worst 
offenders, 
org.apache.hadoop.hive.metastore.model.MStorageDescriptor.inputFormat,outputFormat,
 are actually copies of the corresponding fields of {{StorageDescriptor}}. That 
is, they reference the same string instances. So as soon as you intern the 
strings in {{StorageDescriptor}}, {{MStorageDescriptor}} will stop referencing 
duplicate strings as well.

 

Thus, this patch is good to go from my prospective.

 

> Thrift deserialization of Partition objects should intern fields
> 
>
> Key: HIVE-19041
> URL: https://issues.apache.org/jira/browse/HIVE-19041
> Project: Hive
>  Issue Type: Improvement
>  Components: Metastore
>Affects Versions: 3.0.0, 2.3.2
>Reporter: Vihang Karajgaonkar
>Assignee: Vihang Karajgaonkar
>Priority: Major
> Attachments: HIVE-19041.01.patch, HIVE-19041.02.patch, 
> HIVE-19041.03.patch, HIVE-19041.04.patch
>
>
> When a client is creating large number of partitions, the thrift objects are 
> deserialized into Partition objects. The read method of these objects does 
> not intern the inputformat, location, outputformat which cause large number 
> of duplicate Strings in the HMS memory. We should intern these objects while 
> deserialization to reduce memory pressure. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19178) TestMiniTezCliDriver.testCliDriver[explainanalyze_5] failure

2018-05-10 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19178?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-19178:
---
Resolution: Fixed
Status: Resolved  (was: Patch Available)

Pushed to master, branch-3. Thanks [~vgarg]

> TestMiniTezCliDriver.testCliDriver[explainanalyze_5] failure
> 
>
> Key: HIVE-19178
> URL: https://issues.apache.org/jira/browse/HIVE-19178
> Project: Hive
>  Issue Type: Sub-task
>  Components: Test
>Reporter: Vineet Garg
>Assignee: Jesus Camacho Rodriguez
>Priority: Blocker
> Fix For: 3.0.0
>
> Attachments: HIVE-19178.patch
>
>
> I have verified that this failure is due to HIVE-18825.
> Error stack:
> {code}
> java.lang.IllegalStateException: calling recordValidTxn() more than once in 
> the same txnid:5
>   at org.apache.hadoop.hive.ql.Driver.acquireLocks(Driver.java:1439)
>   at org.apache.hadoop.hive.ql.Driver.lockAndRespond(Driver.java:1624)
>   at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1794)
>   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1538)
>   at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1527)
>   at 
> org.apache.hadoop.hive.ql.parse.ExplainSemanticAnalyzer.analyzeInternal(ExplainSemanticAnalyzer.java:137)
>   at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:287)
>   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:635)
>   at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1655)
>   at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:1602)
>   at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:1597)
>   at 
> org.apache.hadoop.hive.ql.reexec.ReExecDriver.compileAndRespond(ReExecDriver.java:126)
>   at 
> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:200)
>   at 
> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239)
>   at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188)
>   at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402)
>   at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:335)
>   at 
> org.apache.hadoop.hive.ql.QTestUtil.executeClientInternal(QTestUtil.java:1455)
>   at 
> org.apache.hadoop.hive.ql.QTestUtil.executeClient(QTestUtil.java:1429)
>   at 
> org.apache.hadoop.hive.cli.control.CoreCliDriver.runTest(CoreCliDriver.java:177)
>   at 
> org.apache.hadoop.hive.cli.control.CliAdapter.runTest(CliAdapter.java:104)
>   at 
> org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver(TestMiniTezCliDriver.java:59)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
>   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
>   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>   at 
> org.apache.hadoop.hive.cli.control.CliAdapter$2$1.evaluate(CliAdapter.java:92)
>   at org.junit.rules.RunRules.evaluate(RunRules.java:20)
>   at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
>   at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
>   at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
>   at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
>   at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
>   at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
>   at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
>   at org.junit.runners.Suite.runChild(Suite.java:127)
>   at org.junit.runners.Suite.runChild(Suite.java:26)
>   at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
>   at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
>   at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
>   at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
>   at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
>   at 
> 

[jira] [Updated] (HIVE-19395) OutStream.write hangs with misconfigured bufferSize

2018-05-10 Thread John Doe (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19395?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

John Doe updated HIVE-19395:

Description: 
When the bufferSize is configured to be 0 in the class initialization, the 
while loop in OutStream.write function hangs endlessly.
This is because when the bufferSize is 0, current.remaining is 0, length will 
always > 0.
Here is the code snippet.
{code:java}
   OutStream(String name, int bufferSize, CompressionCodec codec, 
OutputReceiver receiver) throws IOException {
...
this.bufferSize = bufferSize; //bufferSize can be configured with 0
...
  }

private void getNewInputBuffer() throws IOException {
  ...
  current = ByteBuffer.allocate(bufferSize);
  ...
  }

  public void write(byte[] bytes, int offset, int length) throws IOException {
if (current == null) {
  getNewInputBuffer();
}
int remaining = Math.min(current.remaining(), length);
current.put(bytes, offset, remaining);
uncompressedBytes += remaining;
length -= remaining;
while (length != 0) {//length > 0
  spill();
  offset += remaining;
  remaining = Math.min(current.remaining(), length);//current.remaining() 
== 0
  current.put(bytes, offset, remaining);
  uncompressedBytes += remaining;
  length -= remaining;
}
  }
{code}
The similar case is HDFS-13513, HDFS-13514

  was:
When the bufferSize is configured to be 0 in the class initialization, the 
while loop in OutStream.write function hangs endlessly.
This is because when the bufferSize is 0, current.remaining is 0, length will 
always > 0.
Here is the code snippet.
{code:java}
   OutStream(String name, int bufferSize, CompressionCodec codec, 
OutputReceiver receiver) throws IOException {
...
this.bufferSize = bufferSize; //bufferSize can be configured with 0
...
  }

private void getNewInputBuffer() throws IOException {
  ...
  current = ByteBuffer.allocate(bufferSize);
  ...
  }

  public void write(byte[] bytes, int offset, int length) throws IOException {
if (current == null) {
  getNewInputBuffer();
}
int remaining = Math.min(current.remaining(), length);
current.put(bytes, offset, remaining);
uncompressedBytes += remaining;
length -= remaining;
while (length != 0) {//length > 0
  spill();
  offset += remaining;
  remaining = Math.min(current.remaining(), length);//current.remaining() 
== 0
  current.put(bytes, offset, remaining);
  uncompressedBytes += remaining;
  length -= remaining;
}
  }
{code}
The similar case is 
[HDFS-13513|https://issues.apache.org/jira/browse/HDFS-13513], 
[HDFS-13514|https://issues.apache.org/jira/browse/HDFS-13514]


> OutStream.write hangs with misconfigured bufferSize
> ---
>
> Key: HIVE-19395
> URL: https://issues.apache.org/jira/browse/HIVE-19395
> Project: Hive
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 1.0.0
>Reporter: John Doe
>Priority: Minor
>
> When the bufferSize is configured to be 0 in the class initialization, the 
> while loop in OutStream.write function hangs endlessly.
> This is because when the bufferSize is 0, current.remaining is 0, length will 
> always > 0.
> Here is the code snippet.
> {code:java}
>OutStream(String name, int bufferSize, CompressionCodec codec, 
> OutputReceiver receiver) throws IOException {
> ...
> this.bufferSize = bufferSize; //bufferSize can be configured with 0
> ...
>   }
> private void getNewInputBuffer() throws IOException {
>   ...
>   current = ByteBuffer.allocate(bufferSize);
>   ...
>   }
>   public void write(byte[] bytes, int offset, int length) throws IOException {
> if (current == null) {
>   getNewInputBuffer();
> }
> int remaining = Math.min(current.remaining(), length);
> current.put(bytes, offset, remaining);
> uncompressedBytes += remaining;
> length -= remaining;
> while (length != 0) {//length > 0
>   spill();
>   offset += remaining;
>   remaining = Math.min(current.remaining(), length);//current.remaining() 
> == 0
>   current.put(bytes, offset, remaining);
>   uncompressedBytes += remaining;
>   length -= remaining;
> }
>   }
> {code}
> The similar case is HDFS-13513, HDFS-13514



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-18652) Print Spark metrics on console

2018-05-10 Thread Vihang Karajgaonkar (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-18652?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16470943#comment-16470943
 ] 

Vihang Karajgaonkar commented on HIVE-18652:


[~stakiar] thanks for the patch. Overall LGTM. I left some minor suggestions. 
Rest looks good. +1

> Print Spark metrics on console
> --
>
> Key: HIVE-18652
> URL: https://issues.apache.org/jira/browse/HIVE-18652
> Project: Hive
>  Issue Type: Sub-task
>  Components: Spark
>Reporter: Sahil Takiar
>Assignee: Sahil Takiar
>Priority: Major
> Attachments: HIVE-18652.1.patch, HIVE-18652.2.patch
>
>
> For Hive-on-MR, each MR job launched prints out some stats about the job:
> {code}
> INFO  : 2018-02-07 17:51:11,218 Stage-1 map = 0%,  reduce = 0%
> INFO  : 2018-02-07 17:51:18,396 Stage-1 map = 100%,  reduce = 0%, Cumulative 
> CPU 1.87 sec
> INFO  : 2018-02-07 17:51:25,742 Stage-1 map = 100%,  reduce = 100%, 
> Cumulative CPU 4.34 sec
> INFO  : MapReduce Total cumulative CPU time: 4 seconds 340 msec
> INFO  : Ended Job = job_1517865654989_0004
> INFO  : MapReduce Jobs Launched:
> INFO  : Stage-Stage-1: Map: 1  Reduce: 1   Cumulative CPU: 4.34 sec   HDFS 
> Read: 7353 HDFS Write: 151 SUCCESS
> INFO  : Total MapReduce CPU Time Spent: 4 seconds 340 msec
> {code}
> We should do the same for Hive-on-Spark.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19474) Decimal type should be casted as part of the CTAS or INSERT Clause.

2018-05-10 Thread Jesus Camacho Rodriguez (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19474?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16470940#comment-16470940
 ] 

Jesus Camacho Rodriguez commented on HIVE-19474:


[~bslim], could we change the message in {{DruidStorageHandlerUtils}} to "Cast 
to any numeric type supported by Druid: x, y, z, t"? Thanks

> Decimal type should be casted as part of the CTAS or INSERT Clause.
> ---
>
> Key: HIVE-19474
> URL: https://issues.apache.org/jira/browse/HIVE-19474
> Project: Hive
>  Issue Type: Bug
>  Components: Druid integration
>Reporter: slim bouguerra
>Assignee: slim bouguerra
>Priority: Major
>  Labels: druid
> Fix For: 3.0.0
>
> Attachments: HIVE-19474.patch
>
>
> HIVE-18569  introduced a runtime config variable to allow the indexing of 
> Decimal as Double, this leads to kind of messy state, Hive metadata think the 
> column is still decimal while it is stored as double. Since the Hive metadata 
> of the column is Decimal the logical optimizer will not push down aggregates. 
> i tried to fix this by adding some logic to the application but it makes the 
> code very clumsy with lot of branches. Instead i propose to revert  
> HIVE-18569  and let the user introduce an explicit cast this will be better 
> since the metada reflects actual storage type and push down aggregates will 
> kick in and there is no config needed without adding any code or bug.
> cc [~ashutoshc] and [~nishantbangarwa]
> You can see the difference with the following DDL
> {code:java}
> create table test_base_table(`timecolumn` timestamp, `interval_marker` 
> string, `num_l` DECIMAL(10,2));
> insert into test_base_table values ('2015-03-08 00:00:00', 'i1-start', 4.5);
> set hive.druid.approx.result=true;
> CREATE TABLE druid_test_table
> STORED BY 'org.apache.hadoop.hive.druid.DruidStorageHandler'
> TBLPROPERTIES ("druid.segment.granularity" = "DAY")
> AS
> select cast(`timecolumn` as timestamp with local time zone) as `__time`, 
> `interval_marker`, cast(`num_l` as double)
> FROM test_base_table;
> describe druid_test_table;
> explain select sum(num_l), min(num_l) FROM druid_test_table;
> CREATE TABLE druid_test_table_2
> STORED BY 'org.apache.hadoop.hive.druid.DruidStorageHandler'
> TBLPROPERTIES ("druid.segment.granularity" = "DAY")
> AS
> select cast(`timecolumn` as timestamp with local time zone) as `__time`, 
> `interval_marker`, `num_l`
> FROM test_base_table;
> describe druid_test_table_2;
> explain select sum(num_l), min(num_l) FROM druid_test_table_2;
> {code}
>  
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-19492) Update Wiki with load data extension syntax

2018-05-10 Thread Deepak Jaiswal (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19492?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Deepak Jaiswal reassigned HIVE-19492:
-


> Update Wiki with load data extension syntax
> ---
>
> Key: HIVE-19492
> URL: https://issues.apache.org/jira/browse/HIVE-19492
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Deepak Jaiswal
>Assignee: Deepak Jaiswal
>Priority: Major
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-19491) Branch-3 Start using storage-api 2.6.1 once available.

2018-05-10 Thread Deepak Jaiswal (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19491?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Deepak Jaiswal reassigned HIVE-19491:
-


> Branch-3 Start using storage-api 2.6.1 once available.
> --
>
> Key: HIVE-19491
> URL: https://issues.apache.org/jira/browse/HIVE-19491
> Project: Hive
>  Issue Type: Task
>Reporter: Deepak Jaiswal
>Assignee: Deepak Jaiswal
>Priority: Blocker
>
> branch-3 needs storage-api 2.6.1 which is in the process of being released.
>  
> cc. [~vgarg]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19471) bucket_map_join_tez1 and bucket_map_join_tez2 are failing

2018-05-10 Thread Deepak Jaiswal (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19471?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Deepak Jaiswal updated HIVE-19471:
--
Resolution: Fixed
Status: Resolved  (was: Patch Available)

Committed to master and branch-3

thanks for the reviews.

> bucket_map_join_tez1 and  bucket_map_join_tez2 are failing
> --
>
> Key: HIVE-19471
> URL: https://issues.apache.org/jira/browse/HIVE-19471
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Vineet Garg
>Assignee: Deepak Jaiswal
>Priority: Blocker
> Fix For: 3.0.0
>
> Attachments: HIVE-19471.01-branch-3.patch, HIVE-19471.1.patch
>
>
> https://builds.apache.org/job/PreCommit-HIVE-Build/10766/testReport/
> TestMiniLlapLocalCliDriver.testCliDriver[bucket_map_join_tez1]
> TestMiniLlapLocalCliDriver.testCliDriver[bucket_map_join_tez2]
> Both are failing. Probably need golden file update.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19454) Test failure : org.apache.hadoop.hive.ql.TestTxnCommands2.testNonAcidToAcidConversion1 fails with java.lang.AssertionError

2018-05-10 Thread Deepak Jaiswal (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19454?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16470908#comment-16470908
 ] 

Deepak Jaiswal commented on HIVE-19454:
---

If an order is not guaranteed then maybe we should always sort them before 
asserting?

> Test failure : 
> org.apache.hadoop.hive.ql.TestTxnCommands2.testNonAcidToAcidConversion1 fails 
> with java.lang.AssertionError
> --
>
> Key: HIVE-19454
> URL: https://issues.apache.org/jira/browse/HIVE-19454
> Project: Hive
>  Issue Type: Bug
> Environment: uname -a
> Linux pts00433-vm40 3.10.0-862.el7.ppc64le #1 SMP Wed Mar 21 22:20:19 UTC 
> 2018 ppc64le ppc64le ppc64le GNU/Linux
>Reporter: Alisha Prabhu
>Priority: Major
> Attachments: HIVE-19454.1.patch
>
>
> org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testNonAcidToAcidConversion1
>  refers to the same test case.
> Maven command used is : mvn -Dtest=TestTxnCommands2 test
> Error:
> {code:java}
> [INFO] Running org.apache.hadoop.hive.ql.TestTxnCommands2
> [ERROR] Tests run: 44, Failures: 1, Errors: 0, Skipped: 3, Time elapsed: 
> 618.215 s <<< FAILURE! - in org.apache.hadoop.hive.ql.TestTxnCommands2
> [ERROR] 
> testNonAcidToAcidConversion1(org.apache.hadoop.hive.ql.TestTxnCommands2)  
> Time elapsed: 17.557 s  <<< FAILURE!
> java.lang.AssertionError
>     at org.junit.Assert.fail(Assert.java:86)
>     at org.junit.Assert.assertTrue(Assert.java:41)
>     at org.junit.Assert.assertTrue(Assert.java:52)
>     at 
> org.apache.hadoop.hive.ql.TestTxnCommands2.testNonAcidToAcidConversion1(TestTxnCommands2.java:499)
>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>     at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>     at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>     at java.lang.reflect.Method.invoke(Method.java:498)
>     at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
>     at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>     at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
>     at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>     at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
>     at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
>     at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
>     at 
> org.junit.rules.ExpectedException$ExpectedExceptionStatement.evaluate(ExpectedException.java:168)
>     at org.junit.rules.RunRules.evaluate(RunRules.java:20)
>     at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
>     at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
>     at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
>     at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
>     at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
>     at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
>     at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
>     at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
>     at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
>     at 
> org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
>     at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273)
>     at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
>     at 
> org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
>     at 
> org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:379)
>     at 
> org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:340)
>     at 
> org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:125)
>     at 
> org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:413)
> [INFO]
> [INFO] Results:
> [INFO]
> [ERROR] Failures:
> [ERROR]   TestTxnCommands2.testNonAcidToAcidConversion1:499
> [INFO]
> [ERROR] Tests run: 44, Failures: 1, Errors: 0, Skipped: 3
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19108) Vectorization and Parquet: Turning on vectorization in parquet_ppd_decimal.q causes Wrong Query Results

2018-05-10 Thread Vihang Karajgaonkar (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-19108?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16470898#comment-16470898
 ] 

Vihang Karajgaonkar commented on HIVE-19108:


attached branch-3 patch. Had to reassign the Jira to myself inorder to attach 
the file. Reassigned it back to [~jerrychenhf]

> Vectorization and Parquet: Turning on vectorization in parquet_ppd_decimal.q 
> causes Wrong Query Results
> ---
>
> Key: HIVE-19108
> URL: https://issues.apache.org/jira/browse/HIVE-19108
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 3.0.0
>Reporter: Matt McCline
>Assignee: Haifeng Chen
>Priority: Critical
> Attachments: HIVE-19108.01.patch, HIVE-19108.02.patch, 
> HIVE-19108.03.patch, HIVE-19108.04.patch, HIVE-19108.05-branch-3.patch
>
>
> Found in vectorization enable by default experiment.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19108) Vectorization and Parquet: Turning on vectorization in parquet_ppd_decimal.q causes Wrong Query Results

2018-05-10 Thread Vihang Karajgaonkar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19108?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vihang Karajgaonkar updated HIVE-19108:
---
Attachment: HIVE-19108.05-branch-3.patch

> Vectorization and Parquet: Turning on vectorization in parquet_ppd_decimal.q 
> causes Wrong Query Results
> ---
>
> Key: HIVE-19108
> URL: https://issues.apache.org/jira/browse/HIVE-19108
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 3.0.0
>Reporter: Matt McCline
>Assignee: Vihang Karajgaonkar
>Priority: Critical
> Attachments: HIVE-19108.01.patch, HIVE-19108.02.patch, 
> HIVE-19108.03.patch, HIVE-19108.04.patch, HIVE-19108.05-branch-3.patch
>
>
> Found in vectorization enable by default experiment.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-19108) Vectorization and Parquet: Turning on vectorization in parquet_ppd_decimal.q causes Wrong Query Results

2018-05-10 Thread Vihang Karajgaonkar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19108?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vihang Karajgaonkar reassigned HIVE-19108:
--

Assignee: Haifeng Chen  (was: Vihang Karajgaonkar)

> Vectorization and Parquet: Turning on vectorization in parquet_ppd_decimal.q 
> causes Wrong Query Results
> ---
>
> Key: HIVE-19108
> URL: https://issues.apache.org/jira/browse/HIVE-19108
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 3.0.0
>Reporter: Matt McCline
>Assignee: Haifeng Chen
>Priority: Critical
> Attachments: HIVE-19108.01.patch, HIVE-19108.02.patch, 
> HIVE-19108.03.patch, HIVE-19108.04.patch, HIVE-19108.05-branch-3.patch
>
>
> Found in vectorization enable by default experiment.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-19108) Vectorization and Parquet: Turning on vectorization in parquet_ppd_decimal.q causes Wrong Query Results

2018-05-10 Thread Vihang Karajgaonkar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-19108?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vihang Karajgaonkar reassigned HIVE-19108:
--

Assignee: Vihang Karajgaonkar  (was: Haifeng Chen)

> Vectorization and Parquet: Turning on vectorization in parquet_ppd_decimal.q 
> causes Wrong Query Results
> ---
>
> Key: HIVE-19108
> URL: https://issues.apache.org/jira/browse/HIVE-19108
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 3.0.0
>Reporter: Matt McCline
>Assignee: Vihang Karajgaonkar
>Priority: Critical
> Attachments: HIVE-19108.01.patch, HIVE-19108.02.patch, 
> HIVE-19108.03.patch, HIVE-19108.04.patch, HIVE-19108.05-branch-3.patch
>
>
> Found in vectorization enable by default experiment.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-18862) qfiles: prepare .q files for using datasets

2018-05-10 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-18862?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16470853#comment-16470853
 ] 

Hive QA commented on HIVE-18862:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:red}-1{color} | {color:red} patch {color} | {color:red}  0m  9s{color} 
| {color:red} 
/data/hiveptest/logs/PreCommit-HIVE-Build-10806/patches/PreCommit-HIVE-Build-10806.patch
 does not apply to master. Rebase required? Wrong Branch? See 
http://cwiki.apache.org/confluence/display/Hive/HowToContribute for help. 
{color} |
\\
\\
|| Subsystem || Report/Notes ||
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-10806/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> qfiles: prepare .q files for using datasets
> ---
>
> Key: HIVE-18862
> URL: https://issues.apache.org/jira/browse/HIVE-18862
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Laszlo Bodor
>Assignee: Laszlo Bodor
>Priority: Major
> Fix For: 3.1.0
>
> Attachments: HIVE-18862.01.patch, HIVE-18862.02.patch, 
> HIVE-18862.03.patch, HIVE-18862.04.patch, HIVE-18862.05.patch, 
> HIVE-18862.06.patch, HIVE-18862.07.patch, HIVE-18862.08.patch, 
> HIVE-18862.09-branch-3.patch, HIVE-18862.09.patch, 
> HIVE-18862.10-branch-3.patch
>
>
> # Parse .q files for source table usage
>  # Add needed dataset annotations
>  # Remove create table statements from "q_test_init.sql" like files
>  # Handle oncoming issues related to dataset introduction



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


  1   2   >