[jira] [Updated] (HIVE-22092) Fetch is failing with IllegalArgumentException: No ValidTxnList when refetch is done

2019-08-08 Thread mahesh kumar behera (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-22092?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

mahesh kumar behera updated HIVE-22092:
---
Attachment: HIVE-22092.01.patch

> Fetch is failing with IllegalArgumentException: No ValidTxnList when refetch 
> is done
> 
>
> Key: HIVE-22092
> URL: https://issues.apache.org/jira/browse/HIVE-22092
> Project: Hive
>  Issue Type: Bug
>Reporter: mahesh kumar behera
>Assignee: mahesh kumar behera
>Priority: Major
> Attachments: HIVE-22092.01.patch
>
>
> The fetch task is created during query compilation with the config of the 
> driver. That config will have the valid txn list set. Thus the fetch task 
> will have valid txn list set while doing fetch for ACID tables. But when user 
> does a refetch with cusrsor set to first position it reinitializes the fetch 
> task with the driver config (cached in task config). But by that time, the 
> select query would have cleaned up the valid txn list from the config and the 
> fetch will happen with valid txn list as null. This will cause illegal 
> argument exception.
>  
> Caused by: java.lang.IllegalArgumentException: No ValidTxnList fo
>  at org.apache.hadoop.hive.ql.io.AcidUtils.isDirUsable(AcidUtils.java:1418) 
> ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at org.apache.hadoop.hive.ql.io.AcidUtils.getChildState(AcidUtils.java:1393) 
> ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at org.apache.hadoop.hive.ql.io.AcidUtils.getAcidState(AcidUtils.java:1205) 
> ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hadoop.hive.ql.io.HiveInputFormat.processForWriteIds(HiveInputFormat.java:633)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hadoop.hive.ql.io.HiveInputFormat.processPathsForMmRead(HiveInputFormat.java:597)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hadoop.hive.ql.exec.FetchOperator.processCurrPathForMmWriteIds(FetchOperator.java:469)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hadoop.hive.ql.exec.FetchOperator.getNextSplits(FetchOperator.java:396)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hadoop.hive.ql.exec.FetchOperator.getRecordReader(FetchOperator.java:322)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hadoop.hive.ql.exec.FetchOperator.getNextRow(FetchOperator.java:563)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hadoop.hive.ql.exec.FetchOperator.pushRow(FetchOperator.java:532) 
> ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at org.apache.hadoop.hive.ql.exec.FetchTask.fetch(FetchTask.java:146) 
> ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at org.apache.hadoop.hive.ql.Driver.getResults(Driver.java:2362) 
> ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hadoop.hive.ql.reexec.ReExecDriver.getResults(ReExecDriver.java:229)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hive.service.cli.operation.SQLOperation.getNextRowSet(SQLOperation.java:473)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hive.service.cli.operation.OperationManager.getOperationNextRowSet(OperationManager.java:328)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hive.service.cli.session.HiveSessionImpl.fetchResults(HiveSessionImpl.java:952)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at org.apache.hive.service.cli.CLIService.fetchResults(CLIService.java:565) 
> ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.FetchResults(ThriftCLIService.java:792)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$FetchResults.getResult(TCLIService.java:1837)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$FetchResults.getResult(TCLIService.java:1822)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (HIVE-22092) Fetch is failing with IllegalArgumentException: No ValidTxnList when refetch is done

2019-08-08 Thread mahesh kumar behera (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-22092?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

mahesh kumar behera updated HIVE-22092:
---
Status: Patch Available  (was: Open)

> Fetch is failing with IllegalArgumentException: No ValidTxnList when refetch 
> is done
> 
>
> Key: HIVE-22092
> URL: https://issues.apache.org/jira/browse/HIVE-22092
> Project: Hive
>  Issue Type: Bug
>Reporter: mahesh kumar behera
>Assignee: mahesh kumar behera
>Priority: Major
> Attachments: HIVE-22092.01.patch
>
>
> The fetch task is created during query compilation with the config of the 
> driver. That config will have the valid txn list set. Thus the fetch task 
> will have valid txn list set while doing fetch for ACID tables. But when user 
> does a refetch with cusrsor set to first position it reinitializes the fetch 
> task with the driver config (cached in task config). But by that time, the 
> select query would have cleaned up the valid txn list from the config and the 
> fetch will happen with valid txn list as null. This will cause illegal 
> argument exception.
>  
> Caused by: java.lang.IllegalArgumentException: No ValidTxnList fo
>  at org.apache.hadoop.hive.ql.io.AcidUtils.isDirUsable(AcidUtils.java:1418) 
> ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at org.apache.hadoop.hive.ql.io.AcidUtils.getChildState(AcidUtils.java:1393) 
> ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at org.apache.hadoop.hive.ql.io.AcidUtils.getAcidState(AcidUtils.java:1205) 
> ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hadoop.hive.ql.io.HiveInputFormat.processForWriteIds(HiveInputFormat.java:633)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hadoop.hive.ql.io.HiveInputFormat.processPathsForMmRead(HiveInputFormat.java:597)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hadoop.hive.ql.exec.FetchOperator.processCurrPathForMmWriteIds(FetchOperator.java:469)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hadoop.hive.ql.exec.FetchOperator.getNextSplits(FetchOperator.java:396)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hadoop.hive.ql.exec.FetchOperator.getRecordReader(FetchOperator.java:322)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hadoop.hive.ql.exec.FetchOperator.getNextRow(FetchOperator.java:563)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hadoop.hive.ql.exec.FetchOperator.pushRow(FetchOperator.java:532) 
> ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at org.apache.hadoop.hive.ql.exec.FetchTask.fetch(FetchTask.java:146) 
> ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at org.apache.hadoop.hive.ql.Driver.getResults(Driver.java:2362) 
> ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hadoop.hive.ql.reexec.ReExecDriver.getResults(ReExecDriver.java:229)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hive.service.cli.operation.SQLOperation.getNextRowSet(SQLOperation.java:473)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hive.service.cli.operation.OperationManager.getOperationNextRowSet(OperationManager.java:328)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hive.service.cli.session.HiveSessionImpl.fetchResults(HiveSessionImpl.java:952)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at org.apache.hive.service.cli.CLIService.fetchResults(CLIService.java:565) 
> ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.FetchResults(ThriftCLIService.java:792)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$FetchResults.getResult(TCLIService.java:1837)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
>  at 
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$FetchResults.getResult(TCLIService.java:1822)
>  ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (HIVE-22092) Fetch is failing with IllegalArgumentException: No ValidTxnList when refetch is done

2019-08-08 Thread mahesh kumar behera (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-22092?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

mahesh kumar behera updated HIVE-22092:
---
Description: 
The fetch task is created during query compilation with the config of the 
driver. That config will have the valid txn list set. Thus the fetch task will 
have valid txn list set while doing fetch for ACID tables. But when user does a 
refetch with cusrsor set to first position it reinitializes the fetch task with 
the driver config (cached in task config). But by that time, the select query 
would have cleaned up the valid txn list from the config and the fetch will 
happen with valid txn list as null. This will cause illegal argument exception.

 
Caused by: java.lang.IllegalArgumentException: No ValidTxnList fo
 at org.apache.hadoop.hive.ql.io.AcidUtils.isDirUsable(AcidUtils.java:1418) 
~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
 at org.apache.hadoop.hive.ql.io.AcidUtils.getChildState(AcidUtils.java:1393) 
~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
 at org.apache.hadoop.hive.ql.io.AcidUtils.getAcidState(AcidUtils.java:1205) 
~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
 at 
org.apache.hadoop.hive.ql.io.HiveInputFormat.processForWriteIds(HiveInputFormat.java:633)
 ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
 at 
org.apache.hadoop.hive.ql.io.HiveInputFormat.processPathsForMmRead(HiveInputFormat.java:597)
 ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
 at 
org.apache.hadoop.hive.ql.exec.FetchOperator.processCurrPathForMmWriteIds(FetchOperator.java:469)
 ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
 at 
org.apache.hadoop.hive.ql.exec.FetchOperator.getNextSplits(FetchOperator.java:396)
 ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
 at 
org.apache.hadoop.hive.ql.exec.FetchOperator.getRecordReader(FetchOperator.java:322)
 ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
 at 
org.apache.hadoop.hive.ql.exec.FetchOperator.getNextRow(FetchOperator.java:563) 
~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
 at 
org.apache.hadoop.hive.ql.exec.FetchOperator.pushRow(FetchOperator.java:532) 
~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
 at org.apache.hadoop.hive.ql.exec.FetchTask.fetch(FetchTask.java:146) 
~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
 at org.apache.hadoop.hive.ql.Driver.getResults(Driver.java:2362) 
~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
 at 
org.apache.hadoop.hive.ql.reexec.ReExecDriver.getResults(ReExecDriver.java:229) 
~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
 at 
org.apache.hive.service.cli.operation.SQLOperation.getNextRowSet(SQLOperation.java:473)
 ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
 at 
org.apache.hive.service.cli.operation.OperationManager.getOperationNextRowSet(OperationManager.java:328)
 ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
 at 
org.apache.hive.service.cli.session.HiveSessionImpl.fetchResults(HiveSessionImpl.java:952)
 ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
 at org.apache.hive.service.cli.CLIService.fetchResults(CLIService.java:565) 
~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
 at 
org.apache.hive.service.cli.thrift.ThriftCLIService.FetchResults(ThriftCLIService.java:792)
 ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
 at 
org.apache.hive.service.rpc.thrift.TCLIService$Processor$FetchResults.getResult(TCLIService.java:1837)
 ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]
 at 
org.apache.hive.service.rpc.thrift.TCLIService$Processor$FetchResults.getResult(TCLIService.java:1822)
 ~[data_analytics_studio-webapp-1.3.1.1.3.1.0-12.jar:?]

  was:The fetch task is created during query compilation with the config of the 
driver. That config will have the valid txn list set. Thus the fetch task will 
have valid txn list set while doing fetch for ACID tables. But when user does a 
refetch with cusrsor set to first position it reinitializes the fetch task with 
the driver config (cached in task config). But by that time, the select query 
would have cleaned up the valid txn list from the config and the fetch will 
happen with valid txn list as null. This will cause illegal argument exception.


> Fetch is failing with IllegalArgumentException: No ValidTxnList when refetch 
> is done
> 
>
> Key: HIVE-22092
> URL: https://issues.apache.org/jira/browse/HIVE-22092
> Project: Hive
>  Issue Type: Bug
>Reporter: mahesh kumar behera
>Assignee: mahesh kumar behera
>Priority: Major
>
> The fetch task is created during query compilation with the config of the 
> driver. That config will have the valid txn list set. Thus the fetch task 
> will have valid txn list set while doing fetch for ACID tables. But 

[jira] [Updated] (HIVE-22092) Fetch is failing with IllegalArgumentException: No ValidTxnList when refetch is done

2019-08-08 Thread mahesh kumar behera (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-22092?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

mahesh kumar behera updated HIVE-22092:
---
Summary: Fetch is failing with IllegalArgumentException: No ValidTxnList 
when refetch is done  (was: Fetch failing with IllegalArgumentException: No 
ValidTxnList when refetch is done)

> Fetch is failing with IllegalArgumentException: No ValidTxnList when refetch 
> is done
> 
>
> Key: HIVE-22092
> URL: https://issues.apache.org/jira/browse/HIVE-22092
> Project: Hive
>  Issue Type: Bug
>Reporter: mahesh kumar behera
>Assignee: mahesh kumar behera
>Priority: Major
>
> The fetch task is created during query compilation with the config of the 
> driver. That config will have the valid txn list set. Thus the fetch task 
> will have valid txn list set while doing fetch for ACID tables. But when user 
> does a refetch with cusrsor set to first position it reinitializes the fetch 
> task with the driver config (cached in task config). But by that time, the 
> select query would have cleaned up the valid txn list from the config and the 
> fetch will happen with valid txn list as null. This will cause illegal 
> argument exception.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Assigned] (HIVE-22092) Fetch failing with IllegalArgumentException: No ValidTxnList when refetch is done

2019-08-08 Thread mahesh kumar behera (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-22092?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

mahesh kumar behera reassigned HIVE-22092:
--


> Fetch failing with IllegalArgumentException: No ValidTxnList when refetch is 
> done
> -
>
> Key: HIVE-22092
> URL: https://issues.apache.org/jira/browse/HIVE-22092
> Project: Hive
>  Issue Type: Bug
>Reporter: mahesh kumar behera
>Assignee: mahesh kumar behera
>Priority: Major
>
> The fetch task is created during query compilation with the config of the 
> driver. That config will have the valid txn list set. Thus the fetch task 
> will have valid txn list set while doing fetch for ACID tables. But when user 
> does a refetch with cusrsor set to first position it reinitializes the fetch 
> task with the driver config (cached in task config). But by that time, the 
> select query would have cleaned up the valid txn list from the config and the 
> fetch will happen with valid txn list as null. This will cause illegal 
> argument exception.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (HIVE-21393) Refactor exception handling in HiveAccumuloTableInputFormat

2019-08-08 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-21393?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903573#comment-16903573
 ] 

Hive QA commented on HIVE-21393:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12977091/HIVE-21393.1-master.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 16723 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/18291/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/18291/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-18291/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12977091 - PreCommit-HIVE-Build

> Refactor exception handling in HiveAccumuloTableInputFormat
> ---
>
> Key: HIVE-21393
> URL: https://issues.apache.org/jira/browse/HIVE-21393
> Project: Hive
>  Issue Type: Bug
>Reporter: Laszlo Bodor
>Assignee: Sakthi
>Priority: Major
>  Labels: newbie
> Attachments: HIVE-21393.1-master.patch
>
>
> https://github.com/apache/hive/blob/173d9816086e4c43677ae6dc3b1bf85203f894a1/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/mr/HiveAccumuloTableInputFormat.java#L178-L186
> From Java7, these catches could be merged.
> {code:java}
> } catch (AccumuloException e) {
>   log.error("Could not configure AccumuloInputFormat", e);
>   throw new IOException(StringUtils.stringifyException(e));
> } catch (AccumuloSecurityException e) {
>   log.error("Could not configure AccumuloInputFormat", e);
>   throw new IOException(StringUtils.stringifyException(e));
> } catch (SerDeException e) {
>   log.error("Could not configure AccumuloInputFormat", e);
>   throw new IOException(StringUtils.stringifyException(e));
> }
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (HIVE-16587) NPE when inserting complex types with nested null values

2019-08-08 Thread Sankar Hariappan (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-16587?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sankar Hariappan updated HIVE-16587:

   Resolution: Fixed
Fix Version/s: 4.0.0
   Status: Resolved  (was: Patch Available)

Committed to master.
Thanks [~nareshpr] for the contribution!

> NPE when inserting complex types with nested null values
> 
>
> Key: HIVE-16587
> URL: https://issues.apache.org/jira/browse/HIVE-16587
> Project: Hive
>  Issue Type: Bug
>  Components: Query Planning
>Affects Versions: 1.2.1
>Reporter: Jason Dere
>Assignee: Naresh P R
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-16587.1.patch, HIVE-16587.patch
>
>
> {noformat}
> CREATE TABLE complex1 (c0 int, c1 array, c2 map, c3 
> struct>, c4 array f3:array>>)
> insert into complex1
>  select 3, array(1, 2, null), map(1, 'one', 2, null), named_struct('f1', 
> cast(null as int), 'f2', cast(null as string), 'f3', array(1,2,null)), 
> array(named_struct('f1', 11, 'f2', 'two', 'f3', array(2,3,4)))
> {noformat}
> Gives the following error:
> {noformat}
> Caused by: org.apache.hive.service.cli.HiveSQLException: Error while 
> compiling statement: FAILED: NullPointerException null
>   at 
> org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:315)
>   at 
> org.apache.hive.service.cli.operation.SQLOperation.prepare(SQLOperation.java:207)
>   at 
> org.apache.hive.service.cli.operation.SQLOperation.runInternal(SQLOperation.java:291)
>   at 
> org.apache.hive.service.cli.operation.Operation.run(Operation.java:255)
>   at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:531)
>   at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync(HiveSessionImpl.java:517)
>   at sun.reflect.GeneratedMethodAccessor49.invoke(Unknown Source)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:78)
>   at 
> org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:36)
>   at 
> org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:63)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:422)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1807)
>   at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:59)
>   at com.sun.proxy.$Proxy126.executeStatementAsync(Unknown Source)
>   at 
> org.apache.hive.service.cli.CLIService.executeStatementAsync(CLIService.java:310)
>   at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:530)
>   at 
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1437)
>   at 
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1422)
>   at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>   at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>   at 
> org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:56)
>   at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:286)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>   at java.lang.Thread.run(Thread.java:745)
> Caused by: java.lang.NullPointerException: null
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getWritableSize(StatsUtils.java:1144)
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getSizeOfMap(StatsUtils.java:1106)
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getSizeOfComplexTypes(StatsUtils.java:978)
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getAvgColLenOf(StatsUtils.java:916)
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getColStatisticsFromExpression(StatsUtils.java:1371)
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getColStatisticsFromExprMap(StatsUtils.java:1194)
>   at 
> org.apache.hadoop.hive.ql.optimizer.stats.annotation.StatsRulesProcFactory$SelectStatsRule.process(StatsRulesProcFactory.java:187)
>   at 
> org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:90)
>   at 
> 

[jira] [Commented] (HIVE-21393) Refactor exception handling in HiveAccumuloTableInputFormat

2019-08-08 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-21393?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903546#comment-16903546
 ] 

Hive QA commented on HIVE-21393:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
32s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
20s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
12s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
31s{color} | {color:blue} accumulo-handler in master has 21 extant Findbugs 
warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
13s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
19s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
21s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
21s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
12s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
36s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
13s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 12m  6s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-18291/dev-support/hive-personality.sh
 |
| git revision | master / 7fc5a88 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| modules | C: accumulo-handler U: accumulo-handler |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18291/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Refactor exception handling in HiveAccumuloTableInputFormat
> ---
>
> Key: HIVE-21393
> URL: https://issues.apache.org/jira/browse/HIVE-21393
> Project: Hive
>  Issue Type: Bug
>Reporter: Laszlo Bodor
>Assignee: Sakthi
>Priority: Major
>  Labels: newbie
> Attachments: HIVE-21393.1-master.patch
>
>
> https://github.com/apache/hive/blob/173d9816086e4c43677ae6dc3b1bf85203f894a1/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/mr/HiveAccumuloTableInputFormat.java#L178-L186
> From Java7, these catches could be merged.
> {code:java}
> } catch (AccumuloException e) {
>   log.error("Could not configure AccumuloInputFormat", e);
>   throw new IOException(StringUtils.stringifyException(e));
> } catch (AccumuloSecurityException e) {
>   log.error("Could not configure AccumuloInputFormat", e);
>   throw new IOException(StringUtils.stringifyException(e));
> } catch (SerDeException e) {
>   log.error("Could not configure AccumuloInputFormat", e);
>   throw new IOException(StringUtils.stringifyException(e));
> }
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (HIVE-21549) Fix duplicated asf headers

2019-08-08 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-21549?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903538#comment-16903538
 ] 

Hive QA commented on HIVE-21549:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12977088/HIVE-21549.1-master.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 6 failed/errored test(s), 16723 tests 
executed
*Failed tests:*
{noformat}
org.apache.hive.jdbc.TestActivePassiveHA.testActivePassiveHA (batchId=283)
org.apache.hive.jdbc.TestActivePassiveHA.testClientConnectionsOnFailover 
(batchId=283)
org.apache.hive.jdbc.TestActivePassiveHA.testConnectionActivePassiveHAServiceDiscovery
 (batchId=283)
org.apache.hive.jdbc.TestActivePassiveHA.testManualFailover (batchId=283)
org.apache.hive.jdbc.TestActivePassiveHA.testManualFailoverUnauthorized 
(batchId=283)
org.apache.hive.jdbc.TestActivePassiveHA.testNoConnectionOnPassive (batchId=283)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/18290/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/18290/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-18290/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 6 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12977088 - PreCommit-HIVE-Build

> Fix duplicated asf headers
> --
>
> Key: HIVE-21549
> URL: https://issues.apache.org/jira/browse/HIVE-21549
> Project: Hive
>  Issue Type: Bug
>Reporter: Laszlo Bodor
>Assignee: Sakthi
>Priority: Trivial
>  Labels: newbie
> Attachments: HIVE-21549.1-master.patch
>
>
> While fixing apache headers in HIVE-21451, a few occurrences of double asf 
> header found
> {code:java}
>  grep -src Licensed to the Apache Software Foundation . | grep -v ':0$' | 
> grep -v ':1$' | grep java
> ./standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/conf/ConfTemplatePrinter.java:2
> ./common/src/java/org/apache/hadoop/hive/ant/GenHiveTemplate.java:2
> ./ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFTrunc.java:2
> ./ql/src/test/org/apache/hadoop/hive/ql/exec/TestOperatorNames.java:2{code}
> (ConfTemplatePrinter seems like a false positive)



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (HIVE-16587) NPE when inserting complex types with nested null values

2019-08-08 Thread Sankar Hariappan (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-16587?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903532#comment-16903532
 ] 

Sankar Hariappan commented on HIVE-16587:
-

+1 for  [^HIVE-16587.1.patch] 


> NPE when inserting complex types with nested null values
> 
>
> Key: HIVE-16587
> URL: https://issues.apache.org/jira/browse/HIVE-16587
> Project: Hive
>  Issue Type: Bug
>  Components: Query Planning
>Affects Versions: 1.2.1
>Reporter: Jason Dere
>Assignee: Naresh P R
>Priority: Major
> Attachments: HIVE-16587.1.patch, HIVE-16587.patch
>
>
> {noformat}
> CREATE TABLE complex1 (c0 int, c1 array, c2 map, c3 
> struct>, c4 array f3:array>>)
> insert into complex1
>  select 3, array(1, 2, null), map(1, 'one', 2, null), named_struct('f1', 
> cast(null as int), 'f2', cast(null as string), 'f3', array(1,2,null)), 
> array(named_struct('f1', 11, 'f2', 'two', 'f3', array(2,3,4)))
> {noformat}
> Gives the following error:
> {noformat}
> Caused by: org.apache.hive.service.cli.HiveSQLException: Error while 
> compiling statement: FAILED: NullPointerException null
>   at 
> org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:315)
>   at 
> org.apache.hive.service.cli.operation.SQLOperation.prepare(SQLOperation.java:207)
>   at 
> org.apache.hive.service.cli.operation.SQLOperation.runInternal(SQLOperation.java:291)
>   at 
> org.apache.hive.service.cli.operation.Operation.run(Operation.java:255)
>   at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:531)
>   at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync(HiveSessionImpl.java:517)
>   at sun.reflect.GeneratedMethodAccessor49.invoke(Unknown Source)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:78)
>   at 
> org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:36)
>   at 
> org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:63)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:422)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1807)
>   at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:59)
>   at com.sun.proxy.$Proxy126.executeStatementAsync(Unknown Source)
>   at 
> org.apache.hive.service.cli.CLIService.executeStatementAsync(CLIService.java:310)
>   at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:530)
>   at 
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1437)
>   at 
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1422)
>   at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>   at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>   at 
> org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:56)
>   at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:286)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>   at java.lang.Thread.run(Thread.java:745)
> Caused by: java.lang.NullPointerException: null
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getWritableSize(StatsUtils.java:1144)
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getSizeOfMap(StatsUtils.java:1106)
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getSizeOfComplexTypes(StatsUtils.java:978)
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getAvgColLenOf(StatsUtils.java:916)
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getColStatisticsFromExpression(StatsUtils.java:1371)
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getColStatisticsFromExprMap(StatsUtils.java:1194)
>   at 
> org.apache.hadoop.hive.ql.optimizer.stats.annotation.StatsRulesProcFactory$SelectStatsRule.process(StatsRulesProcFactory.java:187)
>   at 
> org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:90)
>   at 
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatchAndReturn(DefaultGraphWalker.java:105)
>   at 
> 

[jira] [Commented] (HIVE-16587) NPE when inserting complex types with nested null values

2019-08-08 Thread Naresh P R (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-16587?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903529#comment-16903529
 ] 

Naresh P R commented on HIVE-16587:
---

[~sankarh] I added new patch addressing the review comments. Could you please 
check ?

> NPE when inserting complex types with nested null values
> 
>
> Key: HIVE-16587
> URL: https://issues.apache.org/jira/browse/HIVE-16587
> Project: Hive
>  Issue Type: Bug
>  Components: Query Planning
>Affects Versions: 1.2.1
>Reporter: Jason Dere
>Assignee: Naresh P R
>Priority: Major
> Attachments: HIVE-16587.1.patch, HIVE-16587.patch
>
>
> {noformat}
> CREATE TABLE complex1 (c0 int, c1 array, c2 map, c3 
> struct>, c4 array f3:array>>)
> insert into complex1
>  select 3, array(1, 2, null), map(1, 'one', 2, null), named_struct('f1', 
> cast(null as int), 'f2', cast(null as string), 'f3', array(1,2,null)), 
> array(named_struct('f1', 11, 'f2', 'two', 'f3', array(2,3,4)))
> {noformat}
> Gives the following error:
> {noformat}
> Caused by: org.apache.hive.service.cli.HiveSQLException: Error while 
> compiling statement: FAILED: NullPointerException null
>   at 
> org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:315)
>   at 
> org.apache.hive.service.cli.operation.SQLOperation.prepare(SQLOperation.java:207)
>   at 
> org.apache.hive.service.cli.operation.SQLOperation.runInternal(SQLOperation.java:291)
>   at 
> org.apache.hive.service.cli.operation.Operation.run(Operation.java:255)
>   at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:531)
>   at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync(HiveSessionImpl.java:517)
>   at sun.reflect.GeneratedMethodAccessor49.invoke(Unknown Source)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:78)
>   at 
> org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:36)
>   at 
> org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:63)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:422)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1807)
>   at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:59)
>   at com.sun.proxy.$Proxy126.executeStatementAsync(Unknown Source)
>   at 
> org.apache.hive.service.cli.CLIService.executeStatementAsync(CLIService.java:310)
>   at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:530)
>   at 
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1437)
>   at 
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1422)
>   at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>   at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>   at 
> org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:56)
>   at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:286)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>   at java.lang.Thread.run(Thread.java:745)
> Caused by: java.lang.NullPointerException: null
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getWritableSize(StatsUtils.java:1144)
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getSizeOfMap(StatsUtils.java:1106)
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getSizeOfComplexTypes(StatsUtils.java:978)
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getAvgColLenOf(StatsUtils.java:916)
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getColStatisticsFromExpression(StatsUtils.java:1371)
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getColStatisticsFromExprMap(StatsUtils.java:1194)
>   at 
> org.apache.hadoop.hive.ql.optimizer.stats.annotation.StatsRulesProcFactory$SelectStatsRule.process(StatsRulesProcFactory.java:187)
>   at 
> org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:90)
>   at 
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatchAndReturn(DefaultGraphWalker.java:105)
>   at 
> 

[jira] [Commented] (HIVE-21549) Fix duplicated asf headers

2019-08-08 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-21549?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903513#comment-16903513
 ] 

Hive QA commented on HIVE-21549:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
15s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
9s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
39s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  4m  
1s{color} | {color:blue} ql in master has 2250 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
59s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
29s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
4s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m  
4s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
38s{color} | {color:green} ql: The patch generated 0 new + 4 unchanged - 1 
fixed = 4 total (was 5) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  4m 
12s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
58s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 24m 14s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-18290/dev-support/hive-personality.sh
 |
| git revision | master / 7fc5a88 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| modules | C: ql U: ql |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18290/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Fix duplicated asf headers
> --
>
> Key: HIVE-21549
> URL: https://issues.apache.org/jira/browse/HIVE-21549
> Project: Hive
>  Issue Type: Bug
>Reporter: Laszlo Bodor
>Assignee: Sakthi
>Priority: Trivial
>  Labels: newbie
> Attachments: HIVE-21549.1-master.patch
>
>
> While fixing apache headers in HIVE-21451, a few occurrences of double asf 
> header found
> {code:java}
>  grep -src Licensed to the Apache Software Foundation . | grep -v ':0$' | 
> grep -v ':1$' | grep java
> ./standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/conf/ConfTemplatePrinter.java:2
> ./common/src/java/org/apache/hadoop/hive/ant/GenHiveTemplate.java:2
> ./ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFTrunc.java:2
> ./ql/src/test/org/apache/hadoop/hive/ql/exec/TestOperatorNames.java:2{code}
> (ConfTemplatePrinter seems like a false positive)



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (HIVE-16587) NPE when inserting complex types with nested null values

2019-08-08 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-16587?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903503#comment-16903503
 ] 

Hive QA commented on HIVE-16587:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12977051/HIVE-16587.1.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 16724 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/18289/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/18289/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-18289/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12977051 - PreCommit-HIVE-Build

> NPE when inserting complex types with nested null values
> 
>
> Key: HIVE-16587
> URL: https://issues.apache.org/jira/browse/HIVE-16587
> Project: Hive
>  Issue Type: Bug
>  Components: Query Planning
>Affects Versions: 1.2.1
>Reporter: Jason Dere
>Assignee: Naresh P R
>Priority: Major
> Attachments: HIVE-16587.1.patch, HIVE-16587.patch
>
>
> {noformat}
> CREATE TABLE complex1 (c0 int, c1 array, c2 map, c3 
> struct>, c4 array f3:array>>)
> insert into complex1
>  select 3, array(1, 2, null), map(1, 'one', 2, null), named_struct('f1', 
> cast(null as int), 'f2', cast(null as string), 'f3', array(1,2,null)), 
> array(named_struct('f1', 11, 'f2', 'two', 'f3', array(2,3,4)))
> {noformat}
> Gives the following error:
> {noformat}
> Caused by: org.apache.hive.service.cli.HiveSQLException: Error while 
> compiling statement: FAILED: NullPointerException null
>   at 
> org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:315)
>   at 
> org.apache.hive.service.cli.operation.SQLOperation.prepare(SQLOperation.java:207)
>   at 
> org.apache.hive.service.cli.operation.SQLOperation.runInternal(SQLOperation.java:291)
>   at 
> org.apache.hive.service.cli.operation.Operation.run(Operation.java:255)
>   at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:531)
>   at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync(HiveSessionImpl.java:517)
>   at sun.reflect.GeneratedMethodAccessor49.invoke(Unknown Source)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:78)
>   at 
> org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:36)
>   at 
> org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:63)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:422)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1807)
>   at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:59)
>   at com.sun.proxy.$Proxy126.executeStatementAsync(Unknown Source)
>   at 
> org.apache.hive.service.cli.CLIService.executeStatementAsync(CLIService.java:310)
>   at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:530)
>   at 
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1437)
>   at 
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1422)
>   at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>   at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>   at 
> org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:56)
>   at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:286)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>   at java.lang.Thread.run(Thread.java:745)
> Caused by: java.lang.NullPointerException: null
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getWritableSize(StatsUtils.java:1144)
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getSizeOfMap(StatsUtils.java:1106)
>   

[jira] [Updated] (HIVE-21393) Refactor exception handling in HiveAccumuloTableInputFormat

2019-08-08 Thread Sakthi (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-21393?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sakthi updated HIVE-21393:
--
Attachment: HIVE-21393.1-master.patch

> Refactor exception handling in HiveAccumuloTableInputFormat
> ---
>
> Key: HIVE-21393
> URL: https://issues.apache.org/jira/browse/HIVE-21393
> Project: Hive
>  Issue Type: Bug
>Reporter: Laszlo Bodor
>Assignee: Sakthi
>Priority: Major
>  Labels: newbie
> Attachments: HIVE-21393.1-master.patch
>
>
> https://github.com/apache/hive/blob/173d9816086e4c43677ae6dc3b1bf85203f894a1/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/mr/HiveAccumuloTableInputFormat.java#L178-L186
> From Java7, these catches could be merged.
> {code:java}
> } catch (AccumuloException e) {
>   log.error("Could not configure AccumuloInputFormat", e);
>   throw new IOException(StringUtils.stringifyException(e));
> } catch (AccumuloSecurityException e) {
>   log.error("Could not configure AccumuloInputFormat", e);
>   throw new IOException(StringUtils.stringifyException(e));
> } catch (SerDeException e) {
>   log.error("Could not configure AccumuloInputFormat", e);
>   throw new IOException(StringUtils.stringifyException(e));
> }
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (HIVE-21393) Refactor exception handling in HiveAccumuloTableInputFormat

2019-08-08 Thread Sakthi (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-21393?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sakthi updated HIVE-21393:
--
Status: Patch Available  (was: In Progress)

> Refactor exception handling in HiveAccumuloTableInputFormat
> ---
>
> Key: HIVE-21393
> URL: https://issues.apache.org/jira/browse/HIVE-21393
> Project: Hive
>  Issue Type: Bug
>Reporter: Laszlo Bodor
>Assignee: Sakthi
>Priority: Major
>  Labels: newbie
> Attachments: HIVE-21393.1-master.patch
>
>
> https://github.com/apache/hive/blob/173d9816086e4c43677ae6dc3b1bf85203f894a1/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/mr/HiveAccumuloTableInputFormat.java#L178-L186
> From Java7, these catches could be merged.
> {code:java}
> } catch (AccumuloException e) {
>   log.error("Could not configure AccumuloInputFormat", e);
>   throw new IOException(StringUtils.stringifyException(e));
> } catch (AccumuloSecurityException e) {
>   log.error("Could not configure AccumuloInputFormat", e);
>   throw new IOException(StringUtils.stringifyException(e));
> } catch (SerDeException e) {
>   log.error("Could not configure AccumuloInputFormat", e);
>   throw new IOException(StringUtils.stringifyException(e));
> }
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Assigned] (HIVE-21393) Refactor exception handling in HiveAccumuloTableInputFormat

2019-08-08 Thread Sakthi (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-21393?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sakthi reassigned HIVE-21393:
-

Assignee: Sakthi

> Refactor exception handling in HiveAccumuloTableInputFormat
> ---
>
> Key: HIVE-21393
> URL: https://issues.apache.org/jira/browse/HIVE-21393
> Project: Hive
>  Issue Type: Bug
>Reporter: Laszlo Bodor
>Assignee: Sakthi
>Priority: Major
>  Labels: newbie
>
> https://github.com/apache/hive/blob/173d9816086e4c43677ae6dc3b1bf85203f894a1/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/mr/HiveAccumuloTableInputFormat.java#L178-L186
> From Java7, these catches could be merged.
> {code:java}
> } catch (AccumuloException e) {
>   log.error("Could not configure AccumuloInputFormat", e);
>   throw new IOException(StringUtils.stringifyException(e));
> } catch (AccumuloSecurityException e) {
>   log.error("Could not configure AccumuloInputFormat", e);
>   throw new IOException(StringUtils.stringifyException(e));
> } catch (SerDeException e) {
>   log.error("Could not configure AccumuloInputFormat", e);
>   throw new IOException(StringUtils.stringifyException(e));
> }
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Work started] (HIVE-21393) Refactor exception handling in HiveAccumuloTableInputFormat

2019-08-08 Thread Sakthi (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-21393?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HIVE-21393 started by Sakthi.
-
> Refactor exception handling in HiveAccumuloTableInputFormat
> ---
>
> Key: HIVE-21393
> URL: https://issues.apache.org/jira/browse/HIVE-21393
> Project: Hive
>  Issue Type: Bug
>Reporter: Laszlo Bodor
>Assignee: Sakthi
>Priority: Major
>  Labels: newbie
>
> https://github.com/apache/hive/blob/173d9816086e4c43677ae6dc3b1bf85203f894a1/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/mr/HiveAccumuloTableInputFormat.java#L178-L186
> From Java7, these catches could be merged.
> {code:java}
> } catch (AccumuloException e) {
>   log.error("Could not configure AccumuloInputFormat", e);
>   throw new IOException(StringUtils.stringifyException(e));
> } catch (AccumuloSecurityException e) {
>   log.error("Could not configure AccumuloInputFormat", e);
>   throw new IOException(StringUtils.stringifyException(e));
> } catch (SerDeException e) {
>   log.error("Could not configure AccumuloInputFormat", e);
>   throw new IOException(StringUtils.stringifyException(e));
> }
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (HIVE-21549) Fix duplicated asf headers

2019-08-08 Thread Sakthi (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-21549?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sakthi updated HIVE-21549:
--
Status: Patch Available  (was: In Progress)

> Fix duplicated asf headers
> --
>
> Key: HIVE-21549
> URL: https://issues.apache.org/jira/browse/HIVE-21549
> Project: Hive
>  Issue Type: Bug
>Reporter: Laszlo Bodor
>Assignee: Sakthi
>Priority: Trivial
>  Labels: newbie
> Attachments: HIVE-21549.1-master.patch
>
>
> While fixing apache headers in HIVE-21451, a few occurrences of double asf 
> header found
> {code:java}
>  grep -src Licensed to the Apache Software Foundation . | grep -v ':0$' | 
> grep -v ':1$' | grep java
> ./standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/conf/ConfTemplatePrinter.java:2
> ./common/src/java/org/apache/hadoop/hive/ant/GenHiveTemplate.java:2
> ./ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFTrunc.java:2
> ./ql/src/test/org/apache/hadoop/hive/ql/exec/TestOperatorNames.java:2{code}
> (ConfTemplatePrinter seems like a false positive)



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (HIVE-21549) Fix duplicated asf headers

2019-08-08 Thread Sakthi (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-21549?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sakthi updated HIVE-21549:
--
Attachment: HIVE-21549.1-master.patch

> Fix duplicated asf headers
> --
>
> Key: HIVE-21549
> URL: https://issues.apache.org/jira/browse/HIVE-21549
> Project: Hive
>  Issue Type: Bug
>Reporter: Laszlo Bodor
>Assignee: Sakthi
>Priority: Trivial
>  Labels: newbie
> Attachments: HIVE-21549.1-master.patch
>
>
> While fixing apache headers in HIVE-21451, a few occurrences of double asf 
> header found
> {code:java}
>  grep -src Licensed to the Apache Software Foundation . | grep -v ':0$' | 
> grep -v ':1$' | grep java
> ./standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/conf/ConfTemplatePrinter.java:2
> ./common/src/java/org/apache/hadoop/hive/ant/GenHiveTemplate.java:2
> ./ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFTrunc.java:2
> ./ql/src/test/org/apache/hadoop/hive/ql/exec/TestOperatorNames.java:2{code}
> (ConfTemplatePrinter seems like a false positive)



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (HIVE-21549) Fix duplicated asf headers

2019-08-08 Thread Sakthi (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-21549?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903485#comment-16903485
 ] 

Sakthi commented on HIVE-21549:
---

Verified that the build succeeds. 

> Fix duplicated asf headers
> --
>
> Key: HIVE-21549
> URL: https://issues.apache.org/jira/browse/HIVE-21549
> Project: Hive
>  Issue Type: Bug
>Reporter: Laszlo Bodor
>Assignee: Sakthi
>Priority: Trivial
>  Labels: newbie
> Attachments: HIVE-21549.1-master.patch
>
>
> While fixing apache headers in HIVE-21451, a few occurrences of double asf 
> header found
> {code:java}
>  grep -src Licensed to the Apache Software Foundation . | grep -v ':0$' | 
> grep -v ':1$' | grep java
> ./standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/conf/ConfTemplatePrinter.java:2
> ./common/src/java/org/apache/hadoop/hive/ant/GenHiveTemplate.java:2
> ./ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFTrunc.java:2
> ./ql/src/test/org/apache/hadoop/hive/ql/exec/TestOperatorNames.java:2{code}
> (ConfTemplatePrinter seems like a false positive)



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Work started] (HIVE-21549) Fix duplicated asf headers

2019-08-08 Thread Sakthi (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-21549?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HIVE-21549 started by Sakthi.
-
> Fix duplicated asf headers
> --
>
> Key: HIVE-21549
> URL: https://issues.apache.org/jira/browse/HIVE-21549
> Project: Hive
>  Issue Type: Bug
>Reporter: Laszlo Bodor
>Assignee: Sakthi
>Priority: Trivial
>  Labels: newbie
>
> While fixing apache headers in HIVE-21451, a few occurrences of double asf 
> header found
> {code:java}
>  grep -src Licensed to the Apache Software Foundation . | grep -v ':0$' | 
> grep -v ':1$' | grep java
> ./standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/conf/ConfTemplatePrinter.java:2
> ./common/src/java/org/apache/hadoop/hive/ant/GenHiveTemplate.java:2
> ./ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFTrunc.java:2
> ./ql/src/test/org/apache/hadoop/hive/ql/exec/TestOperatorNames.java:2{code}
> (ConfTemplatePrinter seems like a false positive)



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Assigned] (HIVE-21549) Fix duplicated asf headers

2019-08-08 Thread Sakthi (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-21549?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sakthi reassigned HIVE-21549:
-

Assignee: Sakthi

> Fix duplicated asf headers
> --
>
> Key: HIVE-21549
> URL: https://issues.apache.org/jira/browse/HIVE-21549
> Project: Hive
>  Issue Type: Bug
>Reporter: Laszlo Bodor
>Assignee: Sakthi
>Priority: Trivial
>  Labels: newbie
>
> While fixing apache headers in HIVE-21451, a few occurrences of double asf 
> header found
> {code:java}
>  grep -src Licensed to the Apache Software Foundation . | grep -v ':0$' | 
> grep -v ':1$' | grep java
> ./standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/conf/ConfTemplatePrinter.java:2
> ./common/src/java/org/apache/hadoop/hive/ant/GenHiveTemplate.java:2
> ./ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDFTrunc.java:2
> ./ql/src/test/org/apache/hadoop/hive/ql/exec/TestOperatorNames.java:2{code}
> (ConfTemplatePrinter seems like a false positive)



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (HIVE-16587) NPE when inserting complex types with nested null values

2019-08-08 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-16587?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903478#comment-16903478
 ] 

Hive QA commented on HIVE-16587:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
45s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
 7s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
5s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
39s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  4m  
6s{color} | {color:blue} ql in master has 2250 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
1s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
28s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
28s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
5s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m  
5s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
38s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  4m 
19s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
59s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
13s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 25m 26s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-18289/dev-support/hive-personality.sh
 |
| git revision | master / 7fc5a88 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| modules | C: ql itests U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18289/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> NPE when inserting complex types with nested null values
> 
>
> Key: HIVE-16587
> URL: https://issues.apache.org/jira/browse/HIVE-16587
> Project: Hive
>  Issue Type: Bug
>  Components: Query Planning
>Affects Versions: 1.2.1
>Reporter: Jason Dere
>Assignee: Naresh P R
>Priority: Major
> Attachments: HIVE-16587.1.patch, HIVE-16587.patch
>
>
> {noformat}
> CREATE TABLE complex1 (c0 int, c1 array, c2 map, c3 
> struct>, c4 array f3:array>>)
> insert into complex1
>  select 3, array(1, 2, null), map(1, 'one', 2, null), named_struct('f1', 
> cast(null as int), 'f2', cast(null as string), 'f3', array(1,2,null)), 
> array(named_struct('f1', 11, 'f2', 'two', 'f3', array(2,3,4)))
> {noformat}
> Gives the following error:
> {noformat}
> Caused by: org.apache.hive.service.cli.HiveSQLException: Error while 
> compiling statement: FAILED: NullPointerException null
>   at 
> org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:315)
>   at 
> org.apache.hive.service.cli.operation.SQLOperation.prepare(SQLOperation.java:207)
>   at 
> org.apache.hive.service.cli.operation.SQLOperation.runInternal(SQLOperation.java:291)
>   at 
> org.apache.hive.service.cli.operation.Operation.run(Operation.java:255)
>   at 
> 

[jira] [Assigned] (HIVE-22091) NPE on HS2 start up due to bad data in FUNCS table

2019-08-08 Thread Eric Lin (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-22091?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eric Lin reassigned HIVE-22091:
---

Assignee: Eric Lin

> NPE on HS2 start up due to bad data in FUNCS table
> --
>
> Key: HIVE-22091
> URL: https://issues.apache.org/jira/browse/HIVE-22091
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 3.1.0
>Reporter: Eric Lin
>Assignee: Eric Lin
>Priority: Major
>
> If FUNCS table contains a stale DB_ID that has no links in DBS table, HS2 
> will fail to start up with NPE error:
> {code:bash}
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
> org.apache.hadoop.hive.ql.metadata.HiveException: 
> MetaException(message:java.lang.NullPointerException)
> at 
> org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:220)
> at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:338)
> at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:299)
> at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:274)
> at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:256)
> at 
> org.apache.hadoop.hive.ql.security.authorization.DefaultHiveAuthorizationProvider.init(DefaultHiveAuthorizationProvider.java:29)
> at 
> org.apache.hadoop.hive.ql.security.authorization.HiveAuthorizationProviderBase.setConf(HiveAuthorizationProviderBase.java:112)
> ... 21 more
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
> MetaException(message:java.lang.NullPointerException)
> at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3646)
> at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:231)
> at 
> org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:215)
> ... 27 more
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (HIVE-22033) HiveServer2: fix delegation token renewal

2019-08-08 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-22033?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903459#comment-16903459
 ] 

Hive QA commented on HIVE-22033:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12977041/HIVE-22033.patch

{color:red}ERROR:{color} -1 due to build exiting with an error

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/18288/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/18288/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-18288/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hiveptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ date '+%Y-%m-%d %T.%3N'
2019-08-09 00:53:24.262
+ [[ -n /usr/lib/jvm/java-8-openjdk-amd64 ]]
+ export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ export 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'MAVEN_OPTS=-Xmx1g '
+ MAVEN_OPTS='-Xmx1g '
+ cd /data/hiveptest/working/
+ tee /data/hiveptest/logs/PreCommit-HIVE-Build-18288/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ git = \s\v\n ]]
+ [[ git = \g\i\t ]]
+ [[ -z master ]]
+ [[ -d apache-github-source-source ]]
+ [[ ! -d apache-github-source-source/.git ]]
+ [[ ! -d apache-github-source-source ]]
+ date '+%Y-%m-%d %T.%3N'
2019-08-09 00:53:24.265
+ cd apache-github-source-source
+ git fetch origin
>From https://github.com/apache/hive
   f9bd0bf..7fc5a88  master -> origin/master
+ git reset --hard HEAD
HEAD is now at f9bd0bf HIVE-22080: Prevent implicit conversion from 
String/char/varchar to double/decimal (Ramesh Kumar Thangarajan, reviewed by 
Jason Dere)
+ git clean -f -d
Removing standalone-metastore/metastore-server/src/gen/
+ git checkout master
Already on 'master'
Your branch is behind 'origin/master' by 2 commits, and can be fast-forwarded.
  (use "git pull" to update your local branch)
+ git reset --hard origin/master
HEAD is now at 7fc5a88 HIVE-22089 : Upgrade jackson to 2.9.9 (Ashutosh Chauhan 
via Jesus Camacho Rodriguez)
+ git merge --ff-only origin/master
Already up-to-date.
+ date '+%Y-%m-%d %T.%3N'
2019-08-09 00:53:25.255
+ rm -rf ../yetus_PreCommit-HIVE-Build-18288
+ mkdir ../yetus_PreCommit-HIVE-Build-18288
+ git gc
+ cp -R . ../yetus_PreCommit-HIVE-Build-18288
+ mkdir /data/hiveptest/logs/PreCommit-HIVE-Build-18288/yetus
+ patchCommandPath=/data/hiveptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hiveptest/working/scratch/build.patch
+ [[ -f /data/hiveptest/working/scratch/build.patch ]]
+ chmod +x /data/hiveptest/working/scratch/smart-apply-patch.sh
+ /data/hiveptest/working/scratch/smart-apply-patch.sh 
/data/hiveptest/working/scratch/build.patch
error: 
a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/security/TokenStoreDelegationTokenSecretManager.java:
 does not exist in index
Going to apply patch with: git apply -p1
+ [[ maven == \m\a\v\e\n ]]
+ rm -rf /data/hiveptest/working/maven/org/apache/hive
+ mvn -B clean install -DskipTests -T 4 -q 
-Dmaven.repo.local=/data/hiveptest/working/maven
protoc-jar: executing: [/tmp/protoc184867962921955469.exe, --version]
libprotoc 2.5.0
protoc-jar: executing: [/tmp/protoc184867962921955469.exe, 
-I/data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-common/src/main/protobuf/org/apache/hadoop/hive/metastore,
 
--java_out=/data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-common/target/generated-sources,
 
/data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-common/src/main/protobuf/org/apache/hadoop/hive/metastore/metastore.proto]
ANTLR Parser Generator  Version 3.5.2
protoc-jar: executing: [/tmp/protoc1106268941952266120.exe, --version]
libprotoc 2.5.0
ANTLR Parser Generator  Version 3.5.2
Output file 
/data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-server/target/generated-sources/org/apache/hadoop/hive/metastore/parser/FilterParser.java
 does not exist: must build 
/data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/parser/Filter.g
org/apache/hadoop/hive/metastore/parser/Filter.g
[ERROR] COMPILATION ERROR : 
[ERROR] 

[jira] [Commented] (HIVE-20683) Add the Ability to push Dynamic Between and Bloom filters to Druid

2019-08-08 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20683?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903456#comment-16903456
 ] 

Hive QA commented on HIVE-20683:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12977038/HIVE-20683.5.patch

{color:green}SUCCESS:{color} +1 due to 2 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 16724 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[stats_part] (batchId=52)
org.apache.hadoop.hive.cli.TestTezPerfConstraintsCliDriver.testCliDriver[query1b]
 (batchId=296)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/18287/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/18287/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-18287/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12977038 - PreCommit-HIVE-Build

> Add the Ability to push Dynamic Between and Bloom filters to Druid
> --
>
> Key: HIVE-20683
> URL: https://issues.apache.org/jira/browse/HIVE-20683
> Project: Hive
>  Issue Type: New Feature
>  Components: Druid integration
>Reporter: Nishant Bangarwa
>Assignee: Nishant Bangarwa
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-20683.1.patch, HIVE-20683.2.patch, 
> HIVE-20683.3.patch, HIVE-20683.4.patch, HIVE-20683.5.patch, HIVE-20683.patch
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> For optimizing joins, Hive generates BETWEEN filter with min-max and BLOOM 
> filter for filtering one side of semi-join.
> Druid 0.13.0 will have support for Bloom filters (Added via 
> https://github.com/apache/incubator-druid/pull/6222)
> Implementation details - 
> # Hive generates and passes the filters as part of 'filterExpr' in TableScan. 
> # DruidQueryBasedRecordReader gets this filter passed as part of the conf. 
> # During execution phase, before sending the query to druid in 
> DruidQueryBasedRecordReader we will deserialize this filter, translate it 
> into a DruidDimFilter and add it to existing DruidQuery.  Tez executor 
> already ensures that when we start reading results from the record reader, 
> all the dynamic values are initialized. 
> # Explaining a druid query also prints the query sent to druid as 
> {{druid.json.query}}. We also need to make sure to update the druid query 
> with the filters. During explain we do not have the actual values for the 
> dynamic values, so instead of values we will print the dynamic expression 
> itself as part of druid query. 
> Note:- This work needs druid to be updated to version 0.13.0



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (HIVE-20683) Add the Ability to push Dynamic Between and Bloom filters to Druid

2019-08-08 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-20683?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903448#comment-16903448
 ] 

Hive QA commented on HIVE-20683:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
29s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
31s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  7m 
28s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  2m 
51s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  4m  
1s{color} | {color:blue} ql in master has 2250 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
28s{color} | {color:blue} druid-handler in master has 3 extant Findbugs 
warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
21s{color} | {color:blue} itests/qtest-druid in master has 7 extant Findbugs 
warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  8m 
36s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
25s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  9m 
40s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  7m 
39s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  7m 
39s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  2m 
53s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
2s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  5m 
11s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  8m  
8s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
13s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 67m 49s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  xml  compile  findbugs  
checkstyle  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-18287/dev-support/hive-personality.sh
 |
| git revision | master / 7fc5a88 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| modules | C: ql druid-handler . itests itests/qtest-druid U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18287/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Add the Ability to push Dynamic Between and Bloom filters to Druid
> --
>
> Key: HIVE-20683
> URL: https://issues.apache.org/jira/browse/HIVE-20683
> Project: Hive
>  Issue Type: New Feature
>  Components: Druid integration
>Reporter: Nishant Bangarwa
>Assignee: Nishant Bangarwa
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-20683.1.patch, HIVE-20683.2.patch, 
> HIVE-20683.3.patch, HIVE-20683.4.patch, HIVE-20683.5.patch, HIVE-20683.patch
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> For optimizing joins, Hive generates BETWEEN filter with min-max and BLOOM 
> filter for filtering one side of semi-join.
> Druid 0.13.0 will have support for Bloom filters (Added via 
> 

[jira] [Updated] (HIVE-22089) Upgrade jackson to 2.9.9

2019-08-08 Thread Ashutosh Chauhan (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-22089?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-22089:

   Resolution: Fixed
Fix Version/s: 4.0.0
   Status: Resolved  (was: Patch Available)

Pushed to master.

> Upgrade jackson to 2.9.9
> 
>
> Key: HIVE-22089
> URL: https://issues.apache.org/jira/browse/HIVE-22089
> Project: Hive
>  Issue Type: Improvement
>Affects Versions: 3.0.0, 3.1.0, 3.1.1
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-22089.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (HIVE-22090) Upgrade jetty to 9.3.27

2019-08-08 Thread Ashutosh Chauhan (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-22090?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashutosh Chauhan updated HIVE-22090:

   Resolution: Fixed
Fix Version/s: 4.0.0
   Status: Resolved  (was: Patch Available)

> Upgrade jetty to 9.3.27
> ---
>
> Key: HIVE-22090
> URL: https://issues.apache.org/jira/browse/HIVE-22090
> Project: Hive
>  Issue Type: Improvement
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-22090.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (HIVE-13457) Create HS2 REST API endpoints for monitoring information

2019-08-08 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-13457?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903408#comment-16903408
 ] 

Hive QA commented on HIVE-13457:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12977034/HIVE-13457.7.patch

{color:red}ERROR:{color} -1 due to build exiting with an error

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/18286/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/18286/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-18286/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hiveptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ date '+%Y-%m-%d %T.%3N'
2019-08-08 23:21:27.093
+ [[ -n /usr/lib/jvm/java-8-openjdk-amd64 ]]
+ export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ export 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'MAVEN_OPTS=-Xmx1g '
+ MAVEN_OPTS='-Xmx1g '
+ cd /data/hiveptest/working/
+ tee /data/hiveptest/logs/PreCommit-HIVE-Build-18286/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ git = \s\v\n ]]
+ [[ git = \g\i\t ]]
+ [[ -z master ]]
+ [[ -d apache-github-source-source ]]
+ [[ ! -d apache-github-source-source/.git ]]
+ [[ ! -d apache-github-source-source ]]
+ date '+%Y-%m-%d %T.%3N'
2019-08-08 23:21:27.096
+ cd apache-github-source-source
+ git fetch origin
+ git reset --hard HEAD
HEAD is now at f9bd0bf HIVE-22080: Prevent implicit conversion from 
String/char/varchar to double/decimal (Ramesh Kumar Thangarajan, reviewed by 
Jason Dere)
+ git clean -f -d
Removing standalone-metastore/metastore-server/src/gen/
+ git checkout master
Already on 'master'
Your branch is up-to-date with 'origin/master'.
+ git reset --hard origin/master
HEAD is now at f9bd0bf HIVE-22080: Prevent implicit conversion from 
String/char/varchar to double/decimal (Ramesh Kumar Thangarajan, reviewed by 
Jason Dere)
+ git merge --ff-only origin/master
Already up-to-date.
+ date '+%Y-%m-%d %T.%3N'
2019-08-08 23:21:28.234
+ rm -rf ../yetus_PreCommit-HIVE-Build-18286
+ mkdir ../yetus_PreCommit-HIVE-Build-18286
+ git gc
+ cp -R . ../yetus_PreCommit-HIVE-Build-18286
+ mkdir /data/hiveptest/logs/PreCommit-HIVE-Build-18286/yetus
+ patchCommandPath=/data/hiveptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hiveptest/working/scratch/build.patch
+ [[ -f /data/hiveptest/working/scratch/build.patch ]]
+ chmod +x /data/hiveptest/working/scratch/smart-apply-patch.sh
+ /data/hiveptest/working/scratch/smart-apply-patch.sh 
/data/hiveptest/working/scratch/build.patch
error: a/service/pom.xml: does not exist in index
error: a/service/src/java/org/apache/hive/service/server/HiveServer2.java: does 
not exist in index
error: 
a/service/src/test/org/apache/hive/service/server/TestHS2HttpServer.java: does 
not exist in index
Going to apply patch with: git apply -p1
+ [[ maven == \m\a\v\e\n ]]
+ rm -rf /data/hiveptest/working/maven/org/apache/hive
+ mvn -B clean install -DskipTests -T 4 -q 
-Dmaven.repo.local=/data/hiveptest/working/maven
protoc-jar: executing: [/tmp/protoc5999553822663847895.exe, --version]
libprotoc 2.5.0
protoc-jar: executing: [/tmp/protoc5999553822663847895.exe, 
-I/data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-common/src/main/protobuf/org/apache/hadoop/hive/metastore,
 
--java_out=/data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-common/target/generated-sources,
 
/data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-common/src/main/protobuf/org/apache/hadoop/hive/metastore/metastore.proto]
ANTLR Parser Generator  Version 3.5.2
protoc-jar: executing: [/tmp/protoc4108263382057123461.exe, --version]
libprotoc 2.5.0
ANTLR Parser Generator  Version 3.5.2
Output file 
/data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-server/target/generated-sources/org/apache/hadoop/hive/metastore/parser/FilterParser.java
 does not exist: must build 
/data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/parser/Filter.g
org/apache/hadoop/hive/metastore/parser/Filter.g
log4j:WARN No appenders could be found for logger (DataNucleus.Persistence).
log4j:WARN Please initialize the log4j system properly.
DataNucleus Enhancer (version 4.1.17) for API "JDO"
DataNucleus Enhancer completed with success for 41 

[jira] [Commented] (HIVE-22090) Upgrade jetty to 9.3.27

2019-08-08 Thread Vineet Garg (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-22090?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903407#comment-16903407
 ] 

Vineet Garg commented on HIVE-22090:


+1

> Upgrade jetty to 9.3.27
> ---
>
> Key: HIVE-22090
> URL: https://issues.apache.org/jira/browse/HIVE-22090
> Project: Hive
>  Issue Type: Improvement
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
>Priority: Major
> Attachments: HIVE-22090.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (HIVE-22089) Upgrade jackson to 2.9.9

2019-08-08 Thread Jesus Camacho Rodriguez (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-22089?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903406#comment-16903406
 ] 

Jesus Camacho Rodriguez commented on HIVE-22089:


+1

> Upgrade jackson to 2.9.9
> 
>
> Key: HIVE-22089
> URL: https://issues.apache.org/jira/browse/HIVE-22089
> Project: Hive
>  Issue Type: Improvement
>Affects Versions: 3.0.0, 3.1.0, 3.1.1
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
>Priority: Major
> Attachments: HIVE-22089.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (HIVE-8950) Add support in ParquetHiveSerde to create table schema from a parquet file

2019-08-08 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-8950?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903402#comment-16903402
 ] 

Hive QA commented on HIVE-8950:
---



Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12977013/HIVE-8950.9.patch

{color:green}SUCCESS:{color} +1 due to 14 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 76 failed/errored test(s), 16727 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestAccumuloCliDriver.testCliDriver[accumulo_custom_key2]
 (batchId=289)
org.apache.hadoop.hive.cli.TestAccumuloCliDriver.testCliDriver[accumulo_custom_key]
 (batchId=289)
org.apache.hadoop.hive.cli.TestAccumuloCliDriver.testCliDriver[accumulo_index] 
(batchId=289)
org.apache.hadoop.hive.cli.TestAccumuloCliDriver.testCliDriver[accumulo_predicate_pushdown]
 (batchId=289)
org.apache.hadoop.hive.cli.TestAccumuloCliDriver.testCliDriver[accumulo_queries]
 (batchId=289)
org.apache.hadoop.hive.cli.TestAccumuloCliDriver.testCliDriver[accumulo_single_sourced_multi_insert]
 (batchId=289)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[cast_datetime_with_sql_2016_format]
 (batchId=87)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[create_view_partitioned] 
(batchId=41)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[manyViewJoin] 
(batchId=39)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[nested_column_pruning] 
(batchId=37)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[parquet_array_map_emptynullvals]
 (batchId=37)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[parquet_map_of_arrays_of_ints]
 (batchId=10)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[parquet_map_of_maps] 
(batchId=72)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[parquet_mixed_partition_formats2]
 (batchId=65)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[parquet_schema_evolution]
 (batchId=83)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[parquet_vectorization_decimal_date]
 (batchId=33)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[parquet_vectorization_part]
 (batchId=79)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[parquet_vectorization_part_project]
 (batchId=39)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[parquet_vectorization_part_varchar]
 (batchId=79)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[schema_evol_par_vec_table_dictionary_encoding]
 (batchId=29)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[schema_evol_par_vec_table_non_dictionary_encoding]
 (batchId=55)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[table_storage] 
(batchId=26)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[vectorized_parquet_types]
 (batchId=73)
org.apache.hadoop.hive.cli.TestHBaseCliDriver.org.apache.hadoop.hive.cli.TestHBaseCliDriver
 (batchId=105)
org.apache.hadoop.hive.cli.TestHBaseCliDriver.org.apache.hadoop.hive.cli.TestHBaseCliDriver
 (batchId=106)
org.apache.hadoop.hive.cli.TestHBaseCliDriver.org.apache.hadoop.hive.cli.TestHBaseCliDriver
 (batchId=107)
org.apache.hadoop.hive.cli.TestHBaseCliDriver.org.apache.hadoop.hive.cli.TestHBaseCliDriver
 (batchId=108)
org.apache.hadoop.hive.cli.TestHBaseCliDriver.org.apache.hadoop.hive.cli.TestHBaseCliDriver
 (batchId=109)
org.apache.hadoop.hive.cli.TestHBaseCliDriver.org.apache.hadoop.hive.cli.TestHBaseCliDriver
 (batchId=110)
org.apache.hadoop.hive.cli.TestHBaseNegativeCliDriver.testCliDriver[cascade_dbdrop]
 (batchId=295)
org.apache.hadoop.hive.cli.TestHBaseNegativeCliDriver.testCliDriver[generatehfiles_require_family_path]
 (batchId=295)
org.apache.hadoop.hive.cli.TestHBaseNegativeCliDriver.testCliDriver[hbase_ddl] 
(batchId=295)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druid_materialized_view_rewrite_ssb]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druid_timeseries]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druid_timestamptz2]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druid_topn] 
(batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_dynamic_partition]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_expressions]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_extractTime]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_floorTime]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_mv] 
(batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_test1]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_test_alter]
 (batchId=198)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_test_insert]
 

[jira] [Commented] (HIVE-8950) Add support in ParquetHiveSerde to create table schema from a parquet file

2019-08-08 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-8950?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903392#comment-16903392
 ] 

Hive QA commented on HIVE-8950:
---

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
39s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
26s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
54s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
18s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
33s{color} | {color:blue} common in master has 62 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  1m 
13s{color} | {color:blue} standalone-metastore/metastore-server in master has 
180 extant Findbugs warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  4m  
1s{color} | {color:blue} ql in master has 2250 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
33s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
27s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  2m 
23s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
54s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m 
54s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
43s{color} | {color:red} ql: The patch generated 47 new + 210 unchanged - 0 
fixed = 257 total (was 210) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  4m 
12s{color} | {color:red} ql generated 1 new + 2250 unchanged - 0 fixed = 2251 
total (was 2250) {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
32s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
15s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 34m 43s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| FindBugs | module:ql |
|  |  Dead store to columnNameDelimiter in 
org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe.initialize(Configuration,
 Properties)  At 
ParquetHiveSerDe.java:org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe.initialize(Configuration,
 Properties)  At ParquetHiveSerDe.java:[line 98] |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-18285/dev-support/hive-personality.sh
 |
| git revision | master / f9bd0bf |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18285/yetus/diff-checkstyle-ql.txt
 |
| findbugs | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18285/yetus/new-findbugs-ql.html
 |
| modules | C: common standalone-metastore/metastore-server ql U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18285/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Add support in ParquetHiveSerde to create table schema from a parquet file
> --
>
> Key: HIVE-8950
> URL: https://issues.apache.org/jira/browse/HIVE-8950
> Project: Hive
>  Issue Type: Improvement
>Reporter: Ashish Singh
>

[jira] [Commented] (HIVE-19587) HeartBeat thread uses cancelled delegation token while connecting to meta on KERBEROS cluster

2019-08-08 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19587?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903373#comment-16903373
 ] 

Hive QA commented on HIVE-19587:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12924738/HIVE-19587.1.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 7 failed/errored test(s), 16723 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.llap.cache.TestBuddyAllocator.testMTT[2] (batchId=360)
org.apache.hadoop.hive.ql.parse.TestReplicationScenariosAcidTables.testMultiDBTxn
 (batchId=269)
org.apache.hadoop.hive.ql.parse.TestReplicationWithTableMigrationEx.testConcurrentOpDuringBootStrapDumpCreateTableReplay
 (batchId=264)
org.apache.hadoop.hive.ql.parse.TestReplicationWithTableMigrationEx.testConcurrentOpDuringBootStrapDumpInsertReplay
 (batchId=264)
org.apache.hadoop.hive.ql.parse.TestReplicationWithTableMigrationEx.testIncLoadPenFlagPropAlterDB
 (batchId=264)
org.apache.hadoop.hive.ql.parse.TestReplicationWithTableMigrationEx.testOnwerPropagation
 (batchId=264)
org.apache.hadoop.hive.ql.parse.TestTableLevelReplicationScenarios.testRenameTableScenariosWithReplacePolicyDMLOperattion
 (batchId=255)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/18284/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/18284/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-18284/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 7 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12924738 - PreCommit-HIVE-Build

> HeartBeat thread uses cancelled delegation token while connecting to meta on 
> KERBEROS cluster
> -
>
> Key: HIVE-19587
> URL: https://issues.apache.org/jira/browse/HIVE-19587
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Affects Versions: 3.0.0, 2.3.2, 3.1.0
>Reporter: Oleksiy Sayankin
>Assignee: Oleksiy Sayankin
>Priority: Blocker
> Attachments: HIVE-19587.1.patch
>
>
> *STEP 1. Create test data*
> {code}
> create table t1 (id int) ROW FORMAT DELIMITED FIELDS TERMINATED BY  ",";
> create table t2 (id int) ROW FORMAT DELIMITED FIELDS TERMINATED BY  ",";
> {code}
> Generate 10 000 000 lines of random data
> {code}
> package com.test.app;
> import java.io.FileNotFoundException;
> import java.io.PrintWriter;
> import java.util.concurrent.ThreadLocalRandom;
> public class App {
>   public static void main(String[] args) throws FileNotFoundException {
> try (PrintWriter out = new PrintWriter("table.data");) {
>   int min = 0;
>   int max = 10_000;
>   int numRows = 10_000_000;
>   for (int i = 0; i <= numRows - 1; i++){
> int randomNum = ThreadLocalRandom.current().nextInt(min, max + 1);
> out.println(randomNum);
>   }
> }
>   }
> }
> {code}
> Upload data to Hive tables
> {code}
> load data local inpath '/home/myuser/table.data' into table t1;
> load data local inpath '/home/myuser/table.data' into table t2;
> {code}
> *STEP 2. Configure transactions in hive-site.xml*
> {code}
> 
>
>   hive.exec.dynamic.partition.mode
>   nonstrict
>
>
>   hive.support.concurrency
>   true
>
>
>   hive.enforce.bucketing
>   true
>
>
>   hive.txn.manager
>   org.apache.hadoop.hive.ql.lockmgr.DbTxnManager
>
>
>   hive.compactor.initiator.on
>   true
>
>
>   hive.compactor.worker.threads
>   1
>
> {code}
> *STEP 3. Configure hive.txn.timeout in hive-site.xml*
> {code}
> 
>
>   hive.txn.timeout
>   10s
>
> {code}
> *STEP 4. Connect via beeline to HS2 with KERBEROS*
> {code}
> !connect 
> jdbc:hive2://node8.cluster:1/default;principal=myuser/node8.cluster@NODE8;ssl=true;sslTrustStore=/opt/myuser/conf/ssl_truststore
> {code}
> {code}
> select count(*) from t1;
> {code}
> *STEP 5. Close connection and reconnect*
> {code}
> !close
> {code}
> {code}
> !connect 
> jdbc:hive2://node8.cluster:1/default;principal=myuser/node8.cluster@NODE8;ssl=true;sslTrustStore=/opt/myuser/conf/ssl_truststore
> {code}
> *STEP 6. Perform long playing query*
> This query lasts about 600s
> {code}
> select count(*) from t1 join t2 on t1.id = t2.id;
> {code}
> *EXPECTED RESULT*
> Query finishes successfully
> *ACTUAL RESULT*
> {code}
> 2018-05-17T13:54:54,921 ERROR 

[jira] [Commented] (HIVE-22083) Values of tag order cannot be null, so it can be "byte" instead of "Byte"

2019-08-08 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-22083?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903289#comment-16903289
 ] 

Hive QA commented on HIVE-22083:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12976993/HIVE-22083.3.patch

{color:green}SUCCESS:{color} +1 due to 2 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 16723 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestSparkPerfCliDriver.testCliDriver[query23] 
(batchId=298)
org.apache.hadoop.hive.cli.TestSparkPerfCliDriver.testCliDriver[query88] 
(batchId=298)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/18283/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/18283/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-18283/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12976993 - PreCommit-HIVE-Build

> Values of tag order cannot be null, so it can be "byte" instead of "Byte"
> -
>
> Key: HIVE-22083
> URL: https://issues.apache.org/jira/browse/HIVE-22083
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Reporter: Ivan Suller
>Assignee: Ivan Suller
>Priority: Minor
> Attachments: HIVE-22083.1.patch, HIVE-22083.2.patch, 
> HIVE-22083.3.patch
>
>
> Values of tag order cannot be null, so it can be "byte" instead of "Byte". 
> Switching between Byte and byte is "cheap" - the Byte objects are cached by 
> the JVM - but it still costs a bit more memory and CPU usage.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (HIVE-22083) Values of tag order cannot be null, so it can be "byte" instead of "Byte"

2019-08-08 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-22083?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903261#comment-16903261
 ] 

Hive QA commented on HIVE-22083:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
14s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
4s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
48s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  4m  
4s{color} | {color:blue} ql in master has 2250 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
0s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
25s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
10s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m 
10s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
44s{color} | {color:red} ql: The patch generated 1 new + 599 unchanged - 53 
fixed = 600 total (was 652) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  4m 
18s{color} | {color:red} ql generated 4 new + 2237 unchanged - 13 fixed = 2241 
total (was 2250) {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
1s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
13s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 24m 41s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| FindBugs | module:ql |
|  |  org.apache.hadoop.hive.ql.plan.HashTableSinkDesc.getTagOrder() may expose 
internal representation by returning HashTableSinkDesc.tagOrder  At 
HashTableSinkDesc.java:by returning HashTableSinkDesc.tagOrder  At 
HashTableSinkDesc.java:[line 263] |
|  |  org.apache.hadoop.hive.ql.plan.HashTableSinkDesc.setTagOrder(byte[]) may 
expose internal representation by storing an externally mutable object into 
HashTableSinkDesc.tagOrder  At HashTableSinkDesc.java:by storing an externally 
mutable object into HashTableSinkDesc.tagOrder  At HashTableSinkDesc.java:[line 
268] |
|  |  org.apache.hadoop.hive.ql.plan.JoinDesc.getTagOrder() may expose internal 
representation by returning JoinDesc.tagOrder  At JoinDesc.java:by returning 
JoinDesc.tagOrder  At JoinDesc.java:[line 419] |
|  |  org.apache.hadoop.hive.ql.plan.JoinDesc.setTagOrder(byte[]) may expose 
internal representation by storing an externally mutable object into 
JoinDesc.tagOrder  At JoinDesc.java:by storing an externally mutable object 
into JoinDesc.tagOrder  At JoinDesc.java:[line 429] |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-18283/dev-support/hive-personality.sh
 |
| git revision | master / f9bd0bf |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18283/yetus/diff-checkstyle-ql.txt
 |
| findbugs | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18283/yetus/new-findbugs-ql.html
 |
| modules | C: ql U: ql |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18283/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Values of tag order cannot be null, so it can be "byte" instead of "Byte"
> -
>
> Key: HIVE-22083
> URL: https://issues.apache.org/jira/browse/HIVE-22083
> Project: Hive
>  

[jira] [Commented] (HIVE-22090) Upgrade jetty to 9.3.27

2019-08-08 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-22090?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903238#comment-16903238
 ] 

Hive QA commented on HIVE-22090:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12976984/HIVE-22090.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 16723 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/18282/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/18282/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-18282/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12976984 - PreCommit-HIVE-Build

> Upgrade jetty to 9.3.27
> ---
>
> Key: HIVE-22090
> URL: https://issues.apache.org/jira/browse/HIVE-22090
> Project: Hive
>  Issue Type: Improvement
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
>Priority: Major
> Attachments: HIVE-22090.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (HIVE-22090) Upgrade jetty to 9.3.27

2019-08-08 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-22090?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903216#comment-16903216
 ] 

Hive QA commented on HIVE-22090:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
29s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  5m 
52s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  6m 
46s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  6m 
41s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  6m  
2s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  6m  
2s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
1s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  6m 
47s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 41m 10s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  xml  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-18282/dev-support/hive-personality.sh
 |
| git revision | master / f9bd0bf |
| Default Java | 1.8.0_111 |
| modules | C: . U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18282/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Upgrade jetty to 9.3.27
> ---
>
> Key: HIVE-22090
> URL: https://issues.apache.org/jira/browse/HIVE-22090
> Project: Hive
>  Issue Type: Improvement
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
>Priority: Major
> Attachments: HIVE-22090.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (HIVE-22089) Upgrade jackson to 2.9.9

2019-08-08 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-22089?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903159#comment-16903159
 ] 

Hive QA commented on HIVE-22089:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12976983/HIVE-22089.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 16723 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/18281/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/18281/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-18281/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12976983 - PreCommit-HIVE-Build

> Upgrade jackson to 2.9.9
> 
>
> Key: HIVE-22089
> URL: https://issues.apache.org/jira/browse/HIVE-22089
> Project: Hive
>  Issue Type: Improvement
>Affects Versions: 3.0.0, 3.1.0, 3.1.1
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
>Priority: Major
> Attachments: HIVE-22089.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (HIVE-22089) Upgrade jackson to 2.9.9

2019-08-08 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-22089?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903133#comment-16903133
 ] 

Hive QA commented on HIVE-22089:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
46s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  5m 
59s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  6m 
48s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  6m 
48s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  5m 
51s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  5m 
51s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
1s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  6m 
50s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
12s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 41m 38s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  xml  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-18281/dev-support/hive-personality.sh
 |
| git revision | master / f9bd0bf |
| Default Java | 1.8.0_111 |
| modules | C: . U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18281/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Upgrade jackson to 2.9.9
> 
>
> Key: HIVE-22089
> URL: https://issues.apache.org/jira/browse/HIVE-22089
> Project: Hive
>  Issue Type: Improvement
>Affects Versions: 3.0.0, 3.1.0, 3.1.1
>Reporter: Ashutosh Chauhan
>Assignee: Ashutosh Chauhan
>Priority: Major
> Attachments: HIVE-22089.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (HIVE-16587) NPE when inserting complex types with nested null values

2019-08-08 Thread Naresh P R (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-16587?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Naresh P R updated HIVE-16587:
--
Attachment: HIVE-16587.1.patch

> NPE when inserting complex types with nested null values
> 
>
> Key: HIVE-16587
> URL: https://issues.apache.org/jira/browse/HIVE-16587
> Project: Hive
>  Issue Type: Bug
>  Components: Query Planning
>Affects Versions: 1.2.1
>Reporter: Jason Dere
>Assignee: Naresh P R
>Priority: Major
> Attachments: HIVE-16587.1.patch, HIVE-16587.patch
>
>
> {noformat}
> CREATE TABLE complex1 (c0 int, c1 array, c2 map, c3 
> struct>, c4 array f3:array>>)
> insert into complex1
>  select 3, array(1, 2, null), map(1, 'one', 2, null), named_struct('f1', 
> cast(null as int), 'f2', cast(null as string), 'f3', array(1,2,null)), 
> array(named_struct('f1', 11, 'f2', 'two', 'f3', array(2,3,4)))
> {noformat}
> Gives the following error:
> {noformat}
> Caused by: org.apache.hive.service.cli.HiveSQLException: Error while 
> compiling statement: FAILED: NullPointerException null
>   at 
> org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:315)
>   at 
> org.apache.hive.service.cli.operation.SQLOperation.prepare(SQLOperation.java:207)
>   at 
> org.apache.hive.service.cli.operation.SQLOperation.runInternal(SQLOperation.java:291)
>   at 
> org.apache.hive.service.cli.operation.Operation.run(Operation.java:255)
>   at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:531)
>   at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync(HiveSessionImpl.java:517)
>   at sun.reflect.GeneratedMethodAccessor49.invoke(Unknown Source)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:78)
>   at 
> org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:36)
>   at 
> org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:63)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:422)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1807)
>   at 
> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:59)
>   at com.sun.proxy.$Proxy126.executeStatementAsync(Unknown Source)
>   at 
> org.apache.hive.service.cli.CLIService.executeStatementAsync(CLIService.java:310)
>   at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:530)
>   at 
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1437)
>   at 
> org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1422)
>   at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>   at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>   at 
> org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:56)
>   at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:286)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>   at java.lang.Thread.run(Thread.java:745)
> Caused by: java.lang.NullPointerException: null
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getWritableSize(StatsUtils.java:1144)
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getSizeOfMap(StatsUtils.java:1106)
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getSizeOfComplexTypes(StatsUtils.java:978)
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getAvgColLenOf(StatsUtils.java:916)
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getColStatisticsFromExpression(StatsUtils.java:1371)
>   at 
> org.apache.hadoop.hive.ql.stats.StatsUtils.getColStatisticsFromExprMap(StatsUtils.java:1194)
>   at 
> org.apache.hadoop.hive.ql.optimizer.stats.annotation.StatsRulesProcFactory$SelectStatsRule.process(StatsRulesProcFactory.java:187)
>   at 
> org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:90)
>   at 
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatchAndReturn(DefaultGraphWalker.java:105)
>   at 
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:89)
>   at 
> 

[jira] [Updated] (HIVE-22033) HiveServer2: fix delegation token renewal

2019-08-08 Thread Szehon Ho (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-22033?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho updated HIVE-22033:
-
Assignee: Ion Alberdi
  Status: Patch Available  (was: Open)

> HiveServer2: fix delegation token renewal
> -
>
> Key: HIVE-22033
> URL: https://issues.apache.org/jira/browse/HIVE-22033
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.3.5
>Reporter: Ion Alberdi
>Assignee: Ion Alberdi
>Priority: Major
> Attachments: HIVE-22033.patch
>
>
> Hello, the issue we faced (and a proposal for a fix) in our hive instances is 
> depicted at
>  [https://github.com/criteo-forks/hive/pull/24]
> Reading the master branch of the project
>  
> [https://github.com/apache/hive/blob/master/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/security/TokenStoreDelegationTokenSecretManager.java#L147]
>  I think the same behavior is replicated there.
> Long story short, *TokenStoreDelegationTokenSecretManager.renewToken*, does 
> not update the expiry date of a given token (as it does not get the updated 
> DelegationTokenInformation from *super.currentTokens*).
> This makes any call to renewToken ineffective (the expiry date of the token 
> is not postponed).



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (HIVE-22002) Insert into table partition fails partially with stats.autogather is on.

2019-08-08 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-22002?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16903000#comment-16903000
 ] 

Hive QA commented on HIVE-22002:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12976977/HIVE-22002.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 21 failed/errored test(s), 16723 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[analyze_table_null_partition]
 (batchId=90)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[drop_partitions_filter4] 
(batchId=51)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[partition_coltype_literals]
 (batchId=14)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[partitions_filter_default]
 (batchId=55)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[smb_join_partition_key] 
(batchId=14)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[stats_only_null] 
(batchId=30)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[vector_non_string_partition]
 (batchId=37)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[dynpart_sort_opt_vectorization]
 (batchId=174)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[dynpart_sort_optimization]
 (batchId=175)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[llap_partitioned]
 (batchId=165)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[stats_only_null]
 (batchId=167)
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver[vector_non_string_partition]
 (batchId=110)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[dynpart_sort_optimization]
 (batchId=136)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[stats_only_null] 
(batchId=125)
org.apache.hadoop.hive.ql.lockmgr.TestDbTxnManager.testExceptions (batchId=350)
org.apache.hadoop.hive.ql.lockmgr.TestDbTxnManager.testJoin (batchId=350)
org.apache.hadoop.hive.ql.lockmgr.TestDbTxnManager.testReadWrite (batchId=350)
org.apache.hadoop.hive.ql.lockmgr.TestDbTxnManager.testSingleReadMultiPartition 
(batchId=350)
org.apache.hadoop.hive.ql.lockmgr.TestDbTxnManager.testSingleReadPartition 
(batchId=350)
org.apache.hadoop.hive.ql.lockmgr.TestDbTxnManager.testSingleWritePartition 
(batchId=350)
org.apache.hive.streaming.TestStreamingDynamicPartitioning.testDPTwoLevelNonStringPartitionColumns
 (batchId=357)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/18280/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/18280/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-18280/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 21 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12976977 - PreCommit-HIVE-Build

> Insert into table partition fails partially with stats.autogather is on.
> 
>
> Key: HIVE-22002
> URL: https://issues.apache.org/jira/browse/HIVE-22002
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2
>Affects Versions: 4.0.0
>Reporter: Naveen Gangam
>Assignee: bencao
>Priority: Major
> Attachments: HIVE-22002.patch, image-2019-07-31-20-02-38-069.png
>
>
> create table test_double(id int) partitioned by (dbtest double); 
> insert into test_double partition(dbtest) values (1,9.9); --> this works
> insert into test_double partition(dbtest) values (1,10); --> this fails 
> But if we change it to
> insert into test_double partition(dbtest) values (1, cast (10 as double)); it 
> succeeds 
> -> the problem is only seen when trying to insert a whole number i.e. 10, 
> 10.0, 15, 14.0 etc. The issue is not seen when inserting a number with 
> decimal values other than 0. So insert of 10.1 goes though. 
> The underlying  from the HMS is 
> {code}
> 2019-07-11T07:58:16,670  [pool-6-thread-196]: server.TThreadPoolServer 
> (TThreadPoolServer.java:run(297)) -  occurred during processing of message. 
> java.lang.IndexOutOfBoundsException: Index: 0 at 
> java.util.Collections$EmptyList.get(Collections.java:4454) ~[?:1.8.0_112] at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.updatePartColumnStatsWithMerge(HiveMetaStore.java:7808)
>  ~[hive-exec-3.1.0.3.1.0.0-78.jar:3.1.0.3.1.0.0-78] at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.set_aggr_stats_for(HiveMetaStore.java:7769)
>  

[jira] [Updated] (HIVE-22033) HiveServer2: fix delegation token renewal

2019-08-08 Thread Ion Alberdi (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-22033?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ion Alberdi updated HIVE-22033:
---
Attachment: HIVE-22033.patch

> HiveServer2: fix delegation token renewal
> -
>
> Key: HIVE-22033
> URL: https://issues.apache.org/jira/browse/HIVE-22033
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.3.5
>Reporter: Ion Alberdi
>Priority: Major
> Attachments: HIVE-22033.patch
>
>
> Hello, the issue we faced (and a proposal for a fix) in our hive instances is 
> depicted at
>  [https://github.com/criteo-forks/hive/pull/24]
> Reading the master branch of the project
>  
> [https://github.com/apache/hive/blob/master/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/security/TokenStoreDelegationTokenSecretManager.java#L147]
>  I think the same behavior is replicated there.
> Long story short, *TokenStoreDelegationTokenSecretManager.renewToken*, does 
> not update the expiry date of a given token (as it does not get the updated 
> DelegationTokenInformation from *super.currentTokens*).
> This makes any call to renewToken ineffective (the expiry date of the token 
> is not postponed).



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (HIVE-22033) HiveServer2: fix delegation token renewal

2019-08-08 Thread Ion Alberdi (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-22033?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16902974#comment-16902974
 ] 

Ion Alberdi commented on HIVE-22033:


FYI PR on master [https://github.com/apache/hive/pull/747]

> HiveServer2: fix delegation token renewal
> -
>
> Key: HIVE-22033
> URL: https://issues.apache.org/jira/browse/HIVE-22033
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 2.3.5
>Reporter: Ion Alberdi
>Priority: Major
>
> Hello, the issue we faced (and a proposal for a fix) in our hive instances is 
> depicted at
>  [https://github.com/criteo-forks/hive/pull/24]
> Reading the master branch of the project
>  
> [https://github.com/apache/hive/blob/master/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/security/TokenStoreDelegationTokenSecretManager.java#L147]
>  I think the same behavior is replicated there.
> Long story short, *TokenStoreDelegationTokenSecretManager.renewToken*, does 
> not update the expiry date of a given token (as it does not get the updated 
> DelegationTokenInformation from *super.currentTokens*).
> This makes any call to renewToken ineffective (the expiry date of the token 
> is not postponed).



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (HIVE-20683) Add the Ability to push Dynamic Between and Bloom filters to Druid

2019-08-08 Thread Nishant Bangarwa (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20683?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nishant Bangarwa updated HIVE-20683:

Attachment: HIVE-20683.5.patch

> Add the Ability to push Dynamic Between and Bloom filters to Druid
> --
>
> Key: HIVE-20683
> URL: https://issues.apache.org/jira/browse/HIVE-20683
> Project: Hive
>  Issue Type: New Feature
>  Components: Druid integration
>Reporter: Nishant Bangarwa
>Assignee: Nishant Bangarwa
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-20683.1.patch, HIVE-20683.2.patch, 
> HIVE-20683.3.patch, HIVE-20683.4.patch, HIVE-20683.5.patch, HIVE-20683.patch
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> For optimizing joins, Hive generates BETWEEN filter with min-max and BLOOM 
> filter for filtering one side of semi-join.
> Druid 0.13.0 will have support for Bloom filters (Added via 
> https://github.com/apache/incubator-druid/pull/6222)
> Implementation details - 
> # Hive generates and passes the filters as part of 'filterExpr' in TableScan. 
> # DruidQueryBasedRecordReader gets this filter passed as part of the conf. 
> # During execution phase, before sending the query to druid in 
> DruidQueryBasedRecordReader we will deserialize this filter, translate it 
> into a DruidDimFilter and add it to existing DruidQuery.  Tez executor 
> already ensures that when we start reading results from the record reader, 
> all the dynamic values are initialized. 
> # Explaining a druid query also prints the query sent to druid as 
> {{druid.json.query}}. We also need to make sure to update the druid query 
> with the filters. During explain we do not have the actual values for the 
> dynamic values, so instead of values we will print the dynamic expression 
> itself as part of druid query. 
> Note:- This work needs druid to be updated to version 0.13.0



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (HIVE-20683) Add the Ability to push Dynamic Between and Bloom filters to Druid

2019-08-08 Thread Nishant Bangarwa (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-20683?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nishant Bangarwa updated HIVE-20683:

Attachment: (was: HIVE-20683.5.patch)

> Add the Ability to push Dynamic Between and Bloom filters to Druid
> --
>
> Key: HIVE-20683
> URL: https://issues.apache.org/jira/browse/HIVE-20683
> Project: Hive
>  Issue Type: New Feature
>  Components: Druid integration
>Reporter: Nishant Bangarwa
>Assignee: Nishant Bangarwa
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-20683.1.patch, HIVE-20683.2.patch, 
> HIVE-20683.3.patch, HIVE-20683.4.patch, HIVE-20683.patch
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> For optimizing joins, Hive generates BETWEEN filter with min-max and BLOOM 
> filter for filtering one side of semi-join.
> Druid 0.13.0 will have support for Bloom filters (Added via 
> https://github.com/apache/incubator-druid/pull/6222)
> Implementation details - 
> # Hive generates and passes the filters as part of 'filterExpr' in TableScan. 
> # DruidQueryBasedRecordReader gets this filter passed as part of the conf. 
> # During execution phase, before sending the query to druid in 
> DruidQueryBasedRecordReader we will deserialize this filter, translate it 
> into a DruidDimFilter and add it to existing DruidQuery.  Tez executor 
> already ensures that when we start reading results from the record reader, 
> all the dynamic values are initialized. 
> # Explaining a druid query also prints the query sent to druid as 
> {{druid.json.query}}. We also need to make sure to update the druid query 
> with the filters. During explain we do not have the actual values for the 
> dynamic values, so instead of values we will print the dynamic expression 
> itself as part of druid query. 
> Note:- This work needs druid to be updated to version 0.13.0



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (HIVE-22002) Insert into table partition fails partially with stats.autogather is on.

2019-08-08 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-22002?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16902946#comment-16902946
 ] 

Hive QA commented on HIVE-22002:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
19s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
8s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
41s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  4m  
4s{color} | {color:blue} ql in master has 2250 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
3s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
27s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
9s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m  
9s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
42s{color} | {color:red} ql: The patch generated 12 new + 224 unchanged - 0 
fixed = 236 total (was 224) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  4m 
11s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} javadoc {color} | {color:red}  1m  
0s{color} | {color:red} ql generated 1 new + 99 unchanged - 1 fixed = 100 total 
(was 100) {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
15s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 24m 38s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-18280/dev-support/hive-personality.sh
 |
| git revision | master / f9bd0bf |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18280/yetus/diff-checkstyle-ql.txt
 |
| javadoc | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18280/yetus/diff-javadoc-javadoc-ql.txt
 |
| modules | C: ql U: ql |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-18280/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Insert into table partition fails partially with stats.autogather is on.
> 
>
> Key: HIVE-22002
> URL: https://issues.apache.org/jira/browse/HIVE-22002
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2
>Affects Versions: 4.0.0
>Reporter: Naveen Gangam
>Assignee: bencao
>Priority: Major
> Attachments: HIVE-22002.patch, image-2019-07-31-20-02-38-069.png
>
>
> create table test_double(id int) partitioned by (dbtest double); 
> insert into test_double partition(dbtest) values (1,9.9); --> this works
> insert into test_double partition(dbtest) values (1,10); --> this fails 
> But if we change it to
> insert into test_double partition(dbtest) values (1, cast (10 as double)); it 
> succeeds 
> -> the problem is only seen when trying to insert a whole number i.e. 10, 
> 10.0, 15, 14.0 etc. The issue is not seen when inserting a number with 
> decimal values other than 0. So insert of 10.1 goes though. 
> The underlying  from the HMS is 
> {code}
> 2019-07-11T07:58:16,670  [pool-6-thread-196]: server.TThreadPoolServer 
> (TThreadPoolServer.java:run(297)) -  occurred during processing of message. 
> java.lang.IndexOutOfBoundsException: Index: 0 at 
> java.util.Collections$EmptyList.get(Collections.java:4454) 

[jira] [Commented] (HIVE-13457) Create HS2 REST API endpoints for monitoring information

2019-08-08 Thread Szehon Ho (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-13457?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16902913#comment-16902913
 ] 

Szehon Ho commented on HIVE-13457:
--

Also we have been using this patch for an internal Hive UI project in Criteo 
and it's been quite stable :)

> Create HS2 REST API endpoints for monitoring information
> 
>
> Key: HIVE-13457
> URL: https://issues.apache.org/jira/browse/HIVE-13457
> Project: Hive
>  Issue Type: Improvement
>Reporter: Szehon Ho
>Assignee: Pawel Szostek
>Priority: Major
> Attachments: HIVE-13457.3.patch, HIVE-13457.4.patch, 
> HIVE-13457.5.patch, HIVE-13457.6.patch, HIVE-13457.6.patch, 
> HIVE-13457.7.patch, HIVE-13457.patch, HIVE-13457.patch
>
>
> Similar to what is exposed in HS2 webui in HIVE-12338, it would be nice if 
> other UI's like admin tools or Hue can access and display this information as 
> well.  Hence, we will create some REST endpoints to expose this information.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (HIVE-13457) Create HS2 REST API endpoints for monitoring information

2019-08-08 Thread Szehon Ho (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-13457?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16902911#comment-16902911
 ] 

Szehon Ho commented on HIVE-13457:
--

So someone recently has already added powermock to the parent pom, I think we 
are good.  I +1 the patch and removed the redudant powermock version from 
service pom

> Create HS2 REST API endpoints for monitoring information
> 
>
> Key: HIVE-13457
> URL: https://issues.apache.org/jira/browse/HIVE-13457
> Project: Hive
>  Issue Type: Improvement
>Reporter: Szehon Ho
>Assignee: Pawel Szostek
>Priority: Major
> Attachments: HIVE-13457.3.patch, HIVE-13457.4.patch, 
> HIVE-13457.5.patch, HIVE-13457.6.patch, HIVE-13457.6.patch, 
> HIVE-13457.7.patch, HIVE-13457.patch, HIVE-13457.patch
>
>
> Similar to what is exposed in HS2 webui in HIVE-12338, it would be nice if 
> other UI's like admin tools or Hue can access and display this information as 
> well.  Hence, we will create some REST endpoints to expose this information.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (HIVE-13457) Create HS2 REST API endpoints for monitoring information

2019-08-08 Thread Szehon Ho (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-13457?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho updated HIVE-13457:
-
Attachment: HIVE-13457.7.patch

> Create HS2 REST API endpoints for monitoring information
> 
>
> Key: HIVE-13457
> URL: https://issues.apache.org/jira/browse/HIVE-13457
> Project: Hive
>  Issue Type: Improvement
>Reporter: Szehon Ho
>Assignee: Pawel Szostek
>Priority: Major
> Attachments: HIVE-13457.3.patch, HIVE-13457.4.patch, 
> HIVE-13457.5.patch, HIVE-13457.6.patch, HIVE-13457.6.patch, 
> HIVE-13457.7.patch, HIVE-13457.patch, HIVE-13457.patch
>
>
> Similar to what is exposed in HS2 webui in HIVE-12338, it would be nice if 
> other UI's like admin tools or Hue can access and display this information as 
> well.  Hence, we will create some REST endpoints to expose this information.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (HIVE-21579) Introduce more complex SQL:2016 datetime formats

2019-08-08 Thread Marta Kuczora (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-21579?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16902840#comment-16902840
 ] 

Marta Kuczora commented on HIVE-21579:
--

+1
Thanks a lot [~klcopp] for the patch.

> Introduce more complex SQL:2016 datetime formats
> 
>
> Key: HIVE-21579
> URL: https://issues.apache.org/jira/browse/HIVE-21579
> Project: Hive
>  Issue Type: Improvement
>Reporter: Karen Coppage
>Assignee: Karen Coppage
>Priority: Major
> Attachments: HIVE-21579.01.patch, HIVE-21579.02.patch
>
>
> Enable Hive to parse the following datetime formats when any 
> combination/subset of these or previously implemented patterns is provided in 
> one string. Also catch combinations that conflict.
>  * MONTH
>  * MON
>  * D
>  * DAY
>  * DY
>  * Q
>  * WW
>  * W
> [https://docs.google.com/document/d/1V7k6-lrPGW7_uhqM-FhKl3QsxwCRy69v2KIxPsGjc1k/edit]



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Assigned] (HIVE-8950) Add support in ParquetHiveSerde to create table schema from a parquet file

2019-08-08 Thread Szehon Ho (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-8950?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho reassigned HIVE-8950:
---

Assignee: Szehon Ho  (was: Ashish Singh)

> Add support in ParquetHiveSerde to create table schema from a parquet file
> --
>
> Key: HIVE-8950
> URL: https://issues.apache.org/jira/browse/HIVE-8950
> Project: Hive
>  Issue Type: Improvement
>Reporter: Ashish Singh
>Assignee: Szehon Ho
>Priority: Major
> Attachments: HIVE-8950.1.patch, HIVE-8950.2.patch, HIVE-8950.3.patch, 
> HIVE-8950.4.patch, HIVE-8950.5.patch, HIVE-8950.6.patch, HIVE-8950.7.patch, 
> HIVE-8950.8.patch, HIVE-8950.9.patch, HIVE-8950.patch
>
>
> PARQUET-76 and PARQUET-47 ask for creating parquet backed tables without 
> having to specify the column names and types. As, parquet files store schema 
> in their footer, it is possible to generate hive schema from parquet file's 
> metadata. This will improve usability of parquet backed tables.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (HIVE-8950) Add support in ParquetHiveSerde to create table schema from a parquet file

2019-08-08 Thread Szehon Ho (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-8950?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho updated HIVE-8950:

Assignee: Ashish Singh  (was: Szehon Ho)
  Status: Patch Available  (was: In Progress)

> Add support in ParquetHiveSerde to create table schema from a parquet file
> --
>
> Key: HIVE-8950
> URL: https://issues.apache.org/jira/browse/HIVE-8950
> Project: Hive
>  Issue Type: Improvement
>Reporter: Ashish Singh
>Assignee: Ashish Singh
>Priority: Major
> Attachments: HIVE-8950.1.patch, HIVE-8950.2.patch, HIVE-8950.3.patch, 
> HIVE-8950.4.patch, HIVE-8950.5.patch, HIVE-8950.6.patch, HIVE-8950.7.patch, 
> HIVE-8950.8.patch, HIVE-8950.9.patch, HIVE-8950.patch
>
>
> PARQUET-76 and PARQUET-47 ask for creating parquet backed tables without 
> having to specify the column names and types. As, parquet files store schema 
> in their footer, it is possible to generate hive schema from parquet file's 
> metadata. This will improve usability of parquet backed tables.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (HIVE-8950) Add support in ParquetHiveSerde to create table schema from a parquet file

2019-08-08 Thread Szehon Ho (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-8950?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho updated HIVE-8950:

Assignee: Ashish Singh  (was: Gaurav Kumar)
  Status: In Progress  (was: Patch Available)

> Add support in ParquetHiveSerde to create table schema from a parquet file
> --
>
> Key: HIVE-8950
> URL: https://issues.apache.org/jira/browse/HIVE-8950
> Project: Hive
>  Issue Type: Improvement
>Reporter: Ashish Singh
>Assignee: Ashish Singh
>Priority: Major
> Attachments: HIVE-8950.1.patch, HIVE-8950.2.patch, HIVE-8950.3.patch, 
> HIVE-8950.4.patch, HIVE-8950.5.patch, HIVE-8950.6.patch, HIVE-8950.7.patch, 
> HIVE-8950.8.patch, HIVE-8950.9.patch, HIVE-8950.patch
>
>
> PARQUET-76 and PARQUET-47 ask for creating parquet backed tables without 
> having to specify the column names and types. As, parquet files store schema 
> in their footer, it is possible to generate hive schema from parquet file's 
> metadata. This will improve usability of parquet backed tables.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (HIVE-8950) Add support in ParquetHiveSerde to create table schema from a parquet file

2019-08-08 Thread Szehon Ho (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-8950?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szehon Ho updated HIVE-8950:

Attachment: HIVE-8950.9.patch

> Add support in ParquetHiveSerde to create table schema from a parquet file
> --
>
> Key: HIVE-8950
> URL: https://issues.apache.org/jira/browse/HIVE-8950
> Project: Hive
>  Issue Type: Improvement
>Reporter: Ashish Singh
>Assignee: Gaurav Kumar
>Priority: Major
> Attachments: HIVE-8950.1.patch, HIVE-8950.2.patch, HIVE-8950.3.patch, 
> HIVE-8950.4.patch, HIVE-8950.5.patch, HIVE-8950.6.patch, HIVE-8950.7.patch, 
> HIVE-8950.8.patch, HIVE-8950.9.patch, HIVE-8950.patch
>
>
> PARQUET-76 and PARQUET-47 ask for creating parquet backed tables without 
> having to specify the column names and types. As, parquet files store schema 
> in their footer, it is possible to generate hive schema from parquet file's 
> metadata. This will improve usability of parquet backed tables.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (HIVE-8950) Add support in ParquetHiveSerde to create table schema from a parquet file

2019-08-08 Thread Szehon Ho (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-8950?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16902816#comment-16902816
 ] 

Szehon Ho commented on HIVE-8950:
-

Hello, I'm a few years late :) but in Criteo we have been using this patch for 
a few months and it's really great.  I'll +1 the patch and try to rebase and 
get it in, if its ok for everyone here

> Add support in ParquetHiveSerde to create table schema from a parquet file
> --
>
> Key: HIVE-8950
> URL: https://issues.apache.org/jira/browse/HIVE-8950
> Project: Hive
>  Issue Type: Improvement
>Reporter: Ashish Singh
>Assignee: Gaurav Kumar
>Priority: Major
> Attachments: HIVE-8950.1.patch, HIVE-8950.2.patch, HIVE-8950.3.patch, 
> HIVE-8950.4.patch, HIVE-8950.5.patch, HIVE-8950.6.patch, HIVE-8950.7.patch, 
> HIVE-8950.8.patch, HIVE-8950.patch
>
>
> PARQUET-76 and PARQUET-47 ask for creating parquet backed tables without 
> having to specify the column names and types. As, parquet files store schema 
> in their footer, it is possible to generate hive schema from parquet file's 
> metadata. This will improve usability of parquet backed tables.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (HIVE-21719) Use Random.nextDouble instead of Math.random

2019-08-08 Thread xiepengjie (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-21719?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16902799#comment-16902799
 ] 

xiepengjie commented on HIVE-21719:
---

What are the problems that can be solved by this patch?

> Use Random.nextDouble instead of Math.random
> 
>
> Key: HIVE-21719
> URL: https://issues.apache.org/jira/browse/HIVE-21719
> Project: Hive
>  Issue Type: Improvement
>Reporter: bd2019us
>Priority: Trivial
> Attachments: HIVE-21719.patch
>
>
> Performance overhead from Math.random can be reduced by using 
> Random.nextDouble instead.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (HIVE-19587) HeartBeat thread uses cancelled delegation token while connecting to meta on KERBEROS cluster

2019-08-08 Thread zengxl (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19587?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16902798#comment-16902798
 ] 

zengxl commented on HIVE-19587:
---

We didn't open the transaction, we also encountered the same problem.
h5. 
hive.cluster.delegation.token.store.class=org.apache.hadoop.hive.thrift.ZooKeeperTokenStore
h5. hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DummyTxnManager

> HeartBeat thread uses cancelled delegation token while connecting to meta on 
> KERBEROS cluster
> -
>
> Key: HIVE-19587
> URL: https://issues.apache.org/jira/browse/HIVE-19587
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Affects Versions: 3.0.0, 2.3.2, 3.1.0
>Reporter: Oleksiy Sayankin
>Assignee: Oleksiy Sayankin
>Priority: Blocker
> Attachments: HIVE-19587.1.patch
>
>
> *STEP 1. Create test data*
> {code}
> create table t1 (id int) ROW FORMAT DELIMITED FIELDS TERMINATED BY  ",";
> create table t2 (id int) ROW FORMAT DELIMITED FIELDS TERMINATED BY  ",";
> {code}
> Generate 10 000 000 lines of random data
> {code}
> package com.test.app;
> import java.io.FileNotFoundException;
> import java.io.PrintWriter;
> import java.util.concurrent.ThreadLocalRandom;
> public class App {
>   public static void main(String[] args) throws FileNotFoundException {
> try (PrintWriter out = new PrintWriter("table.data");) {
>   int min = 0;
>   int max = 10_000;
>   int numRows = 10_000_000;
>   for (int i = 0; i <= numRows - 1; i++){
> int randomNum = ThreadLocalRandom.current().nextInt(min, max + 1);
> out.println(randomNum);
>   }
> }
>   }
> }
> {code}
> Upload data to Hive tables
> {code}
> load data local inpath '/home/myuser/table.data' into table t1;
> load data local inpath '/home/myuser/table.data' into table t2;
> {code}
> *STEP 2. Configure transactions in hive-site.xml*
> {code}
> 
>
>   hive.exec.dynamic.partition.mode
>   nonstrict
>
>
>   hive.support.concurrency
>   true
>
>
>   hive.enforce.bucketing
>   true
>
>
>   hive.txn.manager
>   org.apache.hadoop.hive.ql.lockmgr.DbTxnManager
>
>
>   hive.compactor.initiator.on
>   true
>
>
>   hive.compactor.worker.threads
>   1
>
> {code}
> *STEP 3. Configure hive.txn.timeout in hive-site.xml*
> {code}
> 
>
>   hive.txn.timeout
>   10s
>
> {code}
> *STEP 4. Connect via beeline to HS2 with KERBEROS*
> {code}
> !connect 
> jdbc:hive2://node8.cluster:1/default;principal=myuser/node8.cluster@NODE8;ssl=true;sslTrustStore=/opt/myuser/conf/ssl_truststore
> {code}
> {code}
> select count(*) from t1;
> {code}
> *STEP 5. Close connection and reconnect*
> {code}
> !close
> {code}
> {code}
> !connect 
> jdbc:hive2://node8.cluster:1/default;principal=myuser/node8.cluster@NODE8;ssl=true;sslTrustStore=/opt/myuser/conf/ssl_truststore
> {code}
> *STEP 6. Perform long playing query*
> This query lasts about 600s
> {code}
> select count(*) from t1 join t2 on t1.id = t2.id;
> {code}
> *EXPECTED RESULT*
> Query finishes successfully
> *ACTUAL RESULT*
> {code}
> 2018-05-17T13:54:54,921 ERROR [pool-7-thread-10] transport.TSaslTransport: 
> SASL negotiation failure
> javax.security.sasl.SaslException: DIGEST-MD5: IO error acquiring password
> at 
> com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java:598)
>  
> at 
> com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java:244)
> at 
> org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java:539)
>  
> at 
> org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java:283)
> at 
> org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java:41)
>  
> at 
> org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java:216)
>  
> at 
> org.apache.hadoop.hive.thrift.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java:663)
>  
> at 
> org.apache.hadoop.hive.thrift.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java:660)
>  
> at java.security.AccessController.doPrivileged(Native Method) 
> at javax.security.auth.Subject.doAs(Subject.java:360)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1613)
>  
> at 
> org.apache.hadoop.hive.thrift.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java:660)
>  
> at 
> 

[jira] [Updated] (HIVE-22083) Values of tag order cannot be null, so it can be "byte" instead of "Byte"

2019-08-08 Thread Ivan Suller (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-22083?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ivan Suller updated HIVE-22083:
---
Attachment: HIVE-22083.3.patch

> Values of tag order cannot be null, so it can be "byte" instead of "Byte"
> -
>
> Key: HIVE-22083
> URL: https://issues.apache.org/jira/browse/HIVE-22083
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Reporter: Ivan Suller
>Assignee: Ivan Suller
>Priority: Minor
> Attachments: HIVE-22083.1.patch, HIVE-22083.2.patch, 
> HIVE-22083.3.patch
>
>
> Values of tag order cannot be null, so it can be "byte" instead of "Byte". 
> Switching between Byte and byte is "cheap" - the Byte objects are cached by 
> the JVM - but it still costs a bit more memory and CPU usage.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)