[jira] [Updated] (HIVE-23123) Export/Import of views and materialized views

2020-04-01 Thread Miklos Gergely (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23123?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Miklos Gergely updated HIVE-23123:
--
Description: 
According to 
[https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport] 
import and export can be done by using the
{code:java}
export table ...
import table ... 
{code}
commands. The document doesn't mention views or materialized views at all, and 
in fact we don't support commands like
{code:java}
export view ...
import view ...
export materialized view ...
import materialized view ... 
{code}
they can not be parsed at all. The word table is often used though in a broader 
sense, when it means all table like entities, including views and materialized 
views. For example the various Table classes may represent any of these as well.

If I try to export a view with the export table ... command, it goes fine. A 
_metadata file will be created, but no data directory, which is what we'd 
expect. If I try to import it back, an exception is thrown due to the lack of 
the data dir:
{code:java}
java.lang.AssertionError: null==getPath() for exim_view
 at org.apache.hadoop.hive.ql.metadata.Hive.loadTable(Hive.java:3088)
 at org.apache.hadoop.hive.ql.exec.MoveTask.execute(MoveTask.java:419)
 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:213)
 at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:105)
 at org.apache.hadoop.hive.ql.Executor.launchTask(Executor.java:364)
 at org.apache.hadoop.hive.ql.Executor.launchTasks(Executor.java:335)
 at org.apache.hadoop.hive.ql.Executor.runTasks(Executor.java:246)
 at org.apache.hadoop.hive.ql.Executor.execute(Executor.java:109)
 at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:722)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:491)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:485) 
{code}
Still the view gets imported successfully, as data movement wasn't even 
necessary.

If we try to export a materialized view which is transactional, then this 
exception occurs:
{code:java}
org.apache.hadoop.hive.ql.parse.SemanticException: 
org.apache.hadoop.hive.ql.metadata.InvalidTableException: Table not found 
exim_materialized_view_da21d41a_9fe4_4446_9c72_d251496abf9d
 at 
org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyzeAcidExport(AcidExportSemanticAnalyzer.java:163)
 at 
org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyze(AcidExportSemanticAnalyzer.java:71)
 at 
org.apache.hadoop.hive.ql.parse.RewriteSemanticAnalyzer.analyzeInternal(RewriteSemanticAnalyzer.java:72)
 at 
org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:289)
 at org.apache.hadoop.hive.ql.Compiler.analyze(Compiler.java:220)
 at org.apache.hadoop.hive.ql.Compiler.compile(Compiler.java:104)
 at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:183)
 at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:601)
 at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:547)
 at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:541) 
{code}
So the export process can not handle it, as the temporary table is not getting 
created.

 

The import command handling have a lot of codes dedicated to importing views 
and materialized views, which suggests that we support the importing (and thus 
also suggests implicitly that we support the exporting) of views and 
materialiezed views.

 

So the conclusion is that we have to decide if we support exporting/importing 
of views and materialized views.

If we decide not to support them then:
 - export process should throw an exception if a view or materialized view is 
the subject
 - the codes specific to view imports should be removed

If we decide to support them, then:
 - the commands mentioned above above should be introduced
 - exception should be thrown if not the proper command used (e.g. export view 
on a table)
 - the exceptions mentioned above should be fixed

I prefer not to support them, I don't think we should support the exporting / 
importing of views. The point of exporting / importing is the transfer of data, 
not DDL, it causes more issues than it solves. Our current documentation also 
suggests that it is only supported for tables.

 

  was:
According to 
[https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport] 
import and export can be done by using the

 
{code:java}
export table ...
import table ... 
{code}
commands. The document doesn't mention views or materialized views at all, and 
in fact we don't support commands like

 
{code:java}
export view ...
import view ...
export materialized view ...
import materialized view ... 
{code}
they can not be parsed at all. The word table is often used though in a broader 
sense, when it means all table like entities, including views and materialized 
views. For example the various Table classes may represent 

[jira] [Commented] (HIVE-23098) Allow Operation assertState to Accept a Collection

2020-04-01 Thread David Mollitor (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23098?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072814#comment-17072814
 ] 

David Mollitor commented on HIVE-23098:
---

[~ngangam] Here is another trivial review if you have any spare cycles (lol).

> Allow Operation assertState to Accept a Collection
> --
>
> Key: HIVE-23098
> URL: https://issues.apache.org/jira/browse/HIVE-23098
> Project: Hive
>  Issue Type: Improvement
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Minor
> Attachments: HIVE-23098.1.patch, HIVE-23098.2.patch, 
> HIVE-23098.2.patch, HIVE-23098.2.patch, HIVE-23098.3.patch
>
>
> {code:java|title=Operation.java}
>   protected final void assertState(List states) throws 
> HiveSQLException {
> if (!states.contains(state)) {
>   throw new HiveSQLException("Expected states: " + states.toString() + ", 
> but found "
>   + this.state);
> }
> this.lastAccessTime = System.currentTimeMillis();
>   }
> /*/
> public void someMethod() {
> assertState(new 
> ArrayList(Arrays.asList(OperationState.FINISHED)));
> }
> {code}
> By allowing {{assertState}} to accept a {{Collection}}, one can save an 
> allocation and simplify the code:
> {code:java}
> assertState(Collections.singleton(OperationState.FINISHED));
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Resolved] (HIVE-22484) Remove Calls to printStackTrace

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22484?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor resolved HIVE-22484.
---
Fix Version/s: 4.0.0
   Resolution: Fixed

> Remove Calls to printStackTrace
> ---
>
> Key: HIVE-22484
> URL: https://issues.apache.org/jira/browse/HIVE-22484
> Project: Hive
>  Issue Type: Improvement
>Affects Versions: 3.2.0
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Major
> Fix For: 4.0.0
>
>
> In many cases, the call to {{printStackTrace}} bypasses the logging 
> framework, in other cases, the error stack trace is printed and the exception 
> is re-thrown (log-and-throw is a bad pattern), and then there are some other 
> edge cases.
> Remove this call and replace with calls to the logging framework or remove 
> altogether if exception is wrapped and re-thrown.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23110) Prevent NPE in ReExecDriver if the processing is aborted

2020-04-01 Thread Miklos Gergely (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23110?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072499#comment-17072499
 ] 

Miklos Gergely commented on HIVE-23110:
---

[~prasanth_j] do you have a log for 2) ?

Should we run failure hooks on abort? Right now we don't do that (explicitly 
not), and I thought it was intentional as if a query is aborted, it is not a 
failure. Should we run them?

> Prevent NPE in ReExecDriver if the processing is aborted
> 
>
> Key: HIVE-23110
> URL: https://issues.apache.org/jira/browse/HIVE-23110
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2
>Reporter: Miklos Gergely
>Assignee: Zoltan Haindrich
>Priority: Major
> Attachments: HIVE-23110.01.patch
>
>
> In case of abort the context would be null, and thus the planMapper can not 
> be obtained from it.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23110) Prevent NPE in ReExecDriver if the processing is aborted

2020-04-01 Thread Prasanth Jayachandran (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23110?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072509#comment-17072509
 ] 

Prasanth Jayachandran commented on HIVE-23110:
--

I have partial logs
{code:java}
hiveserver2 <14>1 2020-03-31T20:52:24.702Z 
hiveserver2-0.hiveserver2-service.compute-1585643974-lwrg.svc.cluster.local 
hiveserver2 1 6ba03ff1-251f-4878-81ea-1ba72d36c465 [mdc@18060 class="ql.Driver" 
level="INFO" operationLogLevel="EXECUTION" 
queryId="hive_20200331205007_6397e486-03a9-41ec-a56b-e0c4ff1ff26e" 
sessionId="94e0ab1a-e5ca-4237-9713-235b5dd2559a" 
thread="HiveServer2-Background-Pool: Thread-74"] Executing 
command(queryId=hive_20200331205007_6397e486-03a9-41ec-a56b-e0c4ff1ff26e) has 
been interrupted after 133.75 seconds
hiveserver2 <14>1 2020-03-31T20:52:24.702Z 
hiveserver2-0.hiveserver2-service.compute-1585643974-lwrg.svc.cluster.local 
hiveserver2 1 6ba03ff1-251f-4878-81ea-1ba72d36c465 [mdc@18060 class="ql.Driver" 
level="INFO" operationLogLevel="EXECUTION" 
queryId="hive_20200331205007_6397e486-03a9-41ec-a56b-e0c4ff1ff26e" 
sessionId="94e0ab1a-e5ca-4237-9713-235b5dd2559a" 
thread="HiveServer2-Background-Pool: Thread-74"] OK
hiveserver2 <15>1 2020-03-31T20:52:24.702Z 
hiveserver2-0.hiveserver2-service.compute-1585643974-lwrg.svc.cluster.local 
hiveserver2 1 6ba03ff1-251f-4878-81ea-1ba72d36c465 [mdc@18060 
class="log.PerfLogger" level="DEBUG" operationLogLevel="EXECUTION" 
queryId="hive_20200331205007_6397e486-03a9-41ec-a56b-e0c4ff1ff26e" 
sessionId="94e0ab1a-e5ca-4237-9713-235b5dd2559a" 
thread="HiveServer2-Background-Pool: Thread-74"] 
hiveserver2 <14>1 2020-03-31T20:52:24.711Z 
hiveserver2-0.hiveserver2-service.compute-1585643974-lwrg.svc.cluster.local 
hiveserver2 1 6ba03ff1-251f-4878-81ea-1ba72d36c465 [mdc@18060 
class="common.LogUtils" level="INFO" thread="HiveServer2-Background-Pool: 
Thread-74"] Unregistered logging context.
hiveserver2 <14>1 2020-03-31T20:52:24.702Z 
hiveserver2-0.hiveserver2-service.compute-1585643974-lwrg.svc.cluster.local 
hiveserver2 1 6ba03ff1-251f-4878-81ea-1ba72d36c465 [mdc@18060 
class="lockmgr.DbLockManager" level="INFO" operationLogLevel="EXECUTION" 
queryId="hive_20200331205007_6397e486-03a9-41ec-a56b-e0c4ff1ff26e" 
sessionId="94e0ab1a-e5ca-4237-9713-235b5dd2559a" 
thread="HiveServer2-Background-Pool: Thread-74"] releaseLocks: 
hiveserver2 <15>1 2020-03-31T20:52:24.703Z 
hiveserver2-0.hiveserver2-service.compute-1585643974-lwrg.svc.cluster.local 
hiveserver2 1 6ba03ff1-251f-4878-81ea-1ba72d36c465 [mdc@18060 
class="log.PerfLogger" level="DEBUG" operationLogLevel="EXECUTION" 
queryId="hive_20200331205007_6397e486-03a9-41ec-a56b-e0c4ff1ff26e" 
sessionId="94e0ab1a-e5ca-4237-9713-235b5dd2559a" 
thread="HiveServer2-Background-Pool: Thread-74"] 
hiveserver2 <11>1 2020-03-31T20:52:24.711Z 
hiveserver2-0.hiveserver2-service.compute-1585643974-lwrg.svc.cluster.local 
hiveserver2 1 6ba03ff1-251f-4878-81ea-1ba72d36c465 [mdc@18060 
class="operation.Operation" level="ERROR" operationLogLevel="EXECUTION" 
queryId="hive_20200331205007_6397e486-03a9-41ec-a56b-e0c4ff1ff26e" 
sessionId="94e0ab1a-e5ca-4237-9713-235b5dd2559a" 
thread="HiveServer2-Background-Pool: Thread-74"] Error running hive query: 
org.apache.hive.service.cli.HiveSQLException: Illegal Operation state 
transition from CANCELED to FINISHED
at 
org.apache.hive.service.cli.OperationState.validateTransition(OperationState.java:97)
at 
org.apache.hive.service.cli.OperationState.validateTransition(OperationState.java:103)
at 
org.apache.hive.service.cli.operation.Operation.setState(Operation.java:161)
at 
org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:248)
at 
org.apache.hive.service.cli.operation.SQLOperation.access$700(SQLOperation.java:87)
at 
org.apache.hive.service.cli.operation.SQLOperation$BackgroundWork$1.run(SQLOperation.java:322)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1876)
at 
org.apache.hive.service.cli.operation.SQLOperation$BackgroundWork.run(SQLOperation.java:340)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
hiveserver2 2020-03-31 20:52:24,710 
Log4j2-TF-1-AsyncLogger[AsyncContext@18b4aac2]-1 ERROR 
/tmp/hive/operation_logs/94e0ab1a-e5ca-4237-9713-235b5dd2559a/hive_20200331205007_6397e486-03a9-41ec-a56b-e0c4ff1ff26e
 was closed 

[jira] [Commented] (HIVE-21492) VectorizedParquetRecordReader can't to read parquet file generated using thrift/custom tool

2020-04-01 Thread Ganesha Shreedhara (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-21492?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072517#comment-17072517
 ] 

Ganesha Shreedhara commented on HIVE-21492:
---

[~Ferd] Please review the latest patch.

> VectorizedParquetRecordReader can't to read parquet file generated using 
> thrift/custom tool
> ---
>
> Key: HIVE-21492
> URL: https://issues.apache.org/jira/browse/HIVE-21492
> Project: Hive
>  Issue Type: Bug
>Reporter: Ganesha Shreedhara
>Assignee: Ganesha Shreedhara
>Priority: Major
> Attachments: HIVE-21492.2.patch, HIVE-21492.3.patch, HIVE-21492.patch
>
>
> Taking an example of a parquet table having array of integers as below. 
> {code:java}
> CREATE EXTERNAL TABLE ( list_of_ints` array)
> STORED AS PARQUET 
> LOCATION '{location}';
> {code}
> Parquet file generated using hive will have schema for Type as below:
> {code:java}
> group list_of_ints (LIST) { repeated group bag { optional int32 array;\n};\n} 
> {code}
> Parquet file generated using thrift or any custom tool (using 
> org.apache.parquet.io.api.RecordConsumer)
> may have schema for Type as below:
> {code:java}
> required group list_of_ints (LIST) { repeated int32 list_of_tuple} {code}
> VectorizedParquetRecordReader handles only parquet file generated using hive. 
> It throws the following exception when parquet file generated using thrift is 
> read because of the changes done as part of HIVE-18553 .
> {code:java}
> Caused by: java.lang.ClassCastException: repeated int32 list_of_ints_tuple is 
> not a group
>  at org.apache.parquet.schema.Type.asGroupType(Type.java:207)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.getElementType(VectorizedParquetRecordReader.java:479)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.buildVectorizedParquetReader(VectorizedParquetRecordReader.java:532)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.checkEndOfRowGroup(VectorizedParquetRecordReader.java:440)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.nextBatch(VectorizedParquetRecordReader.java:401)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:353)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:92)
>  at 
> org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.doNext(HiveContextAwareRecordReader.java:365){code}
>  
>  I have done a small change to handle the case where the child type of group 
> type can be PrimitiveType.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23067) Use batch DB calls in TxnHandler for commitTxn and abortTxns

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23067?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072529#comment-17072529
 ] 

Hive QA commented on HIVE-23067:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12998333/HIVE-23067.2.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 18163 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[topnkey_grouping_sets] 
(batchId=1)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/21366/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/21366/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-21366/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12998333 - PreCommit-HIVE-Build

> Use batch DB calls in TxnHandler for commitTxn and abortTxns
> 
>
> Key: HIVE-23067
> URL: https://issues.apache.org/jira/browse/HIVE-23067
> Project: Hive
>  Issue Type: Improvement
>Reporter: Marton Bod
>Assignee: Marton Bod
>Priority: Major
> Attachments: HIVE-23067.1.patch, HIVE-23067.1.patch, 
> HIVE-23067.1.patch, HIVE-23067.1.patch, HIVE-23067.2.patch
>
>
> To reduce the number of database calls and network roundtrips, we could use 
> more batching in TxnHandler, where currently in many places we call insert 
> commands in loops sequentially.
> Some examples:
>  * openTxns (TXNS, REPL_TXN_MAP)
>  * commitTxn (COMPLETED_TXN_COMPONENTS)
>  * replTableWriteIdState (TXN_TO_WRITE_ID)
>  * allocateTableWriteIds (TXN_TO_WRITE_ID)
>  * 
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23110) Prevent NPE in ReExecDriver if the processing is aborted

2020-04-01 Thread Zoltan Haindrich (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23110?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072531#comment-17072531
 ] 

Zoltan Haindrich commented on HIVE-23110:
-

I think this issue is not connect to the NPE from ReExecDriver; because if that 
would be there the catch would have done a [return 
here|https://github.com/apache/hive/blob/d2ad5b061706a1d3cd55e59c769ed4f2af01cdbe/service/src/java/org/apache/hive/service/cli/operation/SQLOperation.java#L240]

I don't think we have guarantees that after cancel an exception like that may 
not happen:
* cleanup is synch method (not that it matters)
* it sets state as first step
* and then it start calling driver with .close and .destroy
* as a result of that internally the driver changes to some aborted state
* however..if the job is already near completion it an exception may not 
happen; and when the actual thread (Thread-74 in the logs) starts getting out - 
it doesn't throw an exception; and that is not in line with the SQLOperation's 
expectations; which causes the illegal transition


[SQLOperation.cleanup|https://github.com/apache/hive/blob/d2ad5b061706a1d3cd55e59c769ed4f2af01cdbe/service/src/java/org/apache/hive/service/cli/operation/SQLOperation.java#L396]
 

 



> Prevent NPE in ReExecDriver if the processing is aborted
> 
>
> Key: HIVE-23110
> URL: https://issues.apache.org/jira/browse/HIVE-23110
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2
>Reporter: Miklos Gergely
>Assignee: Zoltan Haindrich
>Priority: Major
> Attachments: HIVE-23110.01.patch
>
>
> In case of abort the context would be null, and thus the planMapper can not 
> be obtained from it.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-5312) Let HiveServer2 run simultaneously in HTTP (over thrift) and Binary (normal thrift transport) mode

2020-04-01 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-5312?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ádám Szita updated HIVE-5312:
-
Attachment: HIVE-5312.3.patch

> Let HiveServer2 run simultaneously in HTTP (over thrift) and Binary (normal 
> thrift transport) mode 
> ---
>
> Key: HIVE-5312
> URL: https://issues.apache.org/jira/browse/HIVE-5312
> Project: Hive
>  Issue Type: Sub-task
>  Components: HiveServer2
>Reporter: Vaibhav Gumashta
>Assignee: Ádám Szita
>Priority: Major
> Attachments: HIVE-5132.branch-2.3.patch, HIVE-5312.1.patch, 
> HIVE-5312.2.patch, HIVE-5312.3.patch
>
>
> [HIVE-4763|https://issues.apache.org/jira/browse/HIVE-4763] adds support for 
> HTTP transport over thrift. With that, HS2 can be configured to run either 
> using using HTTP or using normal thrift binary transport. Ideally HS2 should 
> be supporting both modes simultaneously and the client should be able to 
> specify the mode used in serving the request.  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23122) LLAP TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor updated HIVE-23122:

Status: Patch Available  (was: Open)

> LLAP TaskExecutorService should log details about task eviction decision 
> details
> 
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-23122.01.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23107) Remove MIN_HISTORY_LEVEL table

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23107?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072736#comment-17072736
 ] 

Hive QA commented on HIVE-23107:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
43s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
45s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
27s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
 2s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  1m  
9s{color} | {color:blue} standalone-metastore/metastore-server in master has 
190 extant Findbugs warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m 
40s{color} | {color:blue} ql in master has 1529 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
12s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
27s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
57s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
26s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m 
26s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
23s{color} | {color:red} standalone-metastore/metastore-server: The patch 
generated 4 new + 669 unchanged - 2 fixed = 673 total (was 671) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
22s{color} | {color:green} standalone-metastore/metastore-server generated 0 
new + 187 unchanged - 3 fixed = 187 total (was 190) {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
46s{color} | {color:green} ql in the patch passed. {color} |
| {color:red}-1{color} | {color:red} javadoc {color} | {color:red}  0m 
20s{color} | {color:red} standalone-metastore_metastore-server generated 1 new 
+ 24 unchanged - 0 fixed = 25 total (was 24) {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 30m 10s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-21369/dev-support/hive-personality.sh
 |
| git revision | master / 94f90bc |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21369/yetus/diff-checkstyle-standalone-metastore_metastore-server.txt
 |
| javadoc | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21369/yetus/diff-javadoc-javadoc-standalone-metastore_metastore-server.txt
 |
| modules | C: standalone-metastore/metastore-server ql U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21369/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Remove MIN_HISTORY_LEVEL table
> --
>
> Key: HIVE-23107
> URL: https://issues.apache.org/jira/browse/HIVE-23107
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Reporter: László Pintér
>Assignee: László Pintér
>Priority: Major
> Attachments: HIVE-23107.01.patch, HIVE-23107.02.patch
>
>
> MIN_HISTORY_LEVEL table is used in two places:
>  * Cleaner uses it to decide if the files can 

[jira] [Updated] (HIVE-21304) Show Bucketing version for ReduceSinkOp in explain extended plan

2020-04-01 Thread Zoltan Haindrich (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21304?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltan Haindrich updated HIVE-21304:

Attachment: HIVE-21304.20.patch

> Show Bucketing version for ReduceSinkOp in explain extended plan
> 
>
> Key: HIVE-21304
> URL: https://issues.apache.org/jira/browse/HIVE-21304
> Project: Hive
>  Issue Type: Bug
>Reporter: Deepak Jaiswal
>Assignee: Zoltan Haindrich
>Priority: Major
> Attachments: HIVE-21304.01.patch, HIVE-21304.02.patch, 
> HIVE-21304.03.patch, HIVE-21304.04.patch, HIVE-21304.05.patch, 
> HIVE-21304.06.patch, HIVE-21304.07.patch, HIVE-21304.08.patch, 
> HIVE-21304.09.patch, HIVE-21304.10.patch, HIVE-21304.11.patch, 
> HIVE-21304.12.patch, HIVE-21304.13.patch, HIVE-21304.14.patch, 
> HIVE-21304.15.patch, HIVE-21304.16.patch, HIVE-21304.17.patch, 
> HIVE-21304.18.patch, HIVE-21304.19.patch, HIVE-21304.20.patch
>
>
> Show Bucketing version for ReduceSinkOp in explain extended plan.
> This helps identify what hashing algorithm is being used by by ReduceSinkOp.
>  
> cc [~vgarg]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23122) LLAP TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor updated HIVE-23122:

Attachment: HIVE-23122.01.patch

> LLAP TaskExecutorService should log details about task eviction decision 
> details
> 
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-23122.01.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23079) Remove Calls to printStackTrace in Module hive-serde

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23079?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor updated HIVE-23079:
--
Fix Version/s: 4.0.0
   Resolution: Fixed
   Status: Resolved  (was: Patch Available)

Pushed to master.  Thanks [~pvary] for the review!

> Remove Calls to printStackTrace in Module hive-serde
> 
>
> Key: HIVE-23079
> URL: https://issues.apache.org/jira/browse/HIVE-23079
> Project: Hive
>  Issue Type: Sub-task
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-23079.1.patch, HIVE-23079.1.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23123) Export/Import of views and materialized views

2020-04-01 Thread Miklos Gergely (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23123?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Miklos Gergely updated HIVE-23123:
--
Description: 
According to 
[https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport] 
import and export can be done by using the

 
{code:java}
export table ...
import table ... 
{code}
commands. The document doesn't mention views or materialized views at all, and 
in fact we don't support commands like

 
{code:java}
export view ...
import view ...
export materialized view ...
import materialized view ... 
{code}
they can not be parsed at all. The word table is often used though in a broader 
sense, when it means all table like entities, including views and materialized 
views. For example the various Table classes may represent any of these as well.

If I try to export a view with the export table ... command, it goes fine. A 
_metadata file will be created, but no data directory, which is what we'd 
expect. If I try to import it back, an exception is thrown due to the lack of 
the data dir:

 
{code:java}
java.lang.AssertionError: null==getPath() for exim_view
 at org.apache.hadoop.hive.ql.metadata.Hive.loadTable(Hive.java:3088)
 at org.apache.hadoop.hive.ql.exec.MoveTask.execute(MoveTask.java:419)
 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:213)
 at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:105)
 at org.apache.hadoop.hive.ql.Executor.launchTask(Executor.java:364)
 at org.apache.hadoop.hive.ql.Executor.launchTasks(Executor.java:335)
 at org.apache.hadoop.hive.ql.Executor.runTasks(Executor.java:246)
 at org.apache.hadoop.hive.ql.Executor.execute(Executor.java:109)
 at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:722)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:491)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:485) 
{code}
Still the view gets imported successfully, as data movement wasn't even 
necessary.

If we try to export a materialized view which is transactional, then this 
exception occurs:

 
{code:java}
org.apache.hadoop.hive.ql.parse.SemanticException: 
org.apache.hadoop.hive.ql.metadata.InvalidTableException: Table not found 
exim_materialized_view_da21d41a_9fe4_4446_9c72_d251496abf9d
 at 
org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyzeAcidExport(AcidExportSemanticAnalyzer.java:163)
 at 
org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyze(AcidExportSemanticAnalyzer.java:71)
 at 
org.apache.hadoop.hive.ql.parse.RewriteSemanticAnalyzer.analyzeInternal(RewriteSemanticAnalyzer.java:72)
 at 
org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:289)
 at org.apache.hadoop.hive.ql.Compiler.analyze(Compiler.java:220)
 at org.apache.hadoop.hive.ql.Compiler.compile(Compiler.java:104)
 at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:183)
 at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:601)
 at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:547)
 at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:541) 
{code}
So the export process can not handle it, as the temporary table is not getting 
created.

 

The import command handling have a lot of codes dedicated to importing views 
and materialized views, which suggests that we support the importing (and thus 
also suggests implicitly that we support the exporting) of views and 
materialiezed views.

 

So the conclusion is that we have to decide if we support exporting/importing 
of views and materialized views.

If we decide not to support them then:
 - export process should throw an exception if a view or materialized view is 
the subject
 - the codes specific to view imports should be removed

If we decide to support them, then:
 - the commands mentioned above above should be introduced
 - exception should be thrown if not the proper command used (e.g. export view 
on a table)
 - the exceptions mentioned above should be fixed

I prefer not to support them, I don't think we should support the exporting / 
importing of views. The point of exporting / importing is the transfer of data, 
not DDL, it causes more issues than it solves. Our current documentation also 
suggests that it is only supported for tables.

 

  was:
According to 
[https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport] 
import and export can be done by using the

 
{code:java}
export table ...
import table ... 
{code}
commands. The document doesn't mention views or materialized views at all, and 
in fact we don't support commands like

 
{code:java}
export view ...
import view ...
export materialized view ...
import materialized view ... 
{code}
they can not be parsed at all. The word table is often used though in a broader 
sense, when it means all table like entities, including views and materialized 
views. For example the various Table classes 

[jira] [Commented] (HIVE-23101) Fix topnkey_grouping_sets

2020-04-01 Thread Peter Vary (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23101?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072549#comment-17072549
 ] 

Peter Vary commented on HIVE-23101:
---

Pushed the patch 1 to master to unblock PreCommits.


> Fix topnkey_grouping_sets
> -
>
> Key: HIVE-23101
> URL: https://issues.apache.org/jira/browse/HIVE-23101
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Krisztian Kasa
>Assignee: Krisztian Kasa
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-23101.1.patch, HIVE-23101.2.patch
>
>
> Example:
> {code}
> Client Execution succeeded but contained differences (error code = 1) after 
> executing topnkey_grouping_sets.q 
> 260c260
> < NULL NULL
> ---
> > NULL 1
> {code}
>  
> Test *topnkey_grouping_sets* fails intermittently.
> Queries which project 2 columns but order by only one of them can have more 
> than one good result set:
> {code:java}
> CREATE TABLE t_test_grouping_sets(
>   a int,
>   b int,
>   c int
> );
> INSERT INTO t_test_grouping_sets VALUES
> (NULL, NULL, NULL),
> (5, 2, 3),
> (10, 11, 12),
> (NULL, NULL, NULL),
> (NULL, NULL, NULL),
> (6, 2, 1),
> (7, 8, 4), (7, 8, 4), (7, 8, 4),
> (5, 1, 2), (5, 1, 2), (5, 1, 2),
> (NULL, NULL, NULL);
> SELECT a, b FROM t_test_grouping_sets GROUP BY GROUPING SETS ((a, b), (a), 
> (b), ()) ORDER BY a LIMIT 10;
> {code}
> {code:java}
> 5 NULL
> 5 2
> 5 1
> 6 2
> 6 NULL
> 7 8
> 7 NULL
> 10NULL
> 1011
> NULL  1
> {code}
> {code:java}
> 5 NULL
> 5 2
> 5 1
> 6 2
> 6 NULL
> 7 8
> 7 NULL
> 10NULL
> 1011
> NULL  NULL
> {code}
> Since we don't order by *b* both result sets are valid.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23067) Use batch DB calls in TxnHandler for commitTxn and abortTxns

2020-04-01 Thread Marton Bod (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23067?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Marton Bod updated HIVE-23067:
--
Attachment: HIVE-23067.2.patch

> Use batch DB calls in TxnHandler for commitTxn and abortTxns
> 
>
> Key: HIVE-23067
> URL: https://issues.apache.org/jira/browse/HIVE-23067
> Project: Hive
>  Issue Type: Improvement
>Reporter: Marton Bod
>Assignee: Marton Bod
>Priority: Major
> Attachments: HIVE-23067.1.patch, HIVE-23067.1.patch, 
> HIVE-23067.1.patch, HIVE-23067.1.patch, HIVE-23067.2.patch, HIVE-23067.2.patch
>
>
> To reduce the number of database calls and network roundtrips, we could use 
> more batching in TxnHandler, where currently in many places we call insert 
> commands in loops sequentially.
> Some examples:
>  * openTxns (TXNS, REPL_TXN_MAP)
>  * commitTxn (COMPLETED_TXN_COMPONENTS)
>  * replTableWriteIdState (TXN_TO_WRITE_ID)
>  * allocateTableWriteIds (TXN_TO_WRITE_ID)
>  * 
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23113) Clean Up HiveCallableStatement

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23113?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072699#comment-17072699
 ] 

Hive QA commented on HIVE-23113:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12998347/HIVE-23113.1.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 18162 tests 
executed
*Failed tests:*
{noformat}
org.apache.hive.beeline.TestBeeLineWithArgs.testRowsAffected (batchId=286)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/21368/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/21368/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-21368/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12998347 - PreCommit-HIVE-Build

> Clean Up HiveCallableStatement
> --
>
> Key: HIVE-23113
> URL: https://issues.apache.org/jira/browse/HIVE-23113
> Project: Hive
>  Issue Type: Improvement
>  Components: JDBC
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Trivial
> Attachments: HIVE-23113.1.patch
>
>
> * Add a useful class comment
>  * Remove all non-javadoc comments
>  * Remove 'TODO' tags
>  * Add {{@override}} tags
>  * Checkstyle formatting



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23039) Checkpointing for repl dump bootstrap phase

2020-04-01 Thread Aasha Medhi (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23039?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aasha Medhi updated HIVE-23039:
---
Attachment: HIVE-23039.02.patch
Status: Patch Available  (was: In Progress)

> Checkpointing for repl dump bootstrap phase
> ---
>
> Key: HIVE-23039
> URL: https://issues.apache.org/jira/browse/HIVE-23039
> Project: Hive
>  Issue Type: Bug
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23039.01.patch, HIVE-23039.02.patch
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23122) LLAP TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor updated HIVE-23122:

Description: 
TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
another. Under the hood, the queue uses a configurable 
[comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
 The currently available comparators typically use the following properties of 
a task(wrapper):
getWithinDagPriority: related to vertex
currentAttemptStartTime
firstAttemptStartTime
knownPending: remaining upstream tasks

The problem is, when an eviction happens, the INFO level message doesn't 
provide any insight about the decision, only attempts ids like below:
{code}
attempt_1585248378306_0010_72_02_96_8 evicted from wait queue in favor of 
attempt_1585248378306_0003_175_02_79_175 because of lower priority
{code}



  was:
TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
another. Under the hood, the queue uses a configurable 
[comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
 The currently available comparators typically use the following properties of 
a task:
getWithinDagPriority: related to vertex



> LLAP TaskExecutorService should log details about task eviction decision 
> details
> 
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-23122.01.patch
>
>
> TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
> another. Under the hood, the queue uses a configurable 
> [comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
>  The currently available comparators typically use the following properties 
> of a task(wrapper):
> getWithinDagPriority: related to vertex
> currentAttemptStartTime
> firstAttemptStartTime
> knownPending: remaining upstream tasks
> The problem is, when an eviction happens, the INFO level message doesn't 
> provide any insight about the decision, only attempts ids like below:
> {code}
> attempt_1585248378306_0010_72_02_96_8 evicted from wait queue in favor of 
> attempt_1585248378306_0003_175_02_79_175 because of lower priority
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23052) Optimize lock enqueueing in TxnHandler

2020-04-01 Thread Marton Bod (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23052?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Marton Bod updated HIVE-23052:
--
Attachment: HIVE-23052.14.patch

> Optimize lock enqueueing in TxnHandler
> --
>
> Key: HIVE-23052
> URL: https://issues.apache.org/jira/browse/HIVE-23052
> Project: Hive
>  Issue Type: Improvement
>Reporter: Marton Bod
>Assignee: Marton Bod
>Priority: Major
> Attachments: HIVE-23052.1.patch, HIVE-23052.10.patch, 
> HIVE-23052.11.patch, HIVE-23052.12.patch, HIVE-23052.13.patch, 
> HIVE-23052.14.patch, HIVE-23052.2.patch, HIVE-23052.3.patch, 
> HIVE-23052.4.patch, HIVE-23052.5.patch, HIVE-23052.6.patch, 
> HIVE-23052.7.patch, HIVE-23052.8.patch, HIVE-23052.8.patch, 
> HIVE-23052.8.patch, HIVE-23052.9.patch
>
>
> * Reduce scope of next_lock_id select-for-update by moving the txn_component 
> inserts before the S4U + inserting the hive_locks entries before the S4U 
> (first with a temp ID, which will be replaced later in a single update). This 
> helps decrease the overall time that the next_lock_id table is locked, 
> thereby increasing concurrency
>  * Insert txn_components in a batch instead of one-by-one (also in 
> TxnHandler::addDynamicPartition)
>  * Increment next_lock_id and update hive_locks table in a single batch 
> statement
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23039) Checkpointing for repl dump bootstrap phase

2020-04-01 Thread Aasha Medhi (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23039?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aasha Medhi updated HIVE-23039:
---
Status: In Progress  (was: Patch Available)

> Checkpointing for repl dump bootstrap phase
> ---
>
> Key: HIVE-23039
> URL: https://issues.apache.org/jira/browse/HIVE-23039
> Project: Hive
>  Issue Type: Bug
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23039.01.patch, HIVE-23039.02.patch
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23113) Clean Up HiveCallableStatement

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23113?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor updated HIVE-23113:
--
Attachment: HIVE-23113.1.patch

> Clean Up HiveCallableStatement
> --
>
> Key: HIVE-23113
> URL: https://issues.apache.org/jira/browse/HIVE-23113
> Project: Hive
>  Issue Type: Improvement
>  Components: JDBC
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Trivial
> Attachments: HIVE-23113.1.patch, HIVE-23113.1.patch
>
>
> * Add a useful class comment
>  * Remove all non-javadoc comments
>  * Remove 'TODO' tags
>  * Add {{@override}} tags
>  * Checkstyle formatting



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23123) Export/Import of views and materialized views

2020-04-01 Thread Miklos Gergely (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23123?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Miklos Gergely updated HIVE-23123:
--
Description: 
According to 
[https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport] 
import and export can be done by using the

 
{code:java}
export table ...
import table ... 
{code}
commands. The document doesn't mention views or materialized views at all, and 
in fact we don't support commands like

 
{code:java}
export view ...
import view ...
export materialized view ...
import materialized view ... 
{code}
they can not be parsed at all. The word table is often used though in a broader 
sense, when it means all table like entities, including views and materialized 
views. For example the various Table classes may represent any of these as well.

If I try to export a view with the export table ... command, it goes fine. A 
_metadata file will be created, but no data directory, which is what we'd 
expect. If I try to import it back, an exception is thrown due to the lack of 
the data dir:

 
{code:java}
java.lang.AssertionError: null==getPath() for exim_view
 at org.apache.hadoop.hive.ql.metadata.Hive.loadTable(Hive.java:3088)
 at org.apache.hadoop.hive.ql.exec.MoveTask.execute(MoveTask.java:419)
 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:213)
 at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:105)
 at org.apache.hadoop.hive.ql.Executor.launchTask(Executor.java:364)
 at org.apache.hadoop.hive.ql.Executor.launchTasks(Executor.java:335)
 at org.apache.hadoop.hive.ql.Executor.runTasks(Executor.java:246)
 at org.apache.hadoop.hive.ql.Executor.execute(Executor.java:109)
 at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:722)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:491)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:485) 
{code}
Still the view gets imported successfully, as data movement wasn't even 
necessary.

If we try to export a materialized view which is transactional, then this 
exception occurs:

 
{code:java}
org.apache.hadoop.hive.ql.parse.SemanticException: 
org.apache.hadoop.hive.ql.metadata.InvalidTableException: Table not found 
exim_materialized_view_da21d41a_9fe4_4446_9c72_d251496abf9d
 at 
org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyzeAcidExport(AcidExportSemanticAnalyzer.java:163)
 at 
org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyze(AcidExportSemanticAnalyzer.java:71)
 at 
org.apache.hadoop.hive.ql.parse.RewriteSemanticAnalyzer.analyzeInternal(RewriteSemanticAnalyzer.java:72)
 at 
org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:289)
 at org.apache.hadoop.hive.ql.Compiler.analyze(Compiler.java:220)
 at org.apache.hadoop.hive.ql.Compiler.compile(Compiler.java:104)
 at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:183)
 at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:601)
 at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:547)
 at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:541) 
{code}
So the export process can not handle it, as the temporary table is not getting 
created.

 

The import command handling have a lot of codes dedicated to importing views 
and materialized views, which suggests that we support the importing (and thus 
also suggests implicitly that we support the exporting) of views and 
materialiezed views.

 

So the conclusion is that we have to decide if we support exporting/importing 
of views and materialized views.

If we decide not to support them then:
 - export process should throw an exception if a view or materialized view is 
the subject
 - the codes specific to view imports should be removed

If we decide to support them, then:
 - the commands mentioned above above should be introduced
 - exception should be thrown if not the proper command used (e.g. export view 
on a table)
 - the exceptions mentioned above should be fixed

I prefer #1, I don't think we should support the exporting / importing of 
views. The point of exporting / importing is the transfer of data, not DDL, it 
causes more issues than it solves. Our current documentation also suggests that 
it is only supported for tables.

 

  was:
According to 
[https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport] 
import and export can be done by using the

 
{code:java}
export table ...
import table ... 
{code}
commands. The document doesn't mention views or materialized views at all, and 
in fact we don't support commands like

 
{code:java}
export view ...
import view ...
export materialized view ...
import materialized view ... 
{code}
they can not be parsed at all. The word table is often used though in a broader 
sense, when it means all table like entities, including views and materialized 
views. For example the various Table classes may represent any 

[jira] [Updated] (HIVE-23123) Export/Import of views and materialized views

2020-04-01 Thread Miklos Gergely (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23123?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Miklos Gergely updated HIVE-23123:
--
Description: 
According to 
[https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport] 
import and export can be done by using the

 
{code:java}
export table ...
import table ... 
{code}
commands. The document doesn't mention views or materialized views at all, and 
in fact we don't support commands like

 
{code:java}
export view ...
import view ...
export materialized view ...
import materialized view ... 
{code}
they can not be parsed at all. The word table is often used though in a broader 
sense, when it means all table like entities, including views and materialized 
views. For example the various Table classes may represent any of these as well.

If I try to export a view with the export table ... command, it goes fine. A 
_metadata file will be created, but no data directory, which is what we'd 
expect. If I try to import it back, an exception is thrown due to the lack of 
the data dir:

 
{code:java}
java.lang.AssertionError: null==getPath() for exim_view
 at org.apache.hadoop.hive.ql.metadata.Hive.loadTable(Hive.java:3088)
 at org.apache.hadoop.hive.ql.exec.MoveTask.execute(MoveTask.java:419)
 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:213)
 at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:105)
 at org.apache.hadoop.hive.ql.Executor.launchTask(Executor.java:364)
 at org.apache.hadoop.hive.ql.Executor.launchTasks(Executor.java:335)
 at org.apache.hadoop.hive.ql.Executor.runTasks(Executor.java:246)
 at org.apache.hadoop.hive.ql.Executor.execute(Executor.java:109)
 at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:722)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:491)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:485) 
{code}
Still the view gets imported successfully, as data movement wasn't even 
necessary.

If we try to export a materialized view which is transactional, then this 
exception occurs:

 
{code:java}
org.apache.hadoop.hive.ql.parse.SemanticException: 
org.apache.hadoop.hive.ql.metadata.InvalidTableException: Table not found 
exim_materialized_view_da21d41a_9fe4_4446_9c72_d251496abf9d
 at 
org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyzeAcidExport(AcidExportSemanticAnalyzer.java:163)
 at 
org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyze(AcidExportSemanticAnalyzer.java:71)
 at 
org.apache.hadoop.hive.ql.parse.RewriteSemanticAnalyzer.analyzeInternal(RewriteSemanticAnalyzer.java:72)
 at 
org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:289)
 at org.apache.hadoop.hive.ql.Compiler.analyze(Compiler.java:220)
 at org.apache.hadoop.hive.ql.Compiler.compile(Compiler.java:104)
 at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:183)
 at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:601)
 at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:547)
 at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:541) 
{code}
So the export process can not handle it, as the temporary table is not getting 
created.

 

The import command handling have a lot of codes dedicated to importing views 
and materialized views, which suggests that we support the importing (and thus 
also suggests implicitly that we support the exporting) of views and 
materialiezed views.

 

So the conclusion is that we have to decide if we support exporting/importing 
of views and materialized views.
 #  If we decide not to support them then:
 # 

 - export process should throw an exception if a view or materialized view is 
the subject
 - the codes specific to view imports should be removed

 # If we decide to support them, then:

 - the commands mentioned above above should be introduced
 - exception should be thrown if not the proper command used (e.g. export view 
on a table)
 - the exceptions mentioned above should be fixed

I prefer #1, I don't think we should support the exporting / importing of 
views. The point of exporting / importing is the transfer of data, not DDL, it 
causes more issues than it solves. Our current documentation also suggests that 
it is only supported for tables.

 

  was:
According to 
https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport 
import and export can be done by using the

 
{code:java}
export table ...
import table ... 
{code}
commands. The document doesn't mention views or materialized views at all, and 
in fact we don't support commands like

 
{code:java}
export view ...
import view ...
export materialized view ...
import materialized view ... 
{code}
they can not be parsed at all. The word table is often used though in a broader 
sense, when it means all table like entities, including views and materialized 
views. For example the various Table classes may 

[jira] [Work logged] (HIVE-22821) Add necessary endpoints for proactive cache eviction

2020-04-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-22821?focusedWorklogId=413925=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-413925
 ]

ASF GitHub Bot logged work on HIVE-22821:
-

Author: ASF GitHub Bot
Created on: 01/Apr/20 09:30
Start Date: 01/Apr/20 09:30
Worklog Time Spent: 10m 
  Work Description: szlta commented on pull request #909: HIVE-22821
URL: https://github.com/apache/hive/pull/909#discussion_r401479045
 
 

 ##
 File path: ql/src/java/org/apache/hadoop/hive/llap/ProactiveEviction.java
 ##
 @@ -0,0 +1,311 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.hadoop.hive.llap;
+
+import java.io.IOException;
+import java.util.Collection;
+import java.util.HashMap;
+import java.util.HashSet;
+import java.util.LinkedHashMap;
+import java.util.LinkedHashSet;
+import java.util.LinkedList;
+import java.util.List;
+import java.util.Map;
+import java.util.Set;
+import java.util.concurrent.ExecutorService;
+import java.util.concurrent.Executors;
+import java.util.concurrent.TimeUnit;
+import javax.net.SocketFactory;
+
+import org.apache.hadoop.conf.Configuration;
+import org.apache.hadoop.hive.common.io.CacheTag;
+import org.apache.hadoop.hive.conf.HiveConf;
+import org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos;
+import org.apache.hadoop.hive.llap.impl.LlapManagementProtocolClientImpl;
+import org.apache.hadoop.hive.llap.registry.LlapServiceInstance;
+import org.apache.hadoop.hive.llap.registry.impl.LlapRegistryService;
+import org.apache.hadoop.io.retry.RetryPolicies;
+import org.apache.hadoop.io.retry.RetryPolicy;
+import org.apache.hadoop.net.NetUtils;
+
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+/**
+ * Through this class the caller (typically HS2) can request eviction of 
buffers from LLAP cache by specifying a DB,
+ * table or partition name/(value). Request sending is implemented here.
+ */
+public final class ProactiveEviction {
+
+  private ProactiveEviction() {
+// Not to be used;
+  }
+
+  /**
+   * Trigger LLAP cache eviction of buffers related to entities residing in 
request parameter.
+   * @param conf
+   * @param request
+   */
+  public static void evict(Configuration conf, Request request) {
+if (!HiveConf.getBoolVar(conf, 
HiveConf.ConfVars.LLAP_IO_PROACTIVE_EVICTION_ENABLED)) {
+  return;
+}
+
+try {
+  LlapRegistryService llapRegistryService = 
LlapRegistryService.getClient(conf);
+  Collection instances = 
llapRegistryService.getInstances().getAll();
+  if (instances.size() == 0) {
+// Not in LLAP mode.
+return;
+  }
+  ExecutorService executorService = Executors.newCachedThreadPool();
 
 Review comment:
   Hi @b-slim can you take a look on the recent commit please?
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 413925)
Time Spent: 5h  (was: 4h 50m)

> Add necessary endpoints for proactive cache eviction
> 
>
> Key: HIVE-22821
> URL: https://issues.apache.org/jira/browse/HIVE-22821
> Project: Hive
>  Issue Type: Sub-task
>  Components: llap
>Reporter: Ádám Szita
>Assignee: Ádám Szita
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-22821.0.patch, HIVE-22821.1.patch, 
> HIVE-22821.2.patch
>
>  Time Spent: 5h
>  Remaining Estimate: 0h
>
> Implement the parts required for iHS2 -> LLAP daemons communication:
>  * protobuf message schema and endpoints
>  * Hive configuration
>  * for use cases:
>  ** dropping db
>  ** dropping table
>  ** dropping partition from a table



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23076) Add batching for openTxn

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23076?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072575#comment-17072575
 ] 

Hive QA commented on HIVE-23076:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
59s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
26s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
23s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  1m 
14s{color} | {color:blue} standalone-metastore/metastore-server in master has 
190 extant Findbugs warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
20s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
32s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
27s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
27s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
22s{color} | {color:red} standalone-metastore/metastore-server: The patch 
generated 1 new + 546 unchanged - 0 fixed = 547 total (was 546) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
24s{color} | {color:green} standalone-metastore/metastore-server generated 0 
new + 189 unchanged - 1 fixed = 189 total (was 190) {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
20s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 15m  8s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-21367/dev-support/hive-personality.sh
 |
| git revision | master / 94f90bc |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21367/yetus/diff-checkstyle-standalone-metastore_metastore-server.txt
 |
| modules | C: standalone-metastore/metastore-server U: 
standalone-metastore/metastore-server |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21367/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Add batching for openTxn
> 
>
> Key: HIVE-23076
> URL: https://issues.apache.org/jira/browse/HIVE-23076
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Reporter: Peter Vary
>Assignee: Peter Vary
>Priority: Major
> Attachments: HIVE-23076.2.patch, HIVE-23076.3.patch, 
> HIVE-23076.4.patch, HIVE-23076.5.patch, HIVE-23076.6.patch, 
> HIVE-23076.7.patch, HIVE-23076.8.patch, HIVE-23076.9.patch, HIVE-23076.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23076) Add batching for openTxn

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23076?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072629#comment-17072629
 ] 

Hive QA commented on HIVE-23076:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12998336/HIVE-23076.9.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 18163 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/21367/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/21367/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-21367/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12998336 - PreCommit-HIVE-Build

> Add batching for openTxn
> 
>
> Key: HIVE-23076
> URL: https://issues.apache.org/jira/browse/HIVE-23076
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Reporter: Peter Vary
>Assignee: Peter Vary
>Priority: Major
> Attachments: HIVE-23076.2.patch, HIVE-23076.3.patch, 
> HIVE-23076.4.patch, HIVE-23076.5.patch, HIVE-23076.6.patch, 
> HIVE-23076.7.patch, HIVE-23076.8.patch, HIVE-23076.9.patch, HIVE-23076.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-21304) Show Bucketing version for ReduceSinkOp in explain extended plan

2020-04-01 Thread Zoltan Haindrich (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21304?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltan Haindrich updated HIVE-21304:

Attachment: HIVE-21304.19.patch

> Show Bucketing version for ReduceSinkOp in explain extended plan
> 
>
> Key: HIVE-21304
> URL: https://issues.apache.org/jira/browse/HIVE-21304
> Project: Hive
>  Issue Type: Bug
>Reporter: Deepak Jaiswal
>Assignee: Zoltan Haindrich
>Priority: Major
> Attachments: HIVE-21304.01.patch, HIVE-21304.02.patch, 
> HIVE-21304.03.patch, HIVE-21304.04.patch, HIVE-21304.05.patch, 
> HIVE-21304.06.patch, HIVE-21304.07.patch, HIVE-21304.08.patch, 
> HIVE-21304.09.patch, HIVE-21304.10.patch, HIVE-21304.11.patch, 
> HIVE-21304.12.patch, HIVE-21304.13.patch, HIVE-21304.14.patch, 
> HIVE-21304.15.patch, HIVE-21304.16.patch, HIVE-21304.17.patch, 
> HIVE-21304.18.patch, HIVE-21304.19.patch
>
>
> Show Bucketing version for ReduceSinkOp in explain extended plan.
> This helps identify what hashing algorithm is being used by by ReduceSinkOp.
>  
> cc [~vgarg]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (HIVE-23122) LLAP TaskSchedulerService should log details about task eviction decision details

2020-04-01 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor reassigned HIVE-23122:
---

Assignee: László Bodor

> LLAP TaskSchedulerService should log details about task eviction decision 
> details
> -
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23122) LLAP TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor updated HIVE-23122:

Summary: LLAP TaskExecutorService should log details about task eviction 
decision details  (was: LLAP TaskSchedulerService should log details about task 
eviction decision details)

> LLAP TaskExecutorService should log details about task eviction decision 
> details
> 
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23122) LLAP TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor updated HIVE-23122:

Description: 
TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
another. Under the hood, the queue uses a configurable 
[comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
 The currently available comparators typically use the following properties of 
a task:
getWithinDagPriority: related to vertex


> LLAP TaskExecutorService should log details about task eviction decision 
> details
> 
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-23122.01.patch
>
>
> TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
> another. Under the hood, the queue uses a configurable 
> [comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
>  The currently available comparators typically use the following properties 
> of a task:
> getWithinDagPriority: related to vertex



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23122) LLAP TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Panagiotis Garefalakis (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072716#comment-17072716
 ] 

Panagiotis Garefalakis commented on HIVE-23122:
---

Thanks for the patch [~abstractdog] – pretty useful – would it make sense to 
add the canFinishForPriority property used by the BaseComparator as well?

> LLAP TaskExecutorService should log details about task eviction decision 
> details
> 
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-23122.01.patch
>
>
> TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
> another. Under the hood, the queue uses a configurable 
> [comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
>  The currently available comparators typically use the following properties 
> of a task(wrapper):
> getWithinDagPriority: related to vertex
> currentAttemptStartTime
> firstAttemptStartTime
> knownPending: remaining upstream tasks
> The problem is, when an eviction happens, the INFO level message doesn't 
> provide any insight about the decision, only attempts ids like below:
> {code}
> attempt_1585248378306_0010_72_02_96_8 evicted from wait queue in favor of 
> attempt_1585248378306_0003_175_02_79_175 because of lower priority
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23113) Clean Up HiveCallableStatement

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23113?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072649#comment-17072649
 ] 

Hive QA commented on HIVE-23113:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
59s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
18s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
12s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
29s{color} | {color:blue} jdbc in master has 16 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
13s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
38s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
18s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
18s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
11s{color} | {color:green} jdbc: The patch generated 0 new + 0 unchanged - 1 
fixed = 0 total (was 1) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
35s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
13s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 12m 55s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-21368/dev-support/hive-personality.sh
 |
| git revision | master / 94f90bc |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.1 |
| modules | C: jdbc U: jdbc |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21368/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Clean Up HiveCallableStatement
> --
>
> Key: HIVE-23113
> URL: https://issues.apache.org/jira/browse/HIVE-23113
> Project: Hive
>  Issue Type: Improvement
>  Components: JDBC
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Trivial
> Attachments: HIVE-23113.1.patch
>
>
> * Add a useful class comment
>  * Remove all non-javadoc comments
>  * Remove 'TODO' tags
>  * Add {{@override}} tags
>  * Checkstyle formatting



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (HIVE-23121) Re-examine TestWarehouseExternalDir to see if it uses HMS translation.

2020-04-01 Thread Naveen Gangam (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23121?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Naveen Gangam reassigned HIVE-23121:



> Re-examine TestWarehouseExternalDir to see if it uses HMS translation.
> --
>
> Key: HIVE-23121
> URL: https://issues.apache.org/jira/browse/HIVE-23121
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Naveen Gangam
>Assignee: Naveen Gangam
>Priority: Major
>
> TestWarehouseExternalDir currently passes with just one change related to 
> HIVE-22995. But that change was assuming it was using HMS Translation to 
> convert non-acid managed table to external. 
> Ensure that it still does.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-23039) Checkpointing for repl dump bootstrap phase

2020-04-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23039?focusedWorklogId=413961=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-413961
 ]

ASF GitHub Bot logged work on HIVE-23039:
-

Author: ASF GitHub Bot
Created on: 01/Apr/20 10:23
Start Date: 01/Apr/20 10:23
Worklog Time Spent: 10m 
  Work Description: anishek commented on pull request #965: HIVE-23039 
Checkpointing for repl dump bootstrap phase
URL: https://github.com/apache/hive/pull/965#discussion_r401496956
 
 

 ##
 File path: ql/src/java/org/apache/hadoop/hive/ql/exec/repl/util/ReplUtils.java
 ##
 @@ -296,4 +300,17 @@ public static boolean includeAcidTableInDump(HiveConf 
conf) {
   public static boolean tableIncludedInReplScope(ReplScope replScope, String 
tableName) {
 return ((replScope == null) || 
replScope.tableIncludedInReplScope(tableName));
   }
+
+  public static boolean dataCopyCompleted(Path toPath, HiveConf conf) throws 
IOException {
+FileSystem dstFs = null;
+dstFs = toPath.getFileSystem(conf);
+if (dstFs.exists(new Path(toPath, ReplUtils.COPY_ACKNOWLEDGEMENT))) {
+  return true;
+}
+return false;
+  }
+
+  public static void setDataCopyComplete(Path toPath, HiveConf conf) throws 
SemanticException {
 
 Review comment:
   rename method to ackCopy() ?
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 413961)
Time Spent: 50m  (was: 40m)

> Checkpointing for repl dump bootstrap phase
> ---
>
> Key: HIVE-23039
> URL: https://issues.apache.org/jira/browse/HIVE-23039
> Project: Hive
>  Issue Type: Bug
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23039.01.patch
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-23039) Checkpointing for repl dump bootstrap phase

2020-04-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23039?focusedWorklogId=413957=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-413957
 ]

ASF GitHub Bot logged work on HIVE-23039:
-

Author: ASF GitHub Bot
Created on: 01/Apr/20 10:23
Start Date: 01/Apr/20 10:23
Worklog Time Spent: 10m 
  Work Description: anishek commented on pull request #965: HIVE-23039 
Checkpointing for repl dump bootstrap phase
URL: https://github.com/apache/hive/pull/965#discussion_r401486847
 
 

 ##
 File path: 
itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosExternalTables.java
 ##
 @@ -906,6 +908,131 @@ public void replicationWithTableNameContainsKeywords() 
throws Throwable {
 .verifyReplTargetProperty(replicatedDbName);
   }
 
+  @Test
+  public void testCheckPointing() throws Throwable {
+List withClauseOptions = externalTableBasePathWithClause();
+WarehouseInstance.Tuple bootstrapDump = primary.run("use " + primaryDbName)
+.run("CREATE TABLE t1(a string) STORED AS TEXTFILE")
+.run("CREATE EXTERNAL TABLE t2(a string) STORED AS TEXTFILE")
+.run("insert into t1 values (1)")
+.run("insert into t1 values (2)")
+.run("insert into t2 values (11)")
+.run("insert into t2 values (21)")
+.dump(primaryDbName, withClauseOptions);
+
+// verify that the external table info is written correctly for bootstrap
+assertExternalFileInfo(Arrays.asList("t2"), bootstrapDump.dumpLocation, 
primaryDbName);
 
 Review comment:
   might be better to use another location here completely, sometimes since we 
are writing to same dump location and also the db directory is there you might 
have files in locations that are slight off and you wont realize it till you 
run actual prod scenario
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 413957)
Time Spent: 20m  (was: 10m)

> Checkpointing for repl dump bootstrap phase
> ---
>
> Key: HIVE-23039
> URL: https://issues.apache.org/jira/browse/HIVE-23039
> Project: Hive
>  Issue Type: Bug
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23039.01.patch
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-23039) Checkpointing for repl dump bootstrap phase

2020-04-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23039?focusedWorklogId=413959=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-413959
 ]

ASF GitHub Bot logged work on HIVE-23039:
-

Author: ASF GitHub Bot
Created on: 01/Apr/20 10:23
Start Date: 01/Apr/20 10:23
Worklog Time Spent: 10m 
  Work Description: anishek commented on pull request #965: HIVE-23039 
Checkpointing for repl dump bootstrap phase
URL: https://github.com/apache/hive/pull/965#discussion_r401491670
 
 

 ##
 File path: ql/src/java/org/apache/hadoop/hive/ql/plan/ReplCopyWork.java
 ##
 @@ -120,4 +122,12 @@ public boolean isNeedCheckDuplicateCopy() {
   public void setCheckDuplicateCopy(boolean flag) {
 checkDuplicateCopy = flag;
   }
+
+  public boolean isCheckpointEnabled() {
+return checkpointEnabled;
+  }
+
+  public void setCheckpointEnabled(boolean checkpointEnabled) {
 
 Review comment:
   do initialization based on constructor rather than setters ?
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 413959)

> Checkpointing for repl dump bootstrap phase
> ---
>
> Key: HIVE-23039
> URL: https://issues.apache.org/jira/browse/HIVE-23039
> Project: Hive
>  Issue Type: Bug
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23039.01.patch
>
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-23039) Checkpointing for repl dump bootstrap phase

2020-04-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23039?focusedWorklogId=413958=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-413958
 ]

ASF GitHub Bot logged work on HIVE-23039:
-

Author: ASF GitHub Bot
Created on: 01/Apr/20 10:23
Start Date: 01/Apr/20 10:23
Worklog Time Spent: 10m 
  Work Description: anishek commented on pull request #965: HIVE-23039 
Checkpointing for repl dump bootstrap phase
URL: https://github.com/apache/hive/pull/965#discussion_r401496558
 
 

 ##
 File path: ql/src/java/org/apache/hadoop/hive/ql/exec/repl/util/ReplUtils.java
 ##
 @@ -296,4 +300,17 @@ public static boolean includeAcidTableInDump(HiveConf 
conf) {
   public static boolean tableIncludedInReplScope(ReplScope replScope, String 
tableName) {
 return ((replScope == null) || 
replScope.tableIncludedInReplScope(tableName));
   }
+
+  public static boolean dataCopyCompleted(Path toPath, HiveConf conf) throws 
IOException {
+FileSystem dstFs = null;
+dstFs = toPath.getFileSystem(conf);
+if (dstFs.exists(new Path(toPath, ReplUtils.COPY_ACKNOWLEDGEMENT))) {
 
 Review comment:
   do return of condition directly rather than using if clause
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 413958)
Time Spent: 0.5h  (was: 20m)

> Checkpointing for repl dump bootstrap phase
> ---
>
> Key: HIVE-23039
> URL: https://issues.apache.org/jira/browse/HIVE-23039
> Project: Hive
>  Issue Type: Bug
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23039.01.patch
>
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-23039) Checkpointing for repl dump bootstrap phase

2020-04-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23039?focusedWorklogId=413960=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-413960
 ]

ASF GitHub Bot logged work on HIVE-23039:
-

Author: ASF GitHub Bot
Created on: 01/Apr/20 10:23
Start Date: 01/Apr/20 10:23
Worklog Time Spent: 10m 
  Work Description: anishek commented on pull request #965: HIVE-23039 
Checkpointing for repl dump bootstrap phase
URL: https://github.com/apache/hive/pull/965#discussion_r401496301
 
 

 ##
 File path: ql/src/java/org/apache/hadoop/hive/ql/exec/repl/util/ReplUtils.java
 ##
 @@ -98,6 +100,8 @@
   public static final String DUMP_ACKNOWLEDGEMENT = "_finished_dump";
   //Acknowledgement for repl load complete
   public static final String LOAD_ACKNOWLEDGEMENT = "_finished_load";
+  //Acknowledgement for data copy complete. Used for checkpointing
+  public static final String COPY_ACKNOWLEDGEMENT = "_finished_copy";
 
 Review comment:
   looks like we need a constants Class or a Enum class, the utils one seem to 
have a lot of magic variables.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 413960)
Time Spent: 40m  (was: 0.5h)

> Checkpointing for repl dump bootstrap phase
> ---
>
> Key: HIVE-23039
> URL: https://issues.apache.org/jira/browse/HIVE-23039
> Project: Hive
>  Issue Type: Bug
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23039.01.patch
>
>  Time Spent: 40m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-23020) Avoid using _files for replication data copy during incremental run

2020-04-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23020?focusedWorklogId=413976=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-413976
 ]

ASF GitHub Bot logged work on HIVE-23020:
-

Author: ASF GitHub Bot
Created on: 01/Apr/20 10:51
Start Date: 01/Apr/20 10:51
Worklog Time Spent: 10m 
  Work Description: aasha commented on pull request #963: HIVE-23020 : 
Avoid using _files for replication data copy during incr…
URL: https://github.com/apache/hive/pull/963#discussion_r401526351
 
 

 ##
 File path: 
ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/AddPartitionHandler.java
 ##
 @@ -108,10 +108,8 @@ public void handle(Context withinContext) throws 
Exception {
 Iterable files = partitionFilesIter.next().getFiles();
 if (files != null) {
   // encoded filename/checksum of files, write into _files
-  try (BufferedWriter fileListWriter = writer(withinContext, qlPtn)) {
-for (String file : files) {
-  writeFileEntry(qlMdTable.getDbName(), qlMdTable, file, 
fileListWriter, withinContext);
-}
+  for (String file : files) {
 
 Review comment:
   tests for this
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 413976)
Time Spent: 0.5h  (was: 20m)

> Avoid using _files for replication data copy during incremental run
> ---
>
> Key: HIVE-23020
> URL: https://issues.apache.org/jira/browse/HIVE-23020
> Project: Hive
>  Issue Type: Task
>Reporter: PRAVIN KUMAR SINHA
>Assignee: PRAVIN KUMAR SINHA
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23020.01.patch, HIVE-23020.02.patch
>
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (HIVE-23020) Avoid using _files for replication data copy during incremental run

2020-04-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23020?focusedWorklogId=413974=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-413974
 ]

ASF GitHub Bot logged work on HIVE-23020:
-

Author: ASF GitHub Bot
Created on: 01/Apr/20 10:50
Start Date: 01/Apr/20 10:50
Worklog Time Spent: 10m 
  Work Description: aasha commented on pull request #963: HIVE-23020 : 
Avoid using _files for replication data copy during incr…
URL: https://github.com/apache/hive/pull/963#discussion_r401522879
 
 

 ##
 File path: 
ql/src/java/org/apache/hadoop/hive/ql/parse/ImportSemanticAnalyzer.java
 ##
 @@ -505,9 +506,8 @@ private static ImportTableDesc 
getBaseCreateTableDescFromTable(String dbName,
 
 Task copyTask = null;
 if (replicationSpec.isInReplicationScope()) {
-  boolean isImport = 
ReplicationSpec.Type.IMPORT.equals(replicationSpec.getReplSpecType());
   copyTask = ReplCopyTask.getLoadCopyTask(replicationSpec, dataPath, 
destPath, x.getConf(),
-  isAutoPurge, needRecycle, copyToMigratedTxnTable, !isImport);
+  isAutoPurge, needRecycle, copyToMigratedTxnTable, false);
 
 Review comment:
   Why is this set to false? Do we even need this boolean if everyone is 
setting it to false?
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 413974)
Time Spent: 20m  (was: 10m)

> Avoid using _files for replication data copy during incremental run
> ---
>
> Key: HIVE-23020
> URL: https://issues.apache.org/jira/browse/HIVE-23020
> Project: Hive
>  Issue Type: Task
>Reporter: PRAVIN KUMAR SINHA
>Assignee: PRAVIN KUMAR SINHA
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23020.01.patch, HIVE-23020.02.patch
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (HIVE-23123) Export/Import of views and materialized views

2020-04-01 Thread Miklos Gergely (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23123?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Miklos Gergely reassigned HIVE-23123:
-


> Export/Import of views and materialized views
> -
>
> Key: HIVE-23123
> URL: https://issues.apache.org/jira/browse/HIVE-23123
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Miklos Gergely
>Assignee: Miklos Gergely
>Priority: Major
>
> According to 
> https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport 
> import and export can be done by using the
>  
> {code:java}
> export table ...
> import table ... 
> {code}
> commands. The document doesn't mention views or materialized views at all, 
> and in fact we don't support commands like
>  
> {code:java}
> export view ...
> import view ...
> export materialized view ...
> import materialized view ... 
> {code}
> they can not be parsed at all. The word table is often used though in a 
> broader sense, when it means all table like entities, including views and 
> materialized views. For example the various Table classes may represent any 
> of these as well.
> If I try to export a view with the export table ... command, it goes fine. A 
> _metadata file will be created, but no data directory, which is what we'd 
> expect. If I try to import it back, an exception is thrown due to the lack of 
> the data dir:
>  
> {code:java}
> java.lang.AssertionError: null==getPath() for exim_view
>  at org.apache.hadoop.hive.ql.metadata.Hive.loadTable(Hive.java:3088)
>  at org.apache.hadoop.hive.ql.exec.MoveTask.execute(MoveTask.java:419)
>  at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:213)
>  at 
> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:105)
>  at org.apache.hadoop.hive.ql.Executor.launchTask(Executor.java:364)
>  at org.apache.hadoop.hive.ql.Executor.launchTasks(Executor.java:335)
>  at org.apache.hadoop.hive.ql.Executor.runTasks(Executor.java:246)
>  at org.apache.hadoop.hive.ql.Executor.execute(Executor.java:109)
>  at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:722)
>  at org.apache.hadoop.hive.ql.Driver.run(Driver.java:491)
>  at org.apache.hadoop.hive.ql.Driver.run(Driver.java:485) 
> {code}
> Still the view gets imported successfully, as data movement wasn't even 
> necessary.
> If we try to export a materialized view which is transactional, then this 
> exception occurs:
>  
> {code:java}
> org.apache.hadoop.hive.ql.parse.SemanticException: 
> org.apache.hadoop.hive.ql.metadata.InvalidTableException: Table not found 
> exim_materialized_view_da21d41a_9fe4_4446_9c72_d251496abf9d
>  at 
> org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyzeAcidExport(AcidExportSemanticAnalyzer.java:163)
>  at 
> org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyze(AcidExportSemanticAnalyzer.java:71)
>  at 
> org.apache.hadoop.hive.ql.parse.RewriteSemanticAnalyzer.analyzeInternal(RewriteSemanticAnalyzer.java:72)
>  at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:289)
>  at org.apache.hadoop.hive.ql.Compiler.analyze(Compiler.java:220)
>  at org.apache.hadoop.hive.ql.Compiler.compile(Compiler.java:104)
>  at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:183)
>  at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:601)
>  at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:547)
>  at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:541) 
> {code}
> So the export process can not handle it, as the temporary table is not 
> getting created.
>  
> The import command handling have a lot of codes dedicated to importing views 
> and materialized views, which suggests that we support the importing (and 
> thus also suggests implicitly that we support the exporting) of views and 
> materialiezed views.
>  
> So the conclusion is that we have to decide if we support exporting/importing 
> of views and materialized views.
>  #  If we decide not to support them then:
>  - export process should throw an exception if a view or materialized view is 
> the subject
>  - the codes specific to view imports should be removed
>  # If we decide to support them, then:
>  - the commands mentioned above above should be introduced
>  - exception should be thrown if not the proper command used (e.g. export 
> view on a table)
>  - the exceptions mentioned above should be fixed
> I prefer #1, I don't think we should support the exporting / importing of 
> views. The point of exporting / importing is the transfer of data, not DDL, 
> it causes more issues than it solves. Our current documentation also suggests 
> that it is only supported for tables.
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23118) Option for exposing compile time counters as tez counters

2020-04-01 Thread Prasanth Jayachandran (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23118?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072409#comment-17072409
 ] 

Prasanth Jayachandran commented on HIVE-23118:
--

[~Sreenath] These are hive side counters merged with dag counters on the client 
side. These counters will be added to any tez task during hive query 
compilation. I don't think this will be available at tez side as it does not 
attach to any tez context. It will accessible to hive hooks though (hive proto 
hook can dump it). 

> Option for exposing compile time counters as tez counters
> -
>
> Key: HIVE-23118
> URL: https://issues.apache.org/jira/browse/HIVE-23118
> Project: Hive
>  Issue Type: Improvement
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
>Priority: Minor
>  Labels: pull-request-available
> Attachments: HIVE-23118.1.patch
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> TezCounters currently are runtime only. Some compile time information from 
> optimizer can be exposed as counters which can then be used by workload 
> management to make runtime decisions. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23052) Optimize lock enqueueing in TxnHandler

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23052?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072414#comment-17072414
 ] 

Hive QA commented on HIVE-23052:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  9m 
19s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
27s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
22s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  1m 
13s{color} | {color:blue} standalone-metastore/metastore-server in master has 
190 extant Findbugs warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
20s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
32s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
27s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
27s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
22s{color} | {color:red} standalone-metastore/metastore-server: The patch 
generated 36 new + 535 unchanged - 36 fixed = 571 total (was 571) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
19s{color} | {color:green} standalone-metastore/metastore-server generated 0 
new + 189 unchanged - 1 fixed = 189 total (was 190) {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
21s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 15m 25s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-21365/dev-support/hive-personality.sh
 |
| git revision | master / aa142d1 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21365/yetus/diff-checkstyle-standalone-metastore_metastore-server.txt
 |
| modules | C: standalone-metastore/metastore-server U: 
standalone-metastore/metastore-server |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21365/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Optimize lock enqueueing in TxnHandler
> --
>
> Key: HIVE-23052
> URL: https://issues.apache.org/jira/browse/HIVE-23052
> Project: Hive
>  Issue Type: Improvement
>Reporter: Marton Bod
>Assignee: Marton Bod
>Priority: Major
> Attachments: HIVE-23052.1.patch, HIVE-23052.10.patch, 
> HIVE-23052.11.patch, HIVE-23052.12.patch, HIVE-23052.13.patch, 
> HIVE-23052.2.patch, HIVE-23052.3.patch, HIVE-23052.4.patch, 
> HIVE-23052.5.patch, HIVE-23052.6.patch, HIVE-23052.7.patch, 
> HIVE-23052.8.patch, HIVE-23052.8.patch, HIVE-23052.8.patch, HIVE-23052.9.patch
>
>
> * Reduce scope of next_lock_id select-for-update by moving the txn_component 
> inserts before the S4U + inserting the hive_locks entries before the S4U 
> (first with a temp ID, which will be replaced later in a single update). This 
> helps decrease the overall time that the next_lock_id table is locked, 
> thereby increasing concurrency
>  * Insert txn_components in a batch instead of one-by-one (also in 
> TxnHandler::addDynamicPartition)
>  * Increment next_lock_id and update hive_locks table in a single batch 
> statement
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23082) PK/FK stat rescale doesn't work in some cases

2020-04-01 Thread Zoltan Haindrich (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23082?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltan Haindrich updated HIVE-23082:

Attachment: HIVE-23082.03.patch

> PK/FK stat rescale doesn't work in some cases
> -
>
> Key: HIVE-23082
> URL: https://issues.apache.org/jira/browse/HIVE-23082
> Project: Hive
>  Issue Type: Bug
>  Components: Statistics
>Reporter: Zoltan Haindrich
>Assignee: Zoltan Haindrich
>Priority: Major
> Attachments: HIVE-23082.01.patch, HIVE-23082.02.patch, 
> HIVE-23082.03.patch
>
>
> As a result in Joins may retain the original estimate; see MAPJOIN_33 in this 
> plan ; which retained the estimate of SEL_32
> {code}
> ++
> |  Explain   |
> ++
> | Plan optimized by CBO. |
> ||
> | Vertex dependency in root stage|
> | Map 1 <- Map 2 (BROADCAST_EDGE)|
> ||
> | Stage-0|
> |   Fetch Operator   |
> | limit:12   |
> | Stage-1|
> |   Map 1 vectorized |
> |   File Output Operator [FS_36] |
> | Limit [LIM_35] (rows=12 width=4)   |
> |   Number of rows:12|
> |   Select Operator [SEL_34] (rows=5040 width=4) |
> | Output:["_col0"]   |
> | Map Join Operator [MAPJOIN_33] (rows=5040 width=8) |
> |   Conds:SEL_32._col0=RS_30._col0(Inner) |
> | <-Map 2 [BROADCAST_EDGE] vectorized|
> |   BROADCAST [RS_30]|
> | PartitionCols:_col0|
> | Select Operator [SEL_29] (rows=1 width=8) |
> |   Output:["_col0"] |
> |   Filter Operator [FIL_28] (rows=1 width=108) |
> | predicate:((r_reason_id = 'reason 66') and r_reason_sk 
> is not null) |
> | TableScan [TS_3] (rows=2 width=108) |
> |   
> default@rx0,reason,Tbl:COMPLETE,Col:COMPLETE,Output:["r_reason_id","r_reason_sk"]
>  |
> | <-Select Operator [SEL_32] (rows=5040 width=7) |
> | Output:["_col0"]   |
> | Filter Operator [FIL_31] (rows=5040 width=7) |
> |   predicate:sr_reason_sk is not null |
> |   TableScan [TS_0] (rows=5112 width=7) |
> | 
> default@sr0,store_returns,Tbl:COMPLETE,Col:COMPLETE,Output:["sr_reason_sk"] |
> ||
> ++
> {code}
> repro:
> {code}
> set hive.query.results.cache.enabled=false;
> set hive.explain.user=true;
> drop table if exists default.rx0;
> drop table if exists default.sr0;
> create table rx0 (r_reason_id string, r_reason_sk bigint);
> create table sr0 (sr_reason_sk bigint);
> insert into rx0 values ('',1),('GEAA',70);
> insert into sr0 values (NULL),(1),(2),(3),(4),(5),(6),(7),(8),(9),(10),
> (11),(12),(13),(14),(15),(16),(17),(18),(19),(20),(21),(22),(23),(24),(25),
> (26),(27),(28),(29),(30),(31),(32),(33),(34),(35),(36),(37),(38),(39),(40),
> (41),(42),(43),(44),(45),(46),(47),(48),(49),(50),(51),(52),(53),(54),(55),
> (56),(57),(58),(59),(60),(61),(62),(63),(64),(65),(66),(67),(68),(69),(70);
> insert into sr0 select a.* from sr0 a,sr0 b;
> -- |sr0| ~ 5112
> explain select 1
> from default.sr0  store_returns , default.rx0 reason
> where sr_reason_sk = r_reason_sk
>   and r_reason_id = 'reason 66'
> limit 12;
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23095) NDV might be overestimated for a table with ~70 value

2020-04-01 Thread Zoltan Haindrich (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23095?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltan Haindrich updated HIVE-23095:

Attachment: HIVE-23095.04.patch

> NDV might be overestimated for a table with ~70 value
> -
>
> Key: HIVE-23095
> URL: https://issues.apache.org/jira/browse/HIVE-23095
> Project: Hive
>  Issue Type: Bug
>Reporter: Zoltan Haindrich
>Assignee: Zoltan Haindrich
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23095.01.patch, HIVE-23095.02.patch, 
> HIVE-23095.03.patch, HIVE-23095.04.patch
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> uncovered during looking into HIVE-23082
> https://issues.apache.org/jira/browse/HIVE-23082?focusedCommentId=17067773=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#comment-17067773



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23101) Fix topnkey_grouping_sets

2020-04-01 Thread Krisztian Kasa (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23101?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Krisztian Kasa updated HIVE-23101:
--
Status: Patch Available  (was: Open)

> Fix topnkey_grouping_sets
> -
>
> Key: HIVE-23101
> URL: https://issues.apache.org/jira/browse/HIVE-23101
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Krisztian Kasa
>Assignee: Krisztian Kasa
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-23101.1.patch, HIVE-23101.2.patch
>
>
> Example:
> {code}
> Client Execution succeeded but contained differences (error code = 1) after 
> executing topnkey_grouping_sets.q 
> 260c260
> < NULL NULL
> ---
> > NULL 1
> {code}
>  
> Test *topnkey_grouping_sets* fails intermittently.
> Queries which project 2 columns but order by only one of them can have more 
> than one good result set:
> {code:java}
> CREATE TABLE t_test_grouping_sets(
>   a int,
>   b int,
>   c int
> );
> INSERT INTO t_test_grouping_sets VALUES
> (NULL, NULL, NULL),
> (5, 2, 3),
> (10, 11, 12),
> (NULL, NULL, NULL),
> (NULL, NULL, NULL),
> (6, 2, 1),
> (7, 8, 4), (7, 8, 4), (7, 8, 4),
> (5, 1, 2), (5, 1, 2), (5, 1, 2),
> (NULL, NULL, NULL);
> SELECT a, b FROM t_test_grouping_sets GROUP BY GROUPING SETS ((a, b), (a), 
> (b), ()) ORDER BY a LIMIT 10;
> {code}
> {code:java}
> 5 NULL
> 5 2
> 5 1
> 6 2
> 6 NULL
> 7 8
> 7 NULL
> 10NULL
> 1011
> NULL  1
> {code}
> {code:java}
> 5 NULL
> 5 2
> 5 1
> 6 2
> 6 NULL
> 7 8
> 7 NULL
> 10NULL
> 1011
> NULL  NULL
> {code}
> Since we don't order by *b* both result sets are valid.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23101) Fix topnkey_grouping_sets

2020-04-01 Thread Krisztian Kasa (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23101?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Krisztian Kasa updated HIVE-23101:
--
Status: Open  (was: Patch Available)

> Fix topnkey_grouping_sets
> -
>
> Key: HIVE-23101
> URL: https://issues.apache.org/jira/browse/HIVE-23101
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Krisztian Kasa
>Assignee: Krisztian Kasa
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-23101.1.patch, HIVE-23101.2.patch
>
>
> Example:
> {code}
> Client Execution succeeded but contained differences (error code = 1) after 
> executing topnkey_grouping_sets.q 
> 260c260
> < NULL NULL
> ---
> > NULL 1
> {code}
>  
> Test *topnkey_grouping_sets* fails intermittently.
> Queries which project 2 columns but order by only one of them can have more 
> than one good result set:
> {code:java}
> CREATE TABLE t_test_grouping_sets(
>   a int,
>   b int,
>   c int
> );
> INSERT INTO t_test_grouping_sets VALUES
> (NULL, NULL, NULL),
> (5, 2, 3),
> (10, 11, 12),
> (NULL, NULL, NULL),
> (NULL, NULL, NULL),
> (6, 2, 1),
> (7, 8, 4), (7, 8, 4), (7, 8, 4),
> (5, 1, 2), (5, 1, 2), (5, 1, 2),
> (NULL, NULL, NULL);
> SELECT a, b FROM t_test_grouping_sets GROUP BY GROUPING SETS ((a, b), (a), 
> (b), ()) ORDER BY a LIMIT 10;
> {code}
> {code:java}
> 5 NULL
> 5 2
> 5 1
> 6 2
> 6 NULL
> 7 8
> 7 NULL
> 10NULL
> 1011
> NULL  1
> {code}
> {code:java}
> 5 NULL
> 5 2
> 5 1
> 6 2
> 6 NULL
> 7 8
> 7 NULL
> 10NULL
> 1011
> NULL  NULL
> {code}
> Since we don't order by *b* both result sets are valid.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23101) Fix topnkey_grouping_sets

2020-04-01 Thread Krisztian Kasa (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23101?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Krisztian Kasa updated HIVE-23101:
--
Attachment: HIVE-23101.2.patch

> Fix topnkey_grouping_sets
> -
>
> Key: HIVE-23101
> URL: https://issues.apache.org/jira/browse/HIVE-23101
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Krisztian Kasa
>Assignee: Krisztian Kasa
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-23101.1.patch, HIVE-23101.2.patch
>
>
> Example:
> {code}
> Client Execution succeeded but contained differences (error code = 1) after 
> executing topnkey_grouping_sets.q 
> 260c260
> < NULL NULL
> ---
> > NULL 1
> {code}
>  
> Test *topnkey_grouping_sets* fails intermittently.
> Queries which project 2 columns but order by only one of them can have more 
> than one good result set:
> {code:java}
> CREATE TABLE t_test_grouping_sets(
>   a int,
>   b int,
>   c int
> );
> INSERT INTO t_test_grouping_sets VALUES
> (NULL, NULL, NULL),
> (5, 2, 3),
> (10, 11, 12),
> (NULL, NULL, NULL),
> (NULL, NULL, NULL),
> (6, 2, 1),
> (7, 8, 4), (7, 8, 4), (7, 8, 4),
> (5, 1, 2), (5, 1, 2), (5, 1, 2),
> (NULL, NULL, NULL);
> SELECT a, b FROM t_test_grouping_sets GROUP BY GROUPING SETS ((a, b), (a), 
> (b), ()) ORDER BY a LIMIT 10;
> {code}
> {code:java}
> 5 NULL
> 5 2
> 5 1
> 6 2
> 6 NULL
> 7 8
> 7 NULL
> 10NULL
> 1011
> NULL  1
> {code}
> {code:java}
> 5 NULL
> 5 2
> 5 1
> 6 2
> 6 NULL
> 7 8
> 7 NULL
> 10NULL
> 1011
> NULL  NULL
> {code}
> Since we don't order by *b* both result sets are valid.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23119) Test sort_acid should be run by TestMiniLlapLocalCliDriver only

2020-04-01 Thread Krisztian Kasa (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23119?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Krisztian Kasa updated HIVE-23119:
--
Priority: Minor  (was: Major)

> Test sort_acid should be run by TestMiniLlapLocalCliDriver only
> ---
>
> Key: HIVE-23119
> URL: https://issues.apache.org/jira/browse/HIVE-23119
> Project: Hive
>  Issue Type: Task
>  Components: CBO
>Reporter: Krisztian Kasa
>Priority: Minor
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-21492) VectorizedParquetRecordReader can't to read parquet file generated using thrift/custom tool

2020-04-01 Thread Ferdinand Xu (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-21492?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072466#comment-17072466
 ] 

Ferdinand Xu commented on HIVE-21492:
-

Could you update the indents below?
{code:java}
+ return childType.asGroupType().getFields().get(0)
+  .asPrimitiveType();
{code}
 

> VectorizedParquetRecordReader can't to read parquet file generated using 
> thrift/custom tool
> ---
>
> Key: HIVE-21492
> URL: https://issues.apache.org/jira/browse/HIVE-21492
> Project: Hive
>  Issue Type: Bug
>Reporter: Ganesha Shreedhara
>Assignee: Ganesha Shreedhara
>Priority: Major
> Attachments: HIVE-21492.2.patch, HIVE-21492.patch
>
>
> Taking an example of a parquet table having array of integers as below. 
> {code:java}
> CREATE EXTERNAL TABLE ( list_of_ints` array)
> STORED AS PARQUET 
> LOCATION '{location}';
> {code}
> Parquet file generated using hive will have schema for Type as below:
> {code:java}
> group list_of_ints (LIST) { repeated group bag { optional int32 array;\n};\n} 
> {code}
> Parquet file generated using thrift or any custom tool (using 
> org.apache.parquet.io.api.RecordConsumer)
> may have schema for Type as below:
> {code:java}
> required group list_of_ints (LIST) { repeated int32 list_of_tuple} {code}
> VectorizedParquetRecordReader handles only parquet file generated using hive. 
> It throws the following exception when parquet file generated using thrift is 
> read because of the changes done as part of HIVE-18553 .
> {code:java}
> Caused by: java.lang.ClassCastException: repeated int32 list_of_ints_tuple is 
> not a group
>  at org.apache.parquet.schema.Type.asGroupType(Type.java:207)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.getElementType(VectorizedParquetRecordReader.java:479)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.buildVectorizedParquetReader(VectorizedParquetRecordReader.java:532)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.checkEndOfRowGroup(VectorizedParquetRecordReader.java:440)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.nextBatch(VectorizedParquetRecordReader.java:401)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:353)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:92)
>  at 
> org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.doNext(HiveContextAwareRecordReader.java:365){code}
>  
>  I have done a small change to handle the case where the child type of group 
> type can be PrimitiveType.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23119) Test sort_acid should be run by TestMiniLlapLocalCliDriver only

2020-04-01 Thread Krisztian Kasa (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23119?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Krisztian Kasa updated HIVE-23119:
--
Attachment: HIVE-23119.1.patch

> Test sort_acid should be run by TestMiniLlapLocalCliDriver only
> ---
>
> Key: HIVE-23119
> URL: https://issues.apache.org/jira/browse/HIVE-23119
> Project: Hive
>  Issue Type: Task
>  Components: CBO
>Reporter: Krisztian Kasa
>Priority: Minor
> Attachments: HIVE-23119.1.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23119) Test sort_acid should be run by TestMiniLlapLocalCliDriver only

2020-04-01 Thread Krisztian Kasa (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23119?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Krisztian Kasa updated HIVE-23119:
--
Status: Patch Available  (was: Open)

> Test sort_acid should be run by TestMiniLlapLocalCliDriver only
> ---
>
> Key: HIVE-23119
> URL: https://issues.apache.org/jira/browse/HIVE-23119
> Project: Hive
>  Issue Type: Task
>  Components: CBO
>Reporter: Krisztian Kasa
>Priority: Minor
> Attachments: HIVE-23119.1.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (HIVE-23119) Test sort_acid should be run by TestMiniLlapLocalCliDriver only

2020-04-01 Thread Krisztian Kasa (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23119?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Krisztian Kasa reassigned HIVE-23119:
-

Assignee: Krisztian Kasa

> Test sort_acid should be run by TestMiniLlapLocalCliDriver only
> ---
>
> Key: HIVE-23119
> URL: https://issues.apache.org/jira/browse/HIVE-23119
> Project: Hive
>  Issue Type: Task
>  Components: CBO
>Reporter: Krisztian Kasa
>Assignee: Krisztian Kasa
>Priority: Minor
> Attachments: HIVE-23119.1.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-21492) VectorizedParquetRecordReader can't to read parquet file generated using thrift/custom tool

2020-04-01 Thread Ganesha Shreedhara (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-21492?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072471#comment-17072471
 ] 

Ganesha Shreedhara commented on HIVE-21492:
---

Done.

> VectorizedParquetRecordReader can't to read parquet file generated using 
> thrift/custom tool
> ---
>
> Key: HIVE-21492
> URL: https://issues.apache.org/jira/browse/HIVE-21492
> Project: Hive
>  Issue Type: Bug
>Reporter: Ganesha Shreedhara
>Assignee: Ganesha Shreedhara
>Priority: Major
> Attachments: HIVE-21492.2.patch, HIVE-21492.3.patch, HIVE-21492.patch
>
>
> Taking an example of a parquet table having array of integers as below. 
> {code:java}
> CREATE EXTERNAL TABLE ( list_of_ints` array)
> STORED AS PARQUET 
> LOCATION '{location}';
> {code}
> Parquet file generated using hive will have schema for Type as below:
> {code:java}
> group list_of_ints (LIST) { repeated group bag { optional int32 array;\n};\n} 
> {code}
> Parquet file generated using thrift or any custom tool (using 
> org.apache.parquet.io.api.RecordConsumer)
> may have schema for Type as below:
> {code:java}
> required group list_of_ints (LIST) { repeated int32 list_of_tuple} {code}
> VectorizedParquetRecordReader handles only parquet file generated using hive. 
> It throws the following exception when parquet file generated using thrift is 
> read because of the changes done as part of HIVE-18553 .
> {code:java}
> Caused by: java.lang.ClassCastException: repeated int32 list_of_ints_tuple is 
> not a group
>  at org.apache.parquet.schema.Type.asGroupType(Type.java:207)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.getElementType(VectorizedParquetRecordReader.java:479)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.buildVectorizedParquetReader(VectorizedParquetRecordReader.java:532)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.checkEndOfRowGroup(VectorizedParquetRecordReader.java:440)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.nextBatch(VectorizedParquetRecordReader.java:401)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:353)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:92)
>  at 
> org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.doNext(HiveContextAwareRecordReader.java:365){code}
>  
>  I have done a small change to handle the case where the child type of group 
> type can be PrimitiveType.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (HIVE-23120) TopNKey related tests should be run by TestMiniLlapLocalCliDriver only

2020-04-01 Thread Krisztian Kasa (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23120?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Krisztian Kasa reassigned HIVE-23120:
-


> TopNKey related tests should be run by TestMiniLlapLocalCliDriver only
> --
>
> Key: HIVE-23120
> URL: https://issues.apache.org/jira/browse/HIVE-23120
> Project: Hive
>  Issue Type: Task
>  Components: Physical Optimizer
>Reporter: Krisztian Kasa
>Assignee: Krisztian Kasa
>Priority: Minor
>
> TopNKey optimization is only used when the execution framework is Tez.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-21492) VectorizedParquetRecordReader can't to read parquet file generated using thrift/custom tool

2020-04-01 Thread Ganesha Shreedhara (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21492?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ganesha Shreedhara updated HIVE-21492:
--
Status: Patch Available  (was: In Progress)

> VectorizedParquetRecordReader can't to read parquet file generated using 
> thrift/custom tool
> ---
>
> Key: HIVE-21492
> URL: https://issues.apache.org/jira/browse/HIVE-21492
> Project: Hive
>  Issue Type: Bug
>Reporter: Ganesha Shreedhara
>Assignee: Ganesha Shreedhara
>Priority: Major
> Attachments: HIVE-21492.2.patch, HIVE-21492.3.patch, HIVE-21492.patch
>
>
> Taking an example of a parquet table having array of integers as below. 
> {code:java}
> CREATE EXTERNAL TABLE ( list_of_ints` array)
> STORED AS PARQUET 
> LOCATION '{location}';
> {code}
> Parquet file generated using hive will have schema for Type as below:
> {code:java}
> group list_of_ints (LIST) { repeated group bag { optional int32 array;\n};\n} 
> {code}
> Parquet file generated using thrift or any custom tool (using 
> org.apache.parquet.io.api.RecordConsumer)
> may have schema for Type as below:
> {code:java}
> required group list_of_ints (LIST) { repeated int32 list_of_tuple} {code}
> VectorizedParquetRecordReader handles only parquet file generated using hive. 
> It throws the following exception when parquet file generated using thrift is 
> read because of the changes done as part of HIVE-18553 .
> {code:java}
> Caused by: java.lang.ClassCastException: repeated int32 list_of_ints_tuple is 
> not a group
>  at org.apache.parquet.schema.Type.asGroupType(Type.java:207)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.getElementType(VectorizedParquetRecordReader.java:479)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.buildVectorizedParquetReader(VectorizedParquetRecordReader.java:532)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.checkEndOfRowGroup(VectorizedParquetRecordReader.java:440)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.nextBatch(VectorizedParquetRecordReader.java:401)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:353)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:92)
>  at 
> org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.doNext(HiveContextAwareRecordReader.java:365){code}
>  
>  I have done a small change to handle the case where the child type of group 
> type can be PrimitiveType.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-21492) VectorizedParquetRecordReader can't to read parquet file generated using thrift/custom tool

2020-04-01 Thread Ganesha Shreedhara (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21492?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ganesha Shreedhara updated HIVE-21492:
--
Attachment: HIVE-21492.3.patch

> VectorizedParquetRecordReader can't to read parquet file generated using 
> thrift/custom tool
> ---
>
> Key: HIVE-21492
> URL: https://issues.apache.org/jira/browse/HIVE-21492
> Project: Hive
>  Issue Type: Bug
>Reporter: Ganesha Shreedhara
>Assignee: Ganesha Shreedhara
>Priority: Major
> Attachments: HIVE-21492.2.patch, HIVE-21492.3.patch, HIVE-21492.patch
>
>
> Taking an example of a parquet table having array of integers as below. 
> {code:java}
> CREATE EXTERNAL TABLE ( list_of_ints` array)
> STORED AS PARQUET 
> LOCATION '{location}';
> {code}
> Parquet file generated using hive will have schema for Type as below:
> {code:java}
> group list_of_ints (LIST) { repeated group bag { optional int32 array;\n};\n} 
> {code}
> Parquet file generated using thrift or any custom tool (using 
> org.apache.parquet.io.api.RecordConsumer)
> may have schema for Type as below:
> {code:java}
> required group list_of_ints (LIST) { repeated int32 list_of_tuple} {code}
> VectorizedParquetRecordReader handles only parquet file generated using hive. 
> It throws the following exception when parquet file generated using thrift is 
> read because of the changes done as part of HIVE-18553 .
> {code:java}
> Caused by: java.lang.ClassCastException: repeated int32 list_of_ints_tuple is 
> not a group
>  at org.apache.parquet.schema.Type.asGroupType(Type.java:207)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.getElementType(VectorizedParquetRecordReader.java:479)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.buildVectorizedParquetReader(VectorizedParquetRecordReader.java:532)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.checkEndOfRowGroup(VectorizedParquetRecordReader.java:440)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.nextBatch(VectorizedParquetRecordReader.java:401)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:353)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:92)
>  at 
> org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.doNext(HiveContextAwareRecordReader.java:365){code}
>  
>  I have done a small change to handle the case where the child type of group 
> type can be PrimitiveType.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23032) Add batching in Lock generation

2020-04-01 Thread Denys Kuzmenko (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23032?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Denys Kuzmenko updated HIVE-23032:
--
Resolution: Fixed
Status: Resolved  (was: Patch Available)

> Add batching in Lock generation
> ---
>
> Key: HIVE-23032
> URL: https://issues.apache.org/jira/browse/HIVE-23032
> Project: Hive
>  Issue Type: Improvement
>  Components: Locking
>Reporter: Denys Kuzmenko
>Assignee: Denys Kuzmenko
>Priority: Major
> Attachments: HIVE-23032.1.patch, HIVE-23032.2.patch, 
> HIVE-23032.3.patch, HIVE-23032.4.patch, HIVE-23032.5.patch, 
> HIVE-23032.6.patch, HIVE-23032.7.patch, HIVE-23032.8.patch
>
>
> Replace multi-row insert in Oracle with batching. Performance tests showed 
> significant performance improvement after turning batching on.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23095) NDV might be overestimated for a table with ~70 value

2020-04-01 Thread Zoltan Haindrich (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23095?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zoltan Haindrich updated HIVE-23095:

Attachment: HIVE-23095.03.patch

> NDV might be overestimated for a table with ~70 value
> -
>
> Key: HIVE-23095
> URL: https://issues.apache.org/jira/browse/HIVE-23095
> Project: Hive
>  Issue Type: Bug
>Reporter: Zoltan Haindrich
>Assignee: Zoltan Haindrich
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23095.01.patch, HIVE-23095.02.patch, 
> HIVE-23095.03.patch
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> uncovered during looking into HIVE-23082
> https://issues.apache.org/jira/browse/HIVE-23082?focusedCommentId=17067773=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#comment-17067773



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23052) Optimize lock enqueueing in TxnHandler

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23052?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072470#comment-17072470
 ] 

Hive QA commented on HIVE-23052:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12998346/HIVE-23052.13.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 18163 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[topnkey_grouping_sets] 
(batchId=1)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/21365/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/21365/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-21365/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12998346 - PreCommit-HIVE-Build

> Optimize lock enqueueing in TxnHandler
> --
>
> Key: HIVE-23052
> URL: https://issues.apache.org/jira/browse/HIVE-23052
> Project: Hive
>  Issue Type: Improvement
>Reporter: Marton Bod
>Assignee: Marton Bod
>Priority: Major
> Attachments: HIVE-23052.1.patch, HIVE-23052.10.patch, 
> HIVE-23052.11.patch, HIVE-23052.12.patch, HIVE-23052.13.patch, 
> HIVE-23052.2.patch, HIVE-23052.3.patch, HIVE-23052.4.patch, 
> HIVE-23052.5.patch, HIVE-23052.6.patch, HIVE-23052.7.patch, 
> HIVE-23052.8.patch, HIVE-23052.8.patch, HIVE-23052.8.patch, HIVE-23052.9.patch
>
>
> * Reduce scope of next_lock_id select-for-update by moving the txn_component 
> inserts before the S4U + inserting the hive_locks entries before the S4U 
> (first with a temp ID, which will be replaced later in a single update). This 
> helps decrease the overall time that the next_lock_id table is locked, 
> thereby increasing concurrency
>  * Insert txn_components in a batch instead of one-by-one (also in 
> TxnHandler::addDynamicPartition)
>  * Increment next_lock_id and update hive_locks table in a single batch 
> statement
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-21492) VectorizedParquetRecordReader can't to read parquet file generated using thrift/custom tool

2020-04-01 Thread Ganesha Shreedhara (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21492?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ganesha Shreedhara updated HIVE-21492:
--
Status: In Progress  (was: Patch Available)

> VectorizedParquetRecordReader can't to read parquet file generated using 
> thrift/custom tool
> ---
>
> Key: HIVE-21492
> URL: https://issues.apache.org/jira/browse/HIVE-21492
> Project: Hive
>  Issue Type: Bug
>Reporter: Ganesha Shreedhara
>Assignee: Ganesha Shreedhara
>Priority: Major
> Attachments: HIVE-21492.2.patch, HIVE-21492.patch
>
>
> Taking an example of a parquet table having array of integers as below. 
> {code:java}
> CREATE EXTERNAL TABLE ( list_of_ints` array)
> STORED AS PARQUET 
> LOCATION '{location}';
> {code}
> Parquet file generated using hive will have schema for Type as below:
> {code:java}
> group list_of_ints (LIST) { repeated group bag { optional int32 array;\n};\n} 
> {code}
> Parquet file generated using thrift or any custom tool (using 
> org.apache.parquet.io.api.RecordConsumer)
> may have schema for Type as below:
> {code:java}
> required group list_of_ints (LIST) { repeated int32 list_of_tuple} {code}
> VectorizedParquetRecordReader handles only parquet file generated using hive. 
> It throws the following exception when parquet file generated using thrift is 
> read because of the changes done as part of HIVE-18553 .
> {code:java}
> Caused by: java.lang.ClassCastException: repeated int32 list_of_ints_tuple is 
> not a group
>  at org.apache.parquet.schema.Type.asGroupType(Type.java:207)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.getElementType(VectorizedParquetRecordReader.java:479)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.buildVectorizedParquetReader(VectorizedParquetRecordReader.java:532)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.checkEndOfRowGroup(VectorizedParquetRecordReader.java:440)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.nextBatch(VectorizedParquetRecordReader.java:401)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:353)
>  at 
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:92)
>  at 
> org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.doNext(HiveContextAwareRecordReader.java:365){code}
>  
>  I have done a small change to handle the case where the child type of group 
> type can be PrimitiveType.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23067) Use batch DB calls in TxnHandler for commitTxn and abortTxns

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23067?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072490#comment-17072490
 ] 

Hive QA commented on HIVE-23067:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  9m 
 8s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
28s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
21s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  1m 
13s{color} | {color:blue} standalone-metastore/metastore-server in master has 
190 extant Findbugs warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
21s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
31s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
26s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
26s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
23s{color} | {color:red} standalone-metastore/metastore-server: The patch 
generated 1 new + 545 unchanged - 1 fixed = 546 total (was 546) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  1m 
22s{color} | {color:red} standalone-metastore/metastore-server generated 1 new 
+ 190 unchanged - 0 fixed = 191 total (was 190) {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
19s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 15m 14s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| FindBugs | module:standalone-metastore/metastore-server |
|  |  A prepared statement is generated from a nonconstant String in 
org.apache.hadoop.hive.metastore.txn.TxnHandler.commitTxn(CommitTxnRequest)   
At TxnHandler.java:from a nonconstant String in 
org.apache.hadoop.hive.metastore.txn.TxnHandler.commitTxn(CommitTxnRequest)   
At TxnHandler.java:[line 1298] |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-21366/dev-support/hive-personality.sh
 |
| git revision | master / aa142d1 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21366/yetus/diff-checkstyle-standalone-metastore_metastore-server.txt
 |
| findbugs | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21366/yetus/new-findbugs-standalone-metastore_metastore-server.html
 |
| modules | C: standalone-metastore/metastore-server U: 
standalone-metastore/metastore-server |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21366/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Use batch DB calls in TxnHandler for commitTxn and abortTxns
> 
>
> Key: HIVE-23067
> URL: https://issues.apache.org/jira/browse/HIVE-23067
> Project: Hive
>  Issue Type: Improvement
>Reporter: Marton Bod
>Assignee: Marton Bod
>Priority: Major
> Attachments: HIVE-23067.1.patch, HIVE-23067.1.patch, 
> HIVE-23067.1.patch, HIVE-23067.1.patch, HIVE-23067.2.patch
>
>
> To reduce the number of database calls and network roundtrips, we could use 
> more batching in TxnHandler, where currently in many places we call insert 
> commands in loops sequentially.
> Some examples:
>  * openTxns (TXNS, REPL_TXN_MAP)
>  * commitTxn 

[jira] [Commented] (HIVE-23107) Remove MIN_HISTORY_LEVEL table

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23107?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072781#comment-17072781
 ] 

Hive QA commented on HIVE-23107:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12998338/HIVE-23107.02.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 160 failed/errored test(s), 18162 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.metastore.txn.TestCompactionTxnHandler.testFindNextToClean
 (batchId=318)
org.apache.hadoop.hive.metastore.txn.TestCompactionTxnHandler.testMarkCleaned 
(batchId=318)
org.apache.hadoop.hive.metastore.txn.TestCompactionTxnHandler.testMarkCleanedCleansTxnsAndTxnComponents
 (batchId=318)
org.apache.hadoop.hive.metastore.txn.TestCompactionTxnHandler.testMarkCompacted 
(batchId=318)
org.apache.hadoop.hive.ql.TestAcidOnTez.testInsertWithRemoveUnion (batchId=253)
org.apache.hadoop.hive.ql.TestAcidOnTez.testMapJoinOnMR (batchId=253)
org.apache.hadoop.hive.ql.TestAcidOnTez.testMapJoinOnTez (batchId=253)
org.apache.hadoop.hive.ql.TestAcidOnTez.testMergeJoinOnMR (batchId=253)
org.apache.hadoop.hive.ql.TestAcidOnTez.testMergeJoinOnTez (batchId=253)
org.apache.hadoop.hive.ql.TestTxnCommands.testNonAcidToAcidConversion01 
(batchId=363)
org.apache.hadoop.hive.ql.TestTxnCommands.testVersioning (batchId=363)
org.apache.hadoop.hive.ql.TestTxnCommands2.testACIDwithSchemaEvolutionAndCompaction
 (batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testCleanerForTxnToWriteId 
(batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testCompactWithDelete (batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testEmptyInTblproperties 
(batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testInitiatorWithMultipleFailedCompactions
 (batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testInsertOverwrite1 (batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testInsertOverwrite2 (batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testNonAcidToAcidConversion02 
(batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testNonAcidToAcidConversion1 
(batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testNonAcidToAcidConversion2 
(batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testNonAcidToAcidConversion3 
(batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testOriginalFileReaderWhenNonAcidConvertedToAcid
 (batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.testSchemaEvolutionCompaction 
(batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2.writeBetweenWorkerAndCleaner 
(batchId=344)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testACIDwithSchemaEvolutionAndCompaction
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testCleanerForTxnToWriteId
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testCompactWithDelete
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testEmptyInTblproperties
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testInitiatorWithMultipleFailedCompactions
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testInsertOverwrite1
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testInsertOverwrite2
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testNonAcidToAcidConversion02
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testNonAcidToAcidConversion1
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testNonAcidToAcidConversion2
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testNonAcidToAcidConversion3
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testOriginalFileReaderWhenNonAcidConvertedToAcid
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testSchemaEvolutionCompaction
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.writeBetweenWorkerAndCleaner
 (batchId=358)
org.apache.hadoop.hive.ql.TestTxnCommands3.testAcidMetaColumsDecode 
(batchId=359)
org.apache.hadoop.hive.ql.TestTxnCommands3.testCleaner2 (batchId=359)
org.apache.hadoop.hive.ql.TestTxnCommands3.testCompactionAbort (batchId=359)
org.apache.hadoop.hive.ql.TestTxnCommands3.testDeleteEventPruningOff 
(batchId=359)
org.apache.hadoop.hive.ql.TestTxnCommands3.testDeleteEventPruningOn 
(batchId=359)
org.apache.hadoop.hive.ql.TestTxnCommands3.testNotEnoughToCompactNeedsCleaning 
(batchId=359)
org.apache.hadoop.hive.ql.TestTxnCommandsForMmTable.testInsertOverwriteForPartitionedMmTable
 (batchId=318)

[jira] [Commented] (HIVE-23109) Query-based compaction omits database

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23109?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072850#comment-17072850
 ] 

Hive QA commented on HIVE-23109:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12998353/HIVE-23109.02.patch

{color:green}SUCCESS:{color} +1 due to 3 test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 18166 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/21370/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/21370/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-21370/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12998353 - PreCommit-HIVE-Build

> Query-based compaction omits database
> -
>
> Key: HIVE-23109
> URL: https://issues.apache.org/jira/browse/HIVE-23109
> Project: Hive
>  Issue Type: Bug
>Reporter: Karen Coppage
>Assignee: Karen Coppage
>Priority: Major
> Attachments: HIVE-23109.01.patch, HIVE-23109.02.patch
>
>
> E.g. MM major compaction query looks like:
> {code:java}
> insert into tmp_table select * from src_table;
> {code}
> it should be
> {code:java}
> insert into tmp_table select * from src_db.src_table;
> {code}
> Therefore compaction fails if db of source table isn't default.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23123) Export/Import of views and materialized views

2020-04-01 Thread Miklos Gergely (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23123?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072779#comment-17072779
 ] 

Miklos Gergely commented on HIVE-23123:
---

[~jcamachorodriguez] please let me know what you think.

> Export/Import of views and materialized views
> -
>
> Key: HIVE-23123
> URL: https://issues.apache.org/jira/browse/HIVE-23123
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Miklos Gergely
>Assignee: Miklos Gergely
>Priority: Major
>
> According to 
> [https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport]
>  import and export can be done by using the
>  
> {code:java}
> export table ...
> import table ... 
> {code}
> commands. The document doesn't mention views or materialized views at all, 
> and in fact we don't support commands like
>  
> {code:java}
> export view ...
> import view ...
> export materialized view ...
> import materialized view ... 
> {code}
> they can not be parsed at all. The word table is often used though in a 
> broader sense, when it means all table like entities, including views and 
> materialized views. For example the various Table classes may represent any 
> of these as well.
> If I try to export a view with the export table ... command, it goes fine. A 
> _metadata file will be created, but no data directory, which is what we'd 
> expect. If I try to import it back, an exception is thrown due to the lack of 
> the data dir:
>  
> {code:java}
> java.lang.AssertionError: null==getPath() for exim_view
>  at org.apache.hadoop.hive.ql.metadata.Hive.loadTable(Hive.java:3088)
>  at org.apache.hadoop.hive.ql.exec.MoveTask.execute(MoveTask.java:419)
>  at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:213)
>  at 
> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:105)
>  at org.apache.hadoop.hive.ql.Executor.launchTask(Executor.java:364)
>  at org.apache.hadoop.hive.ql.Executor.launchTasks(Executor.java:335)
>  at org.apache.hadoop.hive.ql.Executor.runTasks(Executor.java:246)
>  at org.apache.hadoop.hive.ql.Executor.execute(Executor.java:109)
>  at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:722)
>  at org.apache.hadoop.hive.ql.Driver.run(Driver.java:491)
>  at org.apache.hadoop.hive.ql.Driver.run(Driver.java:485) 
> {code}
> Still the view gets imported successfully, as data movement wasn't even 
> necessary.
> If we try to export a materialized view which is transactional, then this 
> exception occurs:
>  
> {code:java}
> org.apache.hadoop.hive.ql.parse.SemanticException: 
> org.apache.hadoop.hive.ql.metadata.InvalidTableException: Table not found 
> exim_materialized_view_da21d41a_9fe4_4446_9c72_d251496abf9d
>  at 
> org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyzeAcidExport(AcidExportSemanticAnalyzer.java:163)
>  at 
> org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyze(AcidExportSemanticAnalyzer.java:71)
>  at 
> org.apache.hadoop.hive.ql.parse.RewriteSemanticAnalyzer.analyzeInternal(RewriteSemanticAnalyzer.java:72)
>  at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:289)
>  at org.apache.hadoop.hive.ql.Compiler.analyze(Compiler.java:220)
>  at org.apache.hadoop.hive.ql.Compiler.compile(Compiler.java:104)
>  at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:183)
>  at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:601)
>  at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:547)
>  at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:541) 
> {code}
> So the export process can not handle it, as the temporary table is not 
> getting created.
>  
> The import command handling have a lot of codes dedicated to importing views 
> and materialized views, which suggests that we support the importing (and 
> thus also suggests implicitly that we support the exporting) of views and 
> materialiezed views.
>  
> So the conclusion is that we have to decide if we support exporting/importing 
> of views and materialized views.
> If we decide not to support them then:
>  - export process should throw an exception if a view or materialized view is 
> the subject
>  - the codes specific to view imports should be removed
> If we decide to support them, then:
>  - the commands mentioned above above should be introduced
>  - exception should be thrown if not the proper command used (e.g. export 
> view on a table)
>  - the exceptions mentioned above should be fixed
> I prefer not to support them, I don't think we should support the exporting / 
> importing of views. The point of exporting / importing is the transfer of 
> data, not DDL, it causes more issues than it solves. Our current 
> documentation also suggests that it is only supported for tables.
>  



[jira] [Updated] (HIVE-23123) Export/Import of views and materialized views

2020-04-01 Thread Miklos Gergely (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23123?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Miklos Gergely updated HIVE-23123:
--
Description: 
According to 
[https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport] 
import and export can be done by using the
{code:java}
export table ...
import table ... 
{code}
commands. The document doesn't mention views or materialized views at all, and 
in fact we don't support commands like
{code:java}
export view ...
import view ...
export materialized view ...
import materialized view ... 
{code}
they can not be parsed at all. The word table is often used though in a broader 
sense, when it means all table like entities, including views and materialized 
views. For example the various Table classes may represent any of these as well.

If I try to export a view with the export table ... command, it goes fine. A 
_metadata file will be created, but no data directory, which is what we'd 
expect. If I try to import it back, an exception is thrown due to the lack of 
the data dir:
{code:java}
java.lang.AssertionError: null==getPath() for exim_view
 at org.apache.hadoop.hive.ql.metadata.Hive.loadTable(Hive.java:3088)
 at org.apache.hadoop.hive.ql.exec.MoveTask.execute(MoveTask.java:419)
 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:213)
 at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:105)
 at org.apache.hadoop.hive.ql.Executor.launchTask(Executor.java:364)
 at org.apache.hadoop.hive.ql.Executor.launchTasks(Executor.java:335)
 at org.apache.hadoop.hive.ql.Executor.runTasks(Executor.java:246)
 at org.apache.hadoop.hive.ql.Executor.execute(Executor.java:109)
 at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:722)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:491)
 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:485) 
{code}
Still the view gets imported successfully, as data movement wasn't even 
necessary.

If we try to export a materialized view which is transactional, then this 
exception occurs:
{code:java}
org.apache.hadoop.hive.ql.parse.SemanticException: 
org.apache.hadoop.hive.ql.metadata.InvalidTableException: Table not found 
exim_materialized_view_da21d41a_9fe4_4446_9c72_d251496abf9d
 at 
org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyzeAcidExport(AcidExportSemanticAnalyzer.java:163)
 at 
org.apache.hadoop.hive.ql.parse.AcidExportSemanticAnalyzer.analyze(AcidExportSemanticAnalyzer.java:71)
 at 
org.apache.hadoop.hive.ql.parse.RewriteSemanticAnalyzer.analyzeInternal(RewriteSemanticAnalyzer.java:72)
 at 
org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:289)
 at org.apache.hadoop.hive.ql.Compiler.analyze(Compiler.java:220)
 at org.apache.hadoop.hive.ql.Compiler.compile(Compiler.java:104)
 at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:183)
 at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:601)
 at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:547)
 at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:541) 
{code}
So the export process can not handle it, as the temporary table is not getting 
created.

 

The import command handling have a lot of codes dedicated to importing views 
and materialized views, which suggests that we support the importing (and thus 
also suggests implicitly that we support the exporting) of views and 
materialiezed views.

 

So the conclusion is that we have to decide if we support exporting/importing 
of views and materialized views.

If we decide not to support them then:
 - export process should throw an exception if a view or materialized view is 
the subject
 - the codes specific to view imports should be removed

If we decide to support them, then:
 - the commands mentioned above above should be introduced
 - exception should be thrown if not the proper command used (e.g. export view 
on a table)
 - the exceptions mentioned above should be fixed

I prefer not to support them, I don't think we should support the exporting / 
importing of views. The point of exporting / importing is the transfer of data, 
not DDL, it causes more issues than it solves. Our current documentation also 
suggests that it is only supported for tables.

  was:
According to 
[https://cwiki.apache.org/confluence/display/Hive/LanguageManual+ImportExport] 
import and export can be done by using the
{code:java}
export table ...
import table ... 
{code}
commands. The document doesn't mention views or materialized views at all, and 
in fact we don't support commands like
{code:java}
export view ...
import view ...
export materialized view ...
import materialized view ... 
{code}
they can not be parsed at all. The word table is often used though in a broader 
sense, when it means all table like entities, including views and materialized 
views. For example the various Table classes may represent any of 

[jira] [Updated] (HIVE-23122) LLAP: TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor updated HIVE-23122:

Summary: LLAP: TaskExecutorService should log details about task eviction 
decision details  (was: LLAP TaskExecutorService should log details about task 
eviction decision details)

> LLAP: TaskExecutorService should log details about task eviction decision 
> details
> -
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-23122.02.patch
>
>
> TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
> another. Under the hood, the queue uses a configurable 
> [comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
>  The currently available comparators typically use the following properties 
> of a task(wrapper):
> getWithinDagPriority: related to vertex
> currentAttemptStartTime
> firstAttemptStartTime
> knownPending: remaining upstream tasks
> The problem is, when an eviction happens, the INFO level message doesn't 
> provide any insight about the decision, only attempts ids like below:
> {code}
> attempt_1585248378306_0010_72_02_96_8 evicted from wait queue in favor of 
> attempt_1585248378306_0003_175_02_79_175 because of lower priority
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23124) Review of SQLOperation Class

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23124?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor updated HIVE-23124:
--
Summary: Review of SQLOperation Class  (was: Review of SQLOperation )

> Review of SQLOperation Class
> 
>
> Key: HIVE-23124
> URL: https://issues.apache.org/jira/browse/HIVE-23124
> Project: Hive
>  Issue Type: Improvement
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Major
> Attachments: HIVE-23124.1.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23089) Add constraint checks to CBO plan

2020-04-01 Thread Krisztian Kasa (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23089?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Krisztian Kasa updated HIVE-23089:
--
Attachment: HIVE-23089.1.patch

> Add constraint checks to CBO plan
> -
>
> Key: HIVE-23089
> URL: https://issues.apache.org/jira/browse/HIVE-23089
> Project: Hive
>  Issue Type: Improvement
>  Components: CBO
>Reporter: Krisztian Kasa
>Assignee: Krisztian Kasa
>Priority: Major
> Attachments: HIVE-23089.1.patch
>
>
> {code}
> create table acid_uami(i int,
>  de decimal(5,2) constraint nn1 not null enforced,
>  vc varchar(128) constraint nn2 not null enforced) clustered 
> by (i) into 2 buckets stored as orc TBLPROPERTIES ('transactional'='true');
> explain
> update acid_uami set de=null where i=1;
> {code}
> Non-CBO path:
> {code:java}
> Map Operator Tree:
> TableScan
> alias: acid_uami
> filterExpr: ((i = 1) and enforce_constraint(vc is not null)) 
> (type: boolean)
> Statistics: Num rows: 1 Data size: 216 Basic stats: COMPLETE 
> Column stats: NONE
> Filter Operator
>   predicate: ((i = 1) and enforce_constraint(vc is not null)) 
> (type: boolean)
> {code}
> CBO path:
> {code:java}
> Map Reduce
>   Map Operator Tree:
>   TableScan
> alias: acid_uami
> filterExpr: (i = 1) (type: boolean)
> Statistics: Num rows: 1 Data size: 216 Basic stats: COMPLETE 
> Column stats: NONE
> Filter Operator
>   predicate: (i = 1) (type: boolean)
> ...
>   Reduce Operator Tree:
> ...
>  Filter Operator
> predicate: enforce_constraint((null is not null and _col3 is not 
> null)) (type: boolean)
> {code}
> In CBO path the enforce_constraint function is added to the plan when CBO 
> plan is already generated and optimized.
> {code}
> HiveSortExchange(distribution=[any], collation=[[0]])
>   HiveProject(row__id=[$5], i=[CAST(1):INTEGER], _o__c2=[null:NULL], vc=[$2])
> HiveFilter(condition=[=($0, 1)])
>   HiveTableScan(table=[[default, acid_uami]], table:alias=[acid_uami])
> {code} 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23109) Query-based compaction omits database

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23109?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072809#comment-17072809
 ] 

Hive QA commented on HIVE-23109:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
39s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
49s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
46s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
57s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m 
43s{color} | {color:blue} ql in master has 1528 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
43s{color} | {color:blue} itests/hive-unit in master has 2 extant Findbugs 
warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
21s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
27s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  2m 
 9s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
43s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m 
43s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
40s{color} | {color:green} ql: The patch generated 0 new + 0 unchanged - 1 
fixed = 0 total (was 1) {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
17s{color} | {color:green} The patch hive-unit passed checkstyle {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  4m 
31s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
20s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
15s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 30m  9s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-21370/dev-support/hive-personality.sh
 |
| git revision | master / 484d182 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.1 |
| modules | C: ql itests/hive-unit U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21370/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Query-based compaction omits database
> -
>
> Key: HIVE-23109
> URL: https://issues.apache.org/jira/browse/HIVE-23109
> Project: Hive
>  Issue Type: Bug
>Reporter: Karen Coppage
>Assignee: Karen Coppage
>Priority: Major
> Attachments: HIVE-23109.01.patch, HIVE-23109.02.patch
>
>
> E.g. MM major compaction query looks like:
> {code:java}
> insert into tmp_table select * from src_table;
> {code}
> it should be
> {code:java}
> insert into tmp_table select * from src_db.src_table;
> {code}
> Therefore compaction fails if db of source table isn't default.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23114) Insert overwrite with dynamic partitioning is not working correctly with direct insert

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23114?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072883#comment-17072883
 ] 

Hive QA commented on HIVE-23114:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
42s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
57s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
0s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
44s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m 
45s{color} | {color:blue} ql in master has 1528 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
54s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
29s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
26s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
2s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m  
2s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
44s{color} | {color:red} ql: The patch generated 1 new + 314 unchanged - 0 
fixed = 315 total (was 314) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  3m 
54s{color} | {color:red} ql generated 1 new + 1528 unchanged - 0 fixed = 1529 
total (was 1528) {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
55s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 25m 25s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| FindBugs | module:ql |
|  |  The field 
org.apache.hadoop.hive.ql.exec.FileSinkOperator.dynamicPartitionSpecs is 
transient but isn't set by deserialization  In FileSinkOperator.java:but isn't 
set by deserialization  In FileSinkOperator.java |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-21371/dev-support/hive-personality.sh
 |
| git revision | master / 709235c |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.1 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21371/yetus/diff-checkstyle-ql.txt
 |
| findbugs | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21371/yetus/new-findbugs-ql.html
 |
| modules | C: ql itests U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21371/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Insert overwrite with dynamic partitioning is not working correctly with 
> direct insert
> --
>
> Key: HIVE-23114
> URL: https://issues.apache.org/jira/browse/HIVE-23114
> Project: Hive
>  Issue Type: Bug
>Reporter: Marta Kuczora
>Assignee: Marta Kuczora
>Priority: Major
> Attachments: HIVE-23114.1.patch
>
>
> This is a follow-up Jira for the 
> [conversation|https://issues.apache.org/jira/browse/HIVE-21164?focusedCommentId=17059280=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-17059280]
>  in HIVE-21164
>  Doing an insert overwrite from a multi-insert statement with dynamic 
> partitioning will give wrong results for ACID tables 

[jira] [Updated] (HIVE-23124) Review of SQLOperation Class

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23124?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor updated HIVE-23124:
--
Description: 
* Use ConcurrentHashMap instead of synchronized methods to improve 
multi-threaded access
 * Use JDK 8 facilities where applicable
 * General cleanup
 * Better log messages and Exception messages
 * Use {{switch}} statement instead of if/else blocks
 * Checkstyle fixes

> Review of SQLOperation Class
> 
>
> Key: HIVE-23124
> URL: https://issues.apache.org/jira/browse/HIVE-23124
> Project: Hive
>  Issue Type: Improvement
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Major
> Attachments: HIVE-23124.1.patch
>
>
> * Use ConcurrentHashMap instead of synchronized methods to improve 
> multi-threaded access
>  * Use JDK 8 facilities where applicable
>  * General cleanup
>  * Better log messages and Exception messages
>  * Use {{switch}} statement instead of if/else blocks
>  * Checkstyle fixes



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Resolved] (HIVE-18956) AvroSerDe Race Condition

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-18956?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor resolved HIVE-18956.
---
Resolution: Won't Fix

> AvroSerDe Race Condition
> 
>
> Key: HIVE-18956
> URL: https://issues.apache.org/jira/browse/HIVE-18956
> Project: Hive
>  Issue Type: Bug
>  Components: Serializers/Deserializers
>Affects Versions: 3.0.0, 2.3.2
>Reporter: David Mollitor
>Priority: Trivial
>
> {code}
>   @Override
>   public Writable serialize(Object o, ObjectInspector objectInspector) throws 
> SerDeException {
> if(badSchema) {
>   throw new BadSchemaException();
> }
> return getSerializer().serialize(o, objectInspector, columnNames, 
> columnTypes, schema);
>   }
>   @Override
>   public Object deserialize(Writable writable) throws SerDeException {
> if(badSchema) {
>   throw new BadSchemaException();
> }
> return getDeserializer().deserialize(columnNames, columnTypes, writable, 
> schema);
>   }
> ...
>   private AvroDeserializer getDeserializer() {
> if(avroDeserializer == null) {
>   avroDeserializer = new AvroDeserializer();
> }
> return avroDeserializer;
>   }
>   private AvroSerializer getSerializer() {
> if(avroSerializer == null) {
>   avroSerializer = new AvroSerializer();
> }
> return avroSerializer;
>   }
> {code}
> {{getDeserializer}} and {{getSerializer}} methods are not thread safe, so 
> neither are {{deserialize}} and {{serialize}} methods.  It probably didn't 
> matter with MapReduce, but now that we have Spark/Tez, it may be an issue.
> You could visualize a scenario where three threads all enter 
> {{getSerializer}} and all see that {{avroSerializer}} is _null_ and create 
> three instances, then they would fight to assign the new object to the 
> {{avroSerializer}} variable.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23122) LLAP TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor updated HIVE-23122:

Attachment: HIVE-23122.02.patch

> LLAP TaskExecutorService should log details about task eviction decision 
> details
> 
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-23122.02.patch
>
>
> TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
> another. Under the hood, the queue uses a configurable 
> [comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
>  The currently available comparators typically use the following properties 
> of a task(wrapper):
> getWithinDagPriority: related to vertex
> currentAttemptStartTime
> firstAttemptStartTime
> knownPending: remaining upstream tasks
> The problem is, when an eviction happens, the INFO level message doesn't 
> provide any insight about the decision, only attempts ids like below:
> {code}
> attempt_1585248378306_0010_72_02_96_8 evicted from wait queue in favor of 
> attempt_1585248378306_0003_175_02_79_175 because of lower priority
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23122) LLAP TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor updated HIVE-23122:

Attachment: (was: HIVE-23122.01.patch)

> LLAP TaskExecutorService should log details about task eviction decision 
> details
> 
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-23122.02.patch
>
>
> TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
> another. Under the hood, the queue uses a configurable 
> [comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
>  The currently available comparators typically use the following properties 
> of a task(wrapper):
> getWithinDagPriority: related to vertex
> currentAttemptStartTime
> firstAttemptStartTime
> knownPending: remaining upstream tasks
> The problem is, when an eviction happens, the INFO level message doesn't 
> provide any insight about the decision, only attempts ids like below:
> {code}
> attempt_1585248378306_0010_72_02_96_8 evicted from wait queue in favor of 
> attempt_1585248378306_0003_175_02_79_175 because of lower priority
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23109) Query-based compaction omits database

2020-04-01 Thread Jira


[ 
https://issues.apache.org/jira/browse/HIVE-23109?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072869#comment-17072869
 ] 

László Pintér commented on HIVE-23109:
--

+1

> Query-based compaction omits database
> -
>
> Key: HIVE-23109
> URL: https://issues.apache.org/jira/browse/HIVE-23109
> Project: Hive
>  Issue Type: Bug
>Reporter: Karen Coppage
>Assignee: Karen Coppage
>Priority: Major
> Attachments: HIVE-23109.01.patch, HIVE-23109.02.patch
>
>
> E.g. MM major compaction query looks like:
> {code:java}
> insert into tmp_table select * from src_table;
> {code}
> it should be
> {code:java}
> insert into tmp_table select * from src_db.src_table;
> {code}
> Therefore compaction fails if db of source table isn't default.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23122) LLAP: TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Panagiotis Garefalakis (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072785#comment-17072785
 ] 

Panagiotis Garefalakis commented on HIVE-23122:
---

Thanks [~abstractdog]

+1 tests pending

> LLAP: TaskExecutorService should log details about task eviction decision 
> details
> -
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-23122.02.patch
>
>
> TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
> another. Under the hood, the queue uses a configurable 
> [comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
>  The currently available comparators typically use the following properties 
> of a task(wrapper):
> getWithinDagPriority: related to vertex
> currentAttemptStartTime
> firstAttemptStartTime
> knownPending: remaining upstream tasks
> The problem is, when an eviction happens, the INFO level message doesn't 
> provide any insight about the decision, only attempts ids like below:
> {code}
> attempt_1585248378306_0010_72_02_96_8 evicted from wait queue in favor of 
> attempt_1585248378306_0003_175_02_79_175 because of lower priority
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23089) Add constraint checks to CBO plan

2020-04-01 Thread Krisztian Kasa (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23089?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Krisztian Kasa updated HIVE-23089:
--
Status: Patch Available  (was: Open)

> Add constraint checks to CBO plan
> -
>
> Key: HIVE-23089
> URL: https://issues.apache.org/jira/browse/HIVE-23089
> Project: Hive
>  Issue Type: Improvement
>  Components: CBO
>Reporter: Krisztian Kasa
>Assignee: Krisztian Kasa
>Priority: Major
> Attachments: HIVE-23089.1.patch
>
>
> {code}
> create table acid_uami(i int,
>  de decimal(5,2) constraint nn1 not null enforced,
>  vc varchar(128) constraint nn2 not null enforced) clustered 
> by (i) into 2 buckets stored as orc TBLPROPERTIES ('transactional'='true');
> explain
> update acid_uami set de=null where i=1;
> {code}
> Non-CBO path:
> {code:java}
> Map Operator Tree:
> TableScan
> alias: acid_uami
> filterExpr: ((i = 1) and enforce_constraint(vc is not null)) 
> (type: boolean)
> Statistics: Num rows: 1 Data size: 216 Basic stats: COMPLETE 
> Column stats: NONE
> Filter Operator
>   predicate: ((i = 1) and enforce_constraint(vc is not null)) 
> (type: boolean)
> {code}
> CBO path:
> {code:java}
> Map Reduce
>   Map Operator Tree:
>   TableScan
> alias: acid_uami
> filterExpr: (i = 1) (type: boolean)
> Statistics: Num rows: 1 Data size: 216 Basic stats: COMPLETE 
> Column stats: NONE
> Filter Operator
>   predicate: (i = 1) (type: boolean)
> ...
>   Reduce Operator Tree:
> ...
>  Filter Operator
> predicate: enforce_constraint((null is not null and _col3 is not 
> null)) (type: boolean)
> {code}
> In CBO path the enforce_constraint function is added to the plan when CBO 
> plan is already generated and optimized.
> {code}
> HiveSortExchange(distribution=[any], collation=[[0]])
>   HiveProject(row__id=[$5], i=[CAST(1):INTEGER], _o__c2=[null:NULL], vc=[$2])
> HiveFilter(condition=[=($0, 1)])
>   HiveTableScan(table=[[default, acid_uami]], table:alias=[acid_uami])
> {code} 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (HIVE-23124) Review of SQLOperation

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23124?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor reassigned HIVE-23124:
-

Assignee: David Mollitor

> Review of SQLOperation 
> ---
>
> Key: HIVE-23124
> URL: https://issues.apache.org/jira/browse/HIVE-23124
> Project: Hive
>  Issue Type: Improvement
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Major
> Attachments: HIVE-23124.1.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23124) Review of SQLOperation Class

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23124?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor updated HIVE-23124:
--
Attachment: HIVE-23124.1.patch

> Review of SQLOperation Class
> 
>
> Key: HIVE-23124
> URL: https://issues.apache.org/jira/browse/HIVE-23124
> Project: Hive
>  Issue Type: Improvement
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Major
> Attachments: HIVE-23124.1.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23124) Review of SQLOperation Class

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23124?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor updated HIVE-23124:
--
Status: Patch Available  (was: Open)

> Review of SQLOperation Class
> 
>
> Key: HIVE-23124
> URL: https://issues.apache.org/jira/browse/HIVE-23124
> Project: Hive
>  Issue Type: Improvement
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Major
> Attachments: HIVE-23124.1.patch
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23122) LLAP TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Jira


[ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072783#comment-17072783
 ] 

László Bodor commented on HIVE-23122:
-

thanks for the comments [~pgaref], included canFinishForPriority
removing 01.patch for saving ptest time

> LLAP TaskExecutorService should log details about task eviction decision 
> details
> 
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-23122.02.patch
>
>
> TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
> another. Under the hood, the queue uses a configurable 
> [comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
>  The currently available comparators typically use the following properties 
> of a task(wrapper):
> getWithinDagPriority: related to vertex
> currentAttemptStartTime
> firstAttemptStartTime
> knownPending: remaining upstream tasks
> The problem is, when an eviction happens, the INFO level message doesn't 
> provide any insight about the decision, only attempts ids like below:
> {code}
> attempt_1585248378306_0010_72_02_96_8 evicted from wait queue in favor of 
> attempt_1585248378306_0003_175_02_79_175 because of lower priority
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23122) LLAP: TaskExecutorService should log details about task eviction decision details

2020-04-01 Thread Jira


[ 
https://issues.apache.org/jira/browse/HIVE-23122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072821#comment-17072821
 ] 

László Bodor commented on HIVE-23122:
-

cc: [~rajesh.balamohan]

> LLAP: TaskExecutorService should log details about task eviction decision 
> details
> -
>
> Key: HIVE-23122
> URL: https://issues.apache.org/jira/browse/HIVE-23122
> Project: Hive
>  Issue Type: Improvement
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
> Attachments: HIVE-23122.02.patch
>
>
> TaskExecutorService maintains a waitQueue, and can evict a task in favor of 
> another. Under the hood, the queue uses a configurable 
> [comparator|https://github.com/apache/hive/tree/master/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/comparator].
>  The currently available comparators typically use the following properties 
> of a task(wrapper):
> getWithinDagPriority: related to vertex
> currentAttemptStartTime
> firstAttemptStartTime
> knownPending: remaining upstream tasks
> The problem is, when an eviction happens, the INFO level message doesn't 
> provide any insight about the decision, only attempts ids like below:
> {code}
> attempt_1585248378306_0010_72_02_96_8 evicted from wait queue in favor of 
> attempt_1585248378306_0003_175_02_79_175 because of lower priority
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23114) Insert overwrite with dynamic partitioning is not working correctly with direct insert

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23114?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17072955#comment-17072955
 ] 

Hive QA commented on HIVE-23114:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12998356/HIVE-23114.1.patch

{color:green}SUCCESS:{color} +1 due to 2 test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 18166 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/21371/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/21371/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-21371/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12998356 - PreCommit-HIVE-Build

> Insert overwrite with dynamic partitioning is not working correctly with 
> direct insert
> --
>
> Key: HIVE-23114
> URL: https://issues.apache.org/jira/browse/HIVE-23114
> Project: Hive
>  Issue Type: Bug
>Reporter: Marta Kuczora
>Assignee: Marta Kuczora
>Priority: Major
> Attachments: HIVE-23114.1.patch
>
>
> This is a follow-up Jira for the 
> [conversation|https://issues.apache.org/jira/browse/HIVE-21164?focusedCommentId=17059280=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-17059280]
>  in HIVE-21164
>  Doing an insert overwrite from a multi-insert statement with dynamic 
> partitioning will give wrong results for ACID tables when 
> 'hive.acid.direct.insert.enabled' is true or for insert-only tables.
> Reproduction:
> {noformat}
> set hive.acid.direct.insert.enabled=true;
> set hive.support.concurrency=true;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
> set hive.vectorized.execution.enabled=false;
> set hive.stats.autogather=false;
> create external table multiinsert_test_text (a int, b int, c int) stored as 
> textfile;
> insert into multiinsert_test_text values (, 11, ), (, 22, ), 
> (, 33, ), (, 44, NULL), (, 55, NULL);
> create table multiinsert_test_acid (a int, b int) partitioned by (c int) 
> stored as orc tblproperties('transactional'='true');
> create table multiinsert_test_mm (a int, b int) partitioned by (c int) stored 
> as orc tblproperties('transactional'='true', 
> 'transactional_properties'='insert_only');
> from multiinsert_test_text a
> insert overwrite table multiinsert_test_acid partition (c)
> select
>  a.a,
>  a.b,
>  a.c
>  where a.c is not null
> insert overwrite table multiinsert_test_acid partition (c)
> select
>  a.a,
>  a.b,
>  a.c
> where a.c is null;
> select * from multiinsert_test_acid;
> from multiinsert_test_text a
> insert overwrite table multiinsert_test_mm partition (c)
> select
>  a.a,
>  a.b,
>  a.c
>  where a.c is not null
> insert overwrite table multiinsert_test_mm partition (c)
> select
>  a.a,
>  a.b,
>  a.c
> where a.c is null;
> select * from multiinsert_test_mm;
> {noformat}
> The result of these steps can be different, it depends on the execution order 
> of the FileSinkOperators of the insert overwrite statements. It can happen 
> that an error occurs due to manifest file collision, it can happen that no 
> error occurs but the result will be incorrect.
>  Running the same insert query with an external table of with and ACID table 
> with 'hive.acid.direct.insert.enabled=false' will give the follwing result:
> {noformat}
> 11  
> 22  
> 33  
> 44  NULL
> 55  NULL
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23039) Checkpointing for repl dump bootstrap phase

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23039?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073001#comment-17073001
 ] 

Hive QA commented on HIVE-23039:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m 
46s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
 7s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
44s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
 8s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m 
44s{color} | {color:blue} ql in master has 1528 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
40s{color} | {color:blue} itests/hive-unit in master has 2 extant Findbugs 
warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
22s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
29s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  2m 
 8s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
47s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m 
47s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
40s{color} | {color:red} ql: The patch generated 9 new + 59 unchanged - 0 fixed 
= 68 total (was 59) {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
23s{color} | {color:red} itests/hive-unit: The patch generated 2 new + 688 
unchanged - 0 fixed = 690 total (was 688) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  4m 
41s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
24s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 31m  9s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-21372/dev-support/hive-personality.sh
 |
| git revision | master / 709235c |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.1 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21372/yetus/diff-checkstyle-ql.txt
 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21372/yetus/diff-checkstyle-itests_hive-unit.txt
 |
| modules | C: ql itests/hive-unit U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21372/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Checkpointing for repl dump bootstrap phase
> ---
>
> Key: HIVE-23039
> URL: https://issues.apache.org/jira/browse/HIVE-23039
> Project: Hive
>  Issue Type: Bug
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23039.01.patch, HIVE-23039.02.patch
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23067) Use batch DB calls in TxnHandler for commitTxn and abortTxns

2020-04-01 Thread Marton Bod (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23067?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Marton Bod updated HIVE-23067:
--
Attachment: (was: HIVE-23067.3.patch)

> Use batch DB calls in TxnHandler for commitTxn and abortTxns
> 
>
> Key: HIVE-23067
> URL: https://issues.apache.org/jira/browse/HIVE-23067
> Project: Hive
>  Issue Type: Improvement
>Reporter: Marton Bod
>Assignee: Marton Bod
>Priority: Major
> Attachments: HIVE-23067.1.patch, HIVE-23067.1.patch, 
> HIVE-23067.1.patch, HIVE-23067.1.patch, HIVE-23067.2.patch, 
> HIVE-23067.2.patch, HIVE-23067.4.patch
>
>
> To reduce the number of database calls and network roundtrips, we could use 
> more batching in TxnHandler, where currently in many places we call insert 
> commands in loops sequentially.
> Some examples:
>  * openTxns (TXNS, REPL_TXN_MAP)
>  * commitTxn (COMPLETED_TXN_COMPONENTS)
>  * replTableWriteIdState (TXN_TO_WRITE_ID)
>  * allocateTableWriteIds (TXN_TO_WRITE_ID)
>  * 
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23067) Use batch DB calls in TxnHandler for commitTxn and abortTxns

2020-04-01 Thread Marton Bod (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23067?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Marton Bod updated HIVE-23067:
--
Attachment: HIVE-23067.4.patch

> Use batch DB calls in TxnHandler for commitTxn and abortTxns
> 
>
> Key: HIVE-23067
> URL: https://issues.apache.org/jira/browse/HIVE-23067
> Project: Hive
>  Issue Type: Improvement
>Reporter: Marton Bod
>Assignee: Marton Bod
>Priority: Major
> Attachments: HIVE-23067.1.patch, HIVE-23067.1.patch, 
> HIVE-23067.1.patch, HIVE-23067.1.patch, HIVE-23067.2.patch, 
> HIVE-23067.2.patch, HIVE-23067.4.patch
>
>
> To reduce the number of database calls and network roundtrips, we could use 
> more batching in TxnHandler, where currently in many places we call insert 
> commands in loops sequentially.
> Some examples:
>  * openTxns (TXNS, REPL_TXN_MAP)
>  * commitTxn (COMPLETED_TXN_COMPONENTS)
>  * replTableWriteIdState (TXN_TO_WRITE_ID)
>  * allocateTableWriteIds (TXN_TO_WRITE_ID)
>  * 
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Resolved] (HIVE-18897) Hive is Double-Logging Invalid UDF Error

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-18897?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor resolved HIVE-18897.
---
Resolution: Not A Problem

Based on the current master branch, it looks like the logging was fixed to 
remove the ERROR log in the {{Driver}} class.

> Hive is Double-Logging Invalid UDF Error
> 
>
> Key: HIVE-18897
> URL: https://issues.apache.org/jira/browse/HIVE-18897
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2
>Affects Versions: 3.0.0, 2.4.0
>Reporter: David Mollitor
>Assignee: Ramesh Kumar Thangarajan
>Priority: Minor
>  Labels: newbie, noob
>
> It logs the "invalid" function twice.  Once at ERROR level and once at WARN 
> level.  Please change so that the error is logged once, at the WARN level.  
> The stack trace seems overkill here as well for such a trivial error... 
> usually a user typo or the function needs to be registered.
> {code:java}
> 2018-03-05 07:50:44,473  ERROR org.apache.hadoop.hive.ql.Driver: 
> [HiveServer2-Handler-Pool: Thread-43]: FAILED: SemanticException [Error 
> 10011]: Line 1:7 Invalid function 'aes_encrypt'
> org.apache.hadoop.hive.ql.parse.SemanticException: Line 1:7 Invalid function 
> 'aes_encrypt'
>   at 
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$DefaultExprProcessor.getXpathOrFuncExprNodeDesc(TypeCheckProcFactory.java:836)
>   at 
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory$DefaultExprProcessor.process(TypeCheckProcFactory.java:1176)
>   at 
> org.apache.hadoop.hive.ql.lib.DefaultRuleDispatcher.dispatch(DefaultRuleDispatcher.java:90)
>   at 
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatchAndReturn(DefaultGraphWalker.java:94)
>   at 
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.dispatch(DefaultGraphWalker.java:78)
>   at 
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.walk(DefaultGraphWalker.java:132)
>   at 
> org.apache.hadoop.hive.ql.lib.DefaultGraphWalker.startWalking(DefaultGraphWalker.java:109)
>   at 
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory.genExprNode(TypeCheckProcFactory.java:193)
>   at 
> org.apache.hadoop.hive.ql.parse.TypeCheckProcFactory.genExprNode(TypeCheckProcFactory.java:146)
>   at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genAllExprNodeDesc(SemanticAnalyzer.java:10422)
>   at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:10378)
>   at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genSelectPlan(SemanticAnalyzer.java:3771)
>   at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genSelectPlan(SemanticAnalyzer.java:3550)
>   at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPostGroupByBodyPlan(SemanticAnalyzer.java:8830)
>   at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:8785)
>   at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:9652)
>   at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:9545)
>   at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genOPTree(SemanticAnalyzer.java:10018)
>   at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:10029)
>   at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:9909)
>   at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:223)
>   at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:488)
>   at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1274)
>   at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:1261)
>   at 
> org.apache.hive.service.cli.operation.SQLOperation.prepare(SQLOperation.java:143)
>   at 
> org.apache.hive.service.cli.operation.SQLOperation.runInternal(SQLOperation.java:215)
>   at 
> org.apache.hive.service.cli.operation.Operation.run(Operation.java:337)
>   at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:425)
>   at 
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementAsync(HiveSessionImpl.java:402)
>   at 
> org.apache.hive.service.cli.CLIService.executeStatementAsync(CLIService.java:258)
>   at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:500)
>   at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1313)
>   at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1298)
>   at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>   at 

[jira] [Commented] (HIVE-23106) Cleanup CalcitePlanner genOPTree exception handling

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23106?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073189#comment-17073189
 ] 

Hive QA commented on HIVE-23106:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12998377/HIVE-23106.2.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 18162 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/21374/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/21374/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-21374/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12998377 - PreCommit-HIVE-Build

> Cleanup CalcitePlanner genOPTree exception handling
> ---
>
> Key: HIVE-23106
> URL: https://issues.apache.org/jira/browse/HIVE-23106
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2
>Reporter: John Sherman
>Assignee: John Sherman
>Priority: Minor
> Attachments: HIVE-23106.1.patch, HIVE-23106.2.patch
>
>
> The logic where genOPTree handles exceptions during CBO planning is a bit 
> twisty and could use some cleanup and comments.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23126) Add TXNS table to sysdb

2020-04-01 Thread Ramesh Kumar Thangarajan (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23126?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ramesh Kumar Thangarajan updated HIVE-23126:

Attachment: HIVE-23126.1.patch
Status: Patch Available  (was: Open)

> Add TXNS table to sysdb
> ---
>
> Key: HIVE-23126
> URL: https://issues.apache.org/jira/browse/HIVE-23126
> Project: Hive
>  Issue Type: Task
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
> Attachments: HIVE-23126.1.patch
>
>
> Add TXNS table to sysdb



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23106) Cleanup CalcitePlanner genOPTree exception handling

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23106?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073159#comment-17073159
 ] 

Hive QA commented on HIVE-23106:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  9m 
26s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
1s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
40s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m 
43s{color} | {color:blue} ql in master has 1528 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
55s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
28s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
2s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m  
2s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
42s{color} | {color:green} ql: The patch generated 0 new + 115 unchanged - 7 
fixed = 115 total (was 122) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  4m  
0s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
56s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
15s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 24m 39s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-21374/dev-support/hive-personality.sh
 |
| git revision | master / 709235c |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.1 |
| modules | C: ql U: ql |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21374/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Cleanup CalcitePlanner genOPTree exception handling
> ---
>
> Key: HIVE-23106
> URL: https://issues.apache.org/jira/browse/HIVE-23106
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2
>Reporter: John Sherman
>Assignee: John Sherman
>Priority: Minor
> Attachments: HIVE-23106.1.patch, HIVE-23106.2.patch
>
>
> The logic where genOPTree handles exceptions during CBO planning is a bit 
> twisty and could use some cleanup and comments.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23094) Implement Explain CBO of Update and Delete statements

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23094?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073141#comment-17073141
 ] 

Hive QA commented on HIVE-23094:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12998369/HIVE-23094.1.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 18162 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/21373/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/21373/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-21373/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12998369 - PreCommit-HIVE-Build

> Implement Explain CBO of Update and Delete statements
> -
>
> Key: HIVE-23094
> URL: https://issues.apache.org/jira/browse/HIVE-23094
> Project: Hive
>  Issue Type: Improvement
>  Components: CBO
>Reporter: Krisztian Kasa
>Assignee: Krisztian Kasa
>Priority: Major
> Attachments: HIVE-23094.1.patch
>
>
> {code}
> set hive.support.concurrency=true;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
> create table acidtlb(a int, b int) clustered by (a) into 2 buckets stored as 
> orc TBLPROPERTIES ('transactional'='true');
> explain cbo
> update acidtlb set b=777;
> {code}
> doesn't print CBO plan.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23039) Checkpointing for repl dump bootstrap phase

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23039?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073058#comment-17073058
 ] 

Hive QA commented on HIVE-23039:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12998454/HIVE-23039.02.patch

{color:green}SUCCESS:{color} +1 due to 3 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 8 failed/errored test(s), 18168 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.ql.parse.TestReplWithJsonMessageFormat.testCheckPointing 
(batchId=260)
org.apache.hadoop.hive.ql.parse.TestReplWithJsonMessageFormat.testCheckPointingInDumpFailure
 (batchId=260)
org.apache.hadoop.hive.ql.parse.TestReplicationOnHDFSEncryptedZones.targetAndSourceHaveDifferentEncryptionZoneKeys
 (batchId=283)
org.apache.hadoop.hive.ql.parse.TestReplicationScenarios.testCheckPointing 
(batchId=270)
org.apache.hadoop.hive.ql.parse.TestReplicationScenarios.testCheckPointingInDumpFailure
 (batchId=270)
org.apache.hadoop.hive.ql.parse.TestReplicationScenariosExternalTables.externalTableIncrementalReplication
 (batchId=268)
org.apache.hadoop.hive.ql.parse.TestReplicationScenariosExternalTables.externalTableReplicationWithCustomPaths
 (batchId=268)
org.apache.hadoop.hive.ql.parse.TestReplicationScenariosExternalTables.externalTableWithPartitions
 (batchId=268)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/21372/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/21372/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-21372/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 8 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12998454 - PreCommit-HIVE-Build

> Checkpointing for repl dump bootstrap phase
> ---
>
> Key: HIVE-23039
> URL: https://issues.apache.org/jira/browse/HIVE-23039
> Project: Hive
>  Issue Type: Bug
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23039.01.patch, HIVE-23039.02.patch, 
> HIVE-23039.03.patch
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23039) Checkpointing for repl dump bootstrap phase

2020-04-01 Thread Aasha Medhi (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23039?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aasha Medhi updated HIVE-23039:
---
Attachment: HIVE-23039.03.patch
Status: Patch Available  (was: In Progress)

> Checkpointing for repl dump bootstrap phase
> ---
>
> Key: HIVE-23039
> URL: https://issues.apache.org/jira/browse/HIVE-23039
> Project: Hive
>  Issue Type: Bug
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23039.01.patch, HIVE-23039.02.patch, 
> HIVE-23039.03.patch
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (HIVE-23126) Add TXNS table to sysdb

2020-04-01 Thread Ramesh Kumar Thangarajan (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23126?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ramesh Kumar Thangarajan reassigned HIVE-23126:
---


> Add TXNS table to sysdb
> ---
>
> Key: HIVE-23126
> URL: https://issues.apache.org/jira/browse/HIVE-23126
> Project: Hive
>  Issue Type: Task
>Reporter: Ramesh Kumar Thangarajan
>Assignee: Ramesh Kumar Thangarajan
>Priority: Major
>
> Add TXNS table to sysdb



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-23039) Checkpointing for repl dump bootstrap phase

2020-04-01 Thread Aasha Medhi (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-23039?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aasha Medhi updated HIVE-23039:
---
Status: In Progress  (was: Patch Available)

> Checkpointing for repl dump bootstrap phase
> ---
>
> Key: HIVE-23039
> URL: https://issues.apache.org/jira/browse/HIVE-23039
> Project: Hive
>  Issue Type: Bug
>Reporter: Aasha Medhi
>Assignee: Aasha Medhi
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-23039.01.patch, HIVE-23039.02.patch
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (HIVE-23094) Implement Explain CBO of Update and Delete statements

2020-04-01 Thread Hive QA (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-23094?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17073085#comment-17073085
 ] 

Hive QA commented on HIVE-23094:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  9m 
13s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
1s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
39s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m 
45s{color} | {color:blue} ql in master has 1528 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
53s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
28s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
59s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
59s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
41s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
53s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
54s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 24m 13s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.43-2+deb8u5 (2017-09-19) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-21373/dev-support/hive-personality.sh
 |
| git revision | master / 709235c |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.1 |
| modules | C: ql U: ql |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-21373/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Implement Explain CBO of Update and Delete statements
> -
>
> Key: HIVE-23094
> URL: https://issues.apache.org/jira/browse/HIVE-23094
> Project: Hive
>  Issue Type: Improvement
>  Components: CBO
>Reporter: Krisztian Kasa
>Assignee: Krisztian Kasa
>Priority: Major
> Attachments: HIVE-23094.1.patch
>
>
> {code}
> set hive.support.concurrency=true;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
> create table acidtlb(a int, b int) clustered by (a) into 2 buckets stored as 
> orc TBLPROPERTIES ('transactional'='true');
> explain cbo
> update acidtlb set b=777;
> {code}
> doesn't print CBO plan.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (HIVE-21328) Call To Hadoop Text getBytes() Without Call to getLength()

2020-04-01 Thread David Mollitor (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-21328?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Mollitor updated HIVE-21328:
--
Attachment: HIVE-21328.1.patch

> Call To Hadoop Text getBytes() Without Call to getLength()
> --
>
> Key: HIVE-21328
> URL: https://issues.apache.org/jira/browse/HIVE-21328
> Project: Hive
>  Issue Type: Bug
>  Components: Query Planning
>Affects Versions: 4.0.0, 3.2.0
>Reporter: David Mollitor
>Assignee: David Mollitor
>Priority: Major
> Attachments: HIVE-21328.1.patch
>
>
> I'm not sure if there is actually a bug, but this looks highly suspect:
> {code:java}
>   public Object set(final Object o, final Text text) {
> return new BytesWritable(text == null ? null : text.getBytes());
>   }
> {code}
> https://github.com/apache/hive/blob/master/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/serde/primitive/ParquetStringInspector.java#L104-L106
> There are two components to a Text object.  There are the internal bytes and 
> the length of the bytes.  The two are independent.  I.e., a quick "reset" on 
> the Text object simply sets the internal length counter to zero.  This code 
> is potentially looking at obsolete data that it shouldn't be seeing because 
> it is not considering the length of the Text.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


  1   2   >