[jira] [Commented] (HIVE-16084) SHOW COMPACTIONS should display CompactionID

2017-09-20 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-16084?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174311#comment-16174311
 ] 

Hive QA commented on HIVE-16084:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12888093/HIVE-16084.02.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 9 failed/errored test(s), 11047 tests 
executed
*Failed tests:*
{noformat}
TestAccumuloCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=230)
TestDummy - did not produce a TEST-*.xml file (likely timed out) (batchId=230)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[create_view] (batchId=39)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[udf_mask_hash] 
(batchId=28)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[union_fast_stats]
 (batchId=156)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_explainuser_1]
 (batchId=170)
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver[explainanalyze_2] 
(batchId=100)
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[drop_table_failure2]
 (batchId=89)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query14] 
(batchId=234)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6914/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6914/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6914/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 9 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12888093 - PreCommit-HIVE-Build

> SHOW COMPACTIONS should display CompactionID
> 
>
> Key: HIVE-16084
> URL: https://issues.apache.org/jira/browse/HIVE-16084
> Project: Hive
>  Issue Type: Improvement
>  Components: Transactions
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
>Priority: Minor
> Attachments: HIVE-16084.01.patch, HIVE-16084.02.patch
>
>
> together with HIVE-13353 it will let users search for specific job



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17426) Execution framework in hive to run tasks in parallel

2017-09-20 Thread Thejas M Nair (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17426?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174305#comment-16174305
 ] 

Thejas M Nair commented on HIVE-17426:
--

I think we should make something executable only if it has been cleared to run 
in parallel. That way its going to be more resilient to bugs caused by future 
changes.

+1 since other comments have been addressed.


> Execution framework in hive to run tasks in parallel
> 
>
> Key: HIVE-17426
> URL: https://issues.apache.org/jira/browse/HIVE-17426
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2
>Affects Versions: 3.0.0
>Reporter: anishek
>Assignee: anishek
> Fix For: 3.0.0
>
> Attachments: HIVE-17426.0.patch, HIVE-17426.1.patch, 
> HIVE-17426.2.patch, HIVE-17426.3.patch, HIVE-17426.4.patch, HIVE-17426.5.patch
>
>
> the execution framework currently only runs MR / Spark  Tasks in parallel 
> when {{set hive.exec.parallel=true}}.
> Allow other types of tasks to run in parallel as well to support replication 
> scenarios in hive. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17568) HiveJoinPushTransitivePredicatesRule may exchange predicates which are not valid on the other branch

2017-09-20 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17568?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174261#comment-16174261
 ] 

Hive QA commented on HIVE-17568:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12888072/HIVE-17568.01.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 171 failed/errored test(s), 11047 tests 
executed
*Failed tests:*
{noformat}
TestAccumuloCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=230)
TestDummy - did not produce a TEST-*.xml file (likely timed out) (batchId=230)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[annotate_stats_join_pkfk]
 (batchId=14)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[auto_join11] (batchId=8)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[auto_join12] (batchId=23)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[auto_join13] (batchId=77)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[auto_join14] (batchId=14)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[auto_join16] (batchId=38)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[auto_join27] (batchId=86)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[auto_join4] (batchId=68)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[auto_join5] (batchId=70)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[auto_join8] (batchId=83)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[auto_join_without_localtask]
 (batchId=1)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[cbo_const] (batchId=17)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[cbo_rp_auto_join1] 
(batchId=3)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[cbo_rp_outer_join_ppr] 
(batchId=7)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[comments] (batchId=36)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[correlationoptimizer10] 
(batchId=74)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[correlationoptimizer13] 
(batchId=10)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[correlationoptimizer8] 
(batchId=12)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[correlationoptimizer9] 
(batchId=6)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[create_view] (batchId=39)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[cross_join_merge] 
(batchId=6)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[empty_join] (batchId=76)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[filter_cond_pushdown] 
(batchId=58)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[filter_join_breaktask] 
(batchId=72)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[fouter_join_ppr] 
(batchId=31)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[gby_star] (batchId=16)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[groupby_position] 
(batchId=38)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[index_auto_mult_tables] 
(batchId=82)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[index_auto_mult_tables_compact]
 (batchId=34)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[join11] (batchId=51)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[join12] (batchId=39)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[join13] (batchId=3)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[join14] (batchId=20)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[join16] (batchId=29)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[join34] (batchId=68)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[join35] (batchId=63)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[join38] (batchId=77)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[join4] (batchId=62)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[join5] (batchId=67)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[join8] (batchId=46)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[join_vc] (batchId=4)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[louter_join_ppr] 
(batchId=42)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[mapjoin_mapjoin] 
(batchId=48)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[mergejoins] (batchId=2)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[mergejoins_mixed] 
(batchId=69)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[outer_join_ppr] 
(batchId=19)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[ppd_gby_join] 
(batchId=35)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[ppd_join2] (batchId=40)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[ppd_join3] (batchId=18)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[ppd_join] (batchId=35)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[ppd_outer_join1] 
(batchId=41)

[jira] [Updated] (HIVE-17542) Make HoS CombineEquivalentWorkResolver Configurable

2017-09-20 Thread Sahil Takiar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17542?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sahil Takiar updated HIVE-17542:

   Resolution: Fixed
Fix Version/s: 3.0.0
   Status: Resolved  (was: Patch Available)

> Make HoS CombineEquivalentWorkResolver Configurable
> ---
>
> Key: HIVE-17542
> URL: https://issues.apache.org/jira/browse/HIVE-17542
> Project: Hive
>  Issue Type: Improvement
>  Components: Physical Optimizer, Spark
>Reporter: Sahil Takiar
>Assignee: Sahil Takiar
> Fix For: 3.0.0
>
> Attachments: HIVE-17542.1.patch, HIVE-17542.2.patch, 
> HIVE-17542.3.patch
>
>
> The {{CombineEquivalentWorkResolver}} is run by default. We should make it 
> configurable so that users can disable it in case there are any issues. We 
> can enable it by default to preserve backwards compatibility.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17112) Reduce logging in HiveSparkClientFactory and RemoteHiveSparkClient

2017-09-20 Thread Sahil Takiar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17112?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sahil Takiar updated HIVE-17112:

Resolution: Fixed
Status: Resolved  (was: Patch Available)

Committed to master.

> Reduce logging in HiveSparkClientFactory and RemoteHiveSparkClient
> --
>
> Key: HIVE-17112
> URL: https://issues.apache.org/jira/browse/HIVE-17112
> Project: Hive
>  Issue Type: Improvement
>  Components: Spark
>Reporter: Sahil Takiar
>Assignee: Sahil Takiar
> Attachments: HIVE-17112.1.patch
>
>
> HiveSparkClientFactory has the following line that introduces excess logging:
> {code}
> LOG.info(String.format(
>   "load spark property from %s (%s -> %s).",
>   SPARK_DEFAULT_CONF_FILE, propertyName, 
> LogUtils.maskIfPassword(propertyName,value)));
> {code}
> It basically dumps the entire configuration object to the logs, we can 
> probably change this from INFO to DEBUG.
> Same thing happens in {{RemoteHiveSparkClient#logConfigurations}}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17112) Reduce logging in HiveSparkClientFactory and RemoteHiveSparkClient

2017-09-20 Thread Sahil Takiar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17112?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sahil Takiar updated HIVE-17112:

Fix Version/s: 3.0.0

> Reduce logging in HiveSparkClientFactory and RemoteHiveSparkClient
> --
>
> Key: HIVE-17112
> URL: https://issues.apache.org/jira/browse/HIVE-17112
> Project: Hive
>  Issue Type: Improvement
>  Components: Spark
>Reporter: Sahil Takiar
>Assignee: Sahil Takiar
> Fix For: 3.0.0
>
> Attachments: HIVE-17112.1.patch
>
>
> HiveSparkClientFactory has the following line that introduces excess logging:
> {code}
> LOG.info(String.format(
>   "load spark property from %s (%s -> %s).",
>   SPARK_DEFAULT_CONF_FILE, propertyName, 
> LogUtils.maskIfPassword(propertyName,value)));
> {code}
> It basically dumps the entire configuration object to the logs, we can 
> probably change this from INFO to DEBUG.
> Same thing happens in {{RemoteHiveSparkClient#logConfigurations}}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-14813) Make TransactionBatchImpl.toString() include state of each txn: commit/abort

2017-09-20 Thread Alan Gates (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-14813?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174249#comment-16174249
 ] 

Alan Gates commented on HIVE-14813:
---

If you have to go to the object level anyway why not use an enum that will be 
meaningful to maintainers rather than overloading true/false/null values?  Not 
crucial, just curious.

+1, even if you don't decide to switch the array.

> Make TransactionBatchImpl.toString() include state of each txn: commit/abort
> 
>
> Key: HIVE-14813
> URL: https://issues.apache.org/jira/browse/HIVE-14813
> Project: Hive
>  Issue Type: Bug
>  Components: HCatalog, Transactions
>Affects Versions: 1.2.1
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
> Attachments: HIVE-14813.01.patch
>
>
> perhaps include user abort vs system abort on batch close
> also include if close() was called and if there was an exception
> cc [~deepesh]



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-16360) Improve "No delta files or original files found to compact in" message

2017-09-20 Thread Eugene Koifman (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-16360?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174238#comment-16174238
 ] 

Eugene Koifman commented on HIVE-16360:
---

no related failures

> Improve "No delta files or original files found to compact in" message
> --
>
> Key: HIVE-16360
> URL: https://issues.apache.org/jira/browse/HIVE-16360
> Project: Hive
>  Issue Type: Improvement
>  Components: Transactions
>Affects Versions: 1.2.0
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
> Attachments: HIVE-16360.01.patch, HIVE-16360.02.patch
>
>
> CompactorMR.run() produces
> {noformat}
> LOG.error("No delta files or original files found to compact in " + 
> sd.getLocation() + " for compactionId=" + ci.id);
>   {noformat}
> in some cases even where there are plenty of delta directories in the 
> partition.
> The reason is that compactor won't compact anything above the level of an 
> open transaction.  So if there is a very long running txn, it may filter out 
> all deltas in a partition.  This looks very confusing when looking at the 
> logs.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-14813) Make TransactionBatchImpl.toString() include state of each txn: commit/abort

2017-09-20 Thread Eugene Koifman (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-14813?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174210#comment-16174210
 ] 

Eugene Koifman commented on HIVE-14813:
---

because I need 3 distinct values: Open/Closed/Aborted

> Make TransactionBatchImpl.toString() include state of each txn: commit/abort
> 
>
> Key: HIVE-14813
> URL: https://issues.apache.org/jira/browse/HIVE-14813
> Project: Hive
>  Issue Type: Bug
>  Components: HCatalog, Transactions
>Affects Versions: 1.2.1
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
> Attachments: HIVE-14813.01.patch
>
>
> perhaps include user abort vs system abort on batch close
> also include if close() was called and if there was an exception
> cc [~deepesh]



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17308) Improvement in join cardinality estimation

2017-09-20 Thread Vineet Garg (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17308?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174206#comment-16174206
 ] 

Vineet Garg commented on HIVE-17308:


[~leftylev] Done.

> Improvement in join cardinality estimation
> --
>
> Key: HIVE-17308
> URL: https://issues.apache.org/jira/browse/HIVE-17308
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Planning
>Reporter: Vineet Garg
>Assignee: Vineet Garg
> Fix For: 3.0.0
>
> Attachments: HIVE-17308.1.patch, HIVE-17308.2.patch, 
> HIVE-17308.3.patch, HIVE-17308.4.patch, HIVE-17308.5.patch, 
> HIVE-17308.6.patch, HIVE-17308.7.patch, HIVE-17308.8.patch
>
>
> Currently during logical planning join cardinality is estimated assuming no 
> correlation among join keys (This estimation is done using exponential 
> backoff). Physical planning on the other hand consider correlation for multi 
> keys and uses different estimation. We should consider correlation during 
> logical planning as well.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17308) Improvement in join cardinality estimation

2017-09-20 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17308?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174207#comment-16174207
 ] 

Lefty Leverenz commented on HIVE-17308:
---

Wow, you're fast!

> Improvement in join cardinality estimation
> --
>
> Key: HIVE-17308
> URL: https://issues.apache.org/jira/browse/HIVE-17308
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Planning
>Reporter: Vineet Garg
>Assignee: Vineet Garg
> Fix For: 3.0.0
>
> Attachments: HIVE-17308.1.patch, HIVE-17308.2.patch, 
> HIVE-17308.3.patch, HIVE-17308.4.patch, HIVE-17308.5.patch, 
> HIVE-17308.6.patch, HIVE-17308.7.patch, HIVE-17308.8.patch
>
>
> Currently during logical planning join cardinality is estimated assuming no 
> correlation among join keys (This estimation is done using exponential 
> backoff). Physical planning on the other hand consider correlation for multi 
> keys and uses different estimation. We should consider correlation during 
> logical planning as well.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17308) Improvement in join cardinality estimation

2017-09-20 Thread Vineet Garg (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17308?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-17308:
---
Fix Version/s: 3.0.0

> Improvement in join cardinality estimation
> --
>
> Key: HIVE-17308
> URL: https://issues.apache.org/jira/browse/HIVE-17308
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Planning
>Reporter: Vineet Garg
>Assignee: Vineet Garg
> Fix For: 3.0.0
>
> Attachments: HIVE-17308.1.patch, HIVE-17308.2.patch, 
> HIVE-17308.3.patch, HIVE-17308.4.patch, HIVE-17308.5.patch, 
> HIVE-17308.6.patch, HIVE-17308.7.patch, HIVE-17308.8.patch
>
>
> Currently during logical planning join cardinality is estimated assuming no 
> correlation among join keys (This estimation is done using exponential 
> backoff). Physical planning on the other hand consider correlation for multi 
> keys and uses different estimation. We should consider correlation during 
> logical planning as well.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17308) Improvement in join cardinality estimation

2017-09-20 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17308?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174205#comment-16174205
 ] 

Lefty Leverenz commented on HIVE-17308:
---

Nudge:  [~vgarg], please set the fix version of this issue to 3.0.0.  Thanks.

> Improvement in join cardinality estimation
> --
>
> Key: HIVE-17308
> URL: https://issues.apache.org/jira/browse/HIVE-17308
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Planning
>Reporter: Vineet Garg
>Assignee: Vineet Garg
> Attachments: HIVE-17308.1.patch, HIVE-17308.2.patch, 
> HIVE-17308.3.patch, HIVE-17308.4.patch, HIVE-17308.5.patch, 
> HIVE-17308.6.patch, HIVE-17308.7.patch, HIVE-17308.8.patch
>
>
> Currently during logical planning join cardinality is estimated assuming no 
> correlation among join keys (This estimation is done using exponential 
> backoff). Physical planning on the other hand consider correlation for multi 
> keys and uses different estimation. We should consider correlation during 
> logical planning as well.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17308) Improvement in join cardinality estimation

2017-09-20 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17308?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174202#comment-16174202
 ] 

Lefty Leverenz commented on HIVE-17308:
---

Doc note:  This changes the default value of 
*hive.stats.correlated.multi.key.joins* to true.

No TODOC3.0 label is needed because it will be documented for HIVE-16298, which 
created *hive.stats.correlated.multi.key.joins* in the same release (3.0.0).

> Improvement in join cardinality estimation
> --
>
> Key: HIVE-17308
> URL: https://issues.apache.org/jira/browse/HIVE-17308
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Planning
>Reporter: Vineet Garg
>Assignee: Vineet Garg
> Attachments: HIVE-17308.1.patch, HIVE-17308.2.patch, 
> HIVE-17308.3.patch, HIVE-17308.4.patch, HIVE-17308.5.patch, 
> HIVE-17308.6.patch, HIVE-17308.7.patch, HIVE-17308.8.patch
>
>
> Currently during logical planning join cardinality is estimated assuming no 
> correlation among join keys (This estimation is done using exponential 
> backoff). Physical planning on the other hand consider correlation for multi 
> keys and uses different estimation. We should consider correlation during 
> logical planning as well.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-14813) Make TransactionBatchImpl.toString() include state of each txn: commit/abort

2017-09-20 Thread Alan Gates (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-14813?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174200#comment-16174200
 ] 

Alan Gates commented on HIVE-14813:
---

Rather than use a Boolean[] to store the state of each transaction, why not use 
a BitSet?  It should be much more efficient than creating an object for each 
transaction.

> Make TransactionBatchImpl.toString() include state of each txn: commit/abort
> 
>
> Key: HIVE-14813
> URL: https://issues.apache.org/jira/browse/HIVE-14813
> Project: Hive
>  Issue Type: Bug
>  Components: HCatalog, Transactions
>Affects Versions: 1.2.1
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
> Attachments: HIVE-14813.01.patch
>
>
> perhaps include user abort vs system abort on batch close
> also include if close() was called and if there was an exception
> cc [~deepesh]



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Comment Edited] (HIVE-16298) Add config to specify multi-column joins have correlated columns

2017-09-20 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-16298?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15942197#comment-15942197
 ] 

Lefty Leverenz edited comment on HIVE-16298 at 9/21/17 3:16 AM:


Doc note:  This adds *hive.stats.correlated.multi.key.joins* to HiveConf.java, 
so it needs to be documented in the wiki.

* [Configuration Properties -- Statistics | 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-Statistics]

Added a TODOC3.0 label (which might be changed to 2.2 or 2.3 later).

Update (20/Sep/17):  HIVE-17308 changes the default for 
*hive.stats.correlated.multi.key.joins* to true in the same release (3.0.0).


was (Author: le...@hortonworks.com):
Doc note:  This adds *hive.stats.correlated.multi.key.joins* to HiveConf.java, 
so it needs to be documented in the wiki.

* [Configuration Properties -- Statistics | 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-Statistics]

Added a TODOC3.0 label (which might be changed to 2.2 or 2.3 later).

> Add config to specify multi-column joins have correlated columns
> 
>
> Key: HIVE-16298
> URL: https://issues.apache.org/jira/browse/HIVE-16298
> Project: Hive
>  Issue Type: Bug
>Reporter: Jason Dere
>Assignee: Siddharth Seth
>  Labels: TODOC3.0
> Fix For: 2.3.0
>
> Attachments: HIVE-16298.1.patch, HIVE-16298.2.patch
>
>
> The default row estimation for multi-key joins divides the row estimate by 
> the product of the NDVs for each join column, which can cause the row 
> estimate to be low. Try adding a config to assume the columns are correlated, 
> where we only divide the row estimate by the largest NDV.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-16084) SHOW COMPACTIONS should display CompactionID

2017-09-20 Thread Alan Gates (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-16084?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174189#comment-16174189
 ] 

Alan Gates commented on HIVE-16084:
---

+1.  Does this require any documentation changes?

> SHOW COMPACTIONS should display CompactionID
> 
>
> Key: HIVE-16084
> URL: https://issues.apache.org/jira/browse/HIVE-16084
> Project: Hive
>  Issue Type: Improvement
>  Components: Transactions
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
>Priority: Minor
> Attachments: HIVE-16084.01.patch, HIVE-16084.02.patch
>
>
> together with HIVE-13353 it will let users search for specific job



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-16360) Improve "No delta files or original files found to compact in" message

2017-09-20 Thread Alan Gates (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-16360?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174187#comment-16174187
 ] 

Alan Gates commented on HIVE-16360:
---

+1

> Improve "No delta files or original files found to compact in" message
> --
>
> Key: HIVE-16360
> URL: https://issues.apache.org/jira/browse/HIVE-16360
> Project: Hive
>  Issue Type: Improvement
>  Components: Transactions
>Affects Versions: 1.2.0
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
> Attachments: HIVE-16360.01.patch, HIVE-16360.02.patch
>
>
> CompactorMR.run() produces
> {noformat}
> LOG.error("No delta files or original files found to compact in " + 
> sd.getLocation() + " for compactionId=" + ci.id);
>   {noformat}
> in some cases even where there are plenty of delta directories in the 
> partition.
> The reason is that compactor won't compact anything above the level of an 
> open transaction.  So if there is a very long running txn, it may filter out 
> all deltas in a partition.  This looks very confusing when looking at the 
> logs.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Comment Edited] (HIVE-16298) Add config to specify multi-column joins have correlated columns

2017-09-20 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-16298?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15942194#comment-15942194
 ] 

Lefty Leverenz edited comment on HIVE-16298 at 9/21/17 3:13 AM:


[~jdere], since this was committed to master after branch-2 split off on March 
23 (11:06pm PDT) the fix version should say 3.0.0.  If you want it in 2.3.0 
commit it to branch-2.  For 2.2.0 commit to branch-2.2.

Nudge (edit 20/Sep/17):  This is in master for release 3.0.0 but not in 
branch-2.3 for 2.3.0, nor branch-2 for 2.4.0.  So the fix version should be 
changed from 2.3.0 to 3.0.0.


was (Author: le...@hortonworks.com):
[~jdere], since this was committed to master after branch-2 split off on March 
23 (11:06pm PDT) the fix version should say 3.0.0.  If you want it in 2.3.0 
commit it to branch-2.  For 2.2.0 commit to branch-2.2.

> Add config to specify multi-column joins have correlated columns
> 
>
> Key: HIVE-16298
> URL: https://issues.apache.org/jira/browse/HIVE-16298
> Project: Hive
>  Issue Type: Bug
>Reporter: Jason Dere
>Assignee: Siddharth Seth
>  Labels: TODOC3.0
> Fix For: 2.3.0
>
> Attachments: HIVE-16298.1.patch, HIVE-16298.2.patch
>
>
> The default row estimation for multi-key joins divides the row estimate by 
> the product of the NDVs for each join column, which can cause the row 
> estimate to be low. Try adding a config to assume the columns are correlated, 
> where we only divide the row estimate by the largest NDV.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-16360) Improve "No delta files or original files found to compact in" message

2017-09-20 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-16360?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174183#comment-16174183
 ] 

Hive QA commented on HIVE-16360:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12888066/HIVE-16360.02.patch

{color:green}SUCCESS:{color} +1 due to 2 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 9 failed/errored test(s), 11046 tests 
executed
*Failed tests:*
{noformat}
TestAccumuloCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=230)
TestDummy - did not produce a TEST-*.xml file (likely timed out) (batchId=230)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[create_view] (batchId=39)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[udf_mask_hash] 
(batchId=28)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[union_fast_stats]
 (batchId=156)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_explainuser_1]
 (batchId=170)
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[drop_table_failure2]
 (batchId=89)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query14] 
(batchId=234)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query23] 
(batchId=234)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6912/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6912/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6912/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 9 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12888066 - PreCommit-HIVE-Build

> Improve "No delta files or original files found to compact in" message
> --
>
> Key: HIVE-16360
> URL: https://issues.apache.org/jira/browse/HIVE-16360
> Project: Hive
>  Issue Type: Improvement
>  Components: Transactions
>Affects Versions: 1.2.0
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
> Attachments: HIVE-16360.01.patch, HIVE-16360.02.patch
>
>
> CompactorMR.run() produces
> {noformat}
> LOG.error("No delta files or original files found to compact in " + 
> sd.getLocation() + " for compactionId=" + ci.id);
>   {noformat}
> in some cases even where there are plenty of delta directories in the 
> partition.
> The reason is that compactor won't compact anything above the level of an 
> open transaction.  So if there is a very long running txn, it may filter out 
> all deltas in a partition.  This looks very confusing when looking at the 
> logs.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17554) Occurr java.lang.ArithmeticException: / by zero at hplsql component

2017-09-20 Thread ZhangBing Lin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17554?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174171#comment-16174171
 ] 

ZhangBing Lin commented on HIVE-17554:
--

Thank you,[~lirui]

> Occurr java.lang.ArithmeticException: / by zero at hplsql component
> ---
>
> Key: HIVE-17554
> URL: https://issues.apache.org/jira/browse/HIVE-17554
> Project: Hive
>  Issue Type: Bug
>  Components: hpl/sql
>Affects Versions: 3.0.0
>Reporter: ZhangBing Lin
>Assignee: ZhangBing Lin
> Fix For: 3.0.0
>
> Attachments: HIVE-17554.1.patch, HIVE-17554.2.patch
>
>
> When I execute the copyToTable method in hplsql, the following exception 
> occurs:
> java.lang.ArithmeticException: / by zero
>   at org.apache.hive.hplsql.Copy.copyToTable(Copy.java:170)
>   at org.apache.hive.hplsql.Copy.run(Copy.java:93)
>   at org.apache.hive.hplsql.Exec.visitCopy_stmt(Exec.java:1284)
>   at org.apache.hive.hplsql.Exec.visitCopy_stmt(Exec.java:52)
>   at 
> org.apache.hive.hplsql.HplsqlParser$Copy_stmtContext.accept(HplsqlParser.java:12148)
>   at 
> org.antlr.v4.runtime.tree.AbstractParseTreeVisitor.visitChildren(AbstractParseTreeVisitor.java:70)
>   at org.apache.hive.hplsql.Exec.visitStmt(Exec.java:994)
>   at org.apache.hive.hplsql.Exec.visitStmt(Exec.java:52)
>   at 
> org.apache.hive.hplsql.HplsqlParser$StmtContext.accept(HplsqlParser.java:1012)
>   at 
> org.antlr.v4.runtime.tree.AbstractParseTreeVisitor.visitChildren(AbstractParseTreeVisitor.java:70)
>   at 
> org.apache.hive.hplsql.HplsqlBaseVisitor.visitBlock(HplsqlBaseVisitor.java:28)
>   at 
> org.apache.hive.hplsql.HplsqlParser$BlockContext.accept(HplsqlParser.java:446)
>   at 
> org.antlr.v4.runtime.tree.AbstractParseTreeVisitor.visitChildren(AbstractParseTreeVisitor.java:70)
>   at org.apache.hive.hplsql.Exec.visitProgram(Exec.java:901)
>   at org.apache.hive.hplsql.Exec.visitProgram(Exec.java:52)
>   at 
> org.apache.hive.hplsql.HplsqlParser$ProgramContext.accept(HplsqlParser.java:389)
>   at 
> org.antlr.v4.runtime.tree.AbstractParseTreeVisitor.visit(AbstractParseTreeVisitor.java:42)
>   at org.apache.hive.hplsql.Exec.run(Exec.java:760)
>   at org.apache.hive.hplsql.Exec.run(Exec.java:736)
>   at org.apache.hive.hplsql.Hplsql.main(Hplsql.java:23)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:606)
>   at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
>   at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
> View source code:
> if (info) {
>   info(ctx, "COPY completed: " + rows + " row(s), " + timer.format() + ", 
> " + rows/(elapsed/1000) + " rows/sec");
> }
> When elapsed less than one thousand will cause the divisor to be zero



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Comment Edited] (HIVE-17318) Make Hikari CP configurable using hive properties in hive-site.xml

2017-09-20 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17318?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174155#comment-16174155
 ] 

Lefty Leverenz edited comment on HIVE-17318 at 9/21/17 2:49 AM:


Thanks for the docs.  Here are their links:

* [Configuration Properties -- Hive Metastore Connection Pooling Configuration 
| 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-HiveMetastoreConnectionPoolingConfiguration]
* [Configuration Properties -- hive.conf.restricted.list | 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-hive.conf.restricted.list]

I removed the TODOC3.0 label.


was (Author: le...@hortonworks.com):
Thanks for the docs.  Here are their links:

* [Configuration Properties -- Hive Metastore Connection Pooling Configuration 
| 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-HiveMetastoreConnectionPoolingConfiguration]
* [Configuration Properties -- hive.conf.restricted.list | 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-hive.conf.restricted.list]

> Make Hikari CP configurable using hive properties in hive-site.xml
> --
>
> Key: HIVE-17318
> URL: https://issues.apache.org/jira/browse/HIVE-17318
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Barna Zsombor Klara
>Assignee: Barna Zsombor Klara
> Fix For: 3.0.0
>
> Attachments: HIVE-17318.01.patch, HIVE-17318.02.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17554) Occurr java.lang.ArithmeticException: / by zero at hplsql component

2017-09-20 Thread Rui Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17554?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Li updated HIVE-17554:
--
   Resolution: Fixed
Fix Version/s: 3.0.0
   Status: Resolved  (was: Patch Available)

Pushed to master. Thanks [~linzhangbing] for the contribution.

> Occurr java.lang.ArithmeticException: / by zero at hplsql component
> ---
>
> Key: HIVE-17554
> URL: https://issues.apache.org/jira/browse/HIVE-17554
> Project: Hive
>  Issue Type: Bug
>  Components: hpl/sql
>Affects Versions: 3.0.0
>Reporter: ZhangBing Lin
>Assignee: ZhangBing Lin
> Fix For: 3.0.0
>
> Attachments: HIVE-17554.1.patch, HIVE-17554.2.patch
>
>
> When I execute the copyToTable method in hplsql, the following exception 
> occurs:
> java.lang.ArithmeticException: / by zero
>   at org.apache.hive.hplsql.Copy.copyToTable(Copy.java:170)
>   at org.apache.hive.hplsql.Copy.run(Copy.java:93)
>   at org.apache.hive.hplsql.Exec.visitCopy_stmt(Exec.java:1284)
>   at org.apache.hive.hplsql.Exec.visitCopy_stmt(Exec.java:52)
>   at 
> org.apache.hive.hplsql.HplsqlParser$Copy_stmtContext.accept(HplsqlParser.java:12148)
>   at 
> org.antlr.v4.runtime.tree.AbstractParseTreeVisitor.visitChildren(AbstractParseTreeVisitor.java:70)
>   at org.apache.hive.hplsql.Exec.visitStmt(Exec.java:994)
>   at org.apache.hive.hplsql.Exec.visitStmt(Exec.java:52)
>   at 
> org.apache.hive.hplsql.HplsqlParser$StmtContext.accept(HplsqlParser.java:1012)
>   at 
> org.antlr.v4.runtime.tree.AbstractParseTreeVisitor.visitChildren(AbstractParseTreeVisitor.java:70)
>   at 
> org.apache.hive.hplsql.HplsqlBaseVisitor.visitBlock(HplsqlBaseVisitor.java:28)
>   at 
> org.apache.hive.hplsql.HplsqlParser$BlockContext.accept(HplsqlParser.java:446)
>   at 
> org.antlr.v4.runtime.tree.AbstractParseTreeVisitor.visitChildren(AbstractParseTreeVisitor.java:70)
>   at org.apache.hive.hplsql.Exec.visitProgram(Exec.java:901)
>   at org.apache.hive.hplsql.Exec.visitProgram(Exec.java:52)
>   at 
> org.apache.hive.hplsql.HplsqlParser$ProgramContext.accept(HplsqlParser.java:389)
>   at 
> org.antlr.v4.runtime.tree.AbstractParseTreeVisitor.visit(AbstractParseTreeVisitor.java:42)
>   at org.apache.hive.hplsql.Exec.run(Exec.java:760)
>   at org.apache.hive.hplsql.Exec.run(Exec.java:736)
>   at org.apache.hive.hplsql.Hplsql.main(Hplsql.java:23)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:606)
>   at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
>   at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
> View source code:
> if (info) {
>   info(ctx, "COPY completed: " + rows + " row(s), " + timer.format() + ", 
> " + rows/(elapsed/1000) + " rows/sec");
> }
> When elapsed less than one thousand will cause the divisor to be zero



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17318) Make Hikari CP configurable using hive properties in hive-site.xml

2017-09-20 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17318?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174155#comment-16174155
 ] 

Lefty Leverenz commented on HIVE-17318:
---

Thanks for the docs.  Here are their links:

* [Configuration Properties -- Hive Metastore Connection Pooling Configuration 
| 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-HiveMetastoreConnectionPoolingConfiguration]
* [Configuration Properties -- hive.conf.restricted.list | 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-hive.conf.restricted.list]

> Make Hikari CP configurable using hive properties in hive-site.xml
> --
>
> Key: HIVE-17318
> URL: https://issues.apache.org/jira/browse/HIVE-17318
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Barna Zsombor Klara
>Assignee: Barna Zsombor Klara
> Fix For: 3.0.0
>
> Attachments: HIVE-17318.01.patch, HIVE-17318.02.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17319) Make BoneCp configurable using hive properties in hive-site.xml

2017-09-20 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17319?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174154#comment-16174154
 ] 

Lefty Leverenz commented on HIVE-17319:
---

[~zsombor.klara] documented this here:

* [Configuration Properties -- Hive Metastore Connection Pooling Configuration 
| 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-HiveMetastoreConnectionPoolingConfiguration]
* [Configuration Properties -- hive.conf.restricted.list | 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-hive.conf.restricted.list]

Thanks!

> Make BoneCp configurable using hive properties in hive-site.xml
> ---
>
> Key: HIVE-17319
> URL: https://issues.apache.org/jira/browse/HIVE-17319
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Barna Zsombor Klara
>Assignee: Barna Zsombor Klara
> Fix For: 3.0.0
>
> Attachments: HIVE-17319.01.patch, HIVE-17319.02.patch, 
> HIVE-17319.03.patch, HIVE-17319.draft.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Comment Edited] (HIVE-17316) Use String.startsWith for the hidden configuration variables

2017-09-20 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17316?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174127#comment-16174127
 ] 

Lefty Leverenz edited comment on HIVE-17316 at 9/21/17 2:41 AM:


Should this be documented in the description of *hive.conf.restricted.list* or 
*hive.conf.hidden.list*?

* [hive.conf.restricted.list | 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-hive.conf.restricted.list]
* [hive.conf.hidden.list | 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-hive.conf.hidden.list]


was (Author: le...@hortonworks.com):
Should this be documented in the description of *hive.conf.hidden.list* or 
somewhere else?

* [hive.conf.hidden.list | 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-hive.conf.hidden.list]

> Use String.startsWith for the hidden configuration variables
> 
>
> Key: HIVE-17316
> URL: https://issues.apache.org/jira/browse/HIVE-17316
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Barna Zsombor Klara
>Assignee: Barna Zsombor Klara
> Fix For: 3.0.0
>
> Attachments: HIVE-17316.01.patch, HIVE-17316.02.patch, 
> HIVE-17316.03.patch
>
>
> Currently HiveConf variables which should not be displayed to the user need 
> to be enumerated. We should enhance this to be able to hide configuration 
> variables by string prefix not just full equality.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17318) Make Hikari CP configurable using hive properties in hive-site.xml

2017-09-20 Thread Lefty Leverenz (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17318?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lefty Leverenz updated HIVE-17318:
--
Labels:   (was: TODOC3.0)

> Make Hikari CP configurable using hive properties in hive-site.xml
> --
>
> Key: HIVE-17318
> URL: https://issues.apache.org/jira/browse/HIVE-17318
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Barna Zsombor Klara
>Assignee: Barna Zsombor Klara
> Fix For: 3.0.0
>
> Attachments: HIVE-17318.01.patch, HIVE-17318.02.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17554) Occurr java.lang.ArithmeticException: / by zero at hplsql component

2017-09-20 Thread Rui Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17554?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174142#comment-16174142
 ] 

Rui Li commented on HIVE-17554:
---

+1

> Occurr java.lang.ArithmeticException: / by zero at hplsql component
> ---
>
> Key: HIVE-17554
> URL: https://issues.apache.org/jira/browse/HIVE-17554
> Project: Hive
>  Issue Type: Bug
>  Components: hpl/sql
>Affects Versions: 3.0.0
>Reporter: ZhangBing Lin
>Assignee: ZhangBing Lin
> Attachments: HIVE-17554.1.patch, HIVE-17554.2.patch
>
>
> When I execute the copyToTable method in hplsql, the following exception 
> occurs:
> java.lang.ArithmeticException: / by zero
>   at org.apache.hive.hplsql.Copy.copyToTable(Copy.java:170)
>   at org.apache.hive.hplsql.Copy.run(Copy.java:93)
>   at org.apache.hive.hplsql.Exec.visitCopy_stmt(Exec.java:1284)
>   at org.apache.hive.hplsql.Exec.visitCopy_stmt(Exec.java:52)
>   at 
> org.apache.hive.hplsql.HplsqlParser$Copy_stmtContext.accept(HplsqlParser.java:12148)
>   at 
> org.antlr.v4.runtime.tree.AbstractParseTreeVisitor.visitChildren(AbstractParseTreeVisitor.java:70)
>   at org.apache.hive.hplsql.Exec.visitStmt(Exec.java:994)
>   at org.apache.hive.hplsql.Exec.visitStmt(Exec.java:52)
>   at 
> org.apache.hive.hplsql.HplsqlParser$StmtContext.accept(HplsqlParser.java:1012)
>   at 
> org.antlr.v4.runtime.tree.AbstractParseTreeVisitor.visitChildren(AbstractParseTreeVisitor.java:70)
>   at 
> org.apache.hive.hplsql.HplsqlBaseVisitor.visitBlock(HplsqlBaseVisitor.java:28)
>   at 
> org.apache.hive.hplsql.HplsqlParser$BlockContext.accept(HplsqlParser.java:446)
>   at 
> org.antlr.v4.runtime.tree.AbstractParseTreeVisitor.visitChildren(AbstractParseTreeVisitor.java:70)
>   at org.apache.hive.hplsql.Exec.visitProgram(Exec.java:901)
>   at org.apache.hive.hplsql.Exec.visitProgram(Exec.java:52)
>   at 
> org.apache.hive.hplsql.HplsqlParser$ProgramContext.accept(HplsqlParser.java:389)
>   at 
> org.antlr.v4.runtime.tree.AbstractParseTreeVisitor.visit(AbstractParseTreeVisitor.java:42)
>   at org.apache.hive.hplsql.Exec.run(Exec.java:760)
>   at org.apache.hive.hplsql.Exec.run(Exec.java:736)
>   at org.apache.hive.hplsql.Hplsql.main(Hplsql.java:23)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:606)
>   at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
>   at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
> View source code:
> if (info) {
>   info(ctx, "COPY completed: " + rows + " row(s), " + timer.format() + ", 
> " + rows/(elapsed/1000) + " rows/sec");
> }
> When elapsed less than one thousand will cause the divisor to be zero



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17316) Use String.startsWith for the hidden configuration variables

2017-09-20 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17316?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174127#comment-16174127
 ] 

Lefty Leverenz commented on HIVE-17316:
---

Should this be documented in the description of *hive.conf.hidden.list* or 
somewhere else?

* [hive.conf.hidden.list | 
https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-hive.conf.hidden.list]

> Use String.startsWith for the hidden configuration variables
> 
>
> Key: HIVE-17316
> URL: https://issues.apache.org/jira/browse/HIVE-17316
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Barna Zsombor Klara
>Assignee: Barna Zsombor Klara
> Fix For: 3.0.0
>
> Attachments: HIVE-17316.01.patch, HIVE-17316.02.patch, 
> HIVE-17316.03.patch
>
>
> Currently HiveConf variables which should not be displayed to the user need 
> to be enumerated. We should enhance this to be able to hide configuration 
> variables by string prefix not just full equality.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Assigned] (HIVE-17565) NullPointerException occurs when hive.optimize.skewjoin and hive.auto.convert.join are switched on at the same time

2017-09-20 Thread liyunzhang_intel (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17565?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

liyunzhang_intel reassigned HIVE-17565:
---

Assignee: liyunzhang_intel

> NullPointerException occurs when hive.optimize.skewjoin and 
> hive.auto.convert.join are switched on at the same time
> ---
>
> Key: HIVE-17565
> URL: https://issues.apache.org/jira/browse/HIVE-17565
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 1.2.1
>Reporter: Xin Hao
>Assignee: liyunzhang_intel
>
> (A)NullPointerException occurs when hive.optimize.skewjoin and 
> hive.auto.convert.join are switched on at the same time.
> Could pass when hive.optimize.skewjoin=true and hive.auto.convert.join=false.
> (B)Hive Version:
> Found on Apache Hive 1.2.1
> (C)Workload:
> (1)TPCx-BB Q19
> (2) A small case as below,which was actually simplified from Q19:
> SELECT *
> FROM store_returns sr,
> (
>   SELECT d1.d_date_sk
>   FROM date_dim d1, date_dim d2
>   WHERE d1.d_week_seq = d2.d_week_seq
> ) sr_dateFilter
> WHERE sr.sr_returned_date_sk = d_date_sk;
> (D)Exception Error Message:
> Error: java.lang.RuntimeException: java.lang.NullPointerException
> at 
> org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:179)
> at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54)
> at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:453)
> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
> at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:415)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1796)
> at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
> Caused by: java.lang.NullPointerException
> at 
> org.apache.hadoop.hive.ql.exec.MapJoinOperator.loadHashTable(MapJoinOperator.java:194)
> at 
> org.apache.hadoop.hive.ql.exec.MapJoinOperator.cleanUpInputFileChangedOp(MapJoinOperator.java:223)
> at 
> org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1051)
> at 
> org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055)
> at 
> org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055)
> at 
> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:490)
> at 
> org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:170)
> ... 8 more



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17565) NullPointerException occurs when hive.optimize.skewjoin and hive.auto.convert.join are switched on at the same time

2017-09-20 Thread liyunzhang_intel (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17565?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174112#comment-16174112
 ] 

liyunzhang_intel commented on HIVE-17565:
-

HaoXin: this happens on Hive on MR or Hive on Spark?


> NullPointerException occurs when hive.optimize.skewjoin and 
> hive.auto.convert.join are switched on at the same time
> ---
>
> Key: HIVE-17565
> URL: https://issues.apache.org/jira/browse/HIVE-17565
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 1.2.1
>Reporter: Xin Hao
>
> (A)NullPointerException occurs when hive.optimize.skewjoin and 
> hive.auto.convert.join are switched on at the same time.
> Could pass when hive.optimize.skewjoin=true and hive.auto.convert.join=false.
> (B)Hive Version:
> Found on Apache Hive 1.2.1
> (C)Workload:
> (1)TPCx-BB Q19
> (2) A small case as below,which was actually simplified from Q19:
> SELECT *
> FROM store_returns sr,
> (
>   SELECT d1.d_date_sk
>   FROM date_dim d1, date_dim d2
>   WHERE d1.d_week_seq = d2.d_week_seq
> ) sr_dateFilter
> WHERE sr.sr_returned_date_sk = d_date_sk;
> (D)Exception Error Message:
> Error: java.lang.RuntimeException: java.lang.NullPointerException
> at 
> org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:179)
> at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54)
> at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:453)
> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
> at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:415)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1796)
> at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
> Caused by: java.lang.NullPointerException
> at 
> org.apache.hadoop.hive.ql.exec.MapJoinOperator.loadHashTable(MapJoinOperator.java:194)
> at 
> org.apache.hadoop.hive.ql.exec.MapJoinOperator.cleanUpInputFileChangedOp(MapJoinOperator.java:223)
> at 
> org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1051)
> at 
> org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055)
> at 
> org.apache.hadoop.hive.ql.exec.Operator.cleanUpInputFileChanged(Operator.java:1055)
> at 
> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:490)
> at 
> org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:170)
> ... 8 more



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17483) HS2 kill command to kill queries using query id

2017-09-20 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17483?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174109#comment-16174109
 ] 

Hive QA commented on HIVE-17483:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12888033/HIVE-17483.6.patch

{color:green}SUCCESS:{color} +1 due to 6 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 19 failed/errored test(s), 11053 tests 
executed
*Failed tests:*
{noformat}
TestAccumuloCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=230)
TestDummy - did not produce a TEST-*.xml file (likely timed out) (batchId=230)
org.apache.hadoop.hive.cli.TestBeeLineDriver.testCliDriver[mapjoin2] 
(batchId=239)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[create_view] (batchId=39)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[udf_mask_hash] 
(batchId=28)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[union_fast_stats]
 (batchId=156)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[bucketizedhiveinputformat]
 (batchId=170)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_explainuser_1]
 (batchId=170)
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[drop_table_failure2]
 (batchId=89)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query23] 
(batchId=234)
org.apache.hadoop.hive.ql.lockmgr.TestDbTxnManager2.testShowTablesLock 
(batchId=282)
org.apache.hive.service.cli.session.TestHiveSessionImpl.testLeakOperationHandle 
(batchId=223)
org.apache.hive.service.cli.session.TestQueryDisplay.testQueryDisplay 
(batchId=223)
org.apache.hive.service.cli.session.TestQueryDisplay.testWebUI (batchId=223)
org.apache.hive.service.cli.session.TestSessionManagerMetrics.testAbandonedSessionMetrics
 (batchId=197)
org.apache.hive.service.cli.session.TestSessionManagerMetrics.testActiveSessionMetrics
 (batchId=197)
org.apache.hive.service.cli.session.TestSessionManagerMetrics.testActiveSessionTimeMetrics
 (batchId=197)
org.apache.hive.service.cli.session.TestSessionManagerMetrics.testOpenSessionMetrics
 (batchId=197)
org.apache.hive.service.cli.session.TestSessionManagerMetrics.testOpenSessionTimeMetrics
 (batchId=197)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6911/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6911/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6911/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 19 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12888033 - PreCommit-HIVE-Build

> HS2 kill command to kill queries using query id
> ---
>
> Key: HIVE-17483
> URL: https://issues.apache.org/jira/browse/HIVE-17483
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2
>Reporter: Thejas M Nair
>Assignee: Teddy Choi
> Attachments: HIVE-17483.1.patch, HIVE-17483.2.patch, 
> HIVE-17483.2.patch, HIVE-17483.3.patch, HIVE-17483.4.patch, 
> HIVE-17483.5.patch, HIVE-17483.6.patch
>
>
> For administrators, it is important to be able to kill queries if required. 
> Currently, there is no clean way to do it.
> It would help to have a "kill query " command that can be run using 
> odbc/jdbc against a HiveServer2 instance, to kill a query with that queryid 
> running in that instance.
> Authorization will have to be done to ensure that the user that is invoking 
> the API is allowed to perform this action.
> In case of SQL std authorization, this would require admin role.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-15899) Make CTAS with acid target table and insert into acid_tbl select ... union all ... work

2017-09-20 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15899?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174102#comment-16174102
 ] 

Lefty Leverenz commented on HIVE-15899:
---

Okay, thanks Eugene.

> Make CTAS with acid target table and insert into acid_tbl select ... union 
> all ... work
> ---
>
> Key: HIVE-15899
> URL: https://issues.apache.org/jira/browse/HIVE-15899
> Project: Hive
>  Issue Type: Sub-task
>  Components: Transactions
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
> Fix For: 3.0.0
>
> Attachments: HIVE-15899.01.patch, HIVE-15899.02.patch, 
> HIVE-15899.03.patch, HIVE-15899.04.patch, HIVE-15899.05.patch, 
> HIVE-15899.07.patch, HIVE-15899.08.patch, HIVE-15899.09.patch, 
> HIVE-15899.10.patch, HIVE-15899.11.patch, HIVE-15899.12.patch, 
> HIVE-15899.13.patch
>
>
> Consider:
> {noformat}
> create table T stored as ORC TBLPROPERTIES('transactional'='true') as
>   select a, b from A where a <= 5 union all select a, b from B where a >= 
> 5
> {noformat}
> and 
> {noformat}
> create table T (a int, b int) stored as ORC  TBLPROPERTIES 
> ('transactional'='false';
> insert into T(a,b) select a, b from T where a between 1 and 3 group by a, b 
> union all select a, b from A where a between 5 and 7 union all select a, b 
> from B where a >= 9
> {noformat}
> On Tez, there is an optimization that removes Union All operator writes the 
> data into
> subdirectories of T (in this case T is unpartitioned).
> This also happens on MR but requires 
> {noformat}
> hiveConf.setBoolVar(HiveConf.ConfVars.HIVE_OPTIMIZE_UNION_REMOVE, true);
> hiveConf.setVar(HiveConf.ConfVars.HIVEFETCHTASKCONVERSION, "none");
> {noformat} 
> Need to ensure that when target table is Acid, we generate unique ROW__IDs
> When target is not acid, that we can convert it to Acid via Alter Table even 
> when data layout includes subdirectories.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-15899) Make CTAS with acid target table and insert into acid_tbl select ... union all ... work

2017-09-20 Thread Eugene Koifman (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15899?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174098#comment-16174098
 ] 

Eugene Koifman commented on HIVE-15899:
---

I don't think this needs any additional doc.  It's an existing feature that now 
works properly.

> Make CTAS with acid target table and insert into acid_tbl select ... union 
> all ... work
> ---
>
> Key: HIVE-15899
> URL: https://issues.apache.org/jira/browse/HIVE-15899
> Project: Hive
>  Issue Type: Sub-task
>  Components: Transactions
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
> Fix For: 3.0.0
>
> Attachments: HIVE-15899.01.patch, HIVE-15899.02.patch, 
> HIVE-15899.03.patch, HIVE-15899.04.patch, HIVE-15899.05.patch, 
> HIVE-15899.07.patch, HIVE-15899.08.patch, HIVE-15899.09.patch, 
> HIVE-15899.10.patch, HIVE-15899.11.patch, HIVE-15899.12.patch, 
> HIVE-15899.13.patch
>
>
> Consider:
> {noformat}
> create table T stored as ORC TBLPROPERTIES('transactional'='true') as
>   select a, b from A where a <= 5 union all select a, b from B where a >= 
> 5
> {noformat}
> and 
> {noformat}
> create table T (a int, b int) stored as ORC  TBLPROPERTIES 
> ('transactional'='false';
> insert into T(a,b) select a, b from T where a between 1 and 3 group by a, b 
> union all select a, b from A where a between 5 and 7 union all select a, b 
> from B where a >= 9
> {noformat}
> On Tez, there is an optimization that removes Union All operator writes the 
> data into
> subdirectories of T (in this case T is unpartitioned).
> This also happens on MR but requires 
> {noformat}
> hiveConf.setBoolVar(HiveConf.ConfVars.HIVE_OPTIMIZE_UNION_REMOVE, true);
> hiveConf.setVar(HiveConf.ConfVars.HIVEFETCHTASKCONVERSION, "none");
> {noformat} 
> Need to ensure that when target table is Acid, we generate unique ROW__IDs
> When target is not acid, that we can convert it to Acid via Alter Table even 
> when data layout includes subdirectories.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-15899) Make CTAS with acid target table and insert into acid_tbl select ... union all ... work

2017-09-20 Thread Eugene Koifman (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15899?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eugene Koifman updated HIVE-15899:
--
Fix Version/s: 3.0.0

> Make CTAS with acid target table and insert into acid_tbl select ... union 
> all ... work
> ---
>
> Key: HIVE-15899
> URL: https://issues.apache.org/jira/browse/HIVE-15899
> Project: Hive
>  Issue Type: Sub-task
>  Components: Transactions
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
> Fix For: 3.0.0
>
> Attachments: HIVE-15899.01.patch, HIVE-15899.02.patch, 
> HIVE-15899.03.patch, HIVE-15899.04.patch, HIVE-15899.05.patch, 
> HIVE-15899.07.patch, HIVE-15899.08.patch, HIVE-15899.09.patch, 
> HIVE-15899.10.patch, HIVE-15899.11.patch, HIVE-15899.12.patch, 
> HIVE-15899.13.patch
>
>
> Consider:
> {noformat}
> create table T stored as ORC TBLPROPERTIES('transactional'='true') as
>   select a, b from A where a <= 5 union all select a, b from B where a >= 
> 5
> {noformat}
> and 
> {noformat}
> create table T (a int, b int) stored as ORC  TBLPROPERTIES 
> ('transactional'='false';
> insert into T(a,b) select a, b from T where a between 1 and 3 group by a, b 
> union all select a, b from A where a between 5 and 7 union all select a, b 
> from B where a >= 9
> {noformat}
> On Tez, there is an optimization that removes Union All operator writes the 
> data into
> subdirectories of T (in this case T is unpartitioned).
> This also happens on MR but requires 
> {noformat}
> hiveConf.setBoolVar(HiveConf.ConfVars.HIVE_OPTIMIZE_UNION_REMOVE, true);
> hiveConf.setVar(HiveConf.ConfVars.HIVEFETCHTASKCONVERSION, "none");
> {noformat} 
> Need to ensure that when target table is Acid, we generate unique ROW__IDs
> When target is not acid, that we can convert it to Acid via Alter Table even 
> when data layout includes subdirectories.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-9995) ACID compaction tries to compact a single file

2017-09-20 Thread Eugene Koifman (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9995?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174097#comment-16174097
 ] 

Eugene Koifman commented on HIVE-9995:
--

also, if the user is explicitly requesting Minor compaction but there are 
Original files and no base, should we automatically run Major or reject the 
request

also, 
make sure that Initiator request Major if there are Originals but no base



> ACID compaction tries to compact a single file
> --
>
> Key: HIVE-9995
> URL: https://issues.apache.org/jira/browse/HIVE-9995
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Affects Versions: 1.0.0
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
>
> Consider TestWorker.minorWithOpenInMiddle()
> since there is an open txnId=23, this doesn't have any meaningful minor 
> compaction work to do.  The system still tries to compact a single delta file 
> for 21-22 id range, and effectively copies the file onto itself.
> This is 1. inefficient and 2. can potentially affect a reader.
> (from a real cluster)
> Suppose we start with 
> {noformat}
> drwxr-xr-x   - ekoifman staff  0 2016-06-09 16:03 
> /user/hive/warehouse/t/base_016
> -rw-r--r--   1 ekoifman staff602 2016-06-09 16:03 
> /user/hive/warehouse/t/base_016/bucket_0
> drwxr-xr-x   - ekoifman staff  0 2016-06-09 16:07 
> /user/hive/warehouse/t/base_017
> -rw-r--r--   1 ekoifman staff588 2016-06-09 16:07 
> /user/hive/warehouse/t/base_017/bucket_0
> drwxr-xr-x   - ekoifman staff  0 2016-06-09 16:07 
> /user/hive/warehouse/t/delta_017_017_
> -rw-r--r--   1 ekoifman staff514 2016-06-09 16:06 
> /user/hive/warehouse/t/delta_017_017_/bucket_0
> drwxr-xr-x   - ekoifman staff  0 2016-06-09 16:07 
> /user/hive/warehouse/t/delta_018_018_
> -rw-r--r--   1 ekoifman staff612 2016-06-09 16:07 
> /user/hive/warehouse/t/delta_018_018_/bucket_0
> {noformat}
> then do _alter table T compact 'minor';_
> then we end up with 
> {noformat}
> drwxr-xr-x   - ekoifman staff  0 2016-06-09 16:07 
> /user/hive/warehouse/t/base_017
> -rw-r--r--   1 ekoifman staff588 2016-06-09 16:07 
> /user/hive/warehouse/t/base_017/bucket_0
> drwxr-xr-x   - ekoifman staff  0 2016-06-09 16:11 
> /user/hive/warehouse/t/delta_018_018
> -rw-r--r--   1 ekoifman staff500 2016-06-09 16:11 
> /user/hive/warehouse/t/delta_018_018/bucket_0
> drwxr-xr-x   - ekoifman staff  0 2016-06-09 16:07 
> /user/hive/warehouse/t/delta_018_018_
> -rw-r--r--   1 ekoifman staff612 2016-06-09 16:07 
> /user/hive/warehouse/t/delta_018_018_/bucket_0
> {noformat}
> So compaction created a new dir _/user/hive/warehouse/t/delta_018_018_



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-15899) Make CTAS with acid target table and insert into acid_tbl select ... union all ... work

2017-09-20 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15899?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174096#comment-16174096
 ] 

Lefty Leverenz commented on HIVE-15899:
---

Should this be documented in the wiki?

(Also, please set the fix version to 3.0.0.)

> Make CTAS with acid target table and insert into acid_tbl select ... union 
> all ... work
> ---
>
> Key: HIVE-15899
> URL: https://issues.apache.org/jira/browse/HIVE-15899
> Project: Hive
>  Issue Type: Sub-task
>  Components: Transactions
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
> Attachments: HIVE-15899.01.patch, HIVE-15899.02.patch, 
> HIVE-15899.03.patch, HIVE-15899.04.patch, HIVE-15899.05.patch, 
> HIVE-15899.07.patch, HIVE-15899.08.patch, HIVE-15899.09.patch, 
> HIVE-15899.10.patch, HIVE-15899.11.patch, HIVE-15899.12.patch, 
> HIVE-15899.13.patch
>
>
> Consider:
> {noformat}
> create table T stored as ORC TBLPROPERTIES('transactional'='true') as
>   select a, b from A where a <= 5 union all select a, b from B where a >= 
> 5
> {noformat}
> and 
> {noformat}
> create table T (a int, b int) stored as ORC  TBLPROPERTIES 
> ('transactional'='false';
> insert into T(a,b) select a, b from T where a between 1 and 3 group by a, b 
> union all select a, b from A where a between 5 and 7 union all select a, b 
> from B where a >= 9
> {noformat}
> On Tez, there is an optimization that removes Union All operator writes the 
> data into
> subdirectories of T (in this case T is unpartitioned).
> This also happens on MR but requires 
> {noformat}
> hiveConf.setBoolVar(HiveConf.ConfVars.HIVE_OPTIMIZE_UNION_REMOVE, true);
> hiveConf.setVar(HiveConf.ConfVars.HIVEFETCHTASKCONVERSION, "none");
> {noformat} 
> Need to ensure that when target table is Acid, we generate unique ROW__IDs
> When target is not acid, that we can convert it to Acid via Alter Table even 
> when data layout includes subdirectories.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Comment Edited] (HIVE-16669) Fine tune Compaction to take advantage of Acid 2.0

2017-09-20 Thread Eugene Koifman (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-16669?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174083#comment-16174083
 ] 

Eugene Koifman edited comment on HIVE-16669 at 9/21/17 1:09 AM:


OK, this is more than just fine tuning.  Suppose we have 
base_8
delta_9
delete_delta_10 - this affects rows in base_8

Minor compaction (as currently implemented (inherited from Acid 1)) will 
produce delta_9_10 which means all deletes by txn 10 affecting rows in base_8 
are lost.

so HIVE-17089 is effectively incomplete w/o this


was (Author: ekoifman):
OK, this is more than just fine tuning.  Suppose we have 
base_8
delta_9
delete_delta_10 - this affects rows in base_8

Minor compaction (as currently implemented (inherited from Acid 1)) will 
produce delta_9_10 which means all deletes by txn 10 affecting rows in base_8 
are lost.


> Fine tune Compaction to take advantage of Acid 2.0
> --
>
> Key: HIVE-16669
> URL: https://issues.apache.org/jira/browse/HIVE-16669
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
>Priority: Blocker
>
> * There is little point using 2.0 vectorized reader since there is no 
> operator pipeline in compaction
> * If minor compaction just concats delete_delta files together, then the 2 
> stage compaction should always ensure that we have a limited number of Orc 
> readers to do the merging and current OrcRawRecordMerger should be fine
> * ...



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17535) Select 1 EXCEPT Select 1 fails with NPE

2017-09-20 Thread Vineet Garg (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17535?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-17535:
---
Fix Version/s: 3.0.0

> Select 1 EXCEPT Select 1 fails with NPE
> ---
>
> Key: HIVE-17535
> URL: https://issues.apache.org/jira/browse/HIVE-17535
> Project: Hive
>  Issue Type: Bug
>  Components: Query Planning
>Reporter: Vineet Garg
>Assignee: Vineet Garg
> Fix For: 3.0.0
>
> Attachments: HIVE-17535.1.patch, HIVE-17535.2.patch, 
> HIVE-17535.3.patch, HIVE-17535.4.patch, HIVE-17535.5.patch
>
>
> Since Hive CBO isn't able to handle queries with no table e.g. {{select 1}} 
> queries with SET operators fail (intersect requires CBO).



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17535) Select 1 EXCEPT Select 1 fails with NPE

2017-09-20 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17535?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174084#comment-16174084
 ] 

Lefty Leverenz commented on HIVE-17535:
---

[~vgarg], please set the fix version for this issue to 3.0.0.  Thanks.

> Select 1 EXCEPT Select 1 fails with NPE
> ---
>
> Key: HIVE-17535
> URL: https://issues.apache.org/jira/browse/HIVE-17535
> Project: Hive
>  Issue Type: Bug
>  Components: Query Planning
>Reporter: Vineet Garg
>Assignee: Vineet Garg
> Attachments: HIVE-17535.1.patch, HIVE-17535.2.patch, 
> HIVE-17535.3.patch, HIVE-17535.4.patch, HIVE-17535.5.patch
>
>
> Since Hive CBO isn't able to handle queries with no table e.g. {{select 1}} 
> queries with SET operators fail (intersect requires CBO).



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-16669) Fine tune Compaction to take advantage of Acid 2.0

2017-09-20 Thread Eugene Koifman (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-16669?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174083#comment-16174083
 ] 

Eugene Koifman commented on HIVE-16669:
---

OK, this is more than just fine tuning.  Suppose we have 
base_8
delta_9
delete_delta_10 - this affects rows in base_8

Minor compaction (as currently implemented (inherited from Acid 1)) will 
produce delta_9_10 which means all deletes by txn 10 affecting rows in base_8 
are lost.


> Fine tune Compaction to take advantage of Acid 2.0
> --
>
> Key: HIVE-16669
> URL: https://issues.apache.org/jira/browse/HIVE-16669
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
>Priority: Blocker
>
> * There is little point using 2.0 vectorized reader since there is no 
> operator pipeline in compaction
> * If minor compaction just concats delete_delta files together, then the 2 
> stage compaction should always ensure that we have a limited number of Orc 
> readers to do the merging and current OrcRawRecordMerger should be fine
> * ...



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-16669) Fine tune Compaction to take advantage of Acid 2.0

2017-09-20 Thread Eugene Koifman (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-16669?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eugene Koifman updated HIVE-16669:
--
Priority: Blocker  (was: Major)

> Fine tune Compaction to take advantage of Acid 2.0
> --
>
> Key: HIVE-16669
> URL: https://issues.apache.org/jira/browse/HIVE-16669
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
>Priority: Blocker
>
> * There is little point using 2.0 vectorized reader since there is no 
> operator pipeline in compaction
> * If minor compaction just concats delete_delta files together, then the 2 
> stage compaction should always ensure that we have a limited number of Orc 
> readers to do the merging and current OrcRawRecordMerger should be fine
> * ...



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17466) Metastore API to list unique partition-key-value combinations

2017-09-20 Thread Lefty Leverenz (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17466?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174079#comment-16174079
 ] 

Lefty Leverenz commented on HIVE-17466:
---

Okay, although it could be documented on a different wiki page.  We don't have 
a page about using the metastore but we could create one.  Or it could be 
mentioned on the DDL page as a non-SQL alternative to SHOW PARTITIONS.

I'll just assume no doc is needed unless someone says otherwise.

> Metastore API to list unique partition-key-value combinations
> -
>
> Key: HIVE-17466
> URL: https://issues.apache.org/jira/browse/HIVE-17466
> Project: Hive
>  Issue Type: New Feature
>  Components: Metastore
>Affects Versions: 2.2.0, 3.0.0
>Reporter: Mithun Radhakrishnan
>Assignee: Thiruvel Thirumoolan
> Fix For: 3.0.0, 2.4.0
>
> Attachments: HIVE-17466.1.patch, HIVE-17466.2-branch-2.patch, 
> HIVE-17466.2.patch, HIVE-17466.3.patch
>
>
> Raising this on behalf of [~thiruvel], who wrote this initially as part of a 
> tangential "data-discovery" system.
> Programs like Apache Oozie, Apache Falcon (or Yahoo GDM), etc. launch 
> workflows based on the availability of table/partitions. Partitions are 
> currently discovered by listing partitions using (what boils down to) 
> {{HiveMetaStoreClient.listPartitions()}}. This can be slow and cumbersome, 
> given that {{Partition}} objects are heavyweight and carry redundant 
> information. The alternative is to use partition-names, which will need 
> client-side parsing to extract part-key values.
> When checking which hourly partitions for a particular day have been 
> published already, it would be preferable to have an API that pushed down 
> part-key extraction into the {{RawStore}} layer, and returned key-values as 
> the result. This would be similar to how {{SELECT DISTINCT part_key FROM 
> my_table;}} would run, but at the {{HiveMetaStoreClient}} level.
> Here's what we've been using at Yahoo.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17566) Create schema required for workload management.

2017-09-20 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17566?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174045#comment-16174045
 ] 

Hive QA commented on HIVE-17566:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12888032/HIVE-17566.01.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 14 failed/errored test(s), 11046 tests 
executed
*Failed tests:*
{noformat}
TestAccumuloCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=230)
TestDummy - did not produce a TEST-*.xml file (likely timed out) (batchId=230)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[create_view] (batchId=39)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[udf_mask_hash] 
(batchId=28)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[sysdb] 
(batchId=153)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[union_fast_stats]
 (batchId=156)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_explainuser_1]
 (batchId=170)
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[drop_table_failure2]
 (batchId=89)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query14] 
(batchId=234)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query23] 
(batchId=234)
org.apache.hive.beeline.TestSchemaTool.testHiveMetastoreDbPropertiesTable 
(batchId=221)
org.apache.hive.beeline.TestSchemaTool.testMetastoreDbPropertiesAfterUpgrade 
(batchId=221)
org.apache.hive.beeline.TestSchemaTool.testValidateSchemaTables (batchId=221)
org.apache.hive.jdbc.TestSSL.testMetastoreWithSSL (batchId=224)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6910/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6910/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6910/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 14 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12888032 - PreCommit-HIVE-Build

> Create schema required for workload management.
> ---
>
> Key: HIVE-17566
> URL: https://issues.apache.org/jira/browse/HIVE-17566
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Harish Jaiprakash
>Assignee: Harish Jaiprakash
> Attachments: HIVE-17566.01.patch
>
>
> Schema + model changes required for workload management.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17489) Separate client-facing and server-side Kerberos principals, to support HA

2017-09-20 Thread Mithun Radhakrishnan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17489?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mithun Radhakrishnan updated HIVE-17489:

Status: Patch Available  (was: Open)

> Separate client-facing and server-side Kerberos principals, to support HA
> -
>
> Key: HIVE-17489
> URL: https://issues.apache.org/jira/browse/HIVE-17489
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Reporter: Mithun Radhakrishnan
>Assignee: Thiruvel Thirumoolan
> Attachments: HIVE-17489.2-branch-2.patch, HIVE-17489.2.patch, 
> HIVE-17489.2.patch, HIVE-17489.3-branch-2.patch, HIVE-17489.3.patch
>
>
> On deployments of the Hive metastore where a farm of servers is fronted by a 
> VIP, the hostname of the VIP (e.g. {{mycluster-hcat.blue.myth.net}}) will 
> differ from the actual boxen in the farm (.e.g 
> {{mycluster-hcat-\[0..3\].blue.myth.net}}).
> Such a deployment messes up Kerberos auth, with principals like 
> {{hcat/mycluster-hcat.blue.myth@grid.myth.net}}. Host-based checks will 
> disallow servers behind the VIP from using the VIP's hostname in its 
> principal when accessing, say, HDFS.
> The solution would be to decouple the server-side principal (used to access 
> other services like HDFS as a client) from the client-facing principal (used 
> from Hive-client, BeeLine, etc.).



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17489) Separate client-facing and server-side Kerberos principals, to support HA

2017-09-20 Thread Mithun Radhakrishnan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17489?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mithun Radhakrishnan updated HIVE-17489:

Attachment: HIVE-17489.3-branch-2.patch
HIVE-17489.3.patch

Added logic to fall back to using the server-side principals, if client-facing 
principals are not set. This adds backward compatibility, and should sort out 
the failing tests.

> Separate client-facing and server-side Kerberos principals, to support HA
> -
>
> Key: HIVE-17489
> URL: https://issues.apache.org/jira/browse/HIVE-17489
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Reporter: Mithun Radhakrishnan
>Assignee: Thiruvel Thirumoolan
> Attachments: HIVE-17489.2-branch-2.patch, HIVE-17489.2.patch, 
> HIVE-17489.2.patch, HIVE-17489.3-branch-2.patch, HIVE-17489.3.patch
>
>
> On deployments of the Hive metastore where a farm of servers is fronted by a 
> VIP, the hostname of the VIP (e.g. {{mycluster-hcat.blue.myth.net}}) will 
> differ from the actual boxen in the farm (.e.g 
> {{mycluster-hcat-\[0..3\].blue.myth.net}}).
> Such a deployment messes up Kerberos auth, with principals like 
> {{hcat/mycluster-hcat.blue.myth@grid.myth.net}}. Host-based checks will 
> disallow servers behind the VIP from using the VIP's hostname in its 
> principal when accessing, say, HDFS.
> The solution would be to decouple the server-side principal (used to access 
> other services like HDFS as a client) from the client-facing principal (used 
> from Hive-client, BeeLine, etc.).



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17489) Separate client-facing and server-side Kerberos principals, to support HA

2017-09-20 Thread Mithun Radhakrishnan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17489?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mithun Radhakrishnan updated HIVE-17489:

Status: Open  (was: Patch Available)

> Separate client-facing and server-side Kerberos principals, to support HA
> -
>
> Key: HIVE-17489
> URL: https://issues.apache.org/jira/browse/HIVE-17489
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Reporter: Mithun Radhakrishnan
>Assignee: Thiruvel Thirumoolan
> Attachments: HIVE-17489.2-branch-2.patch, HIVE-17489.2.patch, 
> HIVE-17489.2.patch
>
>
> On deployments of the Hive metastore where a farm of servers is fronted by a 
> VIP, the hostname of the VIP (e.g. {{mycluster-hcat.blue.myth.net}}) will 
> differ from the actual boxen in the farm (.e.g 
> {{mycluster-hcat-\[0..3\].blue.myth.net}}).
> Such a deployment messes up Kerberos auth, with principals like 
> {{hcat/mycluster-hcat.blue.myth@grid.myth.net}}. Host-based checks will 
> disallow servers behind the VIP from using the VIP's hostname in its 
> principal when accessing, say, HDFS.
> The solution would be to decouple the server-side principal (used to access 
> other services like HDFS as a client) from the client-facing principal (used 
> from Hive-client, BeeLine, etc.).



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-14813) Make TransactionBatchImpl.toString() include state of each txn: commit/abort

2017-09-20 Thread Eugene Koifman (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-14813?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16174019#comment-16174019
 ] 

Eugene Koifman commented on HIVE-14813:
---

[~alangates] could you review please

> Make TransactionBatchImpl.toString() include state of each txn: commit/abort
> 
>
> Key: HIVE-14813
> URL: https://issues.apache.org/jira/browse/HIVE-14813
> Project: Hive
>  Issue Type: Bug
>  Components: HCatalog, Transactions
>Affects Versions: 1.2.1
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
> Attachments: HIVE-14813.01.patch
>
>
> perhaps include user abort vs system abort on batch close
> also include if close() was called and if there was an exception
> cc [~deepesh]



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17110) BucketCodec should enforce value ranges

2017-09-20 Thread Eugene Koifman (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17110?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eugene Koifman updated HIVE-17110:
--
Attachment: HIVE-17110.02.patch

> BucketCodec should enforce value ranges
> ---
>
> Key: HIVE-17110
> URL: https://issues.apache.org/jira/browse/HIVE-17110
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
> Fix For: 3.0.0
>
> Attachments: HIVE-17110.01.patch, HIVE-17110.02.patch
>
>
> BucketCodec added in HIVE-16832 can only handle statement id/bucket id values 
> in a particular range.  Add an assert to check that values are indeed in that 
> range.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-14813) Make TransactionBatchImpl.toString() include state of each txn: commit/abort

2017-09-20 Thread Eugene Koifman (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-14813?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eugene Koifman updated HIVE-14813:
--
Status: Patch Available  (was: Open)

> Make TransactionBatchImpl.toString() include state of each txn: commit/abort
> 
>
> Key: HIVE-14813
> URL: https://issues.apache.org/jira/browse/HIVE-14813
> Project: Hive
>  Issue Type: Bug
>  Components: HCatalog, Transactions
>Affects Versions: 1.2.1
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
> Attachments: HIVE-14813.01.patch
>
>
> perhaps include user abort vs system abort on batch close
> also include if close() was called and if there was an exception
> cc [~deepesh]



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-14813) Make TransactionBatchImpl.toString() include state of each txn: commit/abort

2017-09-20 Thread Eugene Koifman (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-14813?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eugene Koifman updated HIVE-14813:
--
Attachment: HIVE-14813.01.patch

> Make TransactionBatchImpl.toString() include state of each txn: commit/abort
> 
>
> Key: HIVE-14813
> URL: https://issues.apache.org/jira/browse/HIVE-14813
> Project: Hive
>  Issue Type: Bug
>  Components: HCatalog, Transactions
>Affects Versions: 1.2.1
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
> Attachments: HIVE-14813.01.patch
>
>
> perhaps include user abort vs system abort on batch close
> also include if close() was called and if there was an exception
> cc [~deepesh]



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17561) Move TxnStore and implementations to standalone metastore

2017-09-20 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17561?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16173968#comment-16173968
 ] 

Hive QA commented on HIVE-17561:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12888021/HIVE-17561.patch

{color:green}SUCCESS:{color} +1 due to 5 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 70 failed/errored test(s), 11045 tests 
executed
*Failed tests:*
{noformat}
TestAccumuloCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=229)
TestDummy - did not produce a TEST-*.xml file (likely timed out) (batchId=229)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[create_view] (batchId=39)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[udf_mask_hash] 
(batchId=28)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[acid_bucket_pruning]
 (batchId=142)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[acid_globallimit]
 (batchId=151)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[acid_no_buckets]
 (batchId=159)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[acid_vectorization_missing_cols]
 (batchId=150)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[delete_all_non_partitioned]
 (batchId=151)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[delete_all_partitioned]
 (batchId=151)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[delete_tmp_table]
 (batchId=156)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[delete_where_no_match]
 (batchId=151)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[delete_where_non_partitioned]
 (batchId=153)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[delete_where_partitioned]
 (batchId=154)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[delete_whole_partition]
 (batchId=147)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[dynamic_semijoin_reduction_3]
 (batchId=160)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[dynpart_sort_optimization_acid]
 (batchId=155)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[explainuser_1]
 (batchId=153)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[insert_orig_table]
 (batchId=158)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[insert_update_delete]
 (batchId=162)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[insert_values_dynamic_partitioned]
 (batchId=160)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[insert_values_non_partitioned]
 (batchId=149)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[insert_values_partitioned]
 (batchId=161)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[insert_values_tmp_table]
 (batchId=146)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[join_acid_non_acid]
 (batchId=161)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[schema_evol_orc_acid_part]
 (batchId=151)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[schema_evol_orc_acid_part_update]
 (batchId=159)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[schema_evol_orc_acid_table]
 (batchId=153)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[schema_evol_orc_acid_table_update]
 (batchId=159)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[schema_evol_orc_acidvec_part]
 (batchId=159)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[schema_evol_orc_acidvec_part_update]
 (batchId=148)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[schema_evol_orc_acidvec_table]
 (batchId=162)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[schema_evol_orc_acidvec_table_update]
 (batchId=146)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[sqlmerge] 
(batchId=161)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[union_fast_stats]
 (batchId=156)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[update_after_multiple_inserts]
 (batchId=159)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[update_all_non_partitioned]
 (batchId=147)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[update_all_partitioned]
 (batchId=156)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[update_all_types]
 (batchId=149)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[update_tmp_table]
 (batchId=153)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[update_two_cols]
 (batchId=149)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[update_where_no_match]
 (batchId=149)

[jira] [Updated] (HIVE-17512) Not use doAs if distcp privileged user same as user running hive

2017-09-20 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17512?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-17512:
-
Resolution: Fixed
Status: Resolved  (was: Patch Available)

Patch committed to master.
Thanks for the patch Anishek!


> Not use doAs if distcp privileged user same as user running hive
> 
>
> Key: HIVE-17512
> URL: https://issues.apache.org/jira/browse/HIVE-17512
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2
>Affects Versions: 3.0.0
>Reporter: anishek
>Assignee: anishek
>Priority: Minor
> Fix For: 3.0.0
>
> Attachments: HIVE-17512.1.patch, HIVE-17512.2.patch, 
> HIVE-17512.2.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17570) Fix view deletion related test failures (create_view.q etc)

2017-09-20 Thread Thejas M Nair (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17570?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16173942#comment-16173942
 ] 

Thejas M Nair commented on HIVE-17570:
--

+1 pending tests


> Fix view deletion related test failures (create_view.q etc)
> ---
>
> Key: HIVE-17570
> URL: https://issues.apache.org/jira/browse/HIVE-17570
> Project: Hive
>  Issue Type: Bug
>Reporter: Tao Li
>Assignee: Tao Li
> Attachments: HIVE-17570.1.patch, HIVE-17570.2.patch
>
>
> Fixing the bug introduced by HIVE-17459. Sorry that did not capture that in a 
> timely fashion.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17402) Provide object location in the HMS notification messages

2017-09-20 Thread Alexander Kolbasov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17402?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Kolbasov updated HIVE-17402:
--
Attachment: HIVE-17402.01.patch

> Provide object location in the HMS notification messages
> 
>
> Key: HIVE-17402
> URL: https://issues.apache.org/jira/browse/HIVE-17402
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Affects Versions: 2.2.0
>Reporter: Alexander Kolbasov
>Assignee: Alexander Kolbasov
> Attachments: HIVE-17402.01.patch
>
>
> While working on the Apache Sentry project that uses HMS notifications we 
> noticed that these notifications are using some useful data - e.g. location 
> information for the objects. To get around these, ApacheSentry implemented 
> its own version of events 
> (https://github.com/apache/sentry/tree/master/sentry-binding/sentry-binding-hive-follower/src/main/java/org/apache/sentry/binding/metastore/messaging/json).
> It seems to be a useful information for Hive as well, so why not add it 
> directly into the standard message factory?



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17402) Provide object location in the HMS notification messages

2017-09-20 Thread Alexander Kolbasov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17402?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Kolbasov updated HIVE-17402:
--
Status: Patch Available  (was: Open)

> Provide object location in the HMS notification messages
> 
>
> Key: HIVE-17402
> URL: https://issues.apache.org/jira/browse/HIVE-17402
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Affects Versions: 2.2.0
>Reporter: Alexander Kolbasov
>Assignee: Alexander Kolbasov
> Attachments: HIVE-17402.01.patch
>
>
> While working on the Apache Sentry project that uses HMS notifications we 
> noticed that these notifications are using some useful data - e.g. location 
> information for the objects. To get around these, ApacheSentry implemented 
> its own version of events 
> (https://github.com/apache/sentry/tree/master/sentry-binding/sentry-binding-hive-follower/src/main/java/org/apache/sentry/binding/metastore/messaging/json).
> It seems to be a useful information for Hive as well, so why not add it 
> directly into the standard message factory?



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17402) Provide object location in the HMS notification messages

2017-09-20 Thread Alexander Kolbasov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17402?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Kolbasov updated HIVE-17402:
--
Status: Open  (was: Patch Available)

> Provide object location in the HMS notification messages
> 
>
> Key: HIVE-17402
> URL: https://issues.apache.org/jira/browse/HIVE-17402
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Affects Versions: 2.2.0
>Reporter: Alexander Kolbasov
>Assignee: Alexander Kolbasov
>
> While working on the Apache Sentry project that uses HMS notifications we 
> noticed that these notifications are using some useful data - e.g. location 
> information for the objects. To get around these, ApacheSentry implemented 
> its own version of events 
> (https://github.com/apache/sentry/tree/master/sentry-binding/sentry-binding-hive-follower/src/main/java/org/apache/sentry/binding/metastore/messaging/json).
> It seems to be a useful information for Hive as well, so why not add it 
> directly into the standard message factory?



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17402) Provide object location in the HMS notification messages

2017-09-20 Thread Alexander Kolbasov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17402?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Kolbasov updated HIVE-17402:
--
Attachment: (was: HIVE-17402.01.patch)

> Provide object location in the HMS notification messages
> 
>
> Key: HIVE-17402
> URL: https://issues.apache.org/jira/browse/HIVE-17402
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Affects Versions: 2.2.0
>Reporter: Alexander Kolbasov
>Assignee: Alexander Kolbasov
>
> While working on the Apache Sentry project that uses HMS notifications we 
> noticed that these notifications are using some useful data - e.g. location 
> information for the objects. To get around these, ApacheSentry implemented 
> its own version of events 
> (https://github.com/apache/sentry/tree/master/sentry-binding/sentry-binding-hive-follower/src/main/java/org/apache/sentry/binding/metastore/messaging/json).
> It seems to be a useful information for Hive as well, so why not add it 
> directly into the standard message factory?



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17426) Execution framework in hive to run tasks in parallel

2017-09-20 Thread Daniel Dai (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17426?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16173889#comment-16173889
 ] 

Daniel Dai commented on HIVE-17426:
---

It sounds like there will be no parallel task run with explain/repldump no 
matter the value of canExecuteInParallel(). Then we don't need to make them 
standout. We only need to disable fetch/ddl for now until it is fixed.

> Execution framework in hive to run tasks in parallel
> 
>
> Key: HIVE-17426
> URL: https://issues.apache.org/jira/browse/HIVE-17426
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2
>Affects Versions: 3.0.0
>Reporter: anishek
>Assignee: anishek
> Fix For: 3.0.0
>
> Attachments: HIVE-17426.0.patch, HIVE-17426.1.patch, 
> HIVE-17426.2.patch, HIVE-17426.3.patch, HIVE-17426.4.patch, HIVE-17426.5.patch
>
>
> the execution framework currently only runs MR / Spark  Tasks in parallel 
> when {{set hive.exec.parallel=true}}.
> Allow other types of tasks to run in parallel as well to support replication 
> scenarios in hive. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-16898) Validation of source file after distcp in repl load

2017-09-20 Thread Daniel Dai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-16898?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daniel Dai updated HIVE-16898:
--
Attachment: HIVE-16898.6.patch

> Validation of source file after distcp in repl load 
> 
>
> Key: HIVE-16898
> URL: https://issues.apache.org/jira/browse/HIVE-16898
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2
>Affects Versions: 3.0.0
>Reporter: anishek
>Assignee: Daniel Dai
> Fix For: 3.0.0
>
> Attachments: HIVE-16898.1.patch, HIVE-16898.2.patch, 
> HIVE-16898.3.patch, HIVE-16898.4.patch, HIVE-16898.5.patch, HIVE-16898.6.patch
>
>
> time between deciding the source and destination path for distcp to invoking 
> of distcp can have a change of the source file, hence distcp might copy the 
> wrong file to destination, hence we should an additional check on the 
> checksum of the source file path after distcp finishes to make sure the path 
> didnot change during the copy process. if it has take additional steps to 
> delete the previous file on destination and copy the new source and repeat 
> the same process as above till we copy the correct file. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-16898) Validation of source file after distcp in repl load

2017-09-20 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-16898?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16173877#comment-16173877
 ] 

Hive QA commented on HIVE-16898:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12888016/HIVE-16898.5.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 10 failed/errored test(s), 11049 tests 
executed
*Failed tests:*
{noformat}
TestAccumuloCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=230)
TestDummy - did not produce a TEST-*.xml file (likely timed out) (batchId=230)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[create_view] (batchId=39)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[udf_mask_hash] 
(batchId=28)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[union_fast_stats]
 (batchId=156)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_explainuser_1]
 (batchId=170)
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[drop_table_failure2]
 (batchId=89)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query14] 
(batchId=234)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query23] 
(batchId=234)
org.apache.hive.hcatalog.pig.TestTextFileHCatStorer.testWriteDecimalXY 
(batchId=183)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6908/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6908/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6908/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 10 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12888016 - PreCommit-HIVE-Build

> Validation of source file after distcp in repl load 
> 
>
> Key: HIVE-16898
> URL: https://issues.apache.org/jira/browse/HIVE-16898
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2
>Affects Versions: 3.0.0
>Reporter: anishek
>Assignee: Daniel Dai
> Fix For: 3.0.0
>
> Attachments: HIVE-16898.1.patch, HIVE-16898.2.patch, 
> HIVE-16898.3.patch, HIVE-16898.4.patch, HIVE-16898.5.patch
>
>
> time between deciding the source and destination path for distcp to invoking 
> of distcp can have a change of the source file, hence distcp might copy the 
> wrong file to destination, hence we should an additional check on the 
> checksum of the source file path after distcp finishes to make sure the path 
> didnot change during the copy process. if it has take additional steps to 
> delete the previous file on destination and copy the new source and repeat 
> the same process as above till we copy the correct file. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17568) HiveJoinPushTransitivePredicatesRule may exchange predicates which are not valid on the other branch

2017-09-20 Thread Ashutosh Chauhan (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17568?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16173860#comment-16173860
 ] 

Ashutosh Chauhan commented on HIVE-17568:
-

+1 
Although we may potentially introduce casts, I think its better to not do that 
for such type-unsafe queries.

> HiveJoinPushTransitivePredicatesRule may exchange predicates which are not 
> valid on the other branch
> 
>
> Key: HIVE-17568
> URL: https://issues.apache.org/jira/browse/HIVE-17568
> Project: Hive
>  Issue Type: Bug
>  Components: Logical Optimizer
>Reporter: Zoltan Haindrich
>Assignee: Zoltan Haindrich
> Attachments: HIVE-17568.01.patch
>
>
> Joining 2 tables on at least 1 column which is not the same type ; 
> (integer/double for example).
> The calcite expressions require double/integer inputs which will became 
> invalid if {{HiveJoinPushTransitivePredicatesRule}} pushes them to the other 
> branch.
> query:
> {code}
> create table t1 (v string, k int);
> insert into t1 values ('people', 10), ('strangers', 20), ('parents', 30);
> create table t2 (v string, k double);
> insert into t2 values ('people', 10), ('strangers', 20), ('parents', 30);
> select * from t1 where t1.k in (select t2.k from t2 where t2.v='people') and 
> t1.k<15;
> {code}
> results in:
> {code}
> java.lang.AssertionError: type mismatch:
> type1:
> DOUBLE
> type2:
> INTEGER
>   at org.apache.calcite.util.Litmus$1.fail(Litmus.java:31)
>   at org.apache.calcite.plan.RelOptUtil.eq(RelOptUtil.java:1841)
>   at 
> org.apache.calcite.rex.RexProgramBuilder$RegisterInputShuttle.visitInputRef(RexProgramBuilder.java:941)
>   at 
> org.apache.calcite.rex.RexProgramBuilder$RegisterInputShuttle.visitInputRef(RexProgramBuilder.java:919)
>   at org.apache.calcite.rex.RexInputRef.accept(RexInputRef.java:112)
>   at org.apache.calcite.rex.RexShuttle.visitList(RexShuttle.java:153)
>   at org.apache.calcite.rex.RexShuttle.visitCall(RexShuttle.java:102)
>   at 
> org.apache.calcite.rex.RexProgramBuilder$RegisterShuttle.visitCall(RexProgramBuilder.java:884)
>   at 
> org.apache.calcite.rex.RexProgramBuilder$RegisterShuttle.visitCall(RexProgramBuilder.java:882)
>   at org.apache.calcite.rex.RexCall.accept(RexCall.java:104)
>   at 
> org.apache.calcite.rex.RexProgramBuilder.registerInput(RexProgramBuilder.java:296)
>   at 
> org.apache.calcite.rex.RexProgramBuilder.addCondition(RexProgramBuilder.java:271)
>   at 
> org.apache.calcite.rel.rules.FilterMergeRule.createProgram(FilterMergeRule.java:98)
>   at 
> org.apache.calcite.rel.rules.FilterMergeRule.onMatch(FilterMergeRule.java:67)
> [...]
> {code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17558) Skip non-native/temporary tables for constraint related scenarios

2017-09-20 Thread Daniel Dai (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17558?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16173847#comment-16173847
 ] 

Daniel Dai commented on HIVE-17558:
---

+1 pending test.

> Skip non-native/temporary tables for constraint related scenarios
> -
>
> Key: HIVE-17558
> URL: https://issues.apache.org/jira/browse/HIVE-17558
> Project: Hive
>  Issue Type: Bug
>Reporter: Tao Li
>Assignee: Tao Li
> Attachments: HIVE-17558.1.patch, HIVE-17558.2.patch
>
>
> The change would be similar to HIVE-17422.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17502) Reuse of default session should not throw an exception in LLAP w/ Tez

2017-09-20 Thread Thai Bui (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17502?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16173846#comment-16173846
 ] 

Thai Bui commented on HIVE-17502:
-

I see what you mean. So without HiveSession(Impl) and SessionState to support 
reusing of sessions from the same user, the patch is not complete. 
Interestingly, we have been using this patch for a couple of weeks with 3-5 
concurrent users each making and reusing sessions according to this logics and 
things were fine. I did observe a couple of weird problems but I don't thing 
they are related.

Anyhow, I'm happy to contribute and/or solidify the HiveSession(Impl) and/or 
SessionState to make them immutable and/or stateless more suitable to be reused 
by multiple sessions. For example, the current SessionState is bind to a 
thread-local static object, if it's bind to a unique SessionID locked in a 
database (hive metadata store?), or Zookeeper, or HDFS files, things could have 
been different.

It is also possible to change Hue to only issue 1 query at a time per user but 
the point was to go beyond that to allow a much better user experience using 
Hive 2 w/ LLAP + Hue 4.

Let me know what you guys think, we (our big data & analytics group at 
Bazaarvoice) are happy to contribute. Currently we have to build a custom 
hive-exec jar with this logic and deploy this jar specifically in Ambari for 
this to work. Both worlds are fine but I would prefer to push the patch 
upstream to make it official and potentially solidify HiveSession 
implementation incremental. For example, I think making a new HS2 config option 
`hive.sessions.default-session.reuse=false` could work for this patch. If the 
option is false (by default), the logics stay the same and an exception is 
thrown, if true, then the new patch logics apply, allowing avid users to have 
multiple sessions per user. Understandably, having too many options is 
confusing. If that's the case we'll just close this ticket but thanks for the 
discussion either way!

> Reuse of default session should not throw an exception in LLAP w/ Tez
> -
>
> Key: HIVE-17502
> URL: https://issues.apache.org/jira/browse/HIVE-17502
> Project: Hive
>  Issue Type: Bug
>  Components: llap, Tez
>Affects Versions: 2.1.1, 2.2.0
> Environment: HDP 2.6.1.0-129, Hue 4
>Reporter: Thai Bui
>Assignee: Thai Bui
>
> Hive2 w/ LLAP on Tez doesn't allow a currently used, default session to be 
> skipped mostly because of this line 
> https://github.com/apache/hive/blob/master/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/TezSessionPoolManager.java#L365.
> However, some clients such as Hue 4, allow multiple sessions to be used per 
> user. Under this configuration, a Thrift client will send a request to either 
> reuse or open a new session. The reuse request could include the session id 
> of a currently used snippet being executed in Hue, this causes HS2 to throw 
> an exception:
> {noformat}
> 2017-09-10T17:51:36,548 INFO  [Thread-89]: tez.TezSessionPoolManager 
> (TezSessionPoolManager.java:canWorkWithSameSession(512)) - The current user: 
> hive, session user: hive
> 2017-09-10T17:51:36,549 ERROR [Thread-89]: exec.Task 
> (TezTask.java:execute(232)) - Failed to execute tez graph.
> org.apache.hadoop.hive.ql.metadata.HiveException: The pool session 
> sessionId=5b61a578-6336-41c5-860d-9838166f97fe, queueName=llap, user=hive, 
> doAs=false, isOpen=true, isDefault=true, expires in 591015330ms should have 
> been returned to the pool
>   at 
> org.apache.hadoop.hive.ql.exec.tez.TezSessionPoolManager.canWorkWithSameSession(TezSessionPoolManager.java:534)
>  ~[hive-exec-2.1.0.2.6.1.0-129.jar:2.1.0.2.6.1.0-129]
>   at 
> org.apache.hadoop.hive.ql.exec.tez.TezSessionPoolManager.getSession(TezSessionPoolManager.java:544)
>  ~[hive-exec-2.1.0.2.6.1.0-129.jar:2.1.0.2.6.1.0-129]
>   at org.apache.hadoop.hive.ql.exec.tez.TezTask.execute(TezTask.java:147) 
> [hive-exec-2.1.0.2.6.1.0-129.jar:2.1.0.2.6.1.0-129]
>   at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:197) 
> [hive-exec-2.1.0.2.6.1.0-129.jar:2.1.0.2.6.1.0-129]
>   at 
> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:100) 
> [hive-exec-2.1.0.2.6.1.0-129.jar:2.1.0.2.6.1.0-129]
>   at org.apache.hadoop.hive.ql.exec.TaskRunner.run(TaskRunner.java:79) 
> [hive-exec-2.1.0.2.6.1.0-129.jar:2.1.0.2.6.1.0-129]
> {noformat}
> Note that every query is issued as a single 'hive' user to share the LLAP 
> daemon pool, a set of pre-determined number of AMs is initialized at setup 
> time. Thus, HS2 should allow new sessions from a Thrift client to be used out 
> of the pool, or an existing session to be skipped and an unused session from 
> the pool to be returned. The logic to throw an exception in the  
> 

[jira] [Updated] (HIVE-17535) Select 1 EXCEPT Select 1 fails with NPE

2017-09-20 Thread Vineet Garg (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17535?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-17535:
---
Resolution: Fixed
Status: Resolved  (was: Patch Available)

Pushed to master. Thanks for reviewing [~ashutoshc]

> Select 1 EXCEPT Select 1 fails with NPE
> ---
>
> Key: HIVE-17535
> URL: https://issues.apache.org/jira/browse/HIVE-17535
> Project: Hive
>  Issue Type: Bug
>  Components: Query Planning
>Reporter: Vineet Garg
>Assignee: Vineet Garg
> Attachments: HIVE-17535.1.patch, HIVE-17535.2.patch, 
> HIVE-17535.3.patch, HIVE-17535.4.patch, HIVE-17535.5.patch
>
>
> Since Hive CBO isn't able to handle queries with no table e.g. {{select 1}} 
> queries with SET operators fail (intersect requires CBO).



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17536) StatsUtil::getBasicStatForTable doesn't distinguish b/w absence of statistics or zero stats

2017-09-20 Thread Vineet Garg (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17536?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-17536:
---
Status: Patch Available  (was: Open)

> StatsUtil::getBasicStatForTable doesn't distinguish b/w absence of statistics 
> or zero stats
> ---
>
> Key: HIVE-17536
> URL: https://issues.apache.org/jira/browse/HIVE-17536
> Project: Hive
>  Issue Type: Improvement
>  Components: Statistics
>Reporter: Vineet Garg
>Assignee: Vineet Garg
> Attachments: HIVE-17536.1.patch, HIVE-17536.2.patch, 
> HIVE-17536.3.patch, HIVE-17536.4.patch, HIVE-17536.5.patch
>
>
> This method returns zero for both of the following cases:
> * Statistics are missing in metastore
> * Actual stats e.g. number of rows are zero
> It'll be good for this method to return e.g. -1 in absence of statistics 
> instead of assuming it to be zero.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17536) StatsUtil::getBasicStatForTable doesn't distinguish b/w absence of statistics or zero stats

2017-09-20 Thread Vineet Garg (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17536?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-17536:
---
Status: Open  (was: Patch Available)

> StatsUtil::getBasicStatForTable doesn't distinguish b/w absence of statistics 
> or zero stats
> ---
>
> Key: HIVE-17536
> URL: https://issues.apache.org/jira/browse/HIVE-17536
> Project: Hive
>  Issue Type: Improvement
>  Components: Statistics
>Reporter: Vineet Garg
>Assignee: Vineet Garg
> Attachments: HIVE-17536.1.patch, HIVE-17536.2.patch, 
> HIVE-17536.3.patch, HIVE-17536.4.patch, HIVE-17536.5.patch
>
>
> This method returns zero for both of the following cases:
> * Statistics are missing in metastore
> * Actual stats e.g. number of rows are zero
> It'll be good for this method to return e.g. -1 in absence of statistics 
> instead of assuming it to be zero.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17536) StatsUtil::getBasicStatForTable doesn't distinguish b/w absence of statistics or zero stats

2017-09-20 Thread Vineet Garg (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17536?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vineet Garg updated HIVE-17536:
---
Attachment: HIVE-17536.5.patch

> StatsUtil::getBasicStatForTable doesn't distinguish b/w absence of statistics 
> or zero stats
> ---
>
> Key: HIVE-17536
> URL: https://issues.apache.org/jira/browse/HIVE-17536
> Project: Hive
>  Issue Type: Improvement
>  Components: Statistics
>Reporter: Vineet Garg
>Assignee: Vineet Garg
> Attachments: HIVE-17536.1.patch, HIVE-17536.2.patch, 
> HIVE-17536.3.patch, HIVE-17536.4.patch, HIVE-17536.5.patch
>
>
> This method returns zero for both of the following cases:
> * Statistics are missing in metastore
> * Actual stats e.g. number of rows are zero
> It'll be good for this method to return e.g. -1 in absence of statistics 
> instead of assuming it to be zero.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17570) Fix view deletion related test failures (create_view.q etc)

2017-09-20 Thread Tao Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17570?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16173803#comment-16173803
 ] 

Tao Li commented on HIVE-17570:
---

Just uploaded another iteration.

> Fix view deletion related test failures (create_view.q etc)
> ---
>
> Key: HIVE-17570
> URL: https://issues.apache.org/jira/browse/HIVE-17570
> Project: Hive
>  Issue Type: Bug
>Reporter: Tao Li
>Assignee: Tao Li
> Attachments: HIVE-17570.1.patch, HIVE-17570.2.patch
>
>
> Fixing the bug introduced by HIVE-17459. Sorry that did not capture that in a 
> timely fashion.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17570) Fix view deletion related test failures (create_view.q etc)

2017-09-20 Thread Tao Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17570?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tao Li updated HIVE-17570:
--
Attachment: HIVE-17570.2.patch

> Fix view deletion related test failures (create_view.q etc)
> ---
>
> Key: HIVE-17570
> URL: https://issues.apache.org/jira/browse/HIVE-17570
> Project: Hive
>  Issue Type: Bug
>Reporter: Tao Li
>Assignee: Tao Li
> Attachments: HIVE-17570.1.patch, HIVE-17570.2.patch
>
>
> Fixing the bug introduced by HIVE-17459. Sorry that did not capture that in a 
> timely fashion.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17570) Fix view deletion related test failures (create_view.q etc)

2017-09-20 Thread Tao Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17570?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tao Li updated HIVE-17570:
--
Status: Patch Available  (was: Open)

> Fix view deletion related test failures (create_view.q etc)
> ---
>
> Key: HIVE-17570
> URL: https://issues.apache.org/jira/browse/HIVE-17570
> Project: Hive
>  Issue Type: Bug
>Reporter: Tao Li
>Assignee: Tao Li
> Attachments: HIVE-17570.1.patch, HIVE-17570.2.patch
>
>
> Fixing the bug introduced by HIVE-17459. Sorry that did not capture that in a 
> timely fashion.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17570) Fix view deletion related test failures (create_view.q etc)

2017-09-20 Thread Tao Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17570?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tao Li updated HIVE-17570:
--
Status: Open  (was: Patch Available)

> Fix view deletion related test failures (create_view.q etc)
> ---
>
> Key: HIVE-17570
> URL: https://issues.apache.org/jira/browse/HIVE-17570
> Project: Hive
>  Issue Type: Bug
>Reporter: Tao Li
>Assignee: Tao Li
> Attachments: HIVE-17570.1.patch, HIVE-17570.2.patch
>
>
> Fixing the bug introduced by HIVE-17459. Sorry that did not capture that in a 
> timely fashion.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17558) Skip non-native/temporary tables for constraint related scenarios

2017-09-20 Thread Tao Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17558?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tao Li updated HIVE-17558:
--
Status: Patch Available  (was: Open)

> Skip non-native/temporary tables for constraint related scenarios
> -
>
> Key: HIVE-17558
> URL: https://issues.apache.org/jira/browse/HIVE-17558
> Project: Hive
>  Issue Type: Bug
>Reporter: Tao Li
>Assignee: Tao Li
> Attachments: HIVE-17558.1.patch, HIVE-17558.2.patch
>
>
> The change would be similar to HIVE-17422.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17562) ACID 1.0 + ETL strategy should treat empty compacted files as uncovered deltas

2017-09-20 Thread Prasanth Jayachandran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17562?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran updated HIVE-17562:
-
Attachment: HIVE-17562-branch-2.patch

Reuploading patch with branch name to see if precommit picks up correct branch.

> ACID 1.0 + ETL strategy should treat empty compacted files as uncovered deltas
> --
>
> Key: HIVE-17562
> URL: https://issues.apache.org/jira/browse/HIVE-17562
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Affects Versions: 2.4.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
> Attachments: HIVE-17562.1.branch-2.patch, HIVE-17562-branch-2.patch
>
>
> In branch-2, with ACID 1.0, following sequence will result in incorrect 
> results
> 1) Set split strategy to ETL
> 2) Insert some rows
> 3) Delete all rows
> 4) Alter table compact MAJOR
> 5) Insert some rows
> 6) Select * query will not return any rows that is written at last (step 5)
> The reason for that, compaction essentially voids the first insert in step 2. 
> Now when ETL split strategy is chosen, there will not be any stripes in the 
> base files. So no split gets generated and any subsequent deltas gets ignored.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Assigned] (HIVE-17571) update sql standard authorization config whitelist to include distcp options for replication

2017-09-20 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17571?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair reassigned HIVE-17571:



> update sql standard authorization config whitelist to include distcp options 
> for replication
> 
>
> Key: HIVE-17571
> URL: https://issues.apache.org/jira/browse/HIVE-17571
> Project: Hive
>  Issue Type: Bug
>Reporter: Thejas M Nair
>Assignee: Thejas M Nair
>
> Additional distcp config options (added in HIVE-16686) need to be added to 
> whitelist of configs that can be updated at runtime, for sql standard 
> authorization.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17558) Skip non-native/temporary tables for constraint related scenarios

2017-09-20 Thread Tao Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17558?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tao Li updated HIVE-17558:
--
Attachment: HIVE-17558.2.patch

> Skip non-native/temporary tables for constraint related scenarios
> -
>
> Key: HIVE-17558
> URL: https://issues.apache.org/jira/browse/HIVE-17558
> Project: Hive
>  Issue Type: Bug
>Reporter: Tao Li
>Assignee: Tao Li
> Attachments: HIVE-17558.1.patch, HIVE-17558.2.patch
>
>
> The change would be similar to HIVE-17422.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17558) Skip non-native/temporary tables for constraint related scenarios

2017-09-20 Thread Tao Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17558?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tao Li updated HIVE-17558:
--
Attachment: (was: HIVE-17558.2.patch)

> Skip non-native/temporary tables for constraint related scenarios
> -
>
> Key: HIVE-17558
> URL: https://issues.apache.org/jira/browse/HIVE-17558
> Project: Hive
>  Issue Type: Bug
>Reporter: Tao Li
>Assignee: Tao Li
> Attachments: HIVE-17558.1.patch
>
>
> The change would be similar to HIVE-17422.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17558) Skip non-native/temporary tables for constraint related scenarios

2017-09-20 Thread Tao Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17558?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tao Li updated HIVE-17558:
--
Attachment: (was: HIVE-17558.2.patch)

> Skip non-native/temporary tables for constraint related scenarios
> -
>
> Key: HIVE-17558
> URL: https://issues.apache.org/jira/browse/HIVE-17558
> Project: Hive
>  Issue Type: Bug
>Reporter: Tao Li
>Assignee: Tao Li
> Attachments: HIVE-17558.1.patch, HIVE-17558.2.patch
>
>
> The change would be similar to HIVE-17422.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17558) Skip non-native/temporary tables for constraint related scenarios

2017-09-20 Thread Tao Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17558?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tao Li updated HIVE-17558:
--
Attachment: HIVE-17558.2.patch

> Skip non-native/temporary tables for constraint related scenarios
> -
>
> Key: HIVE-17558
> URL: https://issues.apache.org/jira/browse/HIVE-17558
> Project: Hive
>  Issue Type: Bug
>Reporter: Tao Li
>Assignee: Tao Li
> Attachments: HIVE-17558.1.patch, HIVE-17558.2.patch
>
>
> The change would be similar to HIVE-17422.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17562) ACID 1.0 + ETL strategy should treat empty compacted files as uncovered deltas

2017-09-20 Thread Prasanth Jayachandran (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17562?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16173783#comment-16173783
 ] 

Prasanth Jayachandran commented on HIVE-17562:
--

The patch is applied against master by precommit and hence the compilation 
error. This issue happens only in branch-2. In master, with ACID 2.0 ACID split 
strategy is essentially a no-op. Whereas in branch-2, ACID split strategy now 
treats base files without any stripes as uncovered deltas. The only breaking 
test in branch-2 is testETLCompactedNoStripes which will fail the assertion 
after the second insert query. testBICompactedNoStripes is added just for test 
coverage.



> ACID 1.0 + ETL strategy should treat empty compacted files as uncovered deltas
> --
>
> Key: HIVE-17562
> URL: https://issues.apache.org/jira/browse/HIVE-17562
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Affects Versions: 2.4.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
> Attachments: HIVE-17562.1.branch-2.patch
>
>
> In branch-2, with ACID 1.0, following sequence will result in incorrect 
> results
> 1) Set split strategy to ETL
> 2) Insert some rows
> 3) Delete all rows
> 4) Alter table compact MAJOR
> 5) Insert some rows
> 6) Select * query will not return any rows that is written at last (step 5)
> The reason for that, compaction essentially voids the first insert in step 2. 
> Now when ETL split strategy is chosen, there will not be any stripes in the 
> base files. So no split gets generated and any subsequent deltas gets ignored.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17558) Skip non-native/temporary tables for constraint related scenarios

2017-09-20 Thread Tao Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17558?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tao Li updated HIVE-17558:
--
Status: Open  (was: Patch Available)

> Skip non-native/temporary tables for constraint related scenarios
> -
>
> Key: HIVE-17558
> URL: https://issues.apache.org/jira/browse/HIVE-17558
> Project: Hive
>  Issue Type: Bug
>Reporter: Tao Li
>Assignee: Tao Li
> Attachments: HIVE-17558.1.patch, HIVE-17558.2.patch
>
>
> The change would be similar to HIVE-17422.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17558) Skip non-native/temporary tables for constraint related scenarios

2017-09-20 Thread Tao Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17558?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tao Li updated HIVE-17558:
--
Status: Open  (was: Patch Available)

> Skip non-native/temporary tables for constraint related scenarios
> -
>
> Key: HIVE-17558
> URL: https://issues.apache.org/jira/browse/HIVE-17558
> Project: Hive
>  Issue Type: Bug
>Reporter: Tao Li
>Assignee: Tao Li
> Attachments: HIVE-17558.1.patch, HIVE-17558.2.patch
>
>
> The change would be similar to HIVE-17422.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17558) Skip non-native/temporary tables for constraint related scenarios

2017-09-20 Thread Tao Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17558?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tao Li updated HIVE-17558:
--
Attachment: HIVE-17558.2.patch

> Skip non-native/temporary tables for constraint related scenarios
> -
>
> Key: HIVE-17558
> URL: https://issues.apache.org/jira/browse/HIVE-17558
> Project: Hive
>  Issue Type: Bug
>Reporter: Tao Li
>Assignee: Tao Li
> Attachments: HIVE-17558.1.patch, HIVE-17558.2.patch
>
>
> The change would be similar to HIVE-17422.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17558) Skip non-native/temporary tables for constraint related scenarios

2017-09-20 Thread Tao Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17558?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tao Li updated HIVE-17558:
--
Status: Patch Available  (was: Open)

> Skip non-native/temporary tables for constraint related scenarios
> -
>
> Key: HIVE-17558
> URL: https://issues.apache.org/jira/browse/HIVE-17558
> Project: Hive
>  Issue Type: Bug
>Reporter: Tao Li
>Assignee: Tao Li
> Attachments: HIVE-17558.1.patch, HIVE-17558.2.patch
>
>
> The change would be similar to HIVE-17422.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17402) Provide object location in the HMS notification messages

2017-09-20 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17402?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16173761#comment-16173761
 ] 

Hive QA commented on HIVE-17402:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12888011/HIVE-17402.01.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 14 failed/errored test(s), 11049 tests 
executed
*Failed tests:*
{noformat}
TestAccumuloCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=230)
TestDummy - did not produce a TEST-*.xml file (likely timed out) (batchId=230)
org.apache.hadoop.hive.cli.TestBeeLineDriver.testCliDriver[smb_mapjoin_13] 
(batchId=239)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[create_view] (batchId=39)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[udf_mask_hash] 
(batchId=28)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[union_fast_stats]
 (batchId=156)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_explainuser_1]
 (batchId=170)
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[drop_table_failure2]
 (batchId=89)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query14] 
(batchId=234)
org.apache.hive.hcatalog.api.repl.TestReplicationTask.testCreate (batchId=180)
org.apache.hive.hcatalog.api.repl.exim.TestEximReplicationTasks.testAlterPartition
 (batchId=180)
org.apache.hive.hcatalog.api.repl.exim.TestEximReplicationTasks.testAlterTable 
(batchId=180)
org.apache.hive.hcatalog.api.repl.exim.TestEximReplicationTasks.testCreateTable 
(batchId=180)
org.apache.hive.hcatalog.api.repl.exim.TestEximReplicationTasks.testDropTable 
(batchId=180)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6907/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6907/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6907/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 14 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12888011 - PreCommit-HIVE-Build

> Provide object location in the HMS notification messages
> 
>
> Key: HIVE-17402
> URL: https://issues.apache.org/jira/browse/HIVE-17402
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Affects Versions: 2.2.0
>Reporter: Alexander Kolbasov
>Assignee: Alexander Kolbasov
> Attachments: HIVE-17402.01.patch
>
>
> While working on the Apache Sentry project that uses HMS notifications we 
> noticed that these notifications are using some useful data - e.g. location 
> information for the objects. To get around these, ApacheSentry implemented 
> its own version of events 
> (https://github.com/apache/sentry/tree/master/sentry-binding/sentry-binding-hive-follower/src/main/java/org/apache/sentry/binding/metastore/messaging/json).
> It seems to be a useful information for Hive as well, so why not add it 
> directly into the standard message factory?



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17110) BucketCodec should enforce value ranges

2017-09-20 Thread Eugene Koifman (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17110?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eugene Koifman updated HIVE-17110:
--
Status: Patch Available  (was: Open)

[~gopalv] could you review please

> BucketCodec should enforce value ranges
> ---
>
> Key: HIVE-17110
> URL: https://issues.apache.org/jira/browse/HIVE-17110
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
> Fix For: 3.0.0
>
> Attachments: HIVE-17110.01.patch
>
>
> BucketCodec added in HIVE-16832 can only handle statement id/bucket id values 
> in a particular range.  Add an assert to check that values are indeed in that 
> range.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17110) BucketCodec should enforce value ranges

2017-09-20 Thread Eugene Koifman (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17110?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eugene Koifman updated HIVE-17110:
--
Attachment: HIVE-17110.01.patch

> BucketCodec should enforce value ranges
> ---
>
> Key: HIVE-17110
> URL: https://issues.apache.org/jira/browse/HIVE-17110
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
> Fix For: 3.0.0
>
> Attachments: HIVE-17110.01.patch
>
>
> BucketCodec added in HIVE-16832 can only handle statement id/bucket id values 
> in a particular range.  Add an assert to check that values are indeed in that 
> range.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17570) Fix view deletion related test failures (create_view.q etc)

2017-09-20 Thread Thejas M Nair (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17570?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16173732#comment-16173732
 ] 

Thejas M Nair commented on HIVE-17570:
--

Looks good. Can you also remove the getExpectedType() method that is not used 
anymore after this change ?


> Fix view deletion related test failures (create_view.q etc)
> ---
>
> Key: HIVE-17570
> URL: https://issues.apache.org/jira/browse/HIVE-17570
> Project: Hive
>  Issue Type: Bug
>Reporter: Tao Li
>Assignee: Tao Li
> Attachments: HIVE-17570.1.patch
>
>
> Fixing the bug introduced by HIVE-17459. Sorry that did not capture that in a 
> timely fashion.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17510) Make comparison of filter predicates in q files deterministic

2017-09-20 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17510?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-17510:
---
   Resolution: Fixed
Fix Version/s: 3.0.0
   Status: Resolved  (was: Patch Available)

Regenerated last couple of needed q files and verified unrelated failures. 
Pushed to master, thanks for reviewing [~ashutoshc]!

> Make comparison of filter predicates in q files deterministic
> -
>
> Key: HIVE-17510
> URL: https://issues.apache.org/jira/browse/HIVE-17510
> Project: Hive
>  Issue Type: Bug
>  Components: Testing Infrastructure
>Affects Versions: 3.0.0
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
> Fix For: 3.0.0
>
> Attachments: HIVE-17510.01.patch, HIVE-17510.patch
>
>
> I have been hitting this issue while submitting patches to test HIVE-17432.
> Basically, the order on which the rewriting might create the children of AND 
> operations is not deterministic. Thus, tests might fail because the created 
> golden file is not the same, thought the test should pass because they just 
> simply do not follow same order:
> {code}
> predicate: ((d_year >= 1992) and (d_year <= 1997) and ((c_city = 'UNITED 
> KI1') or (c_city = 'UNITED KI5')) and ((s_city = 'UNITED KI1') or (s_city = 
> 'UNITED KI5'))) (type: boolean)
> {code}
> {code}
> predicate: ((d_year <= 1997) and (d_year >= 1992) and ((c_city = 'UNITED 
> KI1') or (c_city = 'UNITED KI5')) and ((s_city = 'UNITED KI1') or (s_city = 
> 'UNITED KI5'))) (type: boolean)
> {code}
> This patches fixes the issue by sorting the children of some expressions 
> (currently AND and OR children) when we run explain plan and we are running 
> in test mode.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17527) Support replication for rename/move table across database

2017-09-20 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17527?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair updated HIVE-17527:
-
Assignee: Sankar Hariappan  (was: Thejas M Nair)
  Status: Patch Available  (was: Open)

> Support replication for rename/move table across database
> -
>
> Key: HIVE-17527
> URL: https://issues.apache.org/jira/browse/HIVE-17527
> Project: Hive
>  Issue Type: Sub-task
>  Components: repl
>Affects Versions: 2.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>  Labels: DR, pull-request-available, replication
> Fix For: 3.0.0
>
> Attachments: HIVE-17527.01.patch, HIVE-17527.01.patch
>
>
> Rename/move table across database should be supported for replication. The 
> scenario is as follows.
> 1. Create 2 databases (db1 and db2) in source cluster.
> 2. Create the table db1.tbl1.
> 3. Run bootstrap replication for db1 and db2 to target cluster.
> 4. Rename db1.tbl1 to db2.tbl1 in source.
> 5. Run incremental replication for both db1 and db2.
> - db1 dump missed the rename table operation as no event is generated for 
> db1. So, table exist after load.
> - db2 load skips the rename event as the source table is missing in target.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Assigned] (HIVE-17527) Support replication for rename/move table across database

2017-09-20 Thread Thejas M Nair (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17527?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair reassigned HIVE-17527:


Assignee: Thejas M Nair  (was: Sankar Hariappan)

> Support replication for rename/move table across database
> -
>
> Key: HIVE-17527
> URL: https://issues.apache.org/jira/browse/HIVE-17527
> Project: Hive
>  Issue Type: Sub-task
>  Components: repl
>Affects Versions: 2.1.0
>Reporter: Sankar Hariappan
>Assignee: Thejas M Nair
>  Labels: DR, pull-request-available, replication
> Fix For: 3.0.0
>
> Attachments: HIVE-17527.01.patch, HIVE-17527.01.patch
>
>
> Rename/move table across database should be supported for replication. The 
> scenario is as follows.
> 1. Create 2 databases (db1 and db2) in source cluster.
> 2. Create the table db1.tbl1.
> 3. Run bootstrap replication for db1 and db2 to target cluster.
> 4. Rename db1.tbl1 to db2.tbl1 in source.
> 5. Run incremental replication for both db1 and db2.
> - db1 dump missed the rename table operation as no event is generated for 
> db1. So, table exist after load.
> - db2 load skips the rename event as the source table is missing in target.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17554) Occurr java.lang.ArithmeticException: / by zero at hplsql component

2017-09-20 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17554?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16173660#comment-16173660
 ] 

Hive QA commented on HIVE-17554:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12888007/HIVE-17554.2.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 10 failed/errored test(s), 11044 tests 
executed
*Failed tests:*
{noformat}
TestAccumuloCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=230)
TestDummy - did not produce a TEST-*.xml file (likely timed out) (batchId=230)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[create_view] (batchId=39)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[udf_mask_hash] 
(batchId=28)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[union_fast_stats]
 (batchId=156)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_explainuser_1]
 (batchId=170)
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[drop_table_failure2]
 (batchId=89)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query14] 
(batchId=234)
org.apache.hadoop.hive.ql.TestAcidOnTez.testCtasTezUnion (batchId=215)
org.apache.hadoop.hive.ql.TestAcidOnTez.testNonStandardConversion01 
(batchId=215)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6906/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6906/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6906/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 10 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12888007 - PreCommit-HIVE-Build

> Occurr java.lang.ArithmeticException: / by zero at hplsql component
> ---
>
> Key: HIVE-17554
> URL: https://issues.apache.org/jira/browse/HIVE-17554
> Project: Hive
>  Issue Type: Bug
>  Components: hpl/sql
>Affects Versions: 3.0.0
>Reporter: ZhangBing Lin
>Assignee: ZhangBing Lin
> Attachments: HIVE-17554.1.patch, HIVE-17554.2.patch
>
>
> When I execute the copyToTable method in hplsql, the following exception 
> occurs:
> java.lang.ArithmeticException: / by zero
>   at org.apache.hive.hplsql.Copy.copyToTable(Copy.java:170)
>   at org.apache.hive.hplsql.Copy.run(Copy.java:93)
>   at org.apache.hive.hplsql.Exec.visitCopy_stmt(Exec.java:1284)
>   at org.apache.hive.hplsql.Exec.visitCopy_stmt(Exec.java:52)
>   at 
> org.apache.hive.hplsql.HplsqlParser$Copy_stmtContext.accept(HplsqlParser.java:12148)
>   at 
> org.antlr.v4.runtime.tree.AbstractParseTreeVisitor.visitChildren(AbstractParseTreeVisitor.java:70)
>   at org.apache.hive.hplsql.Exec.visitStmt(Exec.java:994)
>   at org.apache.hive.hplsql.Exec.visitStmt(Exec.java:52)
>   at 
> org.apache.hive.hplsql.HplsqlParser$StmtContext.accept(HplsqlParser.java:1012)
>   at 
> org.antlr.v4.runtime.tree.AbstractParseTreeVisitor.visitChildren(AbstractParseTreeVisitor.java:70)
>   at 
> org.apache.hive.hplsql.HplsqlBaseVisitor.visitBlock(HplsqlBaseVisitor.java:28)
>   at 
> org.apache.hive.hplsql.HplsqlParser$BlockContext.accept(HplsqlParser.java:446)
>   at 
> org.antlr.v4.runtime.tree.AbstractParseTreeVisitor.visitChildren(AbstractParseTreeVisitor.java:70)
>   at org.apache.hive.hplsql.Exec.visitProgram(Exec.java:901)
>   at org.apache.hive.hplsql.Exec.visitProgram(Exec.java:52)
>   at 
> org.apache.hive.hplsql.HplsqlParser$ProgramContext.accept(HplsqlParser.java:389)
>   at 
> org.antlr.v4.runtime.tree.AbstractParseTreeVisitor.visit(AbstractParseTreeVisitor.java:42)
>   at org.apache.hive.hplsql.Exec.run(Exec.java:760)
>   at org.apache.hive.hplsql.Exec.run(Exec.java:736)
>   at org.apache.hive.hplsql.Hplsql.main(Hplsql.java:23)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:606)
>   at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
>   at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
> View source code:
> if (info) {
>   info(ctx, "COPY completed: " + rows + " row(s), " + timer.format() + ", 
> " + rows/(elapsed/1000) + " rows/sec");
> }
> When elapsed less than one thousand will cause the divisor to be zero



[jira] [Commented] (HIVE-17563) CodahaleMetrics.JsonFileReporter is not updating hive.service.metrics.file.location

2017-09-20 Thread Sahil Takiar (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17563?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16173641#comment-16173641
 ] 

Sahil Takiar commented on HIVE-17563:
-

Yes, it creates a {{ProxyLocalFileSystem}}. I agree HIVE-13705 is suspicious, I 
think it was done in order to make the {{LocaFileSystem}} consist with the 
{{DistributedFileSystem}}. The {{DistributedFileSystem}} is used for accessing 
HDFS, and when it renames a file and the dst already exists, then it returns 
false. So the {{ProxyLocalFileSystem}} was introduced to make the 
{{LocaFileSystem}} consistent HDFS.

Ideally, we should have just changed the {{LocaFileSystem}} itself, there is 
some discussion in HDFS-10385 on why that was never done.

I think we can just use the Java APIs to rename this file, 
{{File.renameTo(File)}} should be sufficient, but we should check whether the 
call to {{renameTo}} returns {{true}} or {{false}}. If it returns {{false}} we 
can throw an exception.

This is similar to what the {{LocaFileSystem#rename}} method is doing under the 
hood (see 
https://github.com/apache/hadoop/blob/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/RawLocalFileSystem.java#L354).

> CodahaleMetrics.JsonFileReporter is not updating 
> hive.service.metrics.file.location
> ---
>
> Key: HIVE-17563
> URL: https://issues.apache.org/jira/browse/HIVE-17563
> Project: Hive
>  Issue Type: Bug
>Reporter: Sahil Takiar
>Assignee: Alexander Kolbasov
>
> The {{JsonFileReporter}} has a bug where it never updates the content of 
> {{hive.service.metrics.file.location}}. The original implementation relied on 
> {{fs.rename}} to overwrite the file if it already existed, but after 
> HIVE-13705 {{fs.rename}} now returns {{false}} if the destination file exists.
> Also, there is no reason for the {{JsonFileReporter}} to write to any fs 
> besides the local filesystem, so we should make this local fs specific.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17562) ACID 1.0 + ETL strategy should treat empty compacted files as uncovered deltas

2017-09-20 Thread Eugene Koifman (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17562?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16173638#comment-16173638
 ] 

Eugene Koifman commented on HIVE-17562:
---

The general idea makes sense but it's not clear to me that the tests actually 
prove that this is working.

Also, build bot failed with a compilation error

> ACID 1.0 + ETL strategy should treat empty compacted files as uncovered deltas
> --
>
> Key: HIVE-17562
> URL: https://issues.apache.org/jira/browse/HIVE-17562
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Affects Versions: 2.4.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
> Attachments: HIVE-17562.1.branch-2.patch
>
>
> In branch-2, with ACID 1.0, following sequence will result in incorrect 
> results
> 1) Set split strategy to ETL
> 2) Insert some rows
> 3) Delete all rows
> 4) Alter table compact MAJOR
> 5) Insert some rows
> 6) Select * query will not return any rows that is written at last (step 5)
> The reason for that, compaction essentially voids the first insert in step 2. 
> Now when ETL split strategy is chosen, there will not be any stripes in the 
> base files. So no split gets generated and any subsequent deltas gets ignored.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17556) The test udf_mask_hash.q is failing

2017-09-20 Thread Tao Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17556?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16173636#comment-16173636
 ] 

Tao Li commented on HIVE-17556:
---

Thanks [~kuczoram] for the fix and [~kgyrtkirk] for the reivew.

> The test udf_mask_hash.q is failing
> ---
>
> Key: HIVE-17556
> URL: https://issues.apache.org/jira/browse/HIVE-17556
> Project: Hive
>  Issue Type: Bug
>  Components: Test
>Affects Versions: 3.0.0
>Reporter: Marta Kuczora
>Assignee: Marta Kuczora
> Attachments: HIVE-17556.1.patch
>
>
> After committing 
> [HIVE-17515|https://issues.apache.org/jira/browse/HIVE-17515], the 
> udf_mask_hash.q test is failing with the following error message:
> {noformat}
> Error Message
> Client Execution succeeded but contained differences (error code = 1) after 
> executing udf_mask_hash.q 
> 34c34
> < expressions: 
> '8b44d559dc5d60e4453c9b4edf2a455fbce054bb8504cd3eb9b5f391bd239c90' (type: 
> string)
> ---
> > expressions: 'dd78d68ad1b23bde126812482dd70ac6' (type: string)
> 36c36
> < Statistics: Num rows: 1 Data size: 148 Basic stats: COMPLETE 
> Column stats: COMPLETE
> ---
> > Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE 
> > Column stats: COMPLETE
> 61c61
> < 8b44d559dc5d60e4453c9b4edf2a455fbce054bb8504cd3eb9b5f391bd239c90 
> 8b44d559dc5d60e4453c9b4edf2a455fbce054bb8504cd3eb9b5f391bd239c90 
> 30a88603135d3a6f7a66b4f9193da1ab4423aed45fb8fe736c2f2a08977f2bdd  
>   
>   
>   NULL NULL NULL NULL NULL
> ---
> > dd78d68ad1b23bde126812482dd70ac6 dd78d68ad1b23bde126812482dd70ac6 
> > 835735ba20f1297683efca69fabd0fba
> > 
> > 
> > NULL NULL NULL NULL NULL
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17529) Bucket Map Join : Sets incorrect edge type causing execution failure

2017-09-20 Thread Deepak Jaiswal (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17529?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Deepak Jaiswal updated HIVE-17529:
--
Attachment: HIVE-17529.5.patch

Added order by in the failing tests due to mismatched order of results.

> Bucket Map Join : Sets incorrect edge type causing execution failure
> 
>
> Key: HIVE-17529
> URL: https://issues.apache.org/jira/browse/HIVE-17529
> Project: Hive
>  Issue Type: Bug
>Reporter: Deepak Jaiswal
>Assignee: Deepak Jaiswal
> Attachments: HIVE-17529.1.patch, HIVE-17529.2.patch, 
> HIVE-17529.3.patch, HIVE-17529.4.patch, HIVE-17529.5.patch
>
>
> If while traversing the tree to generate tasks, a bucket mapjoin may set its 
> edge as CUSTOM_SIMPLE_EDGE against CUSTOM_EDGE if the bigtable is already not 
> traversed causing Tez to assert and fail the vertex.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17459) View deletion operation failed to replicate on target cluster

2017-09-20 Thread Tao Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17459?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16173628#comment-16173628
 ] 

Tao Li commented on HIVE-17459:
---

Working on HIVE-17570 to fix the bug.

> View deletion operation failed to replicate on target cluster
> -
>
> Key: HIVE-17459
> URL: https://issues.apache.org/jira/browse/HIVE-17459
> Project: Hive
>  Issue Type: Bug
>  Components: repl
>Reporter: Tao Li
>Assignee: Tao Li
> Fix For: 3.0.0
>
> Attachments: HIVE-17459.1.patch, HIVE-17459.2.patch
>
>
> View dropping is not replicated during incremental repl.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17472) Drop-partition for multi-level partition fails, if data does not exist.

2017-09-20 Thread Mithun Radhakrishnan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17472?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mithun Radhakrishnan updated HIVE-17472:

Status: Open  (was: Patch Available)

> Drop-partition for multi-level partition fails, if data does not exist.
> ---
>
> Key: HIVE-17472
> URL: https://issues.apache.org/jira/browse/HIVE-17472
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Affects Versions: 2.2.0, 3.0.0
>Reporter: Mithun Radhakrishnan
>Assignee: Chris Drome
> Attachments: HIVE-17472.1.patch, HIVE-17472.2-branch-2.patch, 
> HIVE-17472.2.patch, HIVE-17472.3-branch-2.2.patch, 
> HIVE-17472.3-branch-2.patch, HIVE-17472.3.patch, 
> HIVE-17472.4-branch-2.2.patch, HIVE-17472.4.patch
>
>
> Raising this on behalf of [~cdrome] and [~selinazh]. 
> Here's how to reproduce the problem:
> {code:sql}
> CREATE TABLE foobar ( foo STRING, bar STRING ) PARTITIONED BY ( dt STRING, 
> region STRING ) STORED AS RCFILE LOCATION '/tmp/foobar';
> ALTER TABLE foobar ADD PARTITION ( dt='1', region='A' ) ;
> dfs -rm -R -skipTrash /tmp/foobar/dt=1;
> ALTER TABLE foobar DROP PARTITION ( dt='1' );
> {code}
> This causes a client-side error as follows:
> {code}
> 15/02/26 23:08:32 ERROR exec.DDLTask: 
> org.apache.hadoop.hive.ql.metadata.HiveException: Unknown error. Please check 
> logs.
> {code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17472) Drop-partition for multi-level partition fails, if data does not exist.

2017-09-20 Thread Mithun Radhakrishnan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17472?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mithun Radhakrishnan updated HIVE-17472:

Attachment: HIVE-17472.4-branch-2.2.patch

> Drop-partition for multi-level partition fails, if data does not exist.
> ---
>
> Key: HIVE-17472
> URL: https://issues.apache.org/jira/browse/HIVE-17472
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Affects Versions: 2.2.0, 3.0.0
>Reporter: Mithun Radhakrishnan
>Assignee: Chris Drome
> Attachments: HIVE-17472.1.patch, HIVE-17472.2-branch-2.patch, 
> HIVE-17472.2.patch, HIVE-17472.3-branch-2.2.patch, 
> HIVE-17472.3-branch-2.patch, HIVE-17472.3.patch, 
> HIVE-17472.4-branch-2.2.patch, HIVE-17472.4.patch
>
>
> Raising this on behalf of [~cdrome] and [~selinazh]. 
> Here's how to reproduce the problem:
> {code:sql}
> CREATE TABLE foobar ( foo STRING, bar STRING ) PARTITIONED BY ( dt STRING, 
> region STRING ) STORED AS RCFILE LOCATION '/tmp/foobar';
> ALTER TABLE foobar ADD PARTITION ( dt='1', region='A' ) ;
> dfs -rm -R -skipTrash /tmp/foobar/dt=1;
> ALTER TABLE foobar DROP PARTITION ( dt='1' );
> {code}
> This causes a client-side error as follows:
> {code}
> 15/02/26 23:08:32 ERROR exec.DDLTask: 
> org.apache.hadoop.hive.ql.metadata.HiveException: Unknown error. Please check 
> logs.
> {code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17570) Fix view deletion related test failures (create_view.q etc)

2017-09-20 Thread Tao Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17570?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tao Li updated HIVE-17570:
--
Attachment: HIVE-17570.1.patch

> Fix view deletion related test failures (create_view.q etc)
> ---
>
> Key: HIVE-17570
> URL: https://issues.apache.org/jira/browse/HIVE-17570
> Project: Hive
>  Issue Type: Bug
>Reporter: Tao Li
>Assignee: Tao Li
> Attachments: HIVE-17570.1.patch
>
>
> Fixing the bug introduced by HIVE-17459. Sorry that did not capture that in a 
> timely fashion.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


  1   2   >