[jira] [Commented] (HIVE-17277) HiveMetastoreClient Log name is wrong

2017-08-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17277?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16131761#comment-16131761
 ] 

Hive QA commented on HIVE-17277:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12882487/HIVE-17277.patch

{color:red}ERROR:{color} -1 due to build exiting with an error

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6450/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6450/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6450/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hiveptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ date '+%Y-%m-%d %T.%3N'
2017-08-18 05:51:50.190
+ [[ -n /usr/lib/jvm/java-8-openjdk-amd64 ]]
+ export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ export 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'MAVEN_OPTS=-Xmx1g '
+ MAVEN_OPTS='-Xmx1g '
+ cd /data/hiveptest/working/
+ tee /data/hiveptest/logs/PreCommit-HIVE-Build-6450/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ git = \s\v\n ]]
+ [[ git = \g\i\t ]]
+ [[ -z master ]]
+ [[ -d apache-github-source-source ]]
+ [[ ! -d apache-github-source-source/.git ]]
+ [[ ! -d apache-github-source-source ]]
+ date '+%Y-%m-%d %T.%3N'
2017-08-18 05:51:50.193
+ cd apache-github-source-source
+ git fetch origin
>From https://github.com/apache/hive
   48be6ae..c9e0940  master -> origin/master
+ git reset --hard HEAD
HEAD is now at 48be6ae HIVE-17347: 
TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning_mapjoin_only] is 
failing every time (Peter Vary reviewed by Xuefu Zhang)
+ git clean -f -d
Removing ql/src/test/org/apache/hadoop/hive/ql/TestTxnCommandsBase.java
Removing ql/src/test/org/apache/hadoop/hive/ql/TestTxnNoBuckets.java
Removing ql/src/test/queries/clientpositive/acid_no_buckets.q
Removing ql/src/test/results/clientpositive/llap/acid_no_buckets.q.out
+ git checkout master
Already on 'master'
Your branch is behind 'origin/master' by 1 commit, and can be fast-forwarded.
  (use "git pull" to update your local branch)
+ git reset --hard origin/master
HEAD is now at c9e0940 HIVE-17194: JDBC: Implement Gzip compression for HTTP 
mode (Gopal V, reviewed by Thejas Nair)
+ git merge --ff-only origin/master
Already up-to-date.
+ date '+%Y-%m-%d %T.%3N'
2017-08-18 05:51:53.202
+ patchCommandPath=/data/hiveptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hiveptest/working/scratch/build.patch
+ [[ -f /data/hiveptest/working/scratch/build.patch ]]
+ chmod +x /data/hiveptest/working/scratch/smart-apply-patch.sh
+ /data/hiveptest/working/scratch/smart-apply-patch.sh 
/data/hiveptest/working/scratch/build.patch
Going to apply patch with: patch -p0
patching file 
metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java
+ [[ maven == \m\a\v\e\n ]]
+ rm -rf /data/hiveptest/working/maven/org/apache/hive
+ mvn -B clean install -DskipTests -T 4 -q 
-Dmaven.repo.local=/data/hiveptest/working/maven
DataNucleus Enhancer (version 4.1.17) for API "JDO"
DataNucleus Enhancer : Classpath
>>  /usr/share/maven/boot/plexus-classworlds-2.x.jar
ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MDatabase
ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MFieldSchema
ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MType
ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MTable
ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MConstraint
ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MSerDeInfo
ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MOrder
ENHANCED (Persistable) : 
org.apache.hadoop.hive.metastore.model.MColumnDescriptor
ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MStringList
ENHANCED (Persistable) : 
org.apache.hadoop.hive.metastore.model.MStorageDescriptor
ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MPartition
ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MIndex
ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MRole
ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MRoleMap
ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MGlobalPrivilege
ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MDBPrivilege
ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MTablePrivilege
ENHANCED 

[jira] [Commented] (HIVE-17205) add functional support

2017-08-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17205?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16131757#comment-16131757
 ] 

Hive QA commented on HIVE-17205:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12882486/HIVE-17205.10.patch

{color:green}SUCCESS:{color} +1 due to 6 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 17 failed/errored test(s), 10982 tests 
executed
*Failed tests:*
{noformat}
TestTxnCommandsBase - did not produce a TEST-*.xml file (likely timed out) 
(batchId=281)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_vectorized_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[create_not_acid] 
(batchId=90)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query14] 
(batchId=235)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query23] 
(batchId=235)
org.apache.hadoop.hive.metastore.TestEmbeddedHiveMetaStore.testTransactionalValidation
 (batchId=206)
org.apache.hadoop.hive.metastore.TestRemoteHiveMetaStore.testTransactionalValidation
 (batchId=208)
org.apache.hadoop.hive.metastore.TestSetUGIOnBothClientServer.testTransactionalValidation
 (batchId=205)
org.apache.hadoop.hive.metastore.TestSetUGIOnOnlyClient.testTransactionalValidation
 (batchId=203)
org.apache.hadoop.hive.metastore.TestSetUGIOnOnlyServer.testTransactionalValidation
 (batchId=213)
org.apache.hadoop.hive.ql.TestTxnCommands.testNonAcidToAcidConversion01 
(batchId=284)
org.apache.hadoop.hive.ql.TestTxnCommands.testTimeOutReaper (batchId=284)
org.apache.hadoop.hive.ql.TestTxnNoBuckets.testNoBuckets (batchId=271)
org.apache.hadoop.hive.ql.TestTxnNoBuckets.testNoBucketsTableCreate 
(batchId=271)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionSpecRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation 
(batchId=180)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6449/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6449/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6449/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 17 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12882486 - PreCommit-HIVE-Build

> add functional support
> --
>
> Key: HIVE-17205
> URL: https://issues.apache.org/jira/browse/HIVE-17205
> Project: Hive
>  Issue Type: Sub-task
>  Components: Transactions
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
> Attachments: HIVE-17205.01.patch, HIVE-17205.02.patch, 
> HIVE-17205.03.patch, HIVE-17205.09.patch, HIVE-17205.10.patch
>
>
> make sure unbucketed tables can be marked transactional=true
> make insert/update/delete/compaction work



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17194) JDBC: Implement Gzip compression for HTTP mode

2017-08-17 Thread Gopal V (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17194?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gopal V updated HIVE-17194:
---
Summary: JDBC: Implement Gzip compression for HTTP mode  (was: JDBC: 
Implement Gzip servlet filter)

> JDBC: Implement Gzip compression for HTTP mode
> --
>
> Key: HIVE-17194
> URL: https://issues.apache.org/jira/browse/HIVE-17194
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2, JDBC
>Affects Versions: 3.0.0
>Reporter: Gopal V
>Assignee: Gopal V
> Attachments: HIVE-17194.1.patch, HIVE-17194.2.patch, 
> HIVE-17194.3.patch
>
>
> {code}
> POST /cliservice HTTP/1.1
> Content-Type: application/x-thrift
> Accept: application/x-thrift
> User-Agent: Java/THttpClient/HC
> Authorization: Basic YW5vbnltb3VzOmFub255bW91cw==
> Content-Length: 71
> Host: localhost:10007
> Connection: Keep-Alive
> Accept-Encoding: gzip,deflate
> X-XSRF-HEADER: true
> {code}
> The Beeline client clearly sends out HTTP compression headers which are 
> ignored by the HTTP service layer in HS2.
> After patch, result looks like
> {code}
> HTTP/1.1 200 OK
> Date: Tue, 01 Aug 2017 01:47:23 GMT
> Content-Type: application/x-thrift
> Vary: Accept-Encoding, User-Agent
> Content-Encoding: gzip
> Transfer-Encoding: chunked
> Server: Jetty(9.3.8.v20160314)
> {code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-13989) Extended ACLs are not handled according to specification

2017-08-17 Thread Chris Drome (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13989?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Drome updated HIVE-13989:
---
Attachment: HIVE-13989.4-branch-2.2.patch

Uploaded new version of branch-2.2 patch.

> Extended ACLs are not handled according to specification
> 
>
> Key: HIVE-13989
> URL: https://issues.apache.org/jira/browse/HIVE-13989
> Project: Hive
>  Issue Type: Bug
>  Components: HCatalog
>Affects Versions: 1.2.1, 2.0.0
>Reporter: Chris Drome
>Assignee: Chris Drome
> Attachments: HIVE-13989.1-branch-1.patch, HIVE-13989.1.patch, 
> HIVE-13989.4-branch-2.2.patch, HIVE-13989-branch-1.patch, 
> HIVE-13989-branch-2.2.patch, HIVE-13989-branch-2.2.patch, 
> HIVE-13989-branch-2.2.patch
>
>
> Hive takes two approaches to working with extended ACLs depending on whether 
> data is being produced via a Hive query or HCatalog APIs. A Hive query will 
> run an FsShell command to recursively set the extended ACLs for a directory 
> sub-tree. HCatalog APIs will attempt to build up the directory sub-tree 
> programmatically and runs some code to set the ACLs to match the parent 
> directory.
> Some incorrect assumptions were made when implementing the extended ACLs 
> support. Refer to https://issues.apache.org/jira/browse/HDFS-4685 for the 
> design documents of extended ACLs in HDFS. These documents model the 
> implementation after the POSIX implementation on Linux, which can be found at 
> http://www.vanemery.com/Linux/ACL/POSIX_ACL_on_Linux.html.
> The code for setting extended ACLs via HCatalog APIs is found in 
> HdfsUtils.java:
> {code}
> if (aclEnabled) {
>   aclStatus =  sourceStatus.getAclStatus();
>   if (aclStatus != null) {
> LOG.trace(aclStatus.toString());
> aclEntries = aclStatus.getEntries();
> removeBaseAclEntries(aclEntries);
> //the ACL api's also expect the tradition user/group/other permission 
> in the form of ACL
> aclEntries.add(newAclEntry(AclEntryScope.ACCESS, AclEntryType.USER, 
> sourcePerm.getUserAction()));
> aclEntries.add(newAclEntry(AclEntryScope.ACCESS, AclEntryType.GROUP, 
> sourcePerm.getGroupAction()));
> aclEntries.add(newAclEntry(AclEntryScope.ACCESS, AclEntryType.OTHER, 
> sourcePerm.getOtherAction()));
>   }
> }
> {code}
> We found that DEFAULT extended ACL rules were not being inherited properly by 
> the directory sub-tree, so the above code is incomplete because it 
> effectively drops the DEFAULT rules. The second problem is with the call to 
> {{sourcePerm.getGroupAction()}}, which is incorrect in the case of extended 
> ACLs. When extended ACLs are used the GROUP permission is replaced with the 
> extended ACL mask. So the above code will apply the wrong permissions to the 
> GROUP. Instead the correct GROUP permissions now need to be pulled from the 
> AclEntry as returned by {{getAclStatus().getEntries()}}. See the 
> implementation of the new method {{getDefaultAclEntries}} for details.
> Similar issues exist with the HCatalog API. None of the API accounts for 
> setting extended ACLs on the directory sub-tree. The changes to the HCatalog 
> API allow the extended ACLs to be passed into the required methods similar to 
> how basic permissions are passed in. When building the directory sub-tree the 
> extended ACLs of the table directory are inherited by all sub-directories, 
> including the DEFAULT rules.
> Replicating the problem:
> Create a table to write data into (I will use acl_test as the destination and 
> words_text as the source) and set the ACLs as follows:
> {noformat}
> $ hdfs dfs -setfacl -m 
> default:user::rwx,default:group::r-x,default:mask::rwx,default:user:hdfs:rwx,group::r-x,user:hdfs:rwx
>  /user/cdrome/hive/acl_test
> $ hdfs dfs -ls -d /user/cdrome/hive/acl_test
> drwxrwx---+  - cdrome hdfs  0 2016-07-13 20:36 
> /user/cdrome/hive/acl_test
> $ hdfs dfs -getfacl -R /user/cdrome/hive/acl_test
> # file: /user/cdrome/hive/acl_test
> # owner: cdrome
> # group: hdfs
> user::rwx
> user:hdfs:rwx
> group::r-x
> mask::rwx
> other::---
> default:user::rwx
> default:user:hdfs:rwx
> default:group::r-x
> default:mask::rwx
> default:other::---
> {noformat}
> Note that the basic GROUP permission is set to {{rwx}} after setting the 
> ACLs. The ACLs explicitly set the DEFAULT rules and a rule specifically for 
> the {{hdfs}} user.
> Run the following query to populate the table:
> {noformat}
> insert into acl_test partition (dt='a', ds='b') select a, b from words_text 
> where dt = 'c';
> {noformat}
> Note that words_text only has a single partition key.
> Now examine the ACLs for the resulting directories:
> {noformat}
> $ hdfs dfs -getfacl -R /user/cdrome/hive/acl_test
> # file: /user/cdrome/hive/acl_test
> # owner: cdrome
> # group: hdfs
> 

[jira] [Commented] (HIVE-17354) Fix "alter view" for incremental replication

2017-08-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17354?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16131712#comment-16131712
 ] 

Hive QA commented on HIVE-17354:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12882482/HIVE-17354.1.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 6 failed/errored test(s), 10976 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[sample_islocalmode_hook] 
(batchId=12)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_vectorized_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query23] 
(batchId=235)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionSpecRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation 
(batchId=180)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6448/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6448/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6448/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 6 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12882482 - PreCommit-HIVE-Build

> Fix "alter view" for incremental replication
> 
>
> Key: HIVE-17354
> URL: https://issues.apache.org/jira/browse/HIVE-17354
> Project: Hive
>  Issue Type: Bug
>  Components: repl
>Reporter: Tao Li
>Assignee: Tao Li
> Attachments: HIVE-17354.1.patch
>
>
> There is a bug that "alter view" operation is resulting in a view creation 
> operation instead of a overwriting/replacement operation.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-16948) Invalid explain when running dynamic partition pruning query in Hive On Spark

2017-08-17 Thread liyunzhang_intel (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-16948?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

liyunzhang_intel updated HIVE-16948:

Attachment: HIVE-16948.7.patch

> Invalid explain when running dynamic partition pruning query in Hive On Spark
> -
>
> Key: HIVE-16948
> URL: https://issues.apache.org/jira/browse/HIVE-16948
> Project: Hive
>  Issue Type: Bug
>Reporter: liyunzhang_intel
>Assignee: liyunzhang_intel
> Attachments: HIVE-16948_1.patch, HIVE-16948.2.patch, 
> HIVE-16948.5.patch, HIVE-16948.6.patch, HIVE-16948.7.patch, HIVE-16948.patch
>
>
> in 
> [union_subquery.q|https://github.com/apache/hive/blob/master/ql/src/test/queries/clientpositive/spark_dynamic_partition_pruning.q#L107]
>  in spark_dynamic_partition_pruning.q
> {code}
> set hive.optimize.ppd=true;
> set hive.ppd.remove.duplicatefilters=true;
> set hive.spark.dynamic.partition.pruning=true;
> set hive.optimize.metadataonly=false;
> set hive.optimize.index.filter=true;
> set hive.strict.checks.cartesian.product=false;
> explain select ds from (select distinct(ds) as ds from srcpart union all 
> select distinct(ds) as ds from srcpart) s where s.ds in (select 
> max(srcpart.ds) from srcpart union all select min(srcpart.ds) from srcpart);
> {code}
> explain 
> {code}
> STAGE DEPENDENCIES:
>   Stage-2 is a root stage
>   Stage-1 depends on stages: Stage-2
>   Stage-0 depends on stages: Stage-1
> STAGE PLANS:
>   Stage: Stage-2
> Spark
>   Edges:
> Reducer 11 <- Map 10 (GROUP, 1)
> Reducer 13 <- Map 12 (GROUP, 1)
>   DagName: root_20170622231525_20a777e5-e659-4138-b605-65f8395e18e2:2
>   Vertices:
> Map 10 
> Map Operator Tree:
> TableScan
>   alias: srcpart
>   Statistics: Num rows: 1 Data size: 23248 Basic stats: 
> PARTIAL Column stats: NONE
>   Select Operator
> expressions: ds (type: string)
> outputColumnNames: ds
> Statistics: Num rows: 1 Data size: 23248 Basic stats: 
> PARTIAL Column stats: NONE
> Group By Operator
>   aggregations: max(ds)
>   mode: hash
>   outputColumnNames: _col0
>   Statistics: Num rows: 1 Data size: 184 Basic stats: 
> COMPLETE Column stats: NONE
>   Reduce Output Operator
> sort order: 
> Statistics: Num rows: 1 Data size: 184 Basic stats: 
> COMPLETE Column stats: NONE
> value expressions: _col0 (type: string)
> Map 12 
> Map Operator Tree:
> TableScan
>   alias: srcpart
>   Statistics: Num rows: 1 Data size: 23248 Basic stats: 
> PARTIAL Column stats: NONE
>   Select Operator
> expressions: ds (type: string)
> outputColumnNames: ds
> Statistics: Num rows: 1 Data size: 23248 Basic stats: 
> PARTIAL Column stats: NONE
> Group By Operator
>   aggregations: min(ds)
>   mode: hash
>   outputColumnNames: _col0
>   Statistics: Num rows: 1 Data size: 184 Basic stats: 
> COMPLETE Column stats: NONE
>   Reduce Output Operator
> sort order: 
> Statistics: Num rows: 1 Data size: 184 Basic stats: 
> COMPLETE Column stats: NONE
> value expressions: _col0 (type: string)
> Reducer 11 
> Reduce Operator Tree:
>   Group By Operator
> aggregations: max(VALUE._col0)
> mode: mergepartial
> outputColumnNames: _col0
> Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE 
> Column stats: NONE
> Filter Operator
>   predicate: _col0 is not null (type: boolean)
>   Statistics: Num rows: 1 Data size: 184 Basic stats: 
> COMPLETE Column stats: NONE
>   Group By Operator
> keys: _col0 (type: string)
> mode: hash
> outputColumnNames: _col0
> Statistics: Num rows: 2 Data size: 368 Basic stats: 
> COMPLETE Column stats: NONE
> Select Operator
>   expressions: _col0 (type: string)
>   outputColumnNames: _col0
>   Statistics: Num rows: 2 Data size: 368 Basic stats: 
> COMPLETE Column stats: NONE
>   Group By Operator
> keys: _col0 (type: string)
> 

[jira] [Commented] (HIVE-16948) Invalid explain when running dynamic partition pruning query in Hive On Spark

2017-08-17 Thread liyunzhang_intel (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-16948?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16131675#comment-16131675
 ] 

liyunzhang_intel commented on HIVE-16948:
-

[~pvary]: I saw your comment on review board.  Thanks for your tracking 
HIVE-17346.

> Invalid explain when running dynamic partition pruning query in Hive On Spark
> -
>
> Key: HIVE-16948
> URL: https://issues.apache.org/jira/browse/HIVE-16948
> Project: Hive
>  Issue Type: Bug
>Reporter: liyunzhang_intel
>Assignee: liyunzhang_intel
> Attachments: HIVE-16948_1.patch, HIVE-16948.2.patch, 
> HIVE-16948.5.patch, HIVE-16948.6.patch, HIVE-16948.7.patch, HIVE-16948.patch
>
>
> in 
> [union_subquery.q|https://github.com/apache/hive/blob/master/ql/src/test/queries/clientpositive/spark_dynamic_partition_pruning.q#L107]
>  in spark_dynamic_partition_pruning.q
> {code}
> set hive.optimize.ppd=true;
> set hive.ppd.remove.duplicatefilters=true;
> set hive.spark.dynamic.partition.pruning=true;
> set hive.optimize.metadataonly=false;
> set hive.optimize.index.filter=true;
> set hive.strict.checks.cartesian.product=false;
> explain select ds from (select distinct(ds) as ds from srcpart union all 
> select distinct(ds) as ds from srcpart) s where s.ds in (select 
> max(srcpart.ds) from srcpart union all select min(srcpart.ds) from srcpart);
> {code}
> explain 
> {code}
> STAGE DEPENDENCIES:
>   Stage-2 is a root stage
>   Stage-1 depends on stages: Stage-2
>   Stage-0 depends on stages: Stage-1
> STAGE PLANS:
>   Stage: Stage-2
> Spark
>   Edges:
> Reducer 11 <- Map 10 (GROUP, 1)
> Reducer 13 <- Map 12 (GROUP, 1)
>   DagName: root_20170622231525_20a777e5-e659-4138-b605-65f8395e18e2:2
>   Vertices:
> Map 10 
> Map Operator Tree:
> TableScan
>   alias: srcpart
>   Statistics: Num rows: 1 Data size: 23248 Basic stats: 
> PARTIAL Column stats: NONE
>   Select Operator
> expressions: ds (type: string)
> outputColumnNames: ds
> Statistics: Num rows: 1 Data size: 23248 Basic stats: 
> PARTIAL Column stats: NONE
> Group By Operator
>   aggregations: max(ds)
>   mode: hash
>   outputColumnNames: _col0
>   Statistics: Num rows: 1 Data size: 184 Basic stats: 
> COMPLETE Column stats: NONE
>   Reduce Output Operator
> sort order: 
> Statistics: Num rows: 1 Data size: 184 Basic stats: 
> COMPLETE Column stats: NONE
> value expressions: _col0 (type: string)
> Map 12 
> Map Operator Tree:
> TableScan
>   alias: srcpart
>   Statistics: Num rows: 1 Data size: 23248 Basic stats: 
> PARTIAL Column stats: NONE
>   Select Operator
> expressions: ds (type: string)
> outputColumnNames: ds
> Statistics: Num rows: 1 Data size: 23248 Basic stats: 
> PARTIAL Column stats: NONE
> Group By Operator
>   aggregations: min(ds)
>   mode: hash
>   outputColumnNames: _col0
>   Statistics: Num rows: 1 Data size: 184 Basic stats: 
> COMPLETE Column stats: NONE
>   Reduce Output Operator
> sort order: 
> Statistics: Num rows: 1 Data size: 184 Basic stats: 
> COMPLETE Column stats: NONE
> value expressions: _col0 (type: string)
> Reducer 11 
> Reduce Operator Tree:
>   Group By Operator
> aggregations: max(VALUE._col0)
> mode: mergepartial
> outputColumnNames: _col0
> Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE 
> Column stats: NONE
> Filter Operator
>   predicate: _col0 is not null (type: boolean)
>   Statistics: Num rows: 1 Data size: 184 Basic stats: 
> COMPLETE Column stats: NONE
>   Group By Operator
> keys: _col0 (type: string)
> mode: hash
> outputColumnNames: _col0
> Statistics: Num rows: 2 Data size: 368 Basic stats: 
> COMPLETE Column stats: NONE
> Select Operator
>   expressions: _col0 (type: string)
>   outputColumnNames: _col0
>   Statistics: Num rows: 2 Data size: 368 Basic stats: 
> COMPLETE Column stats: NONE
> 

[jira] [Commented] (HIVE-17265) Cache merged column stats from retrieved partitions

2017-08-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17265?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16131668#comment-16131668
 ] 

Hive QA commented on HIVE-17265:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12882477/HIVE-17265.03.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 23 failed/errored test(s), 10976 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[dynamic_partition_pruning_2]
 (batchId=143)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[dynamic_semijoin_user_level]
 (batchId=144)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[auto_sortmerge_join_12]
 (batchId=152)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[bucket_map_join_tez1]
 (batchId=162)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[bucket_map_join_tez2]
 (batchId=148)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[dynamic_semijoin_reduction]
 (batchId=155)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[filter_join_breaktask2]
 (batchId=160)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[filter_join_breaktask]
 (batchId=160)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[mergejoin] 
(batchId=158)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[partition_shared_scan]
 (batchId=148)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[semijoin_hint]
 (batchId=151)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[tez_smb_empty]
 (batchId=153)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[tez_smb_main]
 (batchId=151)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[vector_mr_diff_schema_alias]
 (batchId=158)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning_mapjoin_only]
 (batchId=170)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_vectorized_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver[explainanalyze_2] 
(batchId=100)
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver[tez-tag] 
(batchId=99)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query14] 
(batchId=235)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionSpecRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation 
(batchId=180)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6447/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6447/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6447/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 23 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12882477 - PreCommit-HIVE-Build

> Cache merged column stats from retrieved partitions
> ---
>
> Key: HIVE-17265
> URL: https://issues.apache.org/jira/browse/HIVE-17265
> Project: Hive
>  Issue Type: Bug
>  Components: Statistics
>Affects Versions: 3.0.0
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
> Attachments: HIVE-17265.02.patch, HIVE-17265.03.patch, 
> HIVE-17265.patch
>
>
> Currently when we retrieve stats from the metastore for a column in a 
> partitioned table, we will execute the logic to merge the column stats coming 
> from each partition multiple times.
> Even though we avoid multiple calls to metastore if the cache for the stats 
> in enabled, merging the stats for a given column can take a large amount of 
> time if there is a large number of partitions.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17277) HiveMetastoreClient Log name is wrong

2017-08-17 Thread Zac Zhou (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17277?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zac Zhou updated HIVE-17277:

Attachment: (was: HIVE-17277.patch)

> HiveMetastoreClient Log name is wrong
> -
>
> Key: HIVE-17277
> URL: https://issues.apache.org/jira/browse/HIVE-17277
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Reporter: Zac Zhou
>Assignee: Zac Zhou
>Priority: Minor
>
> The name of Log for HiveMetastoreClient is "hive.metastore". It's confused 
> for users to trace hive log



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17347) TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning_mapjoin_only] is failing every time

2017-08-17 Thread Rui Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17347?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Li updated HIVE-17347:
--
   Resolution: Fixed
Fix Version/s: 3.0.0
   Status: Resolved  (was: Patch Available)

Pushed to master. Thanks Peter and Xuefu!

> TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning_mapjoin_only] is 
> failing every time
> 
>
> Key: HIVE-17347
> URL: https://issues.apache.org/jira/browse/HIVE-17347
> Project: Hive
>  Issue Type: Sub-task
>  Components: Test
>Affects Versions: 3.0.0
>Reporter: Peter Vary
>Assignee: Peter Vary
> Fix For: 3.0.0
>
> Attachments: HIVE-17347.patch
>
>
> As [~lirui] identified there was a missing file from this patch: HIVE-17247 - 
> HoS DPP: UDFs on the partition column side does not evaluate correctly



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17346) TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning] is failing every time

2017-08-17 Thread Rui Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17346?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Li updated HIVE-17346:
--
   Resolution: Fixed
Fix Version/s: 3.0.0
   Status: Resolved  (was: Patch Available)

Pushed to master. Thanks Peter and Vlad.

> TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning] is failing 
> every time
> ---
>
> Key: HIVE-17346
> URL: https://issues.apache.org/jira/browse/HIVE-17346
> Project: Hive
>  Issue Type: Sub-task
>  Components: Test
>Affects Versions: 3.0.0
>Reporter: Peter Vary
>Assignee: Peter Vary
> Fix For: 3.0.0
>
> Attachments: HIVE-17346.patch
>
>
> The TestMiniSparkOnYarnCliDriver.testCliDriver - 
> spark_dynamic_partition_pruning is failing with this diff:
> {code}
> Client Execution succeeded but contained differences (error code = 1) after 
> executing spark_dynamic_partition_pruning.q 
> 714c714
> <   filterExpr: ((date = '2008-04-08') and abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null) (type: 
> boolean)
> ---
> >   filterExpr: ((date = '2008-04-08') and ds is not null) 
> > (type: boolean)
> 717c717
> < predicate: ((date = '2008-04-08') and abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null) (type: 
> boolean)
> ---
> > predicate: ((date = '2008-04-08') and ds is not null) 
> > (type: boolean)
> 749c749
> <   filterExpr: abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null (type: 
> boolean)
> ---
> >   filterExpr: ds is not null (type: boolean)
> 751,752c751,753
> <   Filter Operator
> < predicate: abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null (type: 
> boolean)
> ---
> >   Select Operator
> > expressions: ds (type: string)
> > outputColumnNames: _col0
> 754,756c755,758
> < Select Operator
> <   expressions: ds (type: string)
> <   outputColumnNames: _col0
> ---
> > Reduce Output Operator
> >   key expressions: abs(((- 
> > UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> >   sort order: +
> >   Map-reduce partition columns: abs(((- 
> > UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> 758,762d759
> <   Reduce Output Operator
> < key expressions: abs(((- 
> UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> < sort order: +
> < Map-reduce partition columns: abs(((- 
> UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> < Statistics: Num rows: 2000 Data size: 21248 Basic 
> stats: COMPLETE Column stats: NONE
> 767c764
> <  
> Output was too long and had to be truncated...
> {code}
> I think it is caused by:
> HIVE-17148 - Incorrect result for Hive join query with COALESCE in WHERE 
> condition
> [~allgoodok]: Am I right? Is it an intended change and only the golden file 
> regeneration is needed?
> Thanks,
> Peter



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17353) The ResultSets are not accessible if running multiple queries within the same HiveStatement

2017-08-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17353?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16131625#comment-16131625
 ] 

Hive QA commented on HIVE-17353:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12882472/HIVE-17353.1.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 9 failed/errored test(s), 10977 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[columnstats_part_coltype]
 (batchId=159)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning_mapjoin_only]
 (batchId=170)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_vectorized_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query14] 
(batchId=235)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query23] 
(batchId=235)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionSpecRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation 
(batchId=180)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6446/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6446/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6446/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 9 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12882472 - PreCommit-HIVE-Build

> The ResultSets are not accessible if running multiple queries within the same 
> HiveStatement 
> 
>
> Key: HIVE-17353
> URL: https://issues.apache.org/jira/browse/HIVE-17353
> Project: Hive
>  Issue Type: Bug
>  Components: JDBC
>Affects Versions: 3.0.0
>Reporter: Aihua Xu
>Assignee: Aihua Xu
> Attachments: HIVE-17353.1.patch
>
>
> The following queries would fail,
> {noformat}
> ResultSet rs1 =
> stmt.executeQuery("select * from testMultipleResultSets1");
> ResultSet rs2 =
> stmt.executeQuery("select * from testMultipleResultSets2");
> rs1.next();
> rs2.next();
> {noformat}
> with the exception:
> {noformat}
> [HiveServer2-Handler-Pool: Thread-208]: Error fetching results: 
> org.apache.hive.service.cli.HiveSQLException: Invalid OperationHandle: 
> OperationHandle [opType=EXECUTE_STATEMENT, 
> getHandleIdentifier()=8a1c4fe5-e80b-4d9a-b673-78d92b3baaa8]
>   at 
> org.apache.hive.service.cli.operation.OperationManager.getOperation(OperationManager.java:177)
>   at 
> org.apache.hive.service.cli.CLIService.fetchResults(CLIService.java:462)
>   at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.FetchResults(ThriftCLIService.java:691)
>   at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$FetchResults.getResult(TCLIService.java:1553)
>   at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$FetchResults.getResult(TCLIService.java:1538)
>   at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>   at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>   at 
> org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:56)
>   at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:286)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>   at java.lang.Thread.run(Thread.java:748)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Comment Edited] (HIVE-17292) Change TestMiniSparkOnYarnCliDriver test configuration to use the configured cores

2017-08-17 Thread Rui Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17292?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16131618#comment-16131618
 ] 

Rui Li edited comment on HIVE-17292 at 8/18/17 2:22 AM:


Patch LGTM. Let's wait a little bit to get HIVE-17346 and HIVE-17347 in first.


was (Author: lirui):
Path LGTM. Let's wait a little bit to get HIVE-17346 and HIVE-17347 in first.

> Change TestMiniSparkOnYarnCliDriver test configuration to use the configured 
> cores
> --
>
> Key: HIVE-17292
> URL: https://issues.apache.org/jira/browse/HIVE-17292
> Project: Hive
>  Issue Type: Sub-task
>  Components: Spark, Test
>Affects Versions: 3.0.0
>Reporter: Peter Vary
>Assignee: Peter Vary
> Attachments: HIVE-17292.1.patch, HIVE-17292.2.patch, 
> HIVE-17292.3.patch, HIVE-17292.5.patch, HIVE-17292.6.patch
>
>
> Currently the {{hive-site.xml}} for the {{TestMiniSparkOnYarnCliDriver}} test 
> defines 2 cores, and 2 executors, but only 1 is used, because the MiniCluster 
> does not allows the creation of the 3rd container.
> The FairScheduler uses 1GB increments for memory, but the containers would 
> like to use only 512MB. We should change the fairscheduler configuration to 
> use only the requested 512MB



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17292) Change TestMiniSparkOnYarnCliDriver test configuration to use the configured cores

2017-08-17 Thread Rui Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17292?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16131618#comment-16131618
 ] 

Rui Li commented on HIVE-17292:
---

Path LGTM. Let's wait a little bit to get HIVE-17346 and HIVE-17347 in first.

> Change TestMiniSparkOnYarnCliDriver test configuration to use the configured 
> cores
> --
>
> Key: HIVE-17292
> URL: https://issues.apache.org/jira/browse/HIVE-17292
> Project: Hive
>  Issue Type: Sub-task
>  Components: Spark, Test
>Affects Versions: 3.0.0
>Reporter: Peter Vary
>Assignee: Peter Vary
> Attachments: HIVE-17292.1.patch, HIVE-17292.2.patch, 
> HIVE-17292.3.patch, HIVE-17292.5.patch, HIVE-17292.6.patch
>
>
> Currently the {{hive-site.xml}} for the {{TestMiniSparkOnYarnCliDriver}} test 
> defines 2 cores, and 2 executors, but only 1 is used, because the MiniCluster 
> does not allows the creation of the 3rd container.
> The FairScheduler uses 1GB increments for memory, but the containers would 
> like to use only 512MB. We should change the fairscheduler configuration to 
> use only the requested 512MB



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17277) HiveMetastoreClient Log name is wrong

2017-08-17 Thread Zac Zhou (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17277?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zac Zhou updated HIVE-17277:

Attachment: HIVE-17277.patch

Thanks Alan for your comments. Resubmitting the patch to run unit tests

> HiveMetastoreClient Log name is wrong
> -
>
> Key: HIVE-17277
> URL: https://issues.apache.org/jira/browse/HIVE-17277
> Project: Hive
>  Issue Type: Bug
>  Components: Metastore
>Reporter: Zac Zhou
>Assignee: Zac Zhou
>Priority: Minor
> Attachments: HIVE-17277.patch
>
>
> The name of Log for HiveMetastoreClient is "hive.metastore". It's confused 
> for users to trace hive log



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17347) TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning_mapjoin_only] is failing every time

2017-08-17 Thread Rui Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17347?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16131614#comment-16131614
 ] 

Rui Li commented on HIVE-17347:
---

+1

> TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning_mapjoin_only] is 
> failing every time
> 
>
> Key: HIVE-17347
> URL: https://issues.apache.org/jira/browse/HIVE-17347
> Project: Hive
>  Issue Type: Sub-task
>  Components: Test
>Affects Versions: 3.0.0
>Reporter: Peter Vary
>Assignee: Peter Vary
> Attachments: HIVE-17347.patch
>
>
> As [~lirui] identified there was a missing file from this patch: HIVE-17247 - 
> HoS DPP: UDFs on the partition column side does not evaluate correctly



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17346) TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning] is failing every time

2017-08-17 Thread Rui Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17346?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16131613#comment-16131613
 ] 

Rui Li commented on HIVE-17346:
---

Thanks [~pvary] for working on this. +1

> TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning] is failing 
> every time
> ---
>
> Key: HIVE-17346
> URL: https://issues.apache.org/jira/browse/HIVE-17346
> Project: Hive
>  Issue Type: Sub-task
>  Components: Test
>Affects Versions: 3.0.0
>Reporter: Peter Vary
>Assignee: Peter Vary
> Attachments: HIVE-17346.patch
>
>
> The TestMiniSparkOnYarnCliDriver.testCliDriver - 
> spark_dynamic_partition_pruning is failing with this diff:
> {code}
> Client Execution succeeded but contained differences (error code = 1) after 
> executing spark_dynamic_partition_pruning.q 
> 714c714
> <   filterExpr: ((date = '2008-04-08') and abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null) (type: 
> boolean)
> ---
> >   filterExpr: ((date = '2008-04-08') and ds is not null) 
> > (type: boolean)
> 717c717
> < predicate: ((date = '2008-04-08') and abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null) (type: 
> boolean)
> ---
> > predicate: ((date = '2008-04-08') and ds is not null) 
> > (type: boolean)
> 749c749
> <   filterExpr: abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null (type: 
> boolean)
> ---
> >   filterExpr: ds is not null (type: boolean)
> 751,752c751,753
> <   Filter Operator
> < predicate: abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null (type: 
> boolean)
> ---
> >   Select Operator
> > expressions: ds (type: string)
> > outputColumnNames: _col0
> 754,756c755,758
> < Select Operator
> <   expressions: ds (type: string)
> <   outputColumnNames: _col0
> ---
> > Reduce Output Operator
> >   key expressions: abs(((- 
> > UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> >   sort order: +
> >   Map-reduce partition columns: abs(((- 
> > UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> 758,762d759
> <   Reduce Output Operator
> < key expressions: abs(((- 
> UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> < sort order: +
> < Map-reduce partition columns: abs(((- 
> UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> < Statistics: Num rows: 2000 Data size: 21248 Basic 
> stats: COMPLETE Column stats: NONE
> 767c764
> <  
> Output was too long and had to be truncated...
> {code}
> I think it is caused by:
> HIVE-17148 - Incorrect result for Hive join query with COALESCE in WHERE 
> condition
> [~allgoodok]: Am I right? Is it an intended change and only the golden file 
> regeneration is needed?
> Thanks,
> Peter



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17205) add functional support

2017-08-17 Thread Eugene Koifman (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17205?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eugene Koifman updated HIVE-17205:
--
Attachment: HIVE-17205.10.patch

> add functional support
> --
>
> Key: HIVE-17205
> URL: https://issues.apache.org/jira/browse/HIVE-17205
> Project: Hive
>  Issue Type: Sub-task
>  Components: Transactions
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
> Attachments: HIVE-17205.01.patch, HIVE-17205.02.patch, 
> HIVE-17205.03.patch, HIVE-17205.09.patch, HIVE-17205.10.patch
>
>
> make sure unbucketed tables can be marked transactional=true
> make insert/update/delete/compaction work



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17336) Missing class 'org.apache.hadoop.hive.hbase.HiveHBaseTableInputFormat' from Hive on Spark when inserting into hbase based table

2017-08-17 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17336?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16131600#comment-16131600
 ] 

Xuefu Zhang commented on HIVE-17336:


Sorry for getting to this late. Thanks for fixing this, [~aihuaxu]. I think 
making a copy of jobconf is a conscious decision, though I forgot the reason. 
maybe you can find out from the change history.

> Missing class 'org.apache.hadoop.hive.hbase.HiveHBaseTableInputFormat' from 
> Hive on Spark when inserting into hbase based table
> ---
>
> Key: HIVE-17336
> URL: https://issues.apache.org/jira/browse/HIVE-17336
> Project: Hive
>  Issue Type: Bug
>  Components: Spark
>Affects Versions: 3.0.0
>Reporter: Aihua Xu
>Assignee: Aihua Xu
> Fix For: 3.0.0
>
> Attachments: HIVE-17336.1.patch
>
>
> When inserting into a hbase based table from hive on spark, the following 
> exception is thrown 
> {noformat}
> Error while processing statement: FAILED: Execution Error, return code 3 from 
> org.apache.hadoop.hive.ql.exec.spark.SparkTask. 
> org.apache.hive.com.esotericsoftware.kryo.KryoException: Unable to find 
> class: org.apache.hadoop.hive.hbase.HiveHBaseTableInputFormat
> Serialization trace:
> inputFileFormatClass (org.apache.hadoop.hive.ql.plan.TableDesc)
> tableInfo (org.apache.hadoop.hive.ql.plan.FileSinkDesc)
> conf (org.apache.hadoop.hive.ql.exec.FileSinkOperator)
> childOperators (org.apache.hadoop.hive.ql.exec.SelectOperator)
> childOperators (org.apache.hadoop.hive.ql.exec.TableScanOperator)
> aliasToWork (org.apache.hadoop.hive.ql.plan.MapWork)
> invertedWorkGraph (org.apache.hadoop.hive.ql.plan.SparkWork)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.util.DefaultClassResolver.readName(DefaultClassResolver.java:156)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.util.DefaultClassResolver.readClass(DefaultClassResolver.java:133)
>  at org.apache.hive.com.esotericsoftware.kryo.Kryo.readClass(Kryo.java:670)
>  at 
> org.apache.hadoop.hive.ql.exec.SerializationUtilities$KryoWithHooks.readClass(SerializationUtilities.java:183)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.DefaultSerializers$ClassSerializer.read(DefaultSerializers.java:326)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.DefaultSerializers$ClassSerializer.read(DefaultSerializers.java:314)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.Kryo.readObjectOrNull(Kryo.java:759)
>  at 
> org.apache.hadoop.hive.ql.exec.SerializationUtilities$KryoWithHooks.readObjectOrNull(SerializationUtilities.java:201)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.ObjectField.read(ObjectField.java:132)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.FieldSerializer.read(FieldSerializer.java:551)
>  at org.apache.hive.com.esotericsoftware.kryo.Kryo.readObject(Kryo.java:708)
>  at 
> org.apache.hadoop.hive.ql.exec.SerializationUtilities$KryoWithHooks.readObject(SerializationUtilities.java:216)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.ObjectField.read(ObjectField.java:125)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.FieldSerializer.read(FieldSerializer.java:551)
>  at org.apache.hive.com.esotericsoftware.kryo.Kryo.readObject(Kryo.java:708)
>  at 
> org.apache.hadoop.hive.ql.exec.SerializationUtilities$KryoWithHooks.readObject(SerializationUtilities.java:216)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.ObjectField.read(ObjectField.java:125)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.FieldSerializer.read(FieldSerializer.java:551)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:790)
>  at 
> org.apache.hadoop.hive.ql.exec.SerializationUtilities$KryoWithHooks.readClassAndObject(SerializationUtilities.java:178)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.CollectionSerializer.read(CollectionSerializer.java:134)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.CollectionSerializer.read(CollectionSerializer.java:40)
>  at org.apache.hive.com.esotericsoftware.kryo.Kryo.readObject(Kryo.java:708)
>  at 
> org.apache.hadoop.hive.ql.exec.SerializationUtilities$KryoWithHooks.readObject(SerializationUtilities.java:216)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.ObjectField.read(ObjectField.java:125)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.FieldSerializer.read(FieldSerializer.java:551)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:790)
>  at 
> org.apache.hadoop.hive.ql.exec.SerializationUtilities$KryoWithHooks.readClassAndObject(SerializationUtilities.java:178)
>  at 
> 

[jira] [Updated] (HIVE-17354) Fix "alter view" for incremental replication

2017-08-17 Thread Tao Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17354?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tao Li updated HIVE-17354:
--
Status: Patch Available  (was: Open)

> Fix "alter view" for incremental replication
> 
>
> Key: HIVE-17354
> URL: https://issues.apache.org/jira/browse/HIVE-17354
> Project: Hive
>  Issue Type: Bug
>  Components: repl
>Reporter: Tao Li
>Assignee: Tao Li
> Attachments: HIVE-17354.1.patch
>
>
> There is a bug that "alter view" operation is resulting in a view creation 
> operation instead of a overwriting/replacement operation.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17354) Fix "alter view" for incremental replication

2017-08-17 Thread Tao Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17354?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tao Li updated HIVE-17354:
--
Attachment: HIVE-17354.1.patch

> Fix "alter view" for incremental replication
> 
>
> Key: HIVE-17354
> URL: https://issues.apache.org/jira/browse/HIVE-17354
> Project: Hive
>  Issue Type: Bug
>  Components: repl
>Reporter: Tao Li
>Assignee: Tao Li
> Attachments: HIVE-17354.1.patch
>
>
> There is a bug that "alter view" operation is resulting in a view creation 
> operation instead of a overwriting/replacement operation.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17354) Fix "alter view" for incremental replication

2017-08-17 Thread Tao Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17354?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tao Li updated HIVE-17354:
--
Description: There is a bug that "alter view" operation is resulting in a 
view creation operation instead of a overwriting/replacement operation.

> Fix "alter view" for incremental replication
> 
>
> Key: HIVE-17354
> URL: https://issues.apache.org/jira/browse/HIVE-17354
> Project: Hive
>  Issue Type: Bug
>  Components: repl
>Reporter: Tao Li
>Assignee: Tao Li
>
> There is a bug that "alter view" operation is resulting in a view creation 
> operation instead of a overwriting/replacement operation.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17354) Fix "alter view" for incremental replication

2017-08-17 Thread Tao Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17354?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tao Li updated HIVE-17354:
--
Component/s: repl

> Fix "alter view" for incremental replication
> 
>
> Key: HIVE-17354
> URL: https://issues.apache.org/jira/browse/HIVE-17354
> Project: Hive
>  Issue Type: Bug
>  Components: repl
>Reporter: Tao Li
>Assignee: Tao Li
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Assigned] (HIVE-17354) Fix "alter view" for incremental replication

2017-08-17 Thread Tao Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17354?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tao Li reassigned HIVE-17354:
-


> Fix "alter view" for incremental replication
> 
>
> Key: HIVE-17354
> URL: https://issues.apache.org/jira/browse/HIVE-17354
> Project: Hive
>  Issue Type: Bug
>Reporter: Tao Li
>Assignee: Tao Li
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17352) HiveSever2 error with "Illegal Operation state transition from CLOSED to FINISHED"

2017-08-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17352?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16131567#comment-16131567
 ] 

Hive QA commented on HIVE-17352:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12882461/HIVE-17352.1.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 8 failed/errored test(s), 10962 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[unionDistinct_1] 
(batchId=143)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning_mapjoin_only]
 (batchId=170)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_vectorized_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestSparkCliDriver.org.apache.hadoop.hive.cli.TestSparkCliDriver
 (batchId=129)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionSpecRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation 
(batchId=180)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6445/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6445/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6445/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 8 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12882461 - PreCommit-HIVE-Build

> HiveSever2 error with "Illegal Operation state transition from CLOSED to 
> FINISHED"
> --
>
> Key: HIVE-17352
> URL: https://issues.apache.org/jira/browse/HIVE-17352
> Project: Hive
>  Issue Type: Bug
>Reporter: Deepak Jaiswal
>Assignee: Deepak Jaiswal
> Attachments: HIVE-17352.1.patch
>
>
> HiveSever2 error with "Illegal Operation state transition from CLOSED to 
> FINISHED"
> Many cases like CANCELED, TIMEDOUT AND CLOSED are handled. Need to handle 
> FINISHED in runQuery() method.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17265) Cache merged column stats from retrieved partitions

2017-08-17 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17265?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-17265:
---
Attachment: HIVE-17265.03.patch

> Cache merged column stats from retrieved partitions
> ---
>
> Key: HIVE-17265
> URL: https://issues.apache.org/jira/browse/HIVE-17265
> Project: Hive
>  Issue Type: Bug
>  Components: Statistics
>Affects Versions: 3.0.0
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
> Attachments: HIVE-17265.02.patch, HIVE-17265.03.patch, 
> HIVE-17265.patch
>
>
> Currently when we retrieve stats from the metastore for a column in a 
> partitioned table, we will execute the logic to merge the column stats coming 
> from each partition multiple times.
> Even though we avoid multiple calls to metastore if the cache for the stats 
> in enabled, merging the stats for a given column can take a large amount of 
> time if there is a large number of partitions.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17353) The ResultSets are not accessible if running multiple queries within the same HiveStatement

2017-08-17 Thread Aihua Xu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17353?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aihua Xu updated HIVE-17353:

Status: Patch Available  (was: Open)

The executing query is closed when a new query is run in the same statement. 
That seems not necessary since we can run multiple queries in the same session. 

> The ResultSets are not accessible if running multiple queries within the same 
> HiveStatement 
> 
>
> Key: HIVE-17353
> URL: https://issues.apache.org/jira/browse/HIVE-17353
> Project: Hive
>  Issue Type: Bug
>  Components: JDBC
>Affects Versions: 3.0.0
>Reporter: Aihua Xu
>Assignee: Aihua Xu
> Attachments: HIVE-17353.1.patch
>
>
> The following queries would fail,
> {noformat}
> ResultSet rs1 =
> stmt.executeQuery("select * from testMultipleResultSets1");
> ResultSet rs2 =
> stmt.executeQuery("select * from testMultipleResultSets2");
> rs1.next();
> rs2.next();
> {noformat}
> with the exception:
> {noformat}
> [HiveServer2-Handler-Pool: Thread-208]: Error fetching results: 
> org.apache.hive.service.cli.HiveSQLException: Invalid OperationHandle: 
> OperationHandle [opType=EXECUTE_STATEMENT, 
> getHandleIdentifier()=8a1c4fe5-e80b-4d9a-b673-78d92b3baaa8]
>   at 
> org.apache.hive.service.cli.operation.OperationManager.getOperation(OperationManager.java:177)
>   at 
> org.apache.hive.service.cli.CLIService.fetchResults(CLIService.java:462)
>   at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.FetchResults(ThriftCLIService.java:691)
>   at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$FetchResults.getResult(TCLIService.java:1553)
>   at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$FetchResults.getResult(TCLIService.java:1538)
>   at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>   at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>   at 
> org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:56)
>   at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:286)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>   at java.lang.Thread.run(Thread.java:748)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17353) The ResultSets are not accessible if running multiple queries within the same HiveStatement

2017-08-17 Thread Aihua Xu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17353?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aihua Xu updated HIVE-17353:

Attachment: HIVE-17353.1.patch

> The ResultSets are not accessible if running multiple queries within the same 
> HiveStatement 
> 
>
> Key: HIVE-17353
> URL: https://issues.apache.org/jira/browse/HIVE-17353
> Project: Hive
>  Issue Type: Bug
>  Components: JDBC
>Affects Versions: 3.0.0
>Reporter: Aihua Xu
>Assignee: Aihua Xu
> Attachments: HIVE-17353.1.patch
>
>
> The following queries would fail,
> {noformat}
> ResultSet rs1 =
> stmt.executeQuery("select * from testMultipleResultSets1");
> ResultSet rs2 =
> stmt.executeQuery("select * from testMultipleResultSets2");
> rs1.next();
> rs2.next();
> {noformat}
> with the exception:
> {noformat}
> [HiveServer2-Handler-Pool: Thread-208]: Error fetching results: 
> org.apache.hive.service.cli.HiveSQLException: Invalid OperationHandle: 
> OperationHandle [opType=EXECUTE_STATEMENT, 
> getHandleIdentifier()=8a1c4fe5-e80b-4d9a-b673-78d92b3baaa8]
>   at 
> org.apache.hive.service.cli.operation.OperationManager.getOperation(OperationManager.java:177)
>   at 
> org.apache.hive.service.cli.CLIService.fetchResults(CLIService.java:462)
>   at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.FetchResults(ThriftCLIService.java:691)
>   at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$FetchResults.getResult(TCLIService.java:1553)
>   at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$FetchResults.getResult(TCLIService.java:1538)
>   at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>   at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>   at 
> org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:56)
>   at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:286)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>   at java.lang.Thread.run(Thread.java:748)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Assigned] (HIVE-17353) The ResultSets are not accessible if running multiple queries within the same HiveStatement

2017-08-17 Thread Aihua Xu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17353?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aihua Xu reassigned HIVE-17353:
---


> The ResultSets are not accessible if running multiple queries within the same 
> HiveStatement 
> 
>
> Key: HIVE-17353
> URL: https://issues.apache.org/jira/browse/HIVE-17353
> Project: Hive
>  Issue Type: Bug
>  Components: JDBC
>Affects Versions: 3.0.0
>Reporter: Aihua Xu
>Assignee: Aihua Xu
>
> The following queries would fail,
> {noformat}
> ResultSet rs1 =
> stmt.executeQuery("select * from testMultipleResultSets1");
> ResultSet rs2 =
> stmt.executeQuery("select * from testMultipleResultSets2");
> rs1.next();
> rs2.next();
> {noformat}
> with the exception:
> {noformat}
> [HiveServer2-Handler-Pool: Thread-208]: Error fetching results: 
> org.apache.hive.service.cli.HiveSQLException: Invalid OperationHandle: 
> OperationHandle [opType=EXECUTE_STATEMENT, 
> getHandleIdentifier()=8a1c4fe5-e80b-4d9a-b673-78d92b3baaa8]
>   at 
> org.apache.hive.service.cli.operation.OperationManager.getOperation(OperationManager.java:177)
>   at 
> org.apache.hive.service.cli.CLIService.fetchResults(CLIService.java:462)
>   at 
> org.apache.hive.service.cli.thrift.ThriftCLIService.FetchResults(ThriftCLIService.java:691)
>   at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$FetchResults.getResult(TCLIService.java:1553)
>   at 
> org.apache.hive.service.cli.thrift.TCLIService$Processor$FetchResults.getResult(TCLIService.java:1538)
>   at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
>   at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
>   at 
> org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:56)
>   at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:286)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>   at java.lang.Thread.run(Thread.java:748)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17351) use new slider package installation command in run.sh

2017-08-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17351?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16131510#comment-16131510
 ] 

Hive QA commented on HIVE-17351:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12882454/HIVE-17351.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 10 failed/errored test(s), 10976 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[orc_merge6] 
(batchId=152)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning_mapjoin_only]
 (batchId=170)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_vectorized_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver[explainanalyze_2] 
(batchId=100)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query14] 
(batchId=235)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query23] 
(batchId=235)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionSpecRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation 
(batchId=180)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6444/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6444/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6444/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 10 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12882454 - PreCommit-HIVE-Build

> use new slider package installation command in run.sh
> -
>
> Key: HIVE-17351
> URL: https://issues.apache.org/jira/browse/HIVE-17351
> Project: Hive
>  Issue Type: Bug
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-17351.patch
>
>
> The old syntax does not include some perf improvements in newer versions of 
> Slider.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17352) HiveSever2 error with "Illegal Operation state transition from CLOSED to FINISHED"

2017-08-17 Thread Deepak Jaiswal (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17352?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Deepak Jaiswal updated HIVE-17352:
--
Attachment: HIVE-17352.1.patch

Added the case to ignore CLOSED to FINISHED state.

> HiveSever2 error with "Illegal Operation state transition from CLOSED to 
> FINISHED"
> --
>
> Key: HIVE-17352
> URL: https://issues.apache.org/jira/browse/HIVE-17352
> Project: Hive
>  Issue Type: Bug
>Reporter: Deepak Jaiswal
>Assignee: Deepak Jaiswal
> Attachments: HIVE-17352.1.patch
>
>
> HiveSever2 error with "Illegal Operation state transition from CLOSED to 
> FINISHED"
> Many cases like CANCELED, TIMEDOUT AND CLOSED are handled. Need to handle 
> FINISHED in runQuery() method.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17352) HiveSever2 error with "Illegal Operation state transition from CLOSED to FINISHED"

2017-08-17 Thread Deepak Jaiswal (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17352?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Deepak Jaiswal updated HIVE-17352:
--
Status: Patch Available  (was: In Progress)

> HiveSever2 error with "Illegal Operation state transition from CLOSED to 
> FINISHED"
> --
>
> Key: HIVE-17352
> URL: https://issues.apache.org/jira/browse/HIVE-17352
> Project: Hive
>  Issue Type: Bug
>Reporter: Deepak Jaiswal
>Assignee: Deepak Jaiswal
>
> HiveSever2 error with "Illegal Operation state transition from CLOSED to 
> FINISHED"
> Many cases like CANCELED, TIMEDOUT AND CLOSED are handled. Need to handle 
> FINISHED in runQuery() method.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Work started] (HIVE-17352) HiveSever2 error with "Illegal Operation state transition from CLOSED to FINISHED"

2017-08-17 Thread Deepak Jaiswal (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17352?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HIVE-17352 started by Deepak Jaiswal.
-
> HiveSever2 error with "Illegal Operation state transition from CLOSED to 
> FINISHED"
> --
>
> Key: HIVE-17352
> URL: https://issues.apache.org/jira/browse/HIVE-17352
> Project: Hive
>  Issue Type: Bug
>Reporter: Deepak Jaiswal
>Assignee: Deepak Jaiswal
>
> HiveSever2 error with "Illegal Operation state transition from CLOSED to 
> FINISHED"
> Many cases like CANCELED, TIMEDOUT AND CLOSED are handled. Need to handle 
> FINISHED in runQuery() method.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Assigned] (HIVE-17352) HiveSever2 error with "Illegal Operation state transition from CLOSED to FINISHED"

2017-08-17 Thread Deepak Jaiswal (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17352?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Deepak Jaiswal reassigned HIVE-17352:
-


> HiveSever2 error with "Illegal Operation state transition from CLOSED to 
> FINISHED"
> --
>
> Key: HIVE-17352
> URL: https://issues.apache.org/jira/browse/HIVE-17352
> Project: Hive
>  Issue Type: Bug
>Reporter: Deepak Jaiswal
>Assignee: Deepak Jaiswal
>
> HiveSever2 error with "Illegal Operation state transition from CLOSED to 
> FINISHED"
> Many cases like CANCELED, TIMEDOUT AND CLOSED are handled. Need to handle 
> FINISHED in runQuery() method.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17198) Flaky test: TestBeeLineDriver [smb_mapjoin_7]

2017-08-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17198?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16131449#comment-16131449
 ] 

Hive QA commented on HIVE-17198:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12882431/HIVE-17198.2.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 7 failed/errored test(s), 10977 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning_mapjoin_only]
 (batchId=170)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_vectorized_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query14] 
(batchId=235)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionSpecRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation 
(batchId=180)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6443/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6443/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6443/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 7 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12882431 - PreCommit-HIVE-Build

> Flaky test: TestBeeLineDriver [smb_mapjoin_7]
> -
>
> Key: HIVE-17198
> URL: https://issues.apache.org/jira/browse/HIVE-17198
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Janaki Lahorani
>Assignee: Peter Vary
> Attachments: HIVE-17198.2.patch, HIVE-17198.patch
>
>
> Error:
> Exception running or analyzing the results of the query file: 
> org.apache.hive.beeline.QFile@4f7b68ad



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17351) use new slider package installation command in run.sh

2017-08-17 Thread Gopal V (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17351?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16131441#comment-16131441
 ] 

Gopal V commented on HIVE-17351:


[~gsaha]: I assume the old syntax is deprecated since 0.92, but is still there 
in the docs - https://slider.incubator.apache.org/docs/getting_started.html#load

LGTM - +1


> use new slider package installation command in run.sh
> -
>
> Key: HIVE-17351
> URL: https://issues.apache.org/jira/browse/HIVE-17351
> Project: Hive
>  Issue Type: Bug
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-17351.patch
>
>
> The old syntax does not include some perf improvements in newer versions of 
> Slider.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17266) DecimalColumnVector64: Scaled fixed point column vector format

2017-08-17 Thread Gopal V (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17266?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16131433#comment-16131433
 ] 

Gopal V commented on HIVE-17266:


{code}
+  public Decimal64ColumnVector() {
+this(VectorizedRowBatch.DEFAULT_SIZE, 38, 10);
+  }
{code}

>From a quick look, I think (38,10) won't be supported by this column type at 
>all.

> DecimalColumnVector64: Scaled fixed point column vector format
> --
>
> Key: HIVE-17266
> URL: https://issues.apache.org/jira/browse/HIVE-17266
> Project: Hive
>  Issue Type: New Feature
>  Components: storage-api
>Affects Versions: 3.0.0
>Reporter: Gopal V
>Assignee: Owen O'Malley
> Attachments: HIVE-17266.patch
>
>
> I think we should make a new type that looks like:
> class Decimal64ColumnVector extends ColumnVector {
>   long[] vector;
>   int precision;
>   int scale;
> }
> It will be extremely fast and provide a fast conduit to ORC.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17286) Avoid expensive String serialization/deserialization for bitvectors

2017-08-17 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17286?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-17286:
---
   Resolution: Fixed
Fix Version/s: 3.0.0
   Status: Resolved  (was: Patch Available)

Pushed to master, thanks for reviewing [~ashutoshc]!

> Avoid expensive String serialization/deserialization for bitvectors
> ---
>
> Key: HIVE-17286
> URL: https://issues.apache.org/jira/browse/HIVE-17286
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
> Fix For: 3.0.0
>
> Attachments: HIVE-17286.01.patch, HIVE-17286.02.patch, 
> HIVE-17286.03.patch, HIVE-17286.04.patch, HIVE-17286.05.patch, 
> HIVE-17286.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17351) use new slider package installation command in run.sh

2017-08-17 Thread Gour Saha (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17351?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16131412#comment-16131412
 ] 

Gour Saha commented on HIVE-17351:
--

looks good to me

> use new slider package installation command in run.sh
> -
>
> Key: HIVE-17351
> URL: https://issues.apache.org/jira/browse/HIVE-17351
> Project: Hive
>  Issue Type: Bug
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-17351.patch
>
>
> The old syntax does not include some perf improvements in newer versions of 
> Slider.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17241) Change metastore classes to not use the shims

2017-08-17 Thread Alan Gates (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17241?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16131407#comment-16131407
 ] 

Alan Gates commented on HIVE-17241:
---

[~vgumashta], [~thejas], [~vihangk1], if you get a chance could you review 
this?  Thanks.

> Change metastore classes to not use the shims
> -
>
> Key: HIVE-17241
> URL: https://issues.apache.org/jira/browse/HIVE-17241
> Project: Hive
>  Issue Type: Sub-task
>  Components: Metastore
>Reporter: Alan Gates
>Assignee: Alan Gates
> Attachments: HIVE-17241.2.patch, HIVE-17241.patch
>
>
> As part of moving the metastore into a standalone package, it will no longer 
> have access to the shims.  This means we need to either copy them or access 
> the underlying Hadoop operations directly.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17351) use new slider package installation command in run.sh

2017-08-17 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17351?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-17351:

Status: Patch Available  (was: Open)

> use new slider package installation command in run.sh
> -
>
> Key: HIVE-17351
> URL: https://issues.apache.org/jira/browse/HIVE-17351
> Project: Hive
>  Issue Type: Bug
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-17351.patch
>
>
> The old syntax does not include some perf improvements in newer versions of 
> Slider.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17351) use new slider package installation command in run.sh

2017-08-17 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17351?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin updated HIVE-17351:

Attachment: HIVE-17351.patch

[~gopalv] can you take a look?
[~gsaha] does this make sense?

> use new slider package installation command in run.sh
> -
>
> Key: HIVE-17351
> URL: https://issues.apache.org/jira/browse/HIVE-17351
> Project: Hive
>  Issue Type: Bug
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
> Attachments: HIVE-17351.patch
>
>
> The old syntax does not include some perf improvements in newer versions of 
> Slider.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Assigned] (HIVE-17351) use new slider package installation command in run.sh

2017-08-17 Thread Sergey Shelukhin (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17351?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Shelukhin reassigned HIVE-17351:
---


> use new slider package installation command in run.sh
> -
>
> Key: HIVE-17351
> URL: https://issues.apache.org/jira/browse/HIVE-17351
> Project: Hive
>  Issue Type: Bug
>Reporter: Sergey Shelukhin
>Assignee: Sergey Shelukhin
>
> The old syntax does not include some perf improvements in newer versions of 
> Slider.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17286) Avoid expensive String serialization/deserialization for bitvectors

2017-08-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17286?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16131349#comment-16131349
 ] 

Hive QA commented on HIVE-17286:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12882402/HIVE-17286.05.patch

{color:green}SUCCESS:{color} +1 due to 3 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 8 failed/errored test(s), 10976 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning_mapjoin_only]
 (batchId=170)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_vectorized_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver[explainuser_3] 
(batchId=99)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query14] 
(batchId=235)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionSpecRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation 
(batchId=180)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6442/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6442/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6442/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 8 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12882402 - PreCommit-HIVE-Build

> Avoid expensive String serialization/deserialization for bitvectors
> ---
>
> Key: HIVE-17286
> URL: https://issues.apache.org/jira/browse/HIVE-17286
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
> Attachments: HIVE-17286.01.patch, HIVE-17286.02.patch, 
> HIVE-17286.03.patch, HIVE-17286.04.patch, HIVE-17286.05.patch, 
> HIVE-17286.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17198) Flaky test: TestBeeLineDriver [smb_mapjoin_7]

2017-08-17 Thread Peter Vary (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17198?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Peter Vary updated HIVE-17198:
--
Attachment: HIVE-17198.2.patch

Testing it again

> Flaky test: TestBeeLineDriver [smb_mapjoin_7]
> -
>
> Key: HIVE-17198
> URL: https://issues.apache.org/jira/browse/HIVE-17198
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Janaki Lahorani
>Assignee: Peter Vary
> Attachments: HIVE-17198.2.patch, HIVE-17198.patch
>
>
> Error:
> Exception running or analyzing the results of the query file: 
> org.apache.hive.beeline.QFile@4f7b68ad



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17198) Flaky test: TestBeeLineDriver [smb_mapjoin_7]

2017-08-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17198?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16131254#comment-16131254
 ] 

Hive QA commented on HIVE-17198:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12882367/HIVE-17198.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 9 failed/errored test(s), 10971 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning_mapjoin_only]
 (batchId=170)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_vectorized_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query14] 
(batchId=235)
org.apache.hadoop.hive.cli.TestSparkNegativeCliDriver.org.apache.hadoop.hive.cli.TestSparkNegativeCliDriver
 (batchId=242)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionSpecRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation 
(batchId=180)
org.apache.hive.jdbc.TestJdbcWithMiniHS2.testHttpRetryOnServerIdleTimeout 
(batchId=228)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6441/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6441/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6441/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 9 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12882367 - PreCommit-HIVE-Build

> Flaky test: TestBeeLineDriver [smb_mapjoin_7]
> -
>
> Key: HIVE-17198
> URL: https://issues.apache.org/jira/browse/HIVE-17198
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Janaki Lahorani
>Assignee: Peter Vary
> Attachments: HIVE-17198.patch
>
>
> Error:
> Exception running or analyzing the results of the query file: 
> org.apache.hive.beeline.QFile@4f7b68ad



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17344) LocalCache element memory usage is not calculated properly.

2017-08-17 Thread Sergey Shelukhin (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17344?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16131171#comment-16131171
 ] 

Sergey Shelukhin commented on HIVE-17344:
-

How are those bbs allocated? 

> LocalCache element memory usage is not calculated properly.
> ---
>
> Key: HIVE-17344
> URL: https://issues.apache.org/jira/browse/HIVE-17344
> Project: Hive
>  Issue Type: Bug
>Reporter: Janos Gub
>Assignee: Janos Gub
> Attachments: HIVE-17344.patch
>
>
> Orc footer cache has a calculation of memory usage:
> {code:java}
> public int getMemoryUsage() {
>   return bb.remaining() + 100; // 100 is for 2 longs, BB and java overheads 
> (semi-arbitrary).
> }
> {code}
> ByteBuffer.remaining returns the remaining space in the bytebuffer, thus 
> allowing this cache have elements MAXWEIGHT/100 of arbitrary size. I think 
> the correct solution would be bb.capacity.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17348) Remove unnecessary GenSparkUtils.java.orig file

2017-08-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17348?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16131041#comment-16131041
 ] 

Hive QA commented on HIVE-17348:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12882366/HIVE-17348.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 8 failed/errored test(s), 10977 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[columnstats_part_coltype]
 (batchId=159)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning_mapjoin_only]
 (batchId=170)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_vectorized_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query14] 
(batchId=235)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionSpecRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation 
(batchId=180)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6440/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6440/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6440/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 8 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12882366 - PreCommit-HIVE-Build

> Remove unnecessary GenSparkUtils.java.orig file
> ---
>
> Key: HIVE-17348
> URL: https://issues.apache.org/jira/browse/HIVE-17348
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Peter Vary
>Assignee: Peter Vary
> Attachments: HIVE-17348.patch
>
>
> HIVE-17247 added an extra file, which is most probably not needed :)
> [~stakiar]? :)



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17346) TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning] is failing every time

2017-08-17 Thread Vlad Gudikov (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17346?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16131009#comment-16131009
 ] 

Vlad Gudikov commented on HIVE-17346:
-

Yeah, it was an intended change, missed this one. Thanks!

> TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning] is failing 
> every time
> ---
>
> Key: HIVE-17346
> URL: https://issues.apache.org/jira/browse/HIVE-17346
> Project: Hive
>  Issue Type: Sub-task
>  Components: Test
>Affects Versions: 3.0.0
>Reporter: Peter Vary
>Assignee: Peter Vary
> Attachments: HIVE-17346.patch
>
>
> The TestMiniSparkOnYarnCliDriver.testCliDriver - 
> spark_dynamic_partition_pruning is failing with this diff:
> {code}
> Client Execution succeeded but contained differences (error code = 1) after 
> executing spark_dynamic_partition_pruning.q 
> 714c714
> <   filterExpr: ((date = '2008-04-08') and abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null) (type: 
> boolean)
> ---
> >   filterExpr: ((date = '2008-04-08') and ds is not null) 
> > (type: boolean)
> 717c717
> < predicate: ((date = '2008-04-08') and abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null) (type: 
> boolean)
> ---
> > predicate: ((date = '2008-04-08') and ds is not null) 
> > (type: boolean)
> 749c749
> <   filterExpr: abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null (type: 
> boolean)
> ---
> >   filterExpr: ds is not null (type: boolean)
> 751,752c751,753
> <   Filter Operator
> < predicate: abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null (type: 
> boolean)
> ---
> >   Select Operator
> > expressions: ds (type: string)
> > outputColumnNames: _col0
> 754,756c755,758
> < Select Operator
> <   expressions: ds (type: string)
> <   outputColumnNames: _col0
> ---
> > Reduce Output Operator
> >   key expressions: abs(((- 
> > UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> >   sort order: +
> >   Map-reduce partition columns: abs(((- 
> > UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> 758,762d759
> <   Reduce Output Operator
> < key expressions: abs(((- 
> UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> < sort order: +
> < Map-reduce partition columns: abs(((- 
> UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> < Statistics: Num rows: 2000 Data size: 21248 Basic 
> stats: COMPLETE Column stats: NONE
> 767c764
> <  
> Output was too long and had to be truncated...
> {code}
> I think it is caused by:
> HIVE-17148 - Incorrect result for Hive join query with COALESCE in WHERE 
> condition
> [~allgoodok]: Am I right? Is it an intended change and only the golden file 
> regeneration is needed?
> Thanks,
> Peter



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17292) Change TestMiniSparkOnYarnCliDriver test configuration to use the configured cores

2017-08-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17292?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16130932#comment-16130932
 ] 

Hive QA commented on HIVE-17292:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12882365/HIVE-17292.6.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 9 failed/errored test(s), 10977 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[unionDistinct_1] 
(batchId=143)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[columnstats_part_coltype]
 (batchId=159)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_vectorized_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query14] 
(batchId=235)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query23] 
(batchId=235)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionSpecRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation 
(batchId=180)
org.apache.hive.jdbc.TestJdbcWithMiniHS2.testHttpRetryOnServerIdleTimeout 
(batchId=228)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6439/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6439/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6439/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 9 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12882365 - PreCommit-HIVE-Build

> Change TestMiniSparkOnYarnCliDriver test configuration to use the configured 
> cores
> --
>
> Key: HIVE-17292
> URL: https://issues.apache.org/jira/browse/HIVE-17292
> Project: Hive
>  Issue Type: Sub-task
>  Components: Spark, Test
>Affects Versions: 3.0.0
>Reporter: Peter Vary
>Assignee: Peter Vary
> Attachments: HIVE-17292.1.patch, HIVE-17292.2.patch, 
> HIVE-17292.3.patch, HIVE-17292.5.patch, HIVE-17292.6.patch
>
>
> Currently the {{hive-site.xml}} for the {{TestMiniSparkOnYarnCliDriver}} test 
> defines 2 cores, and 2 executors, but only 1 is used, because the MiniCluster 
> does not allows the creation of the 3rd container.
> The FairScheduler uses 1GB increments for memory, but the containers would 
> like to use only 512MB. We should change the fairscheduler configuration to 
> use only the requested 512MB



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17347) TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning_mapjoin_only] is failing every time

2017-08-17 Thread Xuefu Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17347?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16130931#comment-16130931
 ] 

Xuefu Zhang commented on HIVE-17347:


This seems simple enough for me to review as well. +1

> TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning_mapjoin_only] is 
> failing every time
> 
>
> Key: HIVE-17347
> URL: https://issues.apache.org/jira/browse/HIVE-17347
> Project: Hive
>  Issue Type: Sub-task
>  Components: Test
>Affects Versions: 3.0.0
>Reporter: Peter Vary
>Assignee: Peter Vary
> Attachments: HIVE-17347.patch
>
>
> As [~lirui] identified there was a missing file from this patch: HIVE-17247 - 
> HoS DPP: UDFs on the partition column side does not evaluate correctly



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17286) Avoid expensive String serialization/deserialization for bitvectors

2017-08-17 Thread Jesus Camacho Rodriguez (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17286?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-17286:
---
Attachment: HIVE-17286.05.patch

> Avoid expensive String serialization/deserialization for bitvectors
> ---
>
> Key: HIVE-17286
> URL: https://issues.apache.org/jira/browse/HIVE-17286
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
> Attachments: HIVE-17286.01.patch, HIVE-17286.02.patch, 
> HIVE-17286.03.patch, HIVE-17286.04.patch, HIVE-17286.05.patch, 
> HIVE-17286.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17347) TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning_mapjoin_only] is failing every time

2017-08-17 Thread Peter Vary (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17347?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16130851#comment-16130851
 ] 

Peter Vary commented on HIVE-17347:
---

[~lirui], [~stakiar]: Could you please review?

> TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning_mapjoin_only] is 
> failing every time
> 
>
> Key: HIVE-17347
> URL: https://issues.apache.org/jira/browse/HIVE-17347
> Project: Hive
>  Issue Type: Sub-task
>  Components: Test
>Affects Versions: 3.0.0
>Reporter: Peter Vary
>Assignee: Peter Vary
> Attachments: HIVE-17347.patch
>
>
> As [~lirui] identified there was a missing file from this patch: HIVE-17247 - 
> HoS DPP: UDFs on the partition column side does not evaluate correctly



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17347) TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning_mapjoin_only] is failing every time

2017-08-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17347?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16130813#comment-16130813
 ] 

Hive QA commented on HIVE-17347:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12882363/HIVE-17347.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 5 failed/errored test(s), 10977 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_vectorized_dynamic_partition_pruning]
 (batchId=169)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionSpecRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation 
(batchId=180)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6438/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6438/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6438/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 5 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12882363 - PreCommit-HIVE-Build

> TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning_mapjoin_only] is 
> failing every time
> 
>
> Key: HIVE-17347
> URL: https://issues.apache.org/jira/browse/HIVE-17347
> Project: Hive
>  Issue Type: Sub-task
>  Components: Test
>Affects Versions: 3.0.0
>Reporter: Peter Vary
>Assignee: Peter Vary
> Attachments: HIVE-17347.patch
>
>
> As [~lirui] identified there was a missing file from this patch: HIVE-17247 - 
> HoS DPP: UDFs on the partition column side does not evaluate correctly



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17183) Disable rename operations during bootstrap dump

2017-08-17 Thread Sankar Hariappan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17183?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sankar Hariappan updated HIVE-17183:

Description: 
Currently, bootstrap dump shall lead to data loss when any rename happens while 
dump in progress. 
*Scenario:*
- Fetch table names (T1 and T2)
- Dump table T1
- Rename table T2 to T3 generates RENAME event
- Dump table T2 is noop as table doesn’t exist.
- In target after load, it only have T1.
- Apply RENAME event will fail as T2 doesn’t exist in target.

This feature can be supported in next phase development as it need proper 
design to keep track of renamed tables/partitions. 
So, for time being, we shall disable rename operations when bootstrap dump in 
progress to avoid any inconsistent state.

  was:
Currently, bootstrap dump shall lead to data loss when any rename happens while 
dump in progress. This feature can be supported in next phase development as it 
need proper design to keep track of renamed tables/partitions. 
So, for time being, we shall disable rename operations when bootstrap dump in 
progress to avoid any inconsistent state.


> Disable rename operations during bootstrap dump
> ---
>
> Key: HIVE-17183
> URL: https://issues.apache.org/jira/browse/HIVE-17183
> Project: Hive
>  Issue Type: Sub-task
>  Components: repl
>Affects Versions: 2.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>  Labels: DR, replication
> Fix For: 3.0.0
>
> Attachments: HIVE-17183.01.patch, HIVE-17183.02.patch, 
> HIVE-17183.03.patch
>
>
> Currently, bootstrap dump shall lead to data loss when any rename happens 
> while dump in progress. 
> *Scenario:*
> - Fetch table names (T1 and T2)
> - Dump table T1
> - Rename table T2 to T3 generates RENAME event
> - Dump table T2 is noop as table doesn’t exist.
> - In target after load, it only have T1.
> - Apply RENAME event will fail as T2 doesn’t exist in target.
> This feature can be supported in next phase development as it need proper 
> design to keep track of renamed tables/partitions. 
> So, for time being, we shall disable rename operations when bootstrap dump in 
> progress to avoid any inconsistent state.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17336) Missing class 'org.apache.hadoop.hive.hbase.HiveHBaseTableInputFormat' from Hive on Spark when inserting into hbase based table

2017-08-17 Thread Aihua Xu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17336?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aihua Xu updated HIVE-17336:

   Resolution: Fixed
Fix Version/s: 3.0.0
   Status: Resolved  (was: Patch Available)

Committed to master. Thanks Vihang for reviewing.

> Missing class 'org.apache.hadoop.hive.hbase.HiveHBaseTableInputFormat' from 
> Hive on Spark when inserting into hbase based table
> ---
>
> Key: HIVE-17336
> URL: https://issues.apache.org/jira/browse/HIVE-17336
> Project: Hive
>  Issue Type: Bug
>  Components: Spark
>Affects Versions: 3.0.0
>Reporter: Aihua Xu
>Assignee: Aihua Xu
> Fix For: 3.0.0
>
> Attachments: HIVE-17336.1.patch
>
>
> When inserting into a hbase based table from hive on spark, the following 
> exception is thrown 
> {noformat}
> Error while processing statement: FAILED: Execution Error, return code 3 from 
> org.apache.hadoop.hive.ql.exec.spark.SparkTask. 
> org.apache.hive.com.esotericsoftware.kryo.KryoException: Unable to find 
> class: org.apache.hadoop.hive.hbase.HiveHBaseTableInputFormat
> Serialization trace:
> inputFileFormatClass (org.apache.hadoop.hive.ql.plan.TableDesc)
> tableInfo (org.apache.hadoop.hive.ql.plan.FileSinkDesc)
> conf (org.apache.hadoop.hive.ql.exec.FileSinkOperator)
> childOperators (org.apache.hadoop.hive.ql.exec.SelectOperator)
> childOperators (org.apache.hadoop.hive.ql.exec.TableScanOperator)
> aliasToWork (org.apache.hadoop.hive.ql.plan.MapWork)
> invertedWorkGraph (org.apache.hadoop.hive.ql.plan.SparkWork)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.util.DefaultClassResolver.readName(DefaultClassResolver.java:156)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.util.DefaultClassResolver.readClass(DefaultClassResolver.java:133)
>  at org.apache.hive.com.esotericsoftware.kryo.Kryo.readClass(Kryo.java:670)
>  at 
> org.apache.hadoop.hive.ql.exec.SerializationUtilities$KryoWithHooks.readClass(SerializationUtilities.java:183)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.DefaultSerializers$ClassSerializer.read(DefaultSerializers.java:326)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.DefaultSerializers$ClassSerializer.read(DefaultSerializers.java:314)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.Kryo.readObjectOrNull(Kryo.java:759)
>  at 
> org.apache.hadoop.hive.ql.exec.SerializationUtilities$KryoWithHooks.readObjectOrNull(SerializationUtilities.java:201)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.ObjectField.read(ObjectField.java:132)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.FieldSerializer.read(FieldSerializer.java:551)
>  at org.apache.hive.com.esotericsoftware.kryo.Kryo.readObject(Kryo.java:708)
>  at 
> org.apache.hadoop.hive.ql.exec.SerializationUtilities$KryoWithHooks.readObject(SerializationUtilities.java:216)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.ObjectField.read(ObjectField.java:125)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.FieldSerializer.read(FieldSerializer.java:551)
>  at org.apache.hive.com.esotericsoftware.kryo.Kryo.readObject(Kryo.java:708)
>  at 
> org.apache.hadoop.hive.ql.exec.SerializationUtilities$KryoWithHooks.readObject(SerializationUtilities.java:216)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.ObjectField.read(ObjectField.java:125)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.FieldSerializer.read(FieldSerializer.java:551)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:790)
>  at 
> org.apache.hadoop.hive.ql.exec.SerializationUtilities$KryoWithHooks.readClassAndObject(SerializationUtilities.java:178)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.CollectionSerializer.read(CollectionSerializer.java:134)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.CollectionSerializer.read(CollectionSerializer.java:40)
>  at org.apache.hive.com.esotericsoftware.kryo.Kryo.readObject(Kryo.java:708)
>  at 
> org.apache.hadoop.hive.ql.exec.SerializationUtilities$KryoWithHooks.readObject(SerializationUtilities.java:216)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.ObjectField.read(ObjectField.java:125)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.FieldSerializer.read(FieldSerializer.java:551)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:790)
>  at 
> org.apache.hadoop.hive.ql.exec.SerializationUtilities$KryoWithHooks.readClassAndObject(SerializationUtilities.java:178)
>  at 
> org.apache.hive.com.esotericsoftware.kryo.serializers.CollectionSerializer.read(CollectionSerializer.java:134)
>  at 
> 

[jira] [Commented] (HIVE-17346) TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning] is failing every time

2017-08-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17346?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16130705#comment-16130705
 ] 

Hive QA commented on HIVE-17346:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12882359/HIVE-17346.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 8 failed/errored test(s), 10977 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[orc_merge7] 
(batchId=162)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning_mapjoin_only]
 (batchId=170)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_vectorized_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query14] 
(batchId=235)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query23] 
(batchId=235)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionSpecRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation 
(batchId=180)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6437/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6437/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6437/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 8 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12882359 - PreCommit-HIVE-Build

> TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning] is failing 
> every time
> ---
>
> Key: HIVE-17346
> URL: https://issues.apache.org/jira/browse/HIVE-17346
> Project: Hive
>  Issue Type: Sub-task
>  Components: Test
>Affects Versions: 3.0.0
>Reporter: Peter Vary
>Assignee: Peter Vary
> Attachments: HIVE-17346.patch
>
>
> The TestMiniSparkOnYarnCliDriver.testCliDriver - 
> spark_dynamic_partition_pruning is failing with this diff:
> {code}
> Client Execution succeeded but contained differences (error code = 1) after 
> executing spark_dynamic_partition_pruning.q 
> 714c714
> <   filterExpr: ((date = '2008-04-08') and abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null) (type: 
> boolean)
> ---
> >   filterExpr: ((date = '2008-04-08') and ds is not null) 
> > (type: boolean)
> 717c717
> < predicate: ((date = '2008-04-08') and abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null) (type: 
> boolean)
> ---
> > predicate: ((date = '2008-04-08') and ds is not null) 
> > (type: boolean)
> 749c749
> <   filterExpr: abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null (type: 
> boolean)
> ---
> >   filterExpr: ds is not null (type: boolean)
> 751,752c751,753
> <   Filter Operator
> < predicate: abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null (type: 
> boolean)
> ---
> >   Select Operator
> > expressions: ds (type: string)
> > outputColumnNames: _col0
> 754,756c755,758
> < Select Operator
> <   expressions: ds (type: string)
> <   outputColumnNames: _col0
> ---
> > Reduce Output Operator
> >   key expressions: abs(((- 
> > UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> >   sort order: +
> >   Map-reduce partition columns: abs(((- 
> > UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> 758,762d759
> <   Reduce Output Operator
> < key expressions: abs(((- 
> UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> < sort order: +
> < Map-reduce partition columns: abs(((- 
> UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> < Statistics: Num rows: 2000 Data size: 21248 Basic 
> stats: COMPLETE Column stats: NONE
> 767c764
> <  
> Output was too long and had to be truncated...
> {code}
> I think it is caused by:
> HIVE-17148 - Incorrect result 

[jira] [Commented] (HIVE-14747) Remove JAVA paths from profiles by sending them from ptest-client

2017-08-17 Thread JIRA

[ 
https://issues.apache.org/jira/browse/HIVE-14747?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16130680#comment-16130680
 ] 

Sergio Peña commented on HIVE-14747:


Thanks [~zsombor.klara]. The patch looks great.
+1

> Remove JAVA paths from profiles by sending them from ptest-client
> -
>
> Key: HIVE-14747
> URL: https://issues.apache.org/jira/browse/HIVE-14747
> Project: Hive
>  Issue Type: Sub-task
>  Components: Hive, Testing Infrastructure
>Reporter: Sergio Peña
>Assignee: Barna Zsombor Klara
> Attachments: HIVE-14747.01.patch, HIVE-14747.02.patch, 
> HIVE-14747.03.patch, HIVE-14747.04.patch, HIVE-14747.05.patch, 
> HIVE-14747.06.patch
>
>
> Hive ptest uses some properties files per branch that contain information 
> about how to execute the tests.
> This profile includes JAVA paths to build and execute the tests. We should 
> get rid of these by passing such information from Jenkins to the 
> ptest-server. In case a profile needs a different java version, then we can 
> create a specific Jenkins job for that.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17198) Flaky test: TestBeeLineDriver [smb_mapjoin_7]

2017-08-17 Thread Peter Vary (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17198?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Peter Vary updated HIVE-17198:
--
Attachment: HIVE-17198.patch

Attaching a do-nothing patch to test.

> Flaky test: TestBeeLineDriver [smb_mapjoin_7]
> -
>
> Key: HIVE-17198
> URL: https://issues.apache.org/jira/browse/HIVE-17198
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Janaki Lahorani
>Assignee: Peter Vary
> Attachments: HIVE-17198.patch
>
>
> Error:
> Exception running or analyzing the results of the query file: 
> org.apache.hive.beeline.QFile@4f7b68ad



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17198) Flaky test: TestBeeLineDriver [smb_mapjoin_7]

2017-08-17 Thread Peter Vary (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17198?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Peter Vary updated HIVE-17198:
--
Status: Patch Available  (was: Open)

> Flaky test: TestBeeLineDriver [smb_mapjoin_7]
> -
>
> Key: HIVE-17198
> URL: https://issues.apache.org/jira/browse/HIVE-17198
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Janaki Lahorani
>Assignee: Peter Vary
> Attachments: HIVE-17198.patch
>
>
> Error:
> Exception running or analyzing the results of the query file: 
> org.apache.hive.beeline.QFile@4f7b68ad



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Assigned] (HIVE-17198) Flaky test: TestBeeLineDriver [smb_mapjoin_7]

2017-08-17 Thread Peter Vary (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17198?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Peter Vary reassigned HIVE-17198:
-

Assignee: Peter Vary

> Flaky test: TestBeeLineDriver [smb_mapjoin_7]
> -
>
> Key: HIVE-17198
> URL: https://issues.apache.org/jira/browse/HIVE-17198
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Janaki Lahorani
>Assignee: Peter Vary
>
> Error:
> Exception running or analyzing the results of the query file: 
> org.apache.hive.beeline.QFile@4f7b68ad



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17198) Flaky test: TestBeeLineDriver [smb_mapjoin_7]

2017-08-17 Thread Janaki Lahorani (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17198?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16130521#comment-16130521
 ] 

Janaki Lahorani commented on HIVE-17198:


Hi [~pvary],
I found this test to fail intermittently when tests were run for uploaded 
patches.  I think the results are no longer available.  We will have to rerun 
the test to see if it still reproduces.
Thanks,
Janaki.

> Flaky test: TestBeeLineDriver [smb_mapjoin_7]
> -
>
> Key: HIVE-17198
> URL: https://issues.apache.org/jira/browse/HIVE-17198
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Janaki Lahorani
>
> Error:
> Exception running or analyzing the results of the query file: 
> org.apache.hive.beeline.QFile@4f7b68ad



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17348) Remove unnecessary GenSparkUtils.java.orig file

2017-08-17 Thread Peter Vary (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17348?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Peter Vary updated HIVE-17348:
--
Attachment: HIVE-17348.patch

> Remove unnecessary GenSparkUtils.java.orig file
> ---
>
> Key: HIVE-17348
> URL: https://issues.apache.org/jira/browse/HIVE-17348
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Peter Vary
>Assignee: Peter Vary
> Attachments: HIVE-17348.patch
>
>
> HIVE-17247 added an extra file, which is most probably not needed :)
> [~stakiar]? :)



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17348) Remove unnecessary GenSparkUtils.java.orig file

2017-08-17 Thread Peter Vary (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17348?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Peter Vary updated HIVE-17348:
--
Status: Patch Available  (was: Open)

> Remove unnecessary GenSparkUtils.java.orig file
> ---
>
> Key: HIVE-17348
> URL: https://issues.apache.org/jira/browse/HIVE-17348
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Peter Vary
>Assignee: Peter Vary
> Attachments: HIVE-17348.patch
>
>
> HIVE-17247 added an extra file, which is most probably not needed :)
> [~stakiar]? :)



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Assigned] (HIVE-17348) Remove unnecessary GenSparkUtils.java.orig file

2017-08-17 Thread Peter Vary (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17348?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Peter Vary reassigned HIVE-17348:
-


> Remove unnecessary GenSparkUtils.java.orig file
> ---
>
> Key: HIVE-17348
> URL: https://issues.apache.org/jira/browse/HIVE-17348
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Peter Vary
>Assignee: Peter Vary
>
> HIVE-17247 added an extra file, which is most probably not needed :)
> [~stakiar]? :)



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17292) Change TestMiniSparkOnYarnCliDriver test configuration to use the configured cores

2017-08-17 Thread Peter Vary (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17292?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Peter Vary updated HIVE-17292:
--
Attachment: HIVE-17292.6.patch

The patch contains the following changes:
- Changing Hadoop23Shims.java, so the MiniSparkShim will able to provide the 
requested 2 executors.
- Changing QTestUtil.setSparkSession, so we will wait until every executor is 
available, not only the 1st.
- Changing SparkSessionImpl.getMemoryAndCores, so we use the client provided 
paralellism in case of local spark.master too.
- Regenerating golden files (numReducers, and number of files changed in the 
explain plans)

The change contains 2 golden file changes 
(spark_dynamic_partition_pruning_mapjoin_only.q.out, 
spark_dynamic_partition_pruning.q.out) which are containing other neccessary 
changes for a green run, so this patch should be regenerated after their 
corresponding jiras are solved (HIVE-17347, HIVE-17346)

> Change TestMiniSparkOnYarnCliDriver test configuration to use the configured 
> cores
> --
>
> Key: HIVE-17292
> URL: https://issues.apache.org/jira/browse/HIVE-17292
> Project: Hive
>  Issue Type: Sub-task
>  Components: Spark, Test
>Affects Versions: 3.0.0
>Reporter: Peter Vary
>Assignee: Peter Vary
> Attachments: HIVE-17292.1.patch, HIVE-17292.2.patch, 
> HIVE-17292.3.patch, HIVE-17292.5.patch, HIVE-17292.6.patch
>
>
> Currently the {{hive-site.xml}} for the {{TestMiniSparkOnYarnCliDriver}} test 
> defines 2 cores, and 2 executors, but only 1 is used, because the MiniCluster 
> does not allows the creation of the 3rd container.
> The FairScheduler uses 1GB increments for memory, but the containers would 
> like to use only 512MB. We should change the fairscheduler configuration to 
> use only the requested 512MB



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17347) TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning_mapjoin_only] is failing every time

2017-08-17 Thread Peter Vary (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17347?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Peter Vary updated HIVE-17347:
--
Status: Patch Available  (was: Open)

> TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning_mapjoin_only] is 
> failing every time
> 
>
> Key: HIVE-17347
> URL: https://issues.apache.org/jira/browse/HIVE-17347
> Project: Hive
>  Issue Type: Sub-task
>  Components: Test
>Affects Versions: 3.0.0
>Reporter: Peter Vary
>Assignee: Peter Vary
> Attachments: HIVE-17347.patch
>
>
> As [~lirui] identified there was a missing file from this patch: HIVE-17247 - 
> HoS DPP: UDFs on the partition column side does not evaluate correctly



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17347) TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning_mapjoin_only] is failing every time

2017-08-17 Thread Peter Vary (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17347?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Peter Vary updated HIVE-17347:
--
Attachment: HIVE-17347.patch

Regenerated golden file

> TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning_mapjoin_only] is 
> failing every time
> 
>
> Key: HIVE-17347
> URL: https://issues.apache.org/jira/browse/HIVE-17347
> Project: Hive
>  Issue Type: Sub-task
>  Components: Test
>Affects Versions: 3.0.0
>Reporter: Peter Vary
>Assignee: Peter Vary
> Attachments: HIVE-17347.patch
>
>
> As [~lirui] identified there was a missing file from this patch: HIVE-17247 - 
> HoS DPP: UDFs on the partition column side does not evaluate correctly



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17319) Make BoneCp configurable using hive properties in hive-site.xml

2017-08-17 Thread Barna Zsombor Klara (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17319?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Barna Zsombor Klara updated HIVE-17319:
---
Attachment: HIVE-17319.draft.patch

Attaching the first draft of the patch for BoneCP.
[~thejas], [~ekoifman], [~pvary]
I hope this will clarify a bit what I had in mind. If you have the time please 
feel free to comment on the patch here or on reviewboard.

> Make BoneCp configurable using hive properties in hive-site.xml
> ---
>
> Key: HIVE-17319
> URL: https://issues.apache.org/jira/browse/HIVE-17319
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Barna Zsombor Klara
>Assignee: Barna Zsombor Klara
> Attachments: HIVE-17319.draft.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Assigned] (HIVE-17347) TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning_mapjoin_only] is failing every time

2017-08-17 Thread Peter Vary (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17347?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Peter Vary reassigned HIVE-17347:
-


> TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning_mapjoin_only] is 
> failing every time
> 
>
> Key: HIVE-17347
> URL: https://issues.apache.org/jira/browse/HIVE-17347
> Project: Hive
>  Issue Type: Sub-task
>  Components: Test
>Affects Versions: 3.0.0
>Reporter: Peter Vary
>Assignee: Peter Vary
>
> As [~lirui] identified there was a missing file from this patch: HIVE-17247 - 
> HoS DPP: UDFs on the partition column side does not evaluate correctly



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17346) TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning] is failing every time

2017-08-17 Thread Peter Vary (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17346?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Peter Vary updated HIVE-17346:
--
Affects Version/s: 3.0.0

> TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning] is failing 
> every time
> ---
>
> Key: HIVE-17346
> URL: https://issues.apache.org/jira/browse/HIVE-17346
> Project: Hive
>  Issue Type: Sub-task
>  Components: Test
>Affects Versions: 3.0.0
>Reporter: Peter Vary
>Assignee: Peter Vary
> Attachments: HIVE-17346.patch
>
>
> The TestMiniSparkOnYarnCliDriver.testCliDriver - 
> spark_dynamic_partition_pruning is failing with this diff:
> {code}
> Client Execution succeeded but contained differences (error code = 1) after 
> executing spark_dynamic_partition_pruning.q 
> 714c714
> <   filterExpr: ((date = '2008-04-08') and abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null) (type: 
> boolean)
> ---
> >   filterExpr: ((date = '2008-04-08') and ds is not null) 
> > (type: boolean)
> 717c717
> < predicate: ((date = '2008-04-08') and abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null) (type: 
> boolean)
> ---
> > predicate: ((date = '2008-04-08') and ds is not null) 
> > (type: boolean)
> 749c749
> <   filterExpr: abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null (type: 
> boolean)
> ---
> >   filterExpr: ds is not null (type: boolean)
> 751,752c751,753
> <   Filter Operator
> < predicate: abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null (type: 
> boolean)
> ---
> >   Select Operator
> > expressions: ds (type: string)
> > outputColumnNames: _col0
> 754,756c755,758
> < Select Operator
> <   expressions: ds (type: string)
> <   outputColumnNames: _col0
> ---
> > Reduce Output Operator
> >   key expressions: abs(((- 
> > UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> >   sort order: +
> >   Map-reduce partition columns: abs(((- 
> > UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> 758,762d759
> <   Reduce Output Operator
> < key expressions: abs(((- 
> UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> < sort order: +
> < Map-reduce partition columns: abs(((- 
> UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> < Statistics: Num rows: 2000 Data size: 21248 Basic 
> stats: COMPLETE Column stats: NONE
> 767c764
> <  
> Output was too long and had to be truncated...
> {code}
> I think it is caused by:
> HIVE-17148 - Incorrect result for Hive join query with COALESCE in WHERE 
> condition
> [~allgoodok]: Am I right? Is it an intended change and only the golden file 
> regeneration is needed?
> Thanks,
> Peter



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17346) TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning] is failing every time

2017-08-17 Thread Peter Vary (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17346?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Peter Vary updated HIVE-17346:
--
Status: Patch Available  (was: Open)

> TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning] is failing 
> every time
> ---
>
> Key: HIVE-17346
> URL: https://issues.apache.org/jira/browse/HIVE-17346
> Project: Hive
>  Issue Type: Sub-task
>  Components: Test
>Reporter: Peter Vary
>Assignee: Peter Vary
> Attachments: HIVE-17346.patch
>
>
> The TestMiniSparkOnYarnCliDriver.testCliDriver - 
> spark_dynamic_partition_pruning is failing with this diff:
> {code}
> Client Execution succeeded but contained differences (error code = 1) after 
> executing spark_dynamic_partition_pruning.q 
> 714c714
> <   filterExpr: ((date = '2008-04-08') and abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null) (type: 
> boolean)
> ---
> >   filterExpr: ((date = '2008-04-08') and ds is not null) 
> > (type: boolean)
> 717c717
> < predicate: ((date = '2008-04-08') and abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null) (type: 
> boolean)
> ---
> > predicate: ((date = '2008-04-08') and ds is not null) 
> > (type: boolean)
> 749c749
> <   filterExpr: abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null (type: 
> boolean)
> ---
> >   filterExpr: ds is not null (type: boolean)
> 751,752c751,753
> <   Filter Operator
> < predicate: abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null (type: 
> boolean)
> ---
> >   Select Operator
> > expressions: ds (type: string)
> > outputColumnNames: _col0
> 754,756c755,758
> < Select Operator
> <   expressions: ds (type: string)
> <   outputColumnNames: _col0
> ---
> > Reduce Output Operator
> >   key expressions: abs(((- 
> > UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> >   sort order: +
> >   Map-reduce partition columns: abs(((- 
> > UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> 758,762d759
> <   Reduce Output Operator
> < key expressions: abs(((- 
> UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> < sort order: +
> < Map-reduce partition columns: abs(((- 
> UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> < Statistics: Num rows: 2000 Data size: 21248 Basic 
> stats: COMPLETE Column stats: NONE
> 767c764
> <  
> Output was too long and had to be truncated...
> {code}
> I think it is caused by:
> HIVE-17148 - Incorrect result for Hive join query with COALESCE in WHERE 
> condition
> [~allgoodok]: Am I right? Is it an intended change and only the golden file 
> regeneration is needed?
> Thanks,
> Peter



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17346) TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning] is failing every time

2017-08-17 Thread Peter Vary (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17346?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Peter Vary updated HIVE-17346:
--
Attachment: HIVE-17346.patch

Here is the patch, if only golden file regeneration is needed

> TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning] is failing 
> every time
> ---
>
> Key: HIVE-17346
> URL: https://issues.apache.org/jira/browse/HIVE-17346
> Project: Hive
>  Issue Type: Sub-task
>  Components: Test
>Reporter: Peter Vary
>Assignee: Peter Vary
> Attachments: HIVE-17346.patch
>
>
> The TestMiniSparkOnYarnCliDriver.testCliDriver - 
> spark_dynamic_partition_pruning is failing with this diff:
> {code}
> Client Execution succeeded but contained differences (error code = 1) after 
> executing spark_dynamic_partition_pruning.q 
> 714c714
> <   filterExpr: ((date = '2008-04-08') and abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null) (type: 
> boolean)
> ---
> >   filterExpr: ((date = '2008-04-08') and ds is not null) 
> > (type: boolean)
> 717c717
> < predicate: ((date = '2008-04-08') and abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null) (type: 
> boolean)
> ---
> > predicate: ((date = '2008-04-08') and ds is not null) 
> > (type: boolean)
> 749c749
> <   filterExpr: abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null (type: 
> boolean)
> ---
> >   filterExpr: ds is not null (type: boolean)
> 751,752c751,753
> <   Filter Operator
> < predicate: abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null (type: 
> boolean)
> ---
> >   Select Operator
> > expressions: ds (type: string)
> > outputColumnNames: _col0
> 754,756c755,758
> < Select Operator
> <   expressions: ds (type: string)
> <   outputColumnNames: _col0
> ---
> > Reduce Output Operator
> >   key expressions: abs(((- 
> > UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> >   sort order: +
> >   Map-reduce partition columns: abs(((- 
> > UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> 758,762d759
> <   Reduce Output Operator
> < key expressions: abs(((- 
> UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> < sort order: +
> < Map-reduce partition columns: abs(((- 
> UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> < Statistics: Num rows: 2000 Data size: 21248 Basic 
> stats: COMPLETE Column stats: NONE
> 767c764
> <  
> Output was too long and had to be truncated...
> {code}
> I think it is caused by:
> HIVE-17148 - Incorrect result for Hive join query with COALESCE in WHERE 
> condition
> [~allgoodok]: Am I right? Is it an intended change and only the golden file 
> regeneration is needed?
> Thanks,
> Peter



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-16948) Invalid explain when running dynamic partition pruning query in Hive On Spark

2017-08-17 Thread Peter Vary (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-16948?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16130420#comment-16130420
 ] 

Peter Vary commented on HIVE-16948:
---

I do not find my comment on the review board, so I leave a comment here too:
I think {{spark_dynamic_partition_pruning.q.out}} changes are caused by 
HIVE-17148 - "Incorrect result for Hive join query with COALESCE in WHERE 
condition". Created HIVE-17346 to track the progress there

> Invalid explain when running dynamic partition pruning query in Hive On Spark
> -
>
> Key: HIVE-16948
> URL: https://issues.apache.org/jira/browse/HIVE-16948
> Project: Hive
>  Issue Type: Bug
>Reporter: liyunzhang_intel
>Assignee: liyunzhang_intel
> Attachments: HIVE-16948_1.patch, HIVE-16948.2.patch, 
> HIVE-16948.5.patch, HIVE-16948.6.patch, HIVE-16948.patch
>
>
> in 
> [union_subquery.q|https://github.com/apache/hive/blob/master/ql/src/test/queries/clientpositive/spark_dynamic_partition_pruning.q#L107]
>  in spark_dynamic_partition_pruning.q
> {code}
> set hive.optimize.ppd=true;
> set hive.ppd.remove.duplicatefilters=true;
> set hive.spark.dynamic.partition.pruning=true;
> set hive.optimize.metadataonly=false;
> set hive.optimize.index.filter=true;
> set hive.strict.checks.cartesian.product=false;
> explain select ds from (select distinct(ds) as ds from srcpart union all 
> select distinct(ds) as ds from srcpart) s where s.ds in (select 
> max(srcpart.ds) from srcpart union all select min(srcpart.ds) from srcpart);
> {code}
> explain 
> {code}
> STAGE DEPENDENCIES:
>   Stage-2 is a root stage
>   Stage-1 depends on stages: Stage-2
>   Stage-0 depends on stages: Stage-1
> STAGE PLANS:
>   Stage: Stage-2
> Spark
>   Edges:
> Reducer 11 <- Map 10 (GROUP, 1)
> Reducer 13 <- Map 12 (GROUP, 1)
>   DagName: root_20170622231525_20a777e5-e659-4138-b605-65f8395e18e2:2
>   Vertices:
> Map 10 
> Map Operator Tree:
> TableScan
>   alias: srcpart
>   Statistics: Num rows: 1 Data size: 23248 Basic stats: 
> PARTIAL Column stats: NONE
>   Select Operator
> expressions: ds (type: string)
> outputColumnNames: ds
> Statistics: Num rows: 1 Data size: 23248 Basic stats: 
> PARTIAL Column stats: NONE
> Group By Operator
>   aggregations: max(ds)
>   mode: hash
>   outputColumnNames: _col0
>   Statistics: Num rows: 1 Data size: 184 Basic stats: 
> COMPLETE Column stats: NONE
>   Reduce Output Operator
> sort order: 
> Statistics: Num rows: 1 Data size: 184 Basic stats: 
> COMPLETE Column stats: NONE
> value expressions: _col0 (type: string)
> Map 12 
> Map Operator Tree:
> TableScan
>   alias: srcpart
>   Statistics: Num rows: 1 Data size: 23248 Basic stats: 
> PARTIAL Column stats: NONE
>   Select Operator
> expressions: ds (type: string)
> outputColumnNames: ds
> Statistics: Num rows: 1 Data size: 23248 Basic stats: 
> PARTIAL Column stats: NONE
> Group By Operator
>   aggregations: min(ds)
>   mode: hash
>   outputColumnNames: _col0
>   Statistics: Num rows: 1 Data size: 184 Basic stats: 
> COMPLETE Column stats: NONE
>   Reduce Output Operator
> sort order: 
> Statistics: Num rows: 1 Data size: 184 Basic stats: 
> COMPLETE Column stats: NONE
> value expressions: _col0 (type: string)
> Reducer 11 
> Reduce Operator Tree:
>   Group By Operator
> aggregations: max(VALUE._col0)
> mode: mergepartial
> outputColumnNames: _col0
> Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE 
> Column stats: NONE
> Filter Operator
>   predicate: _col0 is not null (type: boolean)
>   Statistics: Num rows: 1 Data size: 184 Basic stats: 
> COMPLETE Column stats: NONE
>   Group By Operator
> keys: _col0 (type: string)
> mode: hash
> outputColumnNames: _col0
> Statistics: Num rows: 2 Data size: 368 Basic stats: 
> COMPLETE Column stats: NONE
> Select Operator
>   expressions: _col0 (type: string)
>  

[jira] [Assigned] (HIVE-17346) TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning] is failing every time

2017-08-17 Thread Peter Vary (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17346?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Peter Vary reassigned HIVE-17346:
-


> TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning] is failing 
> every time
> ---
>
> Key: HIVE-17346
> URL: https://issues.apache.org/jira/browse/HIVE-17346
> Project: Hive
>  Issue Type: Sub-task
>  Components: Test
>Reporter: Peter Vary
>Assignee: Peter Vary
>
> The TestMiniSparkOnYarnCliDriver.testCliDriver - 
> spark_dynamic_partition_pruning is failing with this diff:
> {code}
> Client Execution succeeded but contained differences (error code = 1) after 
> executing spark_dynamic_partition_pruning.q 
> 714c714
> <   filterExpr: ((date = '2008-04-08') and abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null) (type: 
> boolean)
> ---
> >   filterExpr: ((date = '2008-04-08') and ds is not null) 
> > (type: boolean)
> 717c717
> < predicate: ((date = '2008-04-08') and abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null) (type: 
> boolean)
> ---
> > predicate: ((date = '2008-04-08') and ds is not null) 
> > (type: boolean)
> 749c749
> <   filterExpr: abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null (type: 
> boolean)
> ---
> >   filterExpr: ds is not null (type: boolean)
> 751,752c751,753
> <   Filter Operator
> < predicate: abs(((- 
> UDFToLong(concat(UDFToString(day(ds)), '0'))) + 10)) is not null (type: 
> boolean)
> ---
> >   Select Operator
> > expressions: ds (type: string)
> > outputColumnNames: _col0
> 754,756c755,758
> < Select Operator
> <   expressions: ds (type: string)
> <   outputColumnNames: _col0
> ---
> > Reduce Output Operator
> >   key expressions: abs(((- 
> > UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> >   sort order: +
> >   Map-reduce partition columns: abs(((- 
> > UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> 758,762d759
> <   Reduce Output Operator
> < key expressions: abs(((- 
> UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> < sort order: +
> < Map-reduce partition columns: abs(((- 
> UDFToLong(concat(UDFToString(day(_col0)), '0'))) + 10)) (type: bigint)
> < Statistics: Num rows: 2000 Data size: 21248 Basic 
> stats: COMPLETE Column stats: NONE
> 767c764
> <  
> Output was too long and had to be truncated...
> {code}
> I think it is caused by:
> HIVE-17148 - Incorrect result for Hive join query with COALESCE in WHERE 
> condition
> [~allgoodok]: Am I right? Is it an intended change and only the golden file 
> regeneration is needed?
> Thanks,
> Peter



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17344) LocalCache element memory usage is not calculated properly.

2017-08-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17344?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16130380#comment-16130380
 ] 

Hive QA commented on HIVE-17344:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12882323/HIVE-17344.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 10 failed/errored test(s), 10977 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestBeeLineDriver.testCliDriver[smb_mapjoin_1] 
(batchId=240)
org.apache.hadoop.hive.cli.TestBeeLineDriver.testCliDriver[smb_mapjoin_7] 
(batchId=240)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning_mapjoin_only]
 (batchId=170)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_vectorized_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query14] 
(batchId=235)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query23] 
(batchId=235)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionSpecRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation 
(batchId=180)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6436/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6436/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6436/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 10 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12882323 - PreCommit-HIVE-Build

> LocalCache element memory usage is not calculated properly.
> ---
>
> Key: HIVE-17344
> URL: https://issues.apache.org/jira/browse/HIVE-17344
> Project: Hive
>  Issue Type: Bug
>Reporter: Janos Gub
>Assignee: Janos Gub
> Attachments: HIVE-17344.patch
>
>
> Orc footer cache has a calculation of memory usage:
> {code:java}
> public int getMemoryUsage() {
>   return bb.remaining() + 100; // 100 is for 2 longs, BB and java overheads 
> (semi-arbitrary).
> }
> {code}
> ByteBuffer.remaining returns the remaining space in the bytebuffer, thus 
> allowing this cache have elements MAXWEIGHT/100 of arbitrary size. I think 
> the correct solution would be bb.capacity.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Comment Edited] (HIVE-16713) Remove wrong judgments of single and double quotes in "--" comments

2017-08-17 Thread Alister Roberts (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-16713?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16130323#comment-16130323
 ] 

Alister Roberts edited comment on HIVE-16713 at 8/17/17 12:33 PM:
--

This behaviour is a regression - a comment with an embedded apostrophe worked 
fine in Hive 2.1.0, but fails in Hive 2.3.0 for this reason.

It should be noted that the behaviour does not manifest itself when testing 
using the Hive CLI - I guess it does some processing that masks the problem.


was (Author: gatelogic):
This behaviour is a regression - a comment with an embedded apostrophe worked 
fine in Hive 2.1.0, but fails in Hive 2.3.0 for this reason.

> Remove wrong judgments of single and double quotes in "--" comments
> ---
>
> Key: HIVE-16713
> URL: https://issues.apache.org/jira/browse/HIVE-16713
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Liu Chunxiao
>Assignee: Liu Chunxiao
>
> single quote in "--" comments will lead to parse exception
> e.g. sql files which contain the following:
> select  --...user's...
> count(*) from aa;
> will regard the single quote in comment "user's" as a quote in sql statement. 
> So ";" after aa will be regarded as a ";" in single quote and will not be 
> split by the parser.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-16713) Remove wrong judgments of single and double quotes in "--" comments

2017-08-17 Thread Alister Roberts (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-16713?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alister Roberts updated HIVE-16713:
---
Priority: Major  (was: Minor)

> Remove wrong judgments of single and double quotes in "--" comments
> ---
>
> Key: HIVE-16713
> URL: https://issues.apache.org/jira/browse/HIVE-16713
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Liu Chunxiao
>Assignee: Liu Chunxiao
>
> single quote in "--" comments will lead to parse exception
> e.g. sql files which contain the following:
> select  --...user's...
> count(*) from aa;
> will regard the single quote in comment "user's" as a quote in sql statement. 
> So ";" after aa will be regarded as a ";" in single quote and will not be 
> split by the parser.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-16713) Remove wrong judgments of single and double quotes in "--" comments

2017-08-17 Thread Alister Roberts (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-16713?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16130323#comment-16130323
 ] 

Alister Roberts commented on HIVE-16713:


This behaviour is a regression - a comment with an embedded apostrophe worked 
fine in Hive 2.1.0, but fails in Hive 2.3.0 for this reason.

> Remove wrong judgments of single and double quotes in "--" comments
> ---
>
> Key: HIVE-16713
> URL: https://issues.apache.org/jira/browse/HIVE-16713
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Liu Chunxiao
>Assignee: Liu Chunxiao
>Priority: Minor
>
> single quote in "--" comments will lead to parse exception
> e.g. sql files which contain the following:
> select  --...user's...
> count(*) from aa;
> will regard the single quote in comment "user's" as a quote in sql statement. 
> So ";" after aa will be regarded as a ";" in single quote and will not be 
> split by the parser.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17100) Improve HS2 operation logs for REPL commands.

2017-08-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17100?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16130313#comment-16130313
 ] 

Hive QA commented on HIVE-17100:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12882306/HIVE-17100.05.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 8 failed/errored test(s), 10977 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning_mapjoin_only]
 (batchId=170)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_vectorized_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query14] 
(batchId=235)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query23] 
(batchId=235)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionSpecRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation 
(batchId=180)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6435/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6435/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6435/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 8 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12882306 - PreCommit-HIVE-Build

> Improve HS2 operation logs for REPL commands.
> -
>
> Key: HIVE-17100
> URL: https://issues.apache.org/jira/browse/HIVE-17100
> Project: Hive
>  Issue Type: Sub-task
>  Components: HiveServer2, repl
>Affects Versions: 2.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>  Labels: DR, replication
> Fix For: 3.0.0
>
> Attachments: HIVE-17100.01.patch, HIVE-17100.02.patch, 
> HIVE-17100.03.patch, HIVE-17100.04.patch, HIVE-17100.05.patch
>
>
> It is necessary to log the progress the replication tasks in a structured 
> manner as follows.
> *+Bootstrap Dump:+*
> * At the start of bootstrap dump, will add one log with below details.
> {color:#59afe1}* Database Name
> * Dump Type (BOOTSTRAP)
> * (Estimated) Total number of tables/views to dump
> * (Estimated) Total number of functions to dump.
> * Dump Start Time{color}
> * After each table dump, will add a log as follows
> {color:#59afe1}* Table/View Name
> * Type (TABLE/VIEW/MATERIALIZED_VIEW)
> * Table dump end time
> * Table dump progress. Format is Table sequence no/(Estimated) Total number 
> of tables and views.{color}
> * After each function dump, will add a log as follows
> {color:#59afe1}* Function Name
> * Function dump end time
> * Function dump progress. Format is Function sequence no/(Estimated) Total 
> number of functions.{color}
> * After completion of all dumps, will add a log as follows to consolidate the 
> dump.
> {color:#59afe1}* Database Name.
> * Dump Type (BOOTSTRAP).
> * Dump End Time.
> * (Actual) Total number of tables/views dumped.
> * (Actual) Total number of functions dumped.
> * Dump Directory.
> * Last Repl ID of the dump.{color}
> *Note:* The actual and estimated number of tables/functions may not match if 
> any table/function is dropped when dump in progress.
> *+Bootstrap Load:+*
> * At the start of bootstrap load, will add one log with below details.
> {color:#59afe1}* Database Name
> * Dump directory
> * Load Type (BOOTSTRAP)
> * Total number of tables/views to load
> * Total number of functions to load.
> * Load Start Time{color}
> * After each table load, will add a log as follows
> {color:#59afe1}* Table/View Name
> * Type (TABLE/VIEW/MATERIALIZED_VIEW)
> * Table load completion time
> * Table load progress. Format is Table sequence no/Total number of tables and 
> views.{color}
> * After each function load, will add a log as follows
> {color:#59afe1}* Function Name
> * Function load completion time
> * Function load progress. Format is Function sequence no/Total number of 
> functions.{color}
> * After completion of all dumps, will add a log as follows to consolidate the 
> load.
> {color:#59afe1}* Database Name.
> * Load Type (BOOTSTRAP).
> * Load End Time.
> * Total number of tables/views loaded.
> * Total number of functions loaded.
> * Last Repl ID of the loaded 

[jira] [Updated] (HIVE-17344) LocalCache element memory usage is not calculated properly.

2017-08-17 Thread Janos Gub (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17344?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Janos Gub updated HIVE-17344:
-
Description: 
Orc footer cache has a calculation of memory usage:
{code:java}
public int getMemoryUsage() {
  return bb.remaining() + 100; // 100 is for 2 longs, BB and java overheads 
(semi-arbitrary).
}
{code}

ByteBuffer.remaining returns the remaining space in the bytebuffer, thus 
allowing this cache have elements MAXWEIGHT/100 of arbitrary size. I think the 
correct solution would be bb.capacity.

  was:
Orc footer cache has a calculation of memory usage:
{code:java}
public int getMemoryUsage() {
  return bb.remaining() + 100; // 100 is for 2 longs, BB and java overheads 
(semi-arbitrary).
}
{code}

ByteBuffer.remaining returns the remaining space in the bytebuffer, thus 
allowing this cache to grow until MAXWEIGHT/100. I think the correct solution 
would be bb.capacity.


> LocalCache element memory usage is not calculated properly.
> ---
>
> Key: HIVE-17344
> URL: https://issues.apache.org/jira/browse/HIVE-17344
> Project: Hive
>  Issue Type: Bug
>Reporter: Janos Gub
>Assignee: Janos Gub
> Attachments: HIVE-17344.patch
>
>
> Orc footer cache has a calculation of memory usage:
> {code:java}
> public int getMemoryUsage() {
>   return bb.remaining() + 100; // 100 is for 2 longs, BB and java overheads 
> (semi-arbitrary).
> }
> {code}
> ByteBuffer.remaining returns the remaining space in the bytebuffer, thus 
> allowing this cache have elements MAXWEIGHT/100 of arbitrary size. I think 
> the correct solution would be bb.capacity.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17344) LocalCache element memory usage is not calculated properly.

2017-08-17 Thread Zoltan Haindrich (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17344?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16130285#comment-16130285
 ] 

Zoltan Haindrich commented on HIVE-17344:
-

+1 pending tests

> LocalCache element memory usage is not calculated properly.
> ---
>
> Key: HIVE-17344
> URL: https://issues.apache.org/jira/browse/HIVE-17344
> Project: Hive
>  Issue Type: Bug
>Reporter: Janos Gub
>Assignee: Janos Gub
> Attachments: HIVE-17344.patch
>
>
> Orc footer cache has a calculation of memory usage:
> {code:java}
> public int getMemoryUsage() {
>   return bb.remaining() + 100; // 100 is for 2 longs, BB and java overheads 
> (semi-arbitrary).
> }
> {code}
> ByteBuffer.remaining returns the remaining space in the bytebuffer, thus 
> allowing this cache to grow until MAXWEIGHT/100. I think the correct solution 
> would be bb.capacity.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17344) LocalCache element memory usage is not calculated properly.

2017-08-17 Thread Janos Gub (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17344?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Janos Gub updated HIVE-17344:
-
Description: 
Orc footer cache has a calculation of memory usage:
{code:java}
public int getMemoryUsage() {
  return bb.remaining() + 100; // 100 is for 2 longs, BB and java overheads 
(semi-arbitrary).
}
{code}

ByteBuffer.remaining returns the remaining space in the bytebuffer, thus 
allowing this cache to grow until MAXWEIGHT/100. I think the correct solution 
would be bb.capacity.

  was:
Orc footer cache has a calculation of memory usage:
{code:java}
public int getMemoryUsage() {
  return bb.remaining() + 100; // 100 is for 2 longs, BB and java overheads 
(semi-arbitrary).
}
{code}

ByteBuffer.remaining returns the remaining space in the bytebuffer, thus 
allowing this cache to grow until MAXSIZE/100. I think the correct solution 
would be bb.capacity.


> LocalCache element memory usage is not calculated properly.
> ---
>
> Key: HIVE-17344
> URL: https://issues.apache.org/jira/browse/HIVE-17344
> Project: Hive
>  Issue Type: Bug
>Reporter: Janos Gub
>Assignee: Janos Gub
> Attachments: HIVE-17344.patch
>
>
> Orc footer cache has a calculation of memory usage:
> {code:java}
> public int getMemoryUsage() {
>   return bb.remaining() + 100; // 100 is for 2 longs, BB and java overheads 
> (semi-arbitrary).
> }
> {code}
> ByteBuffer.remaining returns the remaining space in the bytebuffer, thus 
> allowing this cache to grow until MAXWEIGHT/100. I think the correct solution 
> would be bb.capacity.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17344) LocalCache element memory usage is not calculated properly.

2017-08-17 Thread Janos Gub (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17344?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Janos Gub updated HIVE-17344:
-
Status: Patch Available  (was: Open)

> LocalCache element memory usage is not calculated properly.
> ---
>
> Key: HIVE-17344
> URL: https://issues.apache.org/jira/browse/HIVE-17344
> Project: Hive
>  Issue Type: Bug
>Reporter: Janos Gub
>Assignee: Janos Gub
> Attachments: HIVE-17344.patch
>
>
> Orc footer cache has a calculation of memory usage:
> {code:java}
> public int getMemoryUsage() {
>   return bb.remaining() + 100; // 100 is for 2 longs, BB and java overheads 
> (semi-arbitrary).
> }
> {code}
> ByteBuffer.remaining returns the remaining space in the bytebuffer, thus 
> allowing this cache to grow until MAXSIZE/100. I think the correct solution 
> would be bb.capacity.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17344) LocalCache element memory usage is not calculated properly.

2017-08-17 Thread Janos Gub (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17344?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Janos Gub updated HIVE-17344:
-
Attachment: HIVE-17344.patch

> LocalCache element memory usage is not calculated properly.
> ---
>
> Key: HIVE-17344
> URL: https://issues.apache.org/jira/browse/HIVE-17344
> Project: Hive
>  Issue Type: Bug
>Reporter: Janos Gub
>Assignee: Janos Gub
> Attachments: HIVE-17344.patch
>
>
> Orc footer cache has a calculation of memory usage:
> {code:java}
> public int getMemoryUsage() {
>   return bb.remaining() + 100; // 100 is for 2 longs, BB and java overheads 
> (semi-arbitrary).
> }
> {code}
> ByteBuffer.remaining returns the remaining space in the bytebuffer, thus 
> allowing this cache to grow until MAXSIZE/100. I think the correct solution 
> would be bb.capacity.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-16133) Footer cache in Tez AM can take too much memory

2017-08-17 Thread Janos Gub (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-16133?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16130250#comment-16130250
 ] 

Janos Gub commented on HIVE-16133:
--

I think there is an error in the computation of the cache size. I opened a new 
jira ticket:https://issues.apache.org/jira/browse/HIVE-17344

> Footer cache in Tez AM can take too much memory
> ---
>
> Key: HIVE-16133
> URL: https://issues.apache.org/jira/browse/HIVE-16133
> Project: Hive
>  Issue Type: Bug
>Reporter: Siddharth Seth
>Assignee: Sergey Shelukhin
>  Labels: TODOC2.2
> Fix For: 2.3.0
>
> Attachments: HIVE-16133.01.patch, HIVE-16133.02.patch, 
> HIVE-16133.02.patch, HIVE-16133.03.patch, HIVE-16133.04.patch, 
> HIVE-16133.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Assigned] (HIVE-17344) LocalCache element memory usage is not calculated properly.

2017-08-17 Thread Janos Gub (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17344?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Janos Gub reassigned HIVE-17344:



> LocalCache element memory usage is not calculated properly.
> ---
>
> Key: HIVE-17344
> URL: https://issues.apache.org/jira/browse/HIVE-17344
> Project: Hive
>  Issue Type: Bug
>Reporter: Janos Gub
>Assignee: Janos Gub
>
> Orc footer cache has a calculation of memory usage:
> {code:java}
> public int getMemoryUsage() {
>   return bb.remaining() + 100; // 100 is for 2 longs, BB and java overheads 
> (semi-arbitrary).
> }
> {code}
> ByteBuffer.remaining returns the remaining space in the bytebuffer, thus 
> allowing this cache to grow until MAXSIZE/100. I think the correct solution 
> would be bb.capacity.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-14747) Remove JAVA paths from profiles by sending them from ptest-client

2017-08-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-14747?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16130242#comment-16130242
 ] 

Hive QA commented on HIVE-14747:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12882289/HIVE-14747.06.patch

{color:green}SUCCESS:{color} +1 due to 7 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 8 failed/errored test(s), 10977 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[unionDistinct_1] 
(batchId=143)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning_mapjoin_only]
 (batchId=170)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_vectorized_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestPerfCliDriver.testCliDriver[query14] 
(batchId=235)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionSpecRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation 
(batchId=180)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6434/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6434/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6434/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 8 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12882289 - PreCommit-HIVE-Build

> Remove JAVA paths from profiles by sending them from ptest-client
> -
>
> Key: HIVE-14747
> URL: https://issues.apache.org/jira/browse/HIVE-14747
> Project: Hive
>  Issue Type: Sub-task
>  Components: Hive, Testing Infrastructure
>Reporter: Sergio Peña
>Assignee: Barna Zsombor Klara
> Attachments: HIVE-14747.01.patch, HIVE-14747.02.patch, 
> HIVE-14747.03.patch, HIVE-14747.04.patch, HIVE-14747.05.patch, 
> HIVE-14747.06.patch
>
>
> Hive ptest uses some properties files per branch that contain information 
> about how to execute the tests.
> This profile includes JAVA paths to build and execute the tests. We should 
> get rid of these by passing such information from Jenkins to the 
> ptest-server. In case a profile needs a different java version, then we can 
> create a specific Jenkins job for that.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17314) LazySimpleSerializeWrite.writeString() contains if with an empty body

2017-08-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17314?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16130154#comment-16130154
 ] 

Hive QA commented on HIVE-17314:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12882285/HIVE-17314.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 7 failed/errored test(s), 10977 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning]
 (batchId=169)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_dynamic_partition_pruning_mapjoin_only]
 (batchId=170)
org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[spark_vectorized_dynamic_partition_pruning]
 (batchId=169)
org.apache.hive.beeline.TestBeeLineWithArgs.testQueryProgressParallel 
(batchId=222)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testPartitionSpecRegistrationWithCustomSchema
 (batchId=180)
org.apache.hive.hcatalog.api.TestHCatClient.testTableSchemaPropagation 
(batchId=180)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6433/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6433/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6433/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 7 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12882285 - PreCommit-HIVE-Build

> LazySimpleSerializeWrite.writeString() contains if with an empty body
> -
>
> Key: HIVE-17314
> URL: https://issues.apache.org/jira/browse/HIVE-17314
> Project: Hive
>  Issue Type: Bug
>Reporter: Oleg Danilov
>Assignee: Oleg Danilov
>Priority: Minor
> Attachments: HIVE-17314.patch
>
>
> Looking at the LazySimpleSerializeWrite.java I found odd 'if':
> Lines 234-235:
> {code:java}
> if (v.equals(nullSequenceBytes)) {
> }
> {code}
> Seems like either something is missed there or this 'if' could be dropped out.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-17100) Improve HS2 operation logs for REPL commands.

2017-08-17 Thread Sankar Hariappan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17100?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sankar Hariappan updated HIVE-17100:

Status: Patch Available  (was: Open)

> Improve HS2 operation logs for REPL commands.
> -
>
> Key: HIVE-17100
> URL: https://issues.apache.org/jira/browse/HIVE-17100
> Project: Hive
>  Issue Type: Sub-task
>  Components: HiveServer2, repl
>Affects Versions: 2.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>  Labels: DR, replication
> Fix For: 3.0.0
>
> Attachments: HIVE-17100.01.patch, HIVE-17100.02.patch, 
> HIVE-17100.03.patch, HIVE-17100.04.patch, HIVE-17100.05.patch
>
>
> It is necessary to log the progress the replication tasks in a structured 
> manner as follows.
> *+Bootstrap Dump:+*
> * At the start of bootstrap dump, will add one log with below details.
> {color:#59afe1}* Database Name
> * Dump Type (BOOTSTRAP)
> * (Estimated) Total number of tables/views to dump
> * (Estimated) Total number of functions to dump.
> * Dump Start Time{color}
> * After each table dump, will add a log as follows
> {color:#59afe1}* Table/View Name
> * Type (TABLE/VIEW/MATERIALIZED_VIEW)
> * Table dump end time
> * Table dump progress. Format is Table sequence no/(Estimated) Total number 
> of tables and views.{color}
> * After each function dump, will add a log as follows
> {color:#59afe1}* Function Name
> * Function dump end time
> * Function dump progress. Format is Function sequence no/(Estimated) Total 
> number of functions.{color}
> * After completion of all dumps, will add a log as follows to consolidate the 
> dump.
> {color:#59afe1}* Database Name.
> * Dump Type (BOOTSTRAP).
> * Dump End Time.
> * (Actual) Total number of tables/views dumped.
> * (Actual) Total number of functions dumped.
> * Dump Directory.
> * Last Repl ID of the dump.{color}
> *Note:* The actual and estimated number of tables/functions may not match if 
> any table/function is dropped when dump in progress.
> *+Bootstrap Load:+*
> * At the start of bootstrap load, will add one log with below details.
> {color:#59afe1}* Database Name
> * Dump directory
> * Load Type (BOOTSTRAP)
> * Total number of tables/views to load
> * Total number of functions to load.
> * Load Start Time{color}
> * After each table load, will add a log as follows
> {color:#59afe1}* Table/View Name
> * Type (TABLE/VIEW/MATERIALIZED_VIEW)
> * Table load completion time
> * Table load progress. Format is Table sequence no/Total number of tables and 
> views.{color}
> * After each function load, will add a log as follows
> {color:#59afe1}* Function Name
> * Function load completion time
> * Function load progress. Format is Function sequence no/Total number of 
> functions.{color}
> * After completion of all dumps, will add a log as follows to consolidate the 
> load.
> {color:#59afe1}* Database Name.
> * Load Type (BOOTSTRAP).
> * Load End Time.
> * Total number of tables/views loaded.
> * Total number of functions loaded.
> * Last Repl ID of the loaded database.{color}
> *+Incremental Dump:+*
> * At the start of database dump, will add one log with below details.
> {color:#59afe1}* Database Name
> * Dump Type (INCREMENTAL)
> * (Estimated) Total number of events to dump.
> * Dump Start Time{color}
> * After each event dump, will add a log as follows
> {color:#59afe1}* Event ID
> * Event Type (CREATE_TABLE, DROP_TABLE, ALTER_TABLE, INSERT etc)
> * Event dump end time
> * Event dump progress. Format is Event sequence no/ (Estimated) Total number 
> of events.{color}
> * After completion of all event dumps, will add a log as follows.
> {color:#59afe1}* Database Name.
> * Dump Type (INCREMENTAL).
> * Dump End Time.
> * (Actual) Total number of events dumped.
> * Dump Directory.
> * Last Repl ID of the dump.{color}
> *Note:* The estimated number of events can be terribly inaccurate with actual 
> number as we don’t have the number of events upfront until we read from 
> metastore NotificationEvents table.
> *+Incremental Load:+*
> * At the start of incremental load, will add one log with below details.
> {color:#59afe1}* Target Database Name 
> * Dump directory
> * Load Type (INCREMENTAL)
> * Total number of events to load
> * Load Start Time{color}
> * After each event load, will add a log as follows
> {color:#59afe1}* Event ID
> * Event Type (CREATE_TABLE, DROP_TABLE, ALTER_TABLE, INSERT etc)
> * Event load end time
> * Event load progress. Format is Event sequence no/ Total number of 
> events.{color}
> * After completion of all event loads, will add a log as follows to 
> consolidate the load.
> {color:#59afe1}* Target Database Name.
> * Load Type (INCREMENTAL).
> * Load End Time.
> * Total number of events loaded.
> * Last Repl ID of the loaded database.{color}



--
This message was sent by 

[jira] [Updated] (HIVE-17100) Improve HS2 operation logs for REPL commands.

2017-08-17 Thread Sankar Hariappan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17100?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sankar Hariappan updated HIVE-17100:

Attachment: HIVE-17100.05.patch

> Improve HS2 operation logs for REPL commands.
> -
>
> Key: HIVE-17100
> URL: https://issues.apache.org/jira/browse/HIVE-17100
> Project: Hive
>  Issue Type: Sub-task
>  Components: HiveServer2, repl
>Affects Versions: 2.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>  Labels: DR, replication
> Fix For: 3.0.0
>
> Attachments: HIVE-17100.01.patch, HIVE-17100.02.patch, 
> HIVE-17100.03.patch, HIVE-17100.04.patch, HIVE-17100.05.patch
>
>
> It is necessary to log the progress the replication tasks in a structured 
> manner as follows.
> *+Bootstrap Dump:+*
> * At the start of bootstrap dump, will add one log with below details.
> {color:#59afe1}* Database Name
> * Dump Type (BOOTSTRAP)
> * (Estimated) Total number of tables/views to dump
> * (Estimated) Total number of functions to dump.
> * Dump Start Time{color}
> * After each table dump, will add a log as follows
> {color:#59afe1}* Table/View Name
> * Type (TABLE/VIEW/MATERIALIZED_VIEW)
> * Table dump end time
> * Table dump progress. Format is Table sequence no/(Estimated) Total number 
> of tables and views.{color}
> * After each function dump, will add a log as follows
> {color:#59afe1}* Function Name
> * Function dump end time
> * Function dump progress. Format is Function sequence no/(Estimated) Total 
> number of functions.{color}
> * After completion of all dumps, will add a log as follows to consolidate the 
> dump.
> {color:#59afe1}* Database Name.
> * Dump Type (BOOTSTRAP).
> * Dump End Time.
> * (Actual) Total number of tables/views dumped.
> * (Actual) Total number of functions dumped.
> * Dump Directory.
> * Last Repl ID of the dump.{color}
> *Note:* The actual and estimated number of tables/functions may not match if 
> any table/function is dropped when dump in progress.
> *+Bootstrap Load:+*
> * At the start of bootstrap load, will add one log with below details.
> {color:#59afe1}* Database Name
> * Dump directory
> * Load Type (BOOTSTRAP)
> * Total number of tables/views to load
> * Total number of functions to load.
> * Load Start Time{color}
> * After each table load, will add a log as follows
> {color:#59afe1}* Table/View Name
> * Type (TABLE/VIEW/MATERIALIZED_VIEW)
> * Table load completion time
> * Table load progress. Format is Table sequence no/Total number of tables and 
> views.{color}
> * After each function load, will add a log as follows
> {color:#59afe1}* Function Name
> * Function load completion time
> * Function load progress. Format is Function sequence no/Total number of 
> functions.{color}
> * After completion of all dumps, will add a log as follows to consolidate the 
> load.
> {color:#59afe1}* Database Name.
> * Load Type (BOOTSTRAP).
> * Load End Time.
> * Total number of tables/views loaded.
> * Total number of functions loaded.
> * Last Repl ID of the loaded database.{color}
> *+Incremental Dump:+*
> * At the start of database dump, will add one log with below details.
> {color:#59afe1}* Database Name
> * Dump Type (INCREMENTAL)
> * (Estimated) Total number of events to dump.
> * Dump Start Time{color}
> * After each event dump, will add a log as follows
> {color:#59afe1}* Event ID
> * Event Type (CREATE_TABLE, DROP_TABLE, ALTER_TABLE, INSERT etc)
> * Event dump end time
> * Event dump progress. Format is Event sequence no/ (Estimated) Total number 
> of events.{color}
> * After completion of all event dumps, will add a log as follows.
> {color:#59afe1}* Database Name.
> * Dump Type (INCREMENTAL).
> * Dump End Time.
> * (Actual) Total number of events dumped.
> * Dump Directory.
> * Last Repl ID of the dump.{color}
> *Note:* The estimated number of events can be terribly inaccurate with actual 
> number as we don’t have the number of events upfront until we read from 
> metastore NotificationEvents table.
> *+Incremental Load:+*
> * At the start of incremental load, will add one log with below details.
> {color:#59afe1}* Target Database Name 
> * Dump directory
> * Load Type (INCREMENTAL)
> * Total number of events to load
> * Load Start Time{color}
> * After each event load, will add a log as follows
> {color:#59afe1}* Event ID
> * Event Type (CREATE_TABLE, DROP_TABLE, ALTER_TABLE, INSERT etc)
> * Event load end time
> * Event load progress. Format is Event sequence no/ Total number of 
> events.{color}
> * After completion of all event loads, will add a log as follows to 
> consolidate the load.
> {color:#59afe1}* Target Database Name.
> * Load Type (INCREMENTAL).
> * Load End Time.
> * Total number of events loaded.
> * Last Repl ID of the loaded database.{color}



--
This message was sent by 

[jira] [Updated] (HIVE-17100) Improve HS2 operation logs for REPL commands.

2017-08-17 Thread Sankar Hariappan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17100?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sankar Hariappan updated HIVE-17100:

Attachment: (was: HIVE-17100.05.patch)

> Improve HS2 operation logs for REPL commands.
> -
>
> Key: HIVE-17100
> URL: https://issues.apache.org/jira/browse/HIVE-17100
> Project: Hive
>  Issue Type: Sub-task
>  Components: HiveServer2, repl
>Affects Versions: 2.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>  Labels: DR, replication
> Fix For: 3.0.0
>
> Attachments: HIVE-17100.01.patch, HIVE-17100.02.patch, 
> HIVE-17100.03.patch, HIVE-17100.04.patch
>
>
> It is necessary to log the progress the replication tasks in a structured 
> manner as follows.
> *+Bootstrap Dump:+*
> * At the start of bootstrap dump, will add one log with below details.
> {color:#59afe1}* Database Name
> * Dump Type (BOOTSTRAP)
> * (Estimated) Total number of tables/views to dump
> * (Estimated) Total number of functions to dump.
> * Dump Start Time{color}
> * After each table dump, will add a log as follows
> {color:#59afe1}* Table/View Name
> * Type (TABLE/VIEW/MATERIALIZED_VIEW)
> * Table dump end time
> * Table dump progress. Format is Table sequence no/(Estimated) Total number 
> of tables and views.{color}
> * After each function dump, will add a log as follows
> {color:#59afe1}* Function Name
> * Function dump end time
> * Function dump progress. Format is Function sequence no/(Estimated) Total 
> number of functions.{color}
> * After completion of all dumps, will add a log as follows to consolidate the 
> dump.
> {color:#59afe1}* Database Name.
> * Dump Type (BOOTSTRAP).
> * Dump End Time.
> * (Actual) Total number of tables/views dumped.
> * (Actual) Total number of functions dumped.
> * Dump Directory.
> * Last Repl ID of the dump.{color}
> *Note:* The actual and estimated number of tables/functions may not match if 
> any table/function is dropped when dump in progress.
> *+Bootstrap Load:+*
> * At the start of bootstrap load, will add one log with below details.
> {color:#59afe1}* Database Name
> * Dump directory
> * Load Type (BOOTSTRAP)
> * Total number of tables/views to load
> * Total number of functions to load.
> * Load Start Time{color}
> * After each table load, will add a log as follows
> {color:#59afe1}* Table/View Name
> * Type (TABLE/VIEW/MATERIALIZED_VIEW)
> * Table load completion time
> * Table load progress. Format is Table sequence no/Total number of tables and 
> views.{color}
> * After each function load, will add a log as follows
> {color:#59afe1}* Function Name
> * Function load completion time
> * Function load progress. Format is Function sequence no/Total number of 
> functions.{color}
> * After completion of all dumps, will add a log as follows to consolidate the 
> load.
> {color:#59afe1}* Database Name.
> * Load Type (BOOTSTRAP).
> * Load End Time.
> * Total number of tables/views loaded.
> * Total number of functions loaded.
> * Last Repl ID of the loaded database.{color}
> *+Incremental Dump:+*
> * At the start of database dump, will add one log with below details.
> {color:#59afe1}* Database Name
> * Dump Type (INCREMENTAL)
> * (Estimated) Total number of events to dump.
> * Dump Start Time{color}
> * After each event dump, will add a log as follows
> {color:#59afe1}* Event ID
> * Event Type (CREATE_TABLE, DROP_TABLE, ALTER_TABLE, INSERT etc)
> * Event dump end time
> * Event dump progress. Format is Event sequence no/ (Estimated) Total number 
> of events.{color}
> * After completion of all event dumps, will add a log as follows.
> {color:#59afe1}* Database Name.
> * Dump Type (INCREMENTAL).
> * Dump End Time.
> * (Actual) Total number of events dumped.
> * Dump Directory.
> * Last Repl ID of the dump.{color}
> *Note:* The estimated number of events can be terribly inaccurate with actual 
> number as we don’t have the number of events upfront until we read from 
> metastore NotificationEvents table.
> *+Incremental Load:+*
> * At the start of incremental load, will add one log with below details.
> {color:#59afe1}* Target Database Name 
> * Dump directory
> * Load Type (INCREMENTAL)
> * Total number of events to load
> * Load Start Time{color}
> * After each event load, will add a log as follows
> {color:#59afe1}* Event ID
> * Event Type (CREATE_TABLE, DROP_TABLE, ALTER_TABLE, INSERT etc)
> * Event load end time
> * Event load progress. Format is Event sequence no/ Total number of 
> events.{color}
> * After completion of all event loads, will add a log as follows to 
> consolidate the load.
> {color:#59afe1}* Target Database Name.
> * Load Type (INCREMENTAL).
> * Load End Time.
> * Total number of events loaded.
> * Last Repl ID of the loaded database.{color}



--
This message was sent by Atlassian JIRA

[jira] [Updated] (HIVE-17100) Improve HS2 operation logs for REPL commands.

2017-08-17 Thread Sankar Hariappan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17100?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sankar Hariappan updated HIVE-17100:

Status: Open  (was: Patch Available)

> Improve HS2 operation logs for REPL commands.
> -
>
> Key: HIVE-17100
> URL: https://issues.apache.org/jira/browse/HIVE-17100
> Project: Hive
>  Issue Type: Sub-task
>  Components: HiveServer2, repl
>Affects Versions: 2.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
>  Labels: DR, replication
> Fix For: 3.0.0
>
> Attachments: HIVE-17100.01.patch, HIVE-17100.02.patch, 
> HIVE-17100.03.patch, HIVE-17100.04.patch, HIVE-17100.05.patch
>
>
> It is necessary to log the progress the replication tasks in a structured 
> manner as follows.
> *+Bootstrap Dump:+*
> * At the start of bootstrap dump, will add one log with below details.
> {color:#59afe1}* Database Name
> * Dump Type (BOOTSTRAP)
> * (Estimated) Total number of tables/views to dump
> * (Estimated) Total number of functions to dump.
> * Dump Start Time{color}
> * After each table dump, will add a log as follows
> {color:#59afe1}* Table/View Name
> * Type (TABLE/VIEW/MATERIALIZED_VIEW)
> * Table dump end time
> * Table dump progress. Format is Table sequence no/(Estimated) Total number 
> of tables and views.{color}
> * After each function dump, will add a log as follows
> {color:#59afe1}* Function Name
> * Function dump end time
> * Function dump progress. Format is Function sequence no/(Estimated) Total 
> number of functions.{color}
> * After completion of all dumps, will add a log as follows to consolidate the 
> dump.
> {color:#59afe1}* Database Name.
> * Dump Type (BOOTSTRAP).
> * Dump End Time.
> * (Actual) Total number of tables/views dumped.
> * (Actual) Total number of functions dumped.
> * Dump Directory.
> * Last Repl ID of the dump.{color}
> *Note:* The actual and estimated number of tables/functions may not match if 
> any table/function is dropped when dump in progress.
> *+Bootstrap Load:+*
> * At the start of bootstrap load, will add one log with below details.
> {color:#59afe1}* Database Name
> * Dump directory
> * Load Type (BOOTSTRAP)
> * Total number of tables/views to load
> * Total number of functions to load.
> * Load Start Time{color}
> * After each table load, will add a log as follows
> {color:#59afe1}* Table/View Name
> * Type (TABLE/VIEW/MATERIALIZED_VIEW)
> * Table load completion time
> * Table load progress. Format is Table sequence no/Total number of tables and 
> views.{color}
> * After each function load, will add a log as follows
> {color:#59afe1}* Function Name
> * Function load completion time
> * Function load progress. Format is Function sequence no/Total number of 
> functions.{color}
> * After completion of all dumps, will add a log as follows to consolidate the 
> load.
> {color:#59afe1}* Database Name.
> * Load Type (BOOTSTRAP).
> * Load End Time.
> * Total number of tables/views loaded.
> * Total number of functions loaded.
> * Last Repl ID of the loaded database.{color}
> *+Incremental Dump:+*
> * At the start of database dump, will add one log with below details.
> {color:#59afe1}* Database Name
> * Dump Type (INCREMENTAL)
> * (Estimated) Total number of events to dump.
> * Dump Start Time{color}
> * After each event dump, will add a log as follows
> {color:#59afe1}* Event ID
> * Event Type (CREATE_TABLE, DROP_TABLE, ALTER_TABLE, INSERT etc)
> * Event dump end time
> * Event dump progress. Format is Event sequence no/ (Estimated) Total number 
> of events.{color}
> * After completion of all event dumps, will add a log as follows.
> {color:#59afe1}* Database Name.
> * Dump Type (INCREMENTAL).
> * Dump End Time.
> * (Actual) Total number of events dumped.
> * Dump Directory.
> * Last Repl ID of the dump.{color}
> *Note:* The estimated number of events can be terribly inaccurate with actual 
> number as we don’t have the number of events upfront until we read from 
> metastore NotificationEvents table.
> *+Incremental Load:+*
> * At the start of incremental load, will add one log with below details.
> {color:#59afe1}* Target Database Name 
> * Dump directory
> * Load Type (INCREMENTAL)
> * Total number of events to load
> * Load Start Time{color}
> * After each event load, will add a log as follows
> {color:#59afe1}* Event ID
> * Event Type (CREATE_TABLE, DROP_TABLE, ALTER_TABLE, INSERT etc)
> * Event load end time
> * Event load progress. Format is Event sequence no/ Total number of 
> events.{color}
> * After completion of all event loads, will add a log as follows to 
> consolidate the load.
> {color:#59afe1}* Target Database Name.
> * Load Type (INCREMENTAL).
> * Load End Time.
> * Total number of events loaded.
> * Last Repl ID of the loaded database.{color}



--
This message was sent by 

[jira] [Assigned] (HIVE-17319) Make BoneCp configurable using hive properties in hive-site.xml

2017-08-17 Thread Barna Zsombor Klara (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17319?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Barna Zsombor Klara reassigned HIVE-17319:
--

Assignee: Barna Zsombor Klara

> Make BoneCp configurable using hive properties in hive-site.xml
> ---
>
> Key: HIVE-17319
> URL: https://issues.apache.org/jira/browse/HIVE-17319
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Barna Zsombor Klara
>Assignee: Barna Zsombor Klara
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Work started] (HIVE-17319) Make BoneCp configurable using hive properties in hive-site.xml

2017-08-17 Thread Barna Zsombor Klara (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-17319?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HIVE-17319 started by Barna Zsombor Klara.
--
> Make BoneCp configurable using hive properties in hive-site.xml
> ---
>
> Key: HIVE-17319
> URL: https://issues.apache.org/jira/browse/HIVE-17319
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Barna Zsombor Klara
>Assignee: Barna Zsombor Klara
>




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17100) Improve HS2 operation logs for REPL commands.

2017-08-17 Thread Hive QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17100?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16130081#comment-16130081
 ] 

Hive QA commented on HIVE-17100:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12882282/HIVE-17100.05.patch

{color:red}ERROR:{color} -1 due to build exiting with an error

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/6432/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/6432/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-6432/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Tests exited with: ExecutionException: java.util.concurrent.ExecutionException: 
org.apache.hive.ptest.execution.ssh.SSHExecutionException: RSyncResult 
[localFile=/data/hiveptest/logs/PreCommit-HIVE-Build-6432/succeeded/33-TestCliDriver-ql_rewrite_gbtoidx.q-json_serde1.q-constantPropWhen.q-and-27-more,
 remoteFile=/home/hiveptest/104.155.155.39-hiveptest-1/logs/, 
getExitCode()=255, getException()=null, getUser()=hiveptest, 
getHost()=104.155.155.39, getInstance()=1]: 'Warning: Permanently added 
'104.155.155.39' (ECDSA) to the list of known hosts.
receiving incremental file list
./
TEST-33-TestCliDriver-ql_rewrite_gbtoidx.q-json_serde1.q-constantPropWhen.q-and-27-more-TEST-org.apache.hadoop.hive.cli.TestCliDriver.xml

  0   0%0.00kB/s0:00:00  
  8,686 100%8.28MB/s0:00:00 (xfr#1, to-chk=5/7)
maven-test.txt

  0   0%0.00kB/s0:00:00  
 48,240 100%1.35MB/s0:00:00 (xfr#2, to-chk=4/7)
logs/
logs/derby.log

  0   0%0.00kB/s0:00:00  
  1,004 100%   28.84kB/s0:00:00 (xfr#3, to-chk=1/7)
logs/hive.log

  0   0%0.00kB/s0:00:00  
 48,103,424   2%   45.74MB/s0:00:34  
107,479,040   6%   51.05MB/s0:00:30  
165,445,632   9%   52.45MB/s0:00:28  
225,116,160  13%   53.44MB/s0:00:26  
284,950,528  16%   56.29MB/s0:00:24  
343,638,016  20%   56.21MB/s0:00:23  
402,194,432  23%   56.36MB/s0:00:22  
461,766,656  27%   56.35MB/s0:00:21  
521,437,184  31%   56.19MB/s0:00:20  
578,584,576  34%   55.83MB/s0:00:19  
637,534,208  37%   55.91MB/s0:00:18  
696,745,984  41%   55.94MB/s0:00:17  
755,859,456  44%   55.90MB/s0:00:16  
814,743,552  48%   56.32MB/s0:00:15  
873,857,024  51%   56.37MB/s0:00:13  
933,265,408  55%   56.31MB/s0:00:12  
993,099,776  59%   56.34MB/s0:00:11  
  1,052,639,232  62%   56.38MB/s0:00:10  
  1,112,211,456  66%   56.36MB/s0:00:09  
  1,172,144,128  69%   56.46MB/s0:00:08  
  1,232,011,264  73%   56.49MB/s0:00:07  
  1,291,583,488  76%   56.50MB/s0:00:06  
  1,351,581,696  80%   56.56MB/s0:00:05  
  1,411,383,296  83%   56.56MB/s0:00:04  
  1,470,857,216  87%   56.48MB/s0:00:03  
  1,530,396,672  90%   56.47MB/s0:00:02  
  1,590,296,576  94%   56.46MB/s0:00:01  Timeout, server 104.155.155.39 not 
responding.

rsync: connection unexpectedly closed (1621310331 bytes received so far) 
[receiver]
rsync error: error in rsync protocol data stream (code 12) at io.c(226) 
[receiver=3.1.1]
rsync: connection unexpectedly closed (447 bytes received so far) [generator]
rsync error: unexplained error (code 255) at io.c(226) [generator=3.1.1]
ssh: connect to host 104.155.155.39 port 22: Connection timed out
rsync: connection unexpectedly closed (0 bytes received so far) [Receiver]
rsync error: unexplained error (code 255) at io.c(226) [Receiver=3.1.1]
ssh: connect to host 104.155.155.39 port 22: Connection timed out
rsync: connection unexpectedly closed (0 bytes received so far) [Receiver]
rsync error: unexplained error (code 255) at io.c(226) [Receiver=3.1.1]
ssh: connect to host 104.155.155.39 port 22: Connection timed out
rsync: connection unexpectedly closed (0 bytes received so far) [Receiver]
rsync error: unexplained error (code 255) at io.c(226) [Receiver=3.1.1]
ssh: connect to host 104.155.155.39 port 22: Connection timed out
rsync: connection unexpectedly closed (0 bytes received so far) [Receiver]
rsync error: unexplained error (code 255) at io.c(226) [Receiver=3.1.1]
'
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12882282 - PreCommit-HIVE-Build

> Improve HS2 operation logs for REPL commands.
> -
>
> Key: HIVE-17100
> URL: https://issues.apache.org/jira/browse/HIVE-17100
> Project: Hive
>  Issue Type: Sub-task
>  Components: HiveServer2, repl
>Affects Versions: 2.1.0
>Reporter: Sankar Hariappan
>Assignee: Sankar Hariappan
> 

[jira] [Commented] (HIVE-17313) Potentially possible 'case fall through' in the ObjectInspectorConverters

2017-08-17 Thread Zoltan Haindrich (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17313?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16130079#comment-16130079
 ] 

Zoltan Haindrich commented on HIVE-17313:
-

did it cause any known problems ?
+1
seems like the jenkins logs are already removedthe 
org.apache.hive.hcatalog.pig.TestTextFileHCatStorer.testWriteDate2  might be 
related

> Potentially possible 'case fall through' in the ObjectInspectorConverters
> -
>
> Key: HIVE-17313
> URL: https://issues.apache.org/jira/browse/HIVE-17313
> Project: Hive
>  Issue Type: Bug
>Reporter: Oleg Danilov
>Assignee: Oleg Danilov
>Priority: Trivial
> Attachments: HIVE-17313.patch
>
>
> Lines 103-110:
> {code:java}
> case STRING:
>   if (outputOI instanceof WritableStringObjectInspector) {
> return new PrimitiveObjectInspectorConverter.TextConverter(
> inputOI);
>   } else if (outputOI instanceof JavaStringObjectInspector) {
> return new PrimitiveObjectInspectorConverter.StringConverter(
> inputOI);
>   }
> case CHAR:
> {code}
> De-facto it should work correctly since outputOI is either an instance of 
> WritableStringObjectInspector or JavaStringObjectInspector, but it would be 
> better to rewrite this case to avoid possible fall through.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (HIVE-14747) Remove JAVA paths from profiles by sending them from ptest-client

2017-08-17 Thread Barna Zsombor Klara (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-14747?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Barna Zsombor Klara updated HIVE-14747:
---
Attachment: HIVE-14747.06.patch

> Remove JAVA paths from profiles by sending them from ptest-client
> -
>
> Key: HIVE-14747
> URL: https://issues.apache.org/jira/browse/HIVE-14747
> Project: Hive
>  Issue Type: Sub-task
>  Components: Hive, Testing Infrastructure
>Reporter: Sergio Peña
>Assignee: Barna Zsombor Klara
> Attachments: HIVE-14747.01.patch, HIVE-14747.02.patch, 
> HIVE-14747.03.patch, HIVE-14747.04.patch, HIVE-14747.05.patch, 
> HIVE-14747.06.patch
>
>
> Hive ptest uses some properties files per branch that contain information 
> about how to execute the tests.
> This profile includes JAVA paths to build and execute the tests. We should 
> get rid of these by passing such information from Jenkins to the 
> ptest-server. In case a profile needs a different java version, then we can 
> create a specific Jenkins job for that.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17331) Path must be used as key type of the pathToAlises

2017-08-17 Thread Zoltan Haindrich (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17331?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16130070#comment-16130070
 ] 

Zoltan Haindrich commented on HIVE-17331:
-

yes...this is really broken by HIVE-12244 - seems like this part does not have 
enough coverage...
Could you intern the path the same time you create it - or reuse the path 
instance ; this would increase readablily a bit.
note: it would seem to be a good idea to move the interning to 
{{addPathToAlias}}...it just makes it more complicated to read all these...and 
it only make sense when it is added... - possibly in a follow up

> Path must be used as key type of the pathToAlises
> -
>
> Key: HIVE-17331
> URL: https://issues.apache.org/jira/browse/HIVE-17331
> Project: Hive
>  Issue Type: Bug
>Reporter: Oleg Danilov
>Assignee: Oleg Danilov
>Priority: Minor
> Attachments: HIVE-17331.patch
>
>
> This code uses String instead of Path as key type of the pathToAliases map, 
> so seems like get(String) always null.
> +*GenMapRedUtils.java*+
> {code:java}
> for (int pos = 0; pos < size; pos++) {
>   String taskTmpDir = taskTmpDirLst.get(pos);
>   TableDesc tt_desc = tt_descLst.get(pos);
>   MapWork mWork = plan.getMapWork();
>   if (mWork.getPathToAliases().get(taskTmpDir) == null) {
> taskTmpDir = taskTmpDir.intern();
> Path taskTmpDirPath = 
> StringInternUtils.internUriStringsInPath(new Path(taskTmpDir));
> mWork.removePathToAlias(taskTmpDirPath);
> mWork.addPathToAlias(taskTmpDirPath, taskTmpDir);
> mWork.addPathToPartitionInfo(taskTmpDirPath, new 
> PartitionDesc(tt_desc, null));
> mWork.getAliasToWork().put(taskTmpDir, topOperators.get(pos));
> {code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (HIVE-17314) LazySimpleSerializeWrite.writeString() contains if with an empty body

2017-08-17 Thread Oleg Danilov (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-17314?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16130065#comment-16130065
 ] 

Oleg Danilov commented on HIVE-17314:
-

[~kgyrtkirk], uploaded.

> LazySimpleSerializeWrite.writeString() contains if with an empty body
> -
>
> Key: HIVE-17314
> URL: https://issues.apache.org/jira/browse/HIVE-17314
> Project: Hive
>  Issue Type: Bug
>Reporter: Oleg Danilov
>Assignee: Oleg Danilov
>Priority: Minor
> Attachments: HIVE-17314.patch
>
>
> Looking at the LazySimpleSerializeWrite.java I found odd 'if':
> Lines 234-235:
> {code:java}
> if (v.equals(nullSequenceBytes)) {
> }
> {code}
> Seems like either something is missed there or this 'if' could be dropped out.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


  1   2   >