[jira] [Commented] (HIVE-9618) Deduplicate RS keys for ptf/windowing
[ https://issues.apache.org/jira/browse/HIVE-9618?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14317994#comment-14317994 ] Hive QA commented on HIVE-9618: --- {color:red}Overall{color}: -1 at least one tests failed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12698317/HIVE-9618.3.patch.txt {color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 7540 tests executed *Failed tests:* {noformat} org.apache.hadoop.hive.cli.TestMinimrCliDriver.testCliDriver_schemeAuthority org.apache.hive.hcatalog.streaming.TestStreaming.testTransactionBatchEmptyCommit {noformat} Test results: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2773/testReport Console output: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2773/console Test logs: http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2773/ Messages: {noformat} Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 2 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12698317 - PreCommit-HIVE-TRUNK-Build Deduplicate RS keys for ptf/windowing - Key: HIVE-9618 URL: https://issues.apache.org/jira/browse/HIVE-9618 Project: Hive Issue Type: Improvement Components: PTF-Windowing Reporter: Navis Assignee: Navis Priority: Trivial Attachments: HIVE-9618.1.patch.txt, HIVE-9618.2.patch.txt, HIVE-9618.3.patch.txt Currently, partition spec containing same column for partition-by and order-by makes duplicated key column for RS. For example, {noformat} explain select p_mfgr, p_name, p_size, rank() over (partition by p_mfgr order by p_name) as r, dense_rank() over (partition by p_mfgr order by p_name) as dr, sum(p_retailprice) over (partition by p_mfgr order by p_name rows between unbounded preceding and current row) as s1 from noop(on noopwithmap(on noop(on part partition by p_mfgr order by p_mfgr, p_name ))) {noformat} partition by p_mfgr order by p_mfgr, p_name makes duplicated key columns like below {noformat} Reduce Output Operator key expressions: p_mfgr (type: string), p_mfgr (type: string), p_name (type: string) sort order: +++ Map-reduce partition columns: p_mfgr (type: string) value expressions: p_size (type: int), p_retailprice (type: double) {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-9618) Deduplicate RS keys for ptf/windowing
[ https://issues.apache.org/jira/browse/HIVE-9618?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14314497#comment-14314497 ] Hive QA commented on HIVE-9618: --- {color:red}Overall{color}: -1 no tests executed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12697668/HIVE-9618.2.patch.txt Test results: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2740/testReport Console output: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2740/console Test logs: http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2740/ Messages: {noformat} Executing org.apache.hive.ptest.execution.PrepPhase Tests exited with: NonZeroExitCodeException Command 'bash /data/hive-ptest/working/scratch/source-prep.sh' failed with exit status 1 and output '+ [[ -n /usr/java/jdk1.7.0_45-cloudera ]] + export JAVA_HOME=/usr/java/jdk1.7.0_45-cloudera + JAVA_HOME=/usr/java/jdk1.7.0_45-cloudera + export PATH=/usr/java/jdk1.7.0_45-cloudera/bin/:/usr/local/apache-maven-3.0.5/bin:/usr/java/jdk1.7.0_45-cloudera/bin:/usr/local/apache-ant-1.9.1/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/sbin:/home/hiveptest/bin + PATH=/usr/java/jdk1.7.0_45-cloudera/bin/:/usr/local/apache-maven-3.0.5/bin:/usr/java/jdk1.7.0_45-cloudera/bin:/usr/local/apache-ant-1.9.1/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/sbin:/home/hiveptest/bin + export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m ' + ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m ' + export 'M2_OPTS=-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost -Dhttp.proxyPort=3128' + M2_OPTS='-Xmx1g -XX:MaxPermSize=256m -Dhttp.proxyHost=localhost -Dhttp.proxyPort=3128' + cd /data/hive-ptest/working/ + tee /data/hive-ptest/logs/PreCommit-HIVE-TRUNK-Build-2740/source-prep.txt + [[ false == \t\r\u\e ]] + mkdir -p maven ivy + [[ svn = \s\v\n ]] + [[ -n '' ]] + [[ -d apache-svn-trunk-source ]] + [[ ! -d apache-svn-trunk-source/.svn ]] + [[ ! -d apache-svn-trunk-source ]] + cd apache-svn-trunk-source + svn revert -R . Reverted 'ql/src/test/org/apache/hadoop/hive/ql/parse/TestIUD.java' Reverted 'ql/src/java/org/apache/hadoop/hive/ql/parse/QBMetaData.java' Reverted 'ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java' Reverted 'ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g' Reverted 'ql/src/java/org/apache/hadoop/hive/ql/parse/QBParseInfo.java' ++ egrep -v '^X|^Performing status on external' ++ awk '{print $2}' ++ svn status --no-ignore + rm -rf target datanucleus.log ant/target shims/target shims/0.20S/target shims/0.23/target shims/aggregator/target shims/common/target shims/scheduler/target packaging/target hbase-handler/target testutils/target jdbc/target metastore/target itests/target itests/thirdparty itests/hcatalog-unit/target itests/test-serde/target itests/qtest/target itests/hive-unit-hadoop2/target itests/hive-minikdc/target itests/hive-jmh/target itests/hive-unit/target itests/custom-serde/target itests/util/target itests/qtest-spark/target hcatalog/target hcatalog/core/target hcatalog/streaming/target hcatalog/server-extensions/target hcatalog/hcatalog-pig-adapter/target hcatalog/webhcat/svr/target hcatalog/webhcat/java-client/target accumulo-handler/target hwi/target common/target common/src/gen spark-client/target contrib/target service/target serde/target beeline/target odbc/target cli/target ql/dependency-reduced-pom.xml ql/target ql/src/test/results/clientpositive/insert_into_with_schema.q.out ql/src/test/queries/clientpositive/insert_into_with_schema.q + svn update Fetching external item into 'hcatalog/src/test/e2e/harness' External at revision 1658765. At revision 1658765. + patchCommandPath=/data/hive-ptest/working/scratch/smart-apply-patch.sh + patchFilePath=/data/hive-ptest/working/scratch/build.patch + [[ -f /data/hive-ptest/working/scratch/build.patch ]] + chmod +x /data/hive-ptest/working/scratch/smart-apply-patch.sh + /data/hive-ptest/working/scratch/smart-apply-patch.sh /data/hive-ptest/working/scratch/build.patch The patch does not appear to apply with p0, p1, or p2 + exit 1 ' {noformat} This message is automatically generated. ATTACHMENT ID: 12697668 - PreCommit-HIVE-TRUNK-Build Deduplicate RS keys for ptf/windowing - Key: HIVE-9618 URL: https://issues.apache.org/jira/browse/HIVE-9618 Project: Hive Issue Type: Improvement Components: PTF-Windowing Reporter: Navis Assignee: Navis Priority: Trivial Attachments: HIVE-9618.1.patch.txt, HIVE-9618.2.patch.txt Currently, partition spec containing same column for partition-by and order-by makes duplicated key column for RS. For example, {noformat} explain select p_mfgr, p_name,
[jira] [Commented] (HIVE-9618) Deduplicate RS keys for ptf/windowing
[ https://issues.apache.org/jira/browse/HIVE-9618?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14313745#comment-14313745 ] Ashutosh Chauhan commented on HIVE-9618: +1 Deduplicate RS keys for ptf/windowing - Key: HIVE-9618 URL: https://issues.apache.org/jira/browse/HIVE-9618 Project: Hive Issue Type: Improvement Components: PTF-Windowing Reporter: Navis Assignee: Navis Priority: Trivial Attachments: HIVE-9618.1.patch.txt, HIVE-9618.2.patch.txt Currently, partition spec containing same column for partition-by and order-by makes duplicated key column for RS. For example, {noformat} explain select p_mfgr, p_name, p_size, rank() over (partition by p_mfgr order by p_name) as r, dense_rank() over (partition by p_mfgr order by p_name) as dr, sum(p_retailprice) over (partition by p_mfgr order by p_name rows between unbounded preceding and current row) as s1 from noop(on noopwithmap(on noop(on part partition by p_mfgr order by p_mfgr, p_name ))) {noformat} partition by p_mfgr order by p_mfgr, p_name makes duplicated key columns like below {noformat} Reduce Output Operator key expressions: p_mfgr (type: string), p_mfgr (type: string), p_name (type: string) sort order: +++ Map-reduce partition columns: p_mfgr (type: string) value expressions: p_size (type: int), p_retailprice (type: double) {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-9618) Deduplicate RS keys for ptf/windowing
[ https://issues.apache.org/jira/browse/HIVE-9618?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14313080#comment-14313080 ] Hive QA commented on HIVE-9618: --- {color:red}Overall{color}: -1 at least one tests failed Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12697428/HIVE-9618.1.patch.txt {color:red}ERROR:{color} -1 due to 7 failed/errored test(s), 7531 tests executed *Failed tests:* {noformat} org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_correlationoptimizer12 org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_ptf_streaming org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver_vectorized_ptf org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_ptf_streaming org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver_vectorized_ptf org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_ptf_streaming org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver_vectorized_ptf {noformat} Test results: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2726/testReport Console output: http://ec2-174-129-184-35.compute-1.amazonaws.com/jenkins/job/PreCommit-HIVE-TRUNK-Build/2726/console Test logs: http://ec2-174-129-184-35.compute-1.amazonaws.com/logs/PreCommit-HIVE-TRUNK-Build-2726/ Messages: {noformat} Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 7 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12697428 - PreCommit-HIVE-TRUNK-Build Deduplicate RS keys for ptf/windowing - Key: HIVE-9618 URL: https://issues.apache.org/jira/browse/HIVE-9618 Project: Hive Issue Type: Improvement Components: PTF-Windowing Reporter: Navis Assignee: Navis Priority: Trivial Attachments: HIVE-9618.1.patch.txt Currently, partition spec containing same column for partition-by and order-by makes duplicated key column for RS. For example, {noformat} explain select p_mfgr, p_name, p_size, rank() over (partition by p_mfgr order by p_name) as r, dense_rank() over (partition by p_mfgr order by p_name) as dr, sum(p_retailprice) over (partition by p_mfgr order by p_name rows between unbounded preceding and current row) as s1 from noop(on noopwithmap(on noop(on part partition by p_mfgr order by p_mfgr, p_name ))) {noformat} partition by p_mfgr order by p_mfgr, p_name makes duplicated key columns like below {noformat} Reduce Output Operator key expressions: p_mfgr (type: string), p_mfgr (type: string), p_name (type: string) sort order: +++ Map-reduce partition columns: p_mfgr (type: string) value expressions: p_size (type: int), p_retailprice (type: double) {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (HIVE-9618) Deduplicate RS keys for ptf/windowing
[ https://issues.apache.org/jira/browse/HIVE-9618?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14312684#comment-14312684 ] Ashutosh Chauhan commented on HIVE-9618: {code} + int index = ExprNodeDescUtils.indexOf(colDef.getExprNode(), partCols); {code} should this be instead : {code} + int index = ExprNodeDescUtils.indexOf(colDef.getExprNode(), orderCols); {code} since idea is not to add same columns twice in orderCols, right ? Deduplicate RS keys for ptf/windowing - Key: HIVE-9618 URL: https://issues.apache.org/jira/browse/HIVE-9618 Project: Hive Issue Type: Improvement Components: PTF-Windowing Reporter: Navis Assignee: Navis Priority: Trivial Attachments: HIVE-9618.1.patch.txt Currently, partition spec containing same column for partition-by and order-by makes duplicated key column for RS. For example, {noformat} explain select p_mfgr, p_name, p_size, rank() over (partition by p_mfgr order by p_name) as r, dense_rank() over (partition by p_mfgr order by p_name) as dr, sum(p_retailprice) over (partition by p_mfgr order by p_name rows between unbounded preceding and current row) as s1 from noop(on noopwithmap(on noop(on part partition by p_mfgr order by p_mfgr, p_name ))) {noformat} partition by p_mfgr order by p_mfgr, p_name makes duplicated key columns like below {noformat} Reduce Output Operator key expressions: p_mfgr (type: string), p_mfgr (type: string), p_name (type: string) sort order: +++ Map-reduce partition columns: p_mfgr (type: string) value expressions: p_size (type: int), p_retailprice (type: double) {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332)