[ https://issues.apache.org/jira/browse/HIVE-21382?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16790224#comment-16790224 ]
Hive QA commented on HIVE-21382: -------------------------------- Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12962042/HIVE-21382.3.patch {color:red}ERROR:{color} -1 due to build exiting with an error Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/16458/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/16458/console Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-16458/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Executing org.apache.hive.ptest.execution.PrepPhase Tests exited with: NonZeroExitCodeException Command 'bash /data/hiveptest/working/scratch/source-prep.sh' failed with exit status 1 and output '+ date '+%Y-%m-%d %T.%3N' 2019-03-12 04:32:12.577 + [[ -n /usr/lib/jvm/java-8-openjdk-amd64 ]] + export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64 + JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64 + export PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games + PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games + export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m ' + ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m ' + export 'MAVEN_OPTS=-Xmx1g ' + MAVEN_OPTS='-Xmx1g ' + cd /data/hiveptest/working/ + tee /data/hiveptest/logs/PreCommit-HIVE-Build-16458/source-prep.txt + [[ false == \t\r\u\e ]] + mkdir -p maven ivy + [[ git = \s\v\n ]] + [[ git = \g\i\t ]] + [[ -z master ]] + [[ -d apache-github-source-source ]] + [[ ! -d apache-github-source-source/.git ]] + [[ ! -d apache-github-source-source ]] + date '+%Y-%m-%d %T.%3N' 2019-03-12 04:32:12.580 + cd apache-github-source-source + git fetch origin + git reset --hard HEAD HEAD is now at 9f2f101 HIVE-21388: Constant UDF is not pushed to JDBCStorage Handler (Jesus Camacho Rodriguez, reviewed by Jason Dere) + git clean -f -d Removing standalone-metastore/metastore-server/src/gen/ + git checkout master Already on 'master' Your branch is up-to-date with 'origin/master'. + git reset --hard origin/master HEAD is now at 9f2f101 HIVE-21388: Constant UDF is not pushed to JDBCStorage Handler (Jesus Camacho Rodriguez, reviewed by Jason Dere) + git merge --ff-only origin/master Already up-to-date. + date '+%Y-%m-%d %T.%3N' 2019-03-12 04:32:13.469 + rm -rf ../yetus_PreCommit-HIVE-Build-16458 + mkdir ../yetus_PreCommit-HIVE-Build-16458 + git gc + cp -R . ../yetus_PreCommit-HIVE-Build-16458 + mkdir /data/hiveptest/logs/PreCommit-HIVE-Build-16458/yetus + patchCommandPath=/data/hiveptest/working/scratch/smart-apply-patch.sh + patchFilePath=/data/hiveptest/working/scratch/build.patch + [[ -f /data/hiveptest/working/scratch/build.patch ]] + chmod +x /data/hiveptest/working/scratch/smart-apply-patch.sh + /data/hiveptest/working/scratch/smart-apply-patch.sh /data/hiveptest/working/scratch/build.patch error: a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/rules/HiveRelFieldTrimmer.java: does not exist in index error: a/ql/src/test/queries/clientpositive/constraints_optimization.q: does not exist in index error: a/ql/src/test/results/clientpositive/llap/constraints_optimization.q.out: does not exist in index error: a/ql/src/test/results/clientpositive/perf/tez/constraints/cbo_query23.q.out: does not exist in index error: a/ql/src/test/results/clientpositive/perf/tez/constraints/cbo_query39.q.out: does not exist in index error: a/ql/src/test/results/clientpositive/perf/tez/constraints/cbo_query64.q.out: does not exist in index error: a/ql/src/test/results/clientpositive/perf/tez/constraints/query23.q.out: does not exist in index error: a/ql/src/test/results/clientpositive/perf/tez/constraints/query39.q.out: does not exist in index error: a/ql/src/test/results/clientpositive/perf/tez/constraints/query64.q.out: does not exist in index Going to apply patch with: git apply -p1 + [[ maven == \m\a\v\e\n ]] + rm -rf /data/hiveptest/working/maven/org/apache/hive + mvn -B clean install -DskipTests -T 4 -q -Dmaven.repo.local=/data/hiveptest/working/maven protoc-jar: executing: [/tmp/protoc5885607783622954764.exe, --version] protoc-jar: executing: [/tmp/protoc5885607783622954764.exe, -I/data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-common/src/main/protobuf/org/apache/hadoop/hive/metastore, --java_out=/data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-common/target/generated-sources, /data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-common/src/main/protobuf/org/apache/hadoop/hive/metastore/metastore.proto] libprotoc 2.5.0 ANTLR Parser Generator Version 3.5.2 [ERROR] Failed to execute goal org.apache.maven.plugins:maven-remote-resources-plugin:1.5:process (process-resource-bundles) on project hive-shims-0.23: Execution process-resource-bundles of goal org.apache.maven.plugins:maven-remote-resources-plugin:1.5:process failed. ConcurrentModificationException -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/PluginExecutionException [ERROR] [ERROR] After correcting the problems, you can resume the build with the command [ERROR] mvn <goals> -rf :hive-shims-0.23 + result=1 + '[' 1 -ne 0 ']' + rm -rf yetus_PreCommit-HIVE-Build-16458 + exit 1 ' {noformat} This message is automatically generated. ATTACHMENT ID: 12962042 - PreCommit-HIVE-Build > Group by keys reduction optimization - keys are not reduced in query23 > ---------------------------------------------------------------------- > > Key: HIVE-21382 > URL: https://issues.apache.org/jira/browse/HIVE-21382 > Project: Hive > Issue Type: Improvement > Reporter: Vineet Garg > Assignee: Vineet Garg > Priority: Major > Attachments: HIVE-21382.1.patch, HIVE-21382.2.patch, > HIVE-21382.2.patch, HIVE-21382.3.patch > > > {code:sql} > explain cbo with frequent_ss_items as > (select substr(i_item_desc,1,30) itemdesc,i_item_sk item_sk,d_date > solddate,count(*) cnt > from store_sales > ,date_dim > ,item > where ss_sold_date_sk = d_date_sk > and ss_item_sk = i_item_sk > and d_year in (1999,1999+1,1999+2,1999+3) > group by substr(i_item_desc,1,30),i_item_sk,d_date > having count(*) >4) > select sum(sales) > from ((select cs_quantity*cs_list_price sales > from catalog_sales > ,date_dim > where d_year = 1999 > and d_moy = 1 > and cs_sold_date_sk = d_date_sk > and cs_item_sk in (select item_sk from frequent_ss_items))) subq > limit 100; > {code} > {code:sql} > HiveSortLimit(fetch=[100]) > HiveProject($f0=[$0]) > HiveAggregate(group=[{}], agg#0=[sum($0)]) > HiveProject(sales=[*(CAST($2):DECIMAL(10, 0), $3)]) > HiveSemiJoin(condition=[=($1, $5)], joinType=[inner]) > HiveJoin(condition=[=($0, $4)], joinType=[inner], algorithm=[none], > cost=[{2.0 rows, 0.0 cpu, 0.0 io}]) > HiveProject(cs_sold_date_sk=[$0], cs_item_sk=[$15], > cs_quantity=[$18], cs_list_price=[$20]) > HiveFilter(condition=[IS NOT NULL($0)]) > HiveTableScan(table=[[perf_constraints, catalog_sales]], > table:alias=[catalog_sales]) > HiveProject(d_date_sk=[$0]) > HiveFilter(condition=[AND(=($6, 1999), =($8, 1))]) > HiveTableScan(table=[[perf_constraints, date_dim]], > table:alias=[date_dim]) > HiveProject(i_item_sk=[$1]) > HiveFilter(condition=[>($3, 4)]) > HiveProject(substr=[$2], i_item_sk=[$1], d_date=[$0], $f3=[$3]) > HiveAggregate(group=[{3, 4, 5}], agg#0=[count()]) > HiveJoin(condition=[=($1, $4)], joinType=[inner], > algorithm=[none], cost=[{2.0 rows, 0.0 cpu, 0.0 io}]) > HiveJoin(condition=[=($0, $2)], joinType=[inner], > algorithm=[none], cost=[{2.0 rows, 0.0 cpu, 0.0 io}]) > HiveProject(ss_sold_date_sk=[$0], ss_item_sk=[$2]) > HiveFilter(condition=[IS NOT NULL($0)]) > HiveTableScan(table=[[perf_constraints, > store_sales]], table:alias=[store_sales]) > HiveProject(d_date_sk=[$0], d_date=[$2]) > HiveFilter(condition=[IN($6, 1999, 2000, 2001, 2002)]) > HiveTableScan(table=[[perf_constraints, date_dim]], > table:alias=[date_dim]) > HiveProject(i_item_sk=[$0], substr=[substr($4, 1, 30)]) > HiveTableScan(table=[[perf_constraints, item]], > table:alias=[item]) > {code} > Right side of HiveSemiJoin has an aggregate which could be reduce to have > only {{i_item_sk}} as group by key since {{i_item_sk}} is primary key. -- This message was sent by Atlassian JIRA (v7.6.3#76005)