[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16664095#comment-16664095 ] Hive QA commented on HIVE-18778: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12945558/HIVE-18778.3.branch-2.patch {color:red}ERROR:{color} -1 due to no test(s) being added or modified. {color:red}ERROR:{color} -1 due to 14 failed/errored test(s), 10705 tests executed *Failed tests:* {noformat} org.apache.hadoop.hive.cli.TestAccumuloCliDriver.testCliDriver[accumulo_queries] (batchId=227) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[avro_tableproperty_optimize] (batchId=22) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[explaindenpendencydiffengs] (batchId=38) org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[llap_smb] (batchId=142) org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[orc_ppd_basic] (batchId=139) org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[table_nonprintable] (batchId=140) org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[join_acid_non_acid] (batchId=158) org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[union_fast_stats] (batchId=153) org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[vectorized_parquet_types] (batchId=155) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[merge_negative_5] (batchId=88) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[explaindenpendencydiffengs] (batchId=115) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[vectorization_input_format_excludes] (batchId=117) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[vectorized_ptf] (batchId=125) org.apache.hive.hcatalog.api.TestHCatClient.testTransportFailure (batchId=176) {noformat} Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/14640/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/14640/console Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-14640/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.YetusPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 14 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12945558 - PreCommit-HIVE-Build > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Fix For: 4.0.0, 3.2.0, 3.1.1, 2.3.4 > > Attachments: HIVE-18778-SparkPositive.patch, > HIVE-18778.1.branch-2.patch, HIVE-18778.1.patch, > HIVE-18778.10.branch-3.patch, HIVE-18778.11.branch-3.1.patch, > HIVE-18778.11.branch-3.patch, HIVE-18778.12.branch-3.1.patch, > HIVE-18778.2.branch-2.patch, HIVE-18778.2.patch, HIVE-18778.3.branch-2.patch, > HIVE-18778.3.patch, HIVE-18778.4.patch, HIVE-18778.5.patch, > HIVE-18778.6.patch, HIVE-18778.7.patch, HIVE-18778.8.patch, > HIVE-18778.9.branch-3.patch, HIVE-18778.9.patch, > HIVE-18778_TestCliDriver.patch, HIVE-18788_SparkNegative.patch, > HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16663285#comment-16663285 ] Hive QA commented on HIVE-18778: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12945502/HIVE-18778.2.branch-2.patch {color:red}ERROR:{color} -1 due to build exiting with an error Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/14635/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/14635/console Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-14635/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Executing org.apache.hive.ptest.execution.PrepPhase Tests exited with: NonZeroExitCodeException Command 'bash /data/hiveptest/working/scratch/source-prep.sh' failed with exit status 1 and output '+ date '+%Y-%m-%d %T.%3N' 2018-10-25 05:58:13.957 + [[ -n /usr/lib/jvm/java-8-openjdk-amd64 ]] + export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64 + JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64 + export PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games + PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games + export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m ' + ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m ' + export 'MAVEN_OPTS=-Xmx1g ' + MAVEN_OPTS='-Xmx1g ' + cd /data/hiveptest/working/ + tee /data/hiveptest/logs/PreCommit-HIVE-Build-14635/source-prep.txt + [[ false == \t\r\u\e ]] + mkdir -p maven ivy + [[ git = \s\v\n ]] + [[ git = \g\i\t ]] + [[ -z branch-2 ]] + [[ -d apache-github-branch-2-source ]] + [[ ! -d apache-github-branch-2-source/.git ]] + [[ ! -d apache-github-branch-2-source ]] + date '+%Y-%m-%d %T.%3N' 2018-10-25 05:58:13.998 + cd apache-github-branch-2-source + git fetch origin >From https://github.com/apache/hive a457012..2c8422b branch-2 -> origin/branch-2 db72a86..7fecb6f branch-1 -> origin/branch-1 2fe5186..12d2b2d branch-2.3 -> origin/branch-2.3 8703a32..09b92d3 branch-3 -> origin/branch-3 c39b5d1..f4e0529 branch-3.1 -> origin/branch-3.1 37c7fd7..a99be34 master -> origin/master * [new tag] release-3.1.1-rc0 -> release-3.1.1-rc0 + git reset --hard HEAD HEAD is now at a457012 HIVE-18767: Some alterPartitions invocations throw 'NumberFormatException: null' (Mass Dosage, reviewed by Peter Vary) + git clean -f -d + git checkout branch-2 Already on 'branch-2' Your branch is behind 'origin/branch-2' by 1 commit, and can be fast-forwarded. (use "git pull" to update your local branch) + git reset --hard origin/branch-2 HEAD is now at 2c8422b HIVE-20420: Provide a fallback authorizer when no other authorizer is in use (Daniel Dai, reviewed by Laszlo Pinter, Thejas Nair) + git merge --ff-only origin/branch-2 Already up-to-date. + date '+%Y-%m-%d %T.%3N' 2018-10-25 05:58:22.206 + rm -rf ../yetus_PreCommit-HIVE-Build-14635 + mkdir ../yetus_PreCommit-HIVE-Build-14635 + git gc + cp -R . ../yetus_PreCommit-HIVE-Build-14635 + mkdir /data/hiveptest/logs/PreCommit-HIVE-Build-14635/yetus + patchCommandPath=/data/hiveptest/working/scratch/smart-apply-patch.sh + patchFilePath=/data/hiveptest/working/scratch/build.patch + [[ -f /data/hiveptest/working/scratch/build.patch ]] + chmod +x /data/hiveptest/working/scratch/smart-apply-patch.sh + /data/hiveptest/working/scratch/smart-apply-patch.sh /data/hiveptest/working/scratch/build.patch error: a/ql/src/java/org/apache/hadoop/hive/ql/hooks/EnforceReadOnlyTables.java: does not exist in index error: a/ql/src/java/org/apache/hadoop/hive/ql/parse/ExplainSemanticAnalyzer.java: does not exist in index Going to apply patch with: git apply -p1 + [[ maven == \m\a\v\e\n ]] + rm -rf /data/hiveptest/working/maven/org/apache/hive + mvn -B clean install -DskipTests -T 4 -q -Dmaven.repo.local=/data/hiveptest/working/maven ANTLR Parser Generator Version 3.5.2 Output file /data/hiveptest/working/apache-github-branch-2-source/metastore/target/generated-sources/antlr3/org/apache/hadoop/hive/metastore/parser/FilterParser.java does not exist: must build /data/hiveptest/working/apache-github-branch-2-source/metastore/src/java/org/apache/hadoop/hive/metastore/parser/Filter.g org/apache/hadoop/hive/metastore/parser/Filter.g DataNucleus Enhancer (version 4.1.17) for API "JDO" DataNucleus Enhancer : Classpath >> /usr/share/maven/boot/plexus-classworlds-2.x.jar ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MDatabase ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MFieldSchema ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MType ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MTable ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MConstraint ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MSerDeInfo ENHANCED (Persistable)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16661281#comment-16661281 ] Daniel Dai commented on HIVE-18778: --- Further make a patch for branch-2. This takes a shortcut to avoid mass golden file updates. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Fix For: 4.0.0, 3.2.0, 3.1.1 > > Attachments: HIVE-18778-SparkPositive.patch, > HIVE-18778.1.branch-2.patch, HIVE-18778.1.patch, > HIVE-18778.10.branch-3.patch, HIVE-18778.11.branch-3.1.patch, > HIVE-18778.11.branch-3.patch, HIVE-18778.12.branch-3.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778.4.patch, > HIVE-18778.5.patch, HIVE-18778.6.patch, HIVE-18778.7.patch, > HIVE-18778.8.patch, HIVE-18778.9.branch-3.patch, HIVE-18778.9.patch, > HIVE-18778_TestCliDriver.patch, HIVE-18788_SparkNegative.patch, > HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16633121#comment-16633121 ] Daniel Dai commented on HIVE-18778: --- Committed to branch-3.1. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.10.branch-3.patch, HIVE-18778.11.branch-3.1.patch, > HIVE-18778.11.branch-3.patch, HIVE-18778.12.branch-3.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778.4.patch, > HIVE-18778.5.patch, HIVE-18778.6.patch, HIVE-18778.7.patch, > HIVE-18778.8.patch, HIVE-18778.9.branch-3.patch, HIVE-18778.9.patch, > HIVE-18778_TestCliDriver.patch, HIVE-18788_SparkNegative.patch, > HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16633037#comment-16633037 ] Hive QA commented on HIVE-18778: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12941769/HIVE-18778.12.branch-3.1.patch {color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified. {color:red}ERROR:{color} -1 due to 103 failed/errored test(s), 14379 tests executed *Failed tests:* {noformat} TestAddPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestAddPartitionsFromPartSpec - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestAdminUser - did not produce a TEST-*.xml file (likely timed out) (batchId=234) TestAggregateStatsCache - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestAlterPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestAppendPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestBeeLineDriver - did not produce a TEST-*.xml file (likely timed out) (batchId=271) TestCachedStore - did not produce a TEST-*.xml file (likely timed out) (batchId=234) TestCatalogCaching - did not produce a TEST-*.xml file (likely timed out) (batchId=234) TestCatalogNonDefaultClient - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestCatalogNonDefaultSvr - did not produce a TEST-*.xml file (likely timed out) (batchId=234) TestCatalogOldClient - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestCatalogs - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestCheckConstraint - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestDataSourceProviderFactory - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestDatabaseName - did not produce a TEST-*.xml file (likely timed out) (batchId=195) TestDatabases - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestDeadline - did not produce a TEST-*.xml file (likely timed out) (batchId=234) TestDefaultConstraint - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestDropPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestDummy - did not produce a TEST-*.xml file (likely timed out) (batchId=271) TestEmbeddedHiveMetaStore - did not produce a TEST-*.xml file (likely timed out) (batchId=229) TestExchangePartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestFMSketchSerialization - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestFilterHooks - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestForeignKey - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestFunctions - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestGetPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestGetTableMeta - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestHLLNoBias - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestHLLSerialization - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestHdfsUtils - did not produce a TEST-*.xml file (likely timed out) (batchId=234) TestHiveAlterHandler - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestHiveMetaStoreGetMetaConf - did not produce a TEST-*.xml file (likely timed out) (batchId=234) TestHiveMetaStorePartitionSpecs - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestHiveMetaStoreSchemaMethods - did not produce a TEST-*.xml file (likely timed out) (batchId=234) TestHiveMetaStoreTimeout - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestHiveMetaStoreTxns - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestHiveMetaStoreWithEnvironmentContext - did not produce a TEST-*.xml file (likely timed out) (batchId=231) TestHiveMetastoreCli - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestHyperLogLog - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestHyperLogLogDense - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestHyperLogLogMerge - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestHyperLogLogSparse - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestJSONMessageDeserializer - did not produce a TEST-*.xml file (likely timed out) (batchId=234) TestListPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestLockRequestBuilder - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestMarkPartition - did not produce a TEST-*.xml file (likely timed out) (batchId=234) TestMarkPartitionRemote - did not produce a TEST-*.xml file (likely timed
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16633017#comment-16633017 ] Hive QA commented on HIVE-18778: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:red}-1{color} | {color:red} patch {color} | {color:red} 0m 27s{color} | {color:red} /data/hiveptest/logs/PreCommit-HIVE-Build-14126/patches/PreCommit-HIVE-Build-14126.patch does not apply to master. Rebase required? Wrong Branch? See http://cwiki.apache.org/confluence/display/Hive/HowToContribute for help. {color} | \\ \\ || Subsystem || Report/Notes || | Console output | http://104.198.109.242/logs//PreCommit-HIVE-Build-14126/yetus.txt | | Powered by | Apache Yetushttp://yetus.apache.org | This message was automatically generated. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.10.branch-3.patch, HIVE-18778.11.branch-3.1.patch, > HIVE-18778.11.branch-3.patch, HIVE-18778.12.branch-3.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778.4.patch, > HIVE-18778.5.patch, HIVE-18778.6.patch, HIVE-18778.7.patch, > HIVE-18778.8.patch, HIVE-18778.9.branch-3.patch, HIVE-18778.9.patch, > HIVE-18778_TestCliDriver.patch, HIVE-18788_SparkNegative.patch, > HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16632115#comment-16632115 ] Hive QA commented on HIVE-18778: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12941574/HIVE-18778.11.branch-3.1.patch {color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified. {color:red}ERROR:{color} -1 due to 286 failed/errored test(s), 14379 tests executed *Failed tests:* {noformat} TestAddPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestAddPartitionsFromPartSpec - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestAdminUser - did not produce a TEST-*.xml file (likely timed out) (batchId=234) TestAggregateStatsCache - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestAlterPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestAppendPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestBeeLineDriver - did not produce a TEST-*.xml file (likely timed out) (batchId=271) TestCachedStore - did not produce a TEST-*.xml file (likely timed out) (batchId=234) TestCatalogCaching - did not produce a TEST-*.xml file (likely timed out) (batchId=234) TestCatalogNonDefaultClient - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestCatalogNonDefaultSvr - did not produce a TEST-*.xml file (likely timed out) (batchId=234) TestCatalogOldClient - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestCatalogs - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestCheckConstraint - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestDataSourceProviderFactory - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestDatabaseName - did not produce a TEST-*.xml file (likely timed out) (batchId=195) TestDatabases - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestDeadline - did not produce a TEST-*.xml file (likely timed out) (batchId=234) TestDefaultConstraint - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestDropPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestDummy - did not produce a TEST-*.xml file (likely timed out) (batchId=271) TestEmbeddedHiveMetaStore - did not produce a TEST-*.xml file (likely timed out) (batchId=229) TestExchangePartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestFMSketchSerialization - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestFilterHooks - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestForeignKey - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestFunctions - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestGetPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestGetTableMeta - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestHLLNoBias - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestHLLSerialization - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestHdfsUtils - did not produce a TEST-*.xml file (likely timed out) (batchId=234) TestHiveAlterHandler - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestHiveMetaStoreGetMetaConf - did not produce a TEST-*.xml file (likely timed out) (batchId=234) TestHiveMetaStorePartitionSpecs - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestHiveMetaStoreSchemaMethods - did not produce a TEST-*.xml file (likely timed out) (batchId=234) TestHiveMetaStoreTimeout - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestHiveMetaStoreTxns - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestHiveMetaStoreWithEnvironmentContext - did not produce a TEST-*.xml file (likely timed out) (batchId=231) TestHiveMetastoreCli - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestHyperLogLog - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestHyperLogLogDense - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestHyperLogLogMerge - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestHyperLogLogSparse - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestJSONMessageDeserializer - did not produce a TEST-*.xml file (likely timed out) (batchId=234) TestListPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestLockRequestBuilder - did not produce a TEST-*.xml file (likely timed out) (batchId=226) TestMarkPartition - did not produce a TEST-*.xml file (likely timed out) (batchId=234) TestMarkPartitionRemote - did not produce a TEST-*.xml file (likely timed
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16632013#comment-16632013 ] Hive QA commented on HIVE-18778: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:red}-1{color} | {color:red} patch {color} | {color:red} 0m 20s{color} | {color:red} /data/hiveptest/logs/PreCommit-HIVE-Build-14103/patches/PreCommit-HIVE-Build-14103.patch does not apply to master. Rebase required? Wrong Branch? See http://cwiki.apache.org/confluence/display/Hive/HowToContribute for help. {color} | \\ \\ || Subsystem || Report/Notes || | Console output | http://104.198.109.242/logs//PreCommit-HIVE-Build-14103/yetus.txt | | Powered by | Apache Yetushttp://yetus.apache.org | This message was automatically generated. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.10.branch-3.patch, HIVE-18778.11.branch-3.1.patch, > HIVE-18778.11.branch-3.patch, HIVE-18778.2.patch, HIVE-18778.3.patch, > HIVE-18778.4.patch, HIVE-18778.5.patch, HIVE-18778.6.patch, > HIVE-18778.7.patch, HIVE-18778.8.patch, HIVE-18778.9.branch-3.patch, > HIVE-18778.9.patch, HIVE-18778_TestCliDriver.patch, > HIVE-18788_SparkNegative.patch, HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16630849#comment-16630849 ] Daniel Dai commented on HIVE-18778: --- sysdb failure is caused by HIVE-20423, mm_all failure is there for a while. Will fix separately. Committed to branch-3. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.10.branch-3.patch, HIVE-18778.11.branch-3.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778.4.patch, > HIVE-18778.5.patch, HIVE-18778.6.patch, HIVE-18778.7.patch, > HIVE-18778.8.patch, HIVE-18778.9.branch-3.patch, HIVE-18778.9.patch, > HIVE-18778_TestCliDriver.patch, HIVE-18788_SparkNegative.patch, > HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16630481#comment-16630481 ] Hive QA commented on HIVE-18778: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12941436/HIVE-18778.11.branch-3.patch {color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified. {color:red}ERROR:{color} -1 due to 99 failed/errored test(s), 14451 tests executed *Failed tests:* {noformat} TestAddPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestAddPartitionsFromPartSpec - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestAdminUser - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestAggregateStatsCache - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestAlterPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestAppendPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestBeeLineDriver - did not produce a TEST-*.xml file (likely timed out) (batchId=273) TestCachedStore - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestCatalogCaching - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestCatalogNonDefaultClient - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestCatalogNonDefaultSvr - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestCatalogOldClient - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestCatalogs - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestCheckConstraint - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestDataSourceProviderFactory - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestDatabases - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestDeadline - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestDefaultConstraint - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestDropPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestDummy - did not produce a TEST-*.xml file (likely timed out) (batchId=273) TestEmbeddedHiveMetaStore - did not produce a TEST-*.xml file (likely timed out) (batchId=231) TestExchangePartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestFMSketchSerialization - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestFilterHooks - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestForeignKey - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestFunctions - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestGetPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestGetTableMeta - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestHLLNoBias - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestHLLSerialization - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestHdfsUtils - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestHiveAlterHandler - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestHiveMetaStoreGetMetaConf - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestHiveMetaStorePartitionSpecs - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestHiveMetaStoreSchemaMethods - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestHiveMetaStoreTimeout - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestHiveMetaStoreTxns - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestHiveMetaStoreWithEnvironmentContext - did not produce a TEST-*.xml file (likely timed out) (batchId=233) TestHiveMetastoreCli - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestHyperLogLog - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestHyperLogLogDense - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestHyperLogLogMerge - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestHyperLogLogSparse - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestJSONMessageDeserializer - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestListPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestLockRequestBuilder - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestMarkPartition - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestMarkPartitionRemote - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestMetaStoreConnectionUrlHook - did not produce a TEST-*.xml file
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16630363#comment-16630363 ] Hive QA commented on HIVE-18778: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:red}-1{color} | {color:red} patch {color} | {color:red} 0m 16s{color} | {color:red} /data/hiveptest/logs/PreCommit-HIVE-Build-14080/patches/PreCommit-HIVE-Build-14080.patch does not apply to master. Rebase required? Wrong Branch? See http://cwiki.apache.org/confluence/display/Hive/HowToContribute for help. {color} | \\ \\ || Subsystem || Report/Notes || | Console output | http://104.198.109.242/logs//PreCommit-HIVE-Build-14080/yetus.txt | | Powered by | Apache Yetushttp://yetus.apache.org | This message was automatically generated. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.10.branch-3.patch, HIVE-18778.11.branch-3.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778.4.patch, > HIVE-18778.5.patch, HIVE-18778.6.patch, HIVE-18778.7.patch, > HIVE-18778.8.patch, HIVE-18778.9.branch-3.patch, HIVE-18778.9.patch, > HIVE-18778_TestCliDriver.patch, HIVE-18788_SparkNegative.patch, > HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16628319#comment-16628319 ] Hive QA commented on HIVE-18778: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12941303/HIVE-18778.10.branch-3.patch {color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified. {color:red}ERROR:{color} -1 due to 109 failed/errored test(s), 14451 tests executed *Failed tests:* {noformat} TestAddPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestAddPartitionsFromPartSpec - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestAdminUser - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestAggregateStatsCache - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestAlterPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestAppendPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestBeeLineDriver - did not produce a TEST-*.xml file (likely timed out) (batchId=273) TestCachedStore - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestCatalogCaching - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestCatalogNonDefaultClient - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestCatalogNonDefaultSvr - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestCatalogOldClient - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestCatalogs - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestCheckConstraint - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestDataSourceProviderFactory - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestDatabases - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestDeadline - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestDefaultConstraint - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestDropPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestDummy - did not produce a TEST-*.xml file (likely timed out) (batchId=273) TestEmbeddedHiveMetaStore - did not produce a TEST-*.xml file (likely timed out) (batchId=231) TestExchangePartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestFMSketchSerialization - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestFilterHooks - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestForeignKey - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestFunctions - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestGetPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestGetTableMeta - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestHLLNoBias - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestHLLSerialization - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestHdfsUtils - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestHiveAlterHandler - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestHiveMetaStoreGetMetaConf - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestHiveMetaStorePartitionSpecs - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestHiveMetaStoreSchemaMethods - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestHiveMetaStoreTimeout - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestHiveMetaStoreTxns - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestHiveMetaStoreWithEnvironmentContext - did not produce a TEST-*.xml file (likely timed out) (batchId=233) TestHiveMetastoreCli - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestHyperLogLog - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestHyperLogLogDense - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestHyperLogLogMerge - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestHyperLogLogSparse - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestJSONMessageDeserializer - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestListPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestLockRequestBuilder - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestMarkPartition - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestMarkPartitionRemote - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestMetaStoreConnectionUrlHook - did not produce a TEST-*.xml file
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16626649#comment-16626649 ] Hive QA commented on HIVE-18778: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12941078/HIVE-18778.9.branch-3.patch {color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified. {color:red}ERROR:{color} -1 due to 199 failed/errored test(s), 14434 tests executed *Failed tests:* {noformat} TestAddPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestAddPartitionsFromPartSpec - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestAdminUser - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestAggregateStatsCache - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestAlterPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestAlterTableMetadata - did not produce a TEST-*.xml file (likely timed out) (batchId=252) TestAppendPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestAutoPurgeTables - did not produce a TEST-*.xml file (likely timed out) (batchId=252) TestBeeLineDriver - did not produce a TEST-*.xml file (likely timed out) (batchId=273) TestCachedStore - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestCatalogCaching - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestCatalogNonDefaultClient - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestCatalogNonDefaultSvr - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestCatalogOldClient - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestCatalogs - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestCheckConstraint - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestDataSourceProviderFactory - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestDatabases - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestDeadline - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestDefaultConstraint - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestDropPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestDummy - did not produce a TEST-*.xml file (likely timed out) (batchId=273) TestEmbeddedHiveMetaStore - did not produce a TEST-*.xml file (likely timed out) (batchId=231) TestExchangePartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestFMSketchSerialization - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestFilterHooks - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestForeignKey - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestFunctions - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestGetPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestGetTableMeta - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestHLLNoBias - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestHLLSerialization - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestHdfsUtils - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestHiveAlterHandler - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestHiveMetaStoreGetMetaConf - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestHiveMetaStorePartitionSpecs - did not produce a TEST-*.xml file (likely timed out) (batchId=230) TestHiveMetaStoreSchemaMethods - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestHiveMetaStoreTimeout - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestHiveMetaStoreTxns - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestHiveMetaStoreWithEnvironmentContext - did not produce a TEST-*.xml file (likely timed out) (batchId=233) TestHiveMetastoreCli - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestHyperLogLog - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestHyperLogLogDense - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestHyperLogLogMerge - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestHyperLogLogSparse - did not produce a TEST-*.xml file (likely timed out) (batchId=238) TestJSONMessageDeserializer - did not produce a TEST-*.xml file (likely timed out) (batchId=236) TestListPartitions - did not produce a TEST-*.xml file (likely timed out) (batchId=228) TestLocationQueries - did not produce a TEST-*.xml file (likely timed out) (batchId=252) TestLockRequestBuilder - did not produce a TEST-*.xml file (likely timed
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16626586#comment-16626586 ] Hive QA commented on HIVE-18778: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:red}-1{color} | {color:red} patch {color} | {color:red} 0m 15s{color} | {color:red} /data/hiveptest/logs/PreCommit-HIVE-Build-14027/patches/PreCommit-HIVE-Build-14027.patch does not apply to master. Rebase required? Wrong Branch? See http://cwiki.apache.org/confluence/display/Hive/HowToContribute for help. {color} | \\ \\ || Subsystem || Report/Notes || | Console output | http://104.198.109.242/logs//PreCommit-HIVE-Build-14027/yetus.txt | | Powered by | Apache Yetushttp://yetus.apache.org | This message was automatically generated. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778.4.patch, > HIVE-18778.5.patch, HIVE-18778.6.patch, HIVE-18778.7.patch, > HIVE-18778.8.patch, HIVE-18778.9.branch-3.patch, HIVE-18778.9.patch, > HIVE-18778_TestCliDriver.patch, HIVE-18788_SparkNegative.patch, > HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16626099#comment-16626099 ] Daniel Dai commented on HIVE-18778: --- Committed to master. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778.4.patch, > HIVE-18778.5.patch, HIVE-18778.6.patch, HIVE-18778.7.patch, > HIVE-18778.8.patch, HIVE-18778.9.patch, HIVE-18778_TestCliDriver.patch, > HIVE-18788_SparkNegative.patch, HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16625733#comment-16625733 ] Thejas M Nair commented on HIVE-18778: -- +1 > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778.4.patch, > HIVE-18778.5.patch, HIVE-18778.6.patch, HIVE-18778.7.patch, > HIVE-18778.8.patch, HIVE-18778.9.patch, HIVE-18778_TestCliDriver.patch, > HIVE-18788_SparkNegative.patch, HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16625407#comment-16625407 ] Hive QA commented on HIVE-18778: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12941000/HIVE-18778.9.patch {color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified. {color:green}SUCCESS:{color} +1 due to 14995 tests passed Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/14010/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/14010/console Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-14010/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.YetusPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase {noformat} This message is automatically generated. ATTACHMENT ID: 12941000 - PreCommit-HIVE-Build > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778.4.patch, > HIVE-18778.5.patch, HIVE-18778.6.patch, HIVE-18778.7.patch, > HIVE-18778.8.patch, HIVE-18778.9.patch, HIVE-18778_TestCliDriver.patch, > HIVE-18788_SparkNegative.patch, HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16625252#comment-16625252 ] Hive QA commented on HIVE-18778: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12940976/HIVE-18778.8.patch {color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified. {color:red}ERROR:{color} -1 due to 10 failed/errored test(s), 14995 tests executed *Failed tests:* {noformat} org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[authorization_show_grant] (batchId=18) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[constant_prop_2] (batchId=30) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[nonmr_fetch] (batchId=21) org.apache.hadoop.hive.cli.TestEncryptedHDFSCliDriver.testCliDriver[encryption_join_with_different_encryption_keys] (batchId=186) org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidkafkamini_basic] (batchId=194) org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_test1] (batchId=194) org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[external_jdbc_table] (batchId=159) org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver[explainuser_3] (batchId=107) org.apache.hadoop.hive.cli.TestSparkNegativeCliDriver.testCliDriver[spark_task_failure] (batchId=266) org.apache.hadoop.hive.ql.exec.spark.TestSparkSessionTimeout.testMultiSparkSessionTimeout (batchId=245) {noformat} Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/14005/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/14005/console Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-14005/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.YetusPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 10 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12940976 - PreCommit-HIVE-Build > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778.4.patch, > HIVE-18778.5.patch, HIVE-18778.6.patch, HIVE-18778.7.patch, > HIVE-18778.8.patch, HIVE-18778_TestCliDriver.patch, > HIVE-18788_SparkNegative.patch, HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16625242#comment-16625242 ] Hive QA commented on HIVE-18778: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || || || || || {color:brown} Prechecks {color} || | {color:red}-1{color} | {color:red} @author {color} | {color:red} 0m 0s{color} | {color:red} The patch appears to contain 6 @author tags which the community has agreed to not allow in code contributions. {color} | || || || || {color:brown} master Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 8s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 6m 22s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 2m 23s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 45s{color} | {color:green} master passed {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 28s{color} | {color:blue} accumulo-handler in master has 21 extant Findbugs warnings. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 22s{color} | {color:blue} contrib in master has 13 extant Findbugs warnings. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 27s{color} | {color:blue} hbase-handler in master has 15 extant Findbugs warnings. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 3m 44s{color} | {color:blue} ql in master has 2326 extant Findbugs warnings. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 47s{color} | {color:green} master passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 10s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 2m 40s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 2m 21s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 2m 21s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 45s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 5m 37s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 49s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 13s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 43m 13s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Optional Tests | asflicense javac javadoc findbugs checkstyle compile | | uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux | | Build tool | maven | | Personality | /data/hiveptest/working/yetus_PreCommit-HIVE-Build-14005/dev-support/hive-personality.sh | | git revision | master / 80fd0ec | | Default Java | 1.8.0_111 | | @author | http://104.198.109.242/logs//PreCommit-HIVE-Build-14005/yetus/author-tags.txt | | findbugs | v3.0.0 | | modules | C: accumulo-handler contrib hbase-handler itests/hive-blobstore ql U: . | | Console output | http://104.198.109.242/logs//PreCommit-HIVE-Build-14005/yetus.txt | | Powered by | Apache Yetushttp://yetus.apache.org | This message was automatically generated. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778.4.patch, > HIVE-18778.5.patch, HIVE-18778.6.patch, HIVE-18778.7.patch, > HIVE-18778.8.patch, HIVE-18778_TestCliDriver.patch, > HIVE-18788_SparkNegative.patch, HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16625179#comment-16625179 ] Hive QA commented on HIVE-18778: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12940952/HIVE-18778.7.patch {color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified. {color:red}ERROR:{color} -1 due to 35 failed/errored test(s), 14992 tests executed *Failed tests:* {noformat} TestMiniDruidCliDriver - did not produce a TEST-*.xml file (likely timed out) (batchId=194) [druidmini_masking.q,druidmini_test1.q,druidkafkamini_basic.q,druidmini_joins.q,druid_timestamptz.q] org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[authorization_show_grant] (batchId=18) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[constant_prop_2] (batchId=30) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[nonmr_fetch] (batchId=21) org.apache.hadoop.hive.cli.TestEncryptedHDFSCliDriver.testCliDriver[encryption_join_with_different_encryption_keys] (batchId=186) org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_dynamic_partition] (batchId=193) org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_expressions] (batchId=193) org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_extractTime] (batchId=192) org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_floorTime] (batchId=192) org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_mv] (batchId=192) org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_test_insert] (batchId=193) org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_test_ts] (batchId=193) org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver[explainuser_3] (batchId=107) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[groupby9] (batchId=112) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[groupby_complex_types] (batchId=140) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[groupby_complex_types_multi_single_reducer] (batchId=150) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[groupby_cube1] (batchId=110) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[groupby_map_ppr] (batchId=113) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[groupby_map_ppr_multi_distinct] (batchId=132) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[groupby_multi_insert_common_distinct] (batchId=143) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[groupby_multi_single_reducer2] (batchId=118) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[groupby_multi_single_reducer3] (batchId=131) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[parquet_vectorization_7] (batchId=148) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[parquet_vectorization_8] (batchId=115) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[parquet_vectorization_9] (batchId=123) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[parquet_vectorization_decimal_date] (batchId=123) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[parquet_vectorization_div0] (batchId=145) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[parquet_vectorization_limit] (batchId=120) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[parquet_vectorization_offset_limit] (batchId=124) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[parquet_vectorization_part_project] (batchId=125) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[parquet_vectorization_pushdown] (batchId=124) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[union] (batchId=110) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[windowing] (batchId=133) org.apache.hadoop.hive.cli.TestSparkNegativeCliDriver.testCliDriver[spark_task_failure] (batchId=266) org.apache.hive.jdbc.TestJdbcWithMiniLlapArrow.testKillQuery (batchId=251) {noformat} Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/14002/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/14002/console Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-14002/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.YetusPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 35 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12940952 - PreCommit-HIVE-Build > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 >
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16625170#comment-16625170 ] Hive QA commented on HIVE-18778: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || || || || || {color:brown} Prechecks {color} || | {color:red}-1{color} | {color:red} @author {color} | {color:red} 0m 0s{color} | {color:red} The patch appears to contain 6 @author tags which the community has agreed to not allow in code contributions. {color} | || || || || {color:brown} master Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 8s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 6m 28s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 2m 21s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 44s{color} | {color:green} master passed {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 29s{color} | {color:blue} accumulo-handler in master has 21 extant Findbugs warnings. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 23s{color} | {color:blue} contrib in master has 13 extant Findbugs warnings. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 33s{color} | {color:blue} hbase-handler in master has 15 extant Findbugs warnings. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 3m 55s{color} | {color:blue} ql in master has 2326 extant Findbugs warnings. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 55s{color} | {color:green} master passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 9s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 2m 44s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 2m 23s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 2m 23s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 49s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 1s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 5m 45s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 48s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 14s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 43m 20s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Optional Tests | asflicense javac javadoc findbugs checkstyle compile | | uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux | | Build tool | maven | | Personality | /data/hiveptest/working/yetus_PreCommit-HIVE-Build-14002/dev-support/hive-personality.sh | | git revision | master / 80fd0ec | | Default Java | 1.8.0_111 | | @author | http://104.198.109.242/logs//PreCommit-HIVE-Build-14002/yetus/author-tags.txt | | findbugs | v3.0.0 | | modules | C: accumulo-handler contrib hbase-handler itests/hive-blobstore ql U: . | | Console output | http://104.198.109.242/logs//PreCommit-HIVE-Build-14002/yetus.txt | | Powered by | Apache Yetushttp://yetus.apache.org | This message was automatically generated. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778.4.patch, > HIVE-18778.5.patch, HIVE-18778.6.patch, HIVE-18778.7.patch, > HIVE-18778_TestCliDriver.patch, HIVE-18788_SparkNegative.patch, > HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16624871#comment-16624871 ] Hive QA commented on HIVE-18778: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12940816/HIVE-18778.6.patch {color:red}ERROR:{color} -1 due to build exiting with an error Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/13985/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/13985/console Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-13985/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Tests exited with: Exception: Patch URL https://issues.apache.org/jira/secure/attachment/12940816/HIVE-18778.6.patch was found in seen patch url's cache and a test was probably run already on it. Aborting... {noformat} This message is automatically generated. ATTACHMENT ID: 12940816 - PreCommit-HIVE-Build > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778.4.patch, > HIVE-18778.5.patch, HIVE-18778.6.patch, HIVE-18778_TestCliDriver.patch, > HIVE-18788_SparkNegative.patch, HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16624738#comment-16624738 ] Hive QA commented on HIVE-18778: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12940816/HIVE-18778.6.patch {color:red}ERROR:{color} -1 due to no test(s) being added or modified. {color:red}ERROR:{color} -1 due to 856 failed/errored test(s), 14991 tests executed *Failed tests:* {noformat} TestMiniDruidCliDriver - did not produce a TEST-*.xml file (likely timed out) (batchId=194) [druidmini_masking.q,druidmini_test1.q,druidkafkamini_basic.q,druidmini_joins.q,druid_timestamptz.q] org.apache.hadoop.hive.cli.TestAccumuloCliDriver.testCliDriver[accumulo_predicate_pushdown] (batchId=256) org.apache.hadoop.hive.cli.TestAccumuloCliDriver.testCliDriver[accumulo_queries] (batchId=256) org.apache.hadoop.hive.cli.TestAccumuloCliDriver.testCliDriver[accumulo_single_sourced_multi_insert] (batchId=256) org.apache.hadoop.hive.cli.TestBeeLineDriver.testCliDriver[explain_outputs] (batchId=264) org.apache.hadoop.hive.cli.TestBeeLineDriver.testCliDriver[mapjoin2] (batchId=264) org.apache.hadoop.hive.cli.TestBeeLineDriver.testCliDriver[materialized_view_create_rewrite] (batchId=264) org.apache.hadoop.hive.cli.TestBeeLineDriver.testCliDriver[select_dummy_source] (batchId=264) org.apache.hadoop.hive.cli.TestBeeLineDriver.testCliDriver[smb_mapjoin_10] (batchId=264) org.apache.hadoop.hive.cli.TestBeeLineDriver.testCliDriver[smb_mapjoin_11] (batchId=264) org.apache.hadoop.hive.cli.TestBeeLineDriver.testCliDriver[smb_mapjoin_12] (batchId=264) org.apache.hadoop.hive.cli.TestBeeLineDriver.testCliDriver[smb_mapjoin_13] (batchId=264) org.apache.hadoop.hive.cli.TestBeeLineDriver.testCliDriver[smb_mapjoin_16] (batchId=264) org.apache.hadoop.hive.cli.TestBeeLineDriver.testCliDriver[smb_mapjoin_1] (batchId=264) org.apache.hadoop.hive.cli.TestBeeLineDriver.testCliDriver[smb_mapjoin_2] (batchId=264) org.apache.hadoop.hive.cli.TestBeeLineDriver.testCliDriver[smb_mapjoin_3] (batchId=264) org.apache.hadoop.hive.cli.TestBeeLineDriver.testCliDriver[smb_mapjoin_7] (batchId=264) org.apache.hadoop.hive.cli.TestBlobstoreCliDriver.testCliDriver[explain] (batchId=267) org.apache.hadoop.hive.cli.TestBlobstoreCliDriver.testCliDriver[insert_into_dynamic_partitions] (batchId=267) org.apache.hadoop.hive.cli.TestBlobstoreCliDriver.testCliDriver[insert_into_table] (batchId=267) org.apache.hadoop.hive.cli.TestBlobstoreCliDriver.testCliDriver[insert_overwrite_directory] (batchId=267) org.apache.hadoop.hive.cli.TestBlobstoreCliDriver.testCliDriver[insert_overwrite_dynamic_partitions] (batchId=267) org.apache.hadoop.hive.cli.TestBlobstoreCliDriver.testCliDriver[insert_overwrite_table] (batchId=267) org.apache.hadoop.hive.cli.TestBlobstoreCliDriver.testCliDriver[write_final_output_blobstore] (batchId=267) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[authorization_show_grant] (batchId=18) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[constant_prop_2] (batchId=30) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[fetch_aggregation] (batchId=71) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[filter_in_or_dup] (batchId=36) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[groupby6_noskew] (batchId=2) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[implicit_decimal] (batchId=72) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[lateral_view_explode2] (batchId=91) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[load_dyn_part4] (batchId=67) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[macro] (batchId=4) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[nonmr_fetch] (batchId=21) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[udf_aes_decrypt] (batchId=58) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[udf_explode] (batchId=47) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[vectorized_mapjoin2] (batchId=25) org.apache.hadoop.hive.cli.TestContribCliDriver.testCliDriver[dboutput] (batchId=259) org.apache.hadoop.hive.cli.TestContribCliDriver.testCliDriver[fileformat_base64] (batchId=259) org.apache.hadoop.hive.cli.TestContribCliDriver.testCliDriver[serde_regex] (batchId=259) org.apache.hadoop.hive.cli.TestContribCliDriver.testCliDriver[serde_typedbytes2] (batchId=259) org.apache.hadoop.hive.cli.TestContribCliDriver.testCliDriver[serde_typedbytes3] (batchId=259) org.apache.hadoop.hive.cli.TestContribCliDriver.testCliDriver[serde_typedbytes4] (batchId=259) org.apache.hadoop.hive.cli.TestContribCliDriver.testCliDriver[serde_typedbytes5] (batchId=259) org.apache.hadoop.hive.cli.TestContribCliDriver.testCliDriver[serde_typedbytes] (batchId=259) org.apache.hadoop.hive.cli.TestContribCliDriver.testCliDriver[udaf_example_avg] (batchId=259)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16624725#comment-16624725 ] Hive QA commented on HIVE-18778: | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 1s{color} | {color:green} The patch does not contain any @author tags. {color} | || || || || {color:brown} master Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 7s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 6m 29s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 27s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 58s{color} | {color:green} master passed {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 30s{color} | {color:blue} hbase-handler in master has 15 extant Findbugs warnings. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 3m 56s{color} | {color:blue} ql in master has 2326 extant Findbugs warnings. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 10s{color} | {color:green} master passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 10s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 46s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 27s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 1m 27s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 2s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 1s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 4m 56s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 11s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 14s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 32m 18s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Optional Tests | asflicense javac javadoc findbugs checkstyle compile | | uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux | | Build tool | maven | | Personality | /data/hiveptest/working/yetus_PreCommit-HIVE-Build-13975/dev-support/hive-personality.sh | | git revision | master / cdba00c | | Default Java | 1.8.0_111 | | findbugs | v3.0.0 | | modules | C: hbase-handler ql U: . | | Console output | http://104.198.109.242/logs//PreCommit-HIVE-Build-13975/yetus.txt | | Powered by | Apache Yetushttp://yetus.apache.org | This message was automatically generated. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778.4.patch, > HIVE-18778.5.patch, HIVE-18778.6.patch, HIVE-18778_TestCliDriver.patch, > HIVE-18788_SparkNegative.patch, HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16616665#comment-16616665 ] Hive QA commented on HIVE-18778: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || || || || || {color:brown} Prechecks {color} || | {color:red}-1{color} | {color:red} @author {color} | {color:red} 0m 0s{color} | {color:red} The patch appears to contain 6 @author tags which the community has agreed to not allow in code contributions. {color} | || || || || {color:brown} master Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 8m 46s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 18s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 41s{color} | {color:green} master passed {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 4m 14s{color} | {color:blue} ql in master has 2311 extant Findbugs warnings. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 3s{color} | {color:green} master passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 28s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 7s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 1m 7s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 40s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 4m 13s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 57s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 14s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 25m 11s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Optional Tests | asflicense javac javadoc findbugs checkstyle compile | | uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux | | Build tool | maven | | Personality | /data/hiveptest/working/yetus_PreCommit-HIVE-Build-13829/dev-support/hive-personality.sh | | git revision | master / 45163ee | | Default Java | 1.8.0_111 | | @author | http://104.198.109.242/logs//PreCommit-HIVE-Build-13829/yetus/author-tags.txt | | findbugs | v3.0.0 | | modules | C: ql U: ql | | Console output | http://104.198.109.242/logs//PreCommit-HIVE-Build-13829/yetus.txt | | Powered by | Apache Yetushttp://yetus.apache.org | This message was automatically generated. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778.4.patch, > HIVE-18778_TestCliDriver.patch, HIVE-18788_SparkNegative.patch, > HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16441563#comment-16441563 ] Naveen Gangam commented on HIVE-18778: -- Thanks [~thejas]. I will sync up with him when he is back. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778_TestCliDriver.patch, > HIVE-18788_SparkNegative.patch, HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16441492#comment-16441492 ] Thejas M Nair commented on HIVE-18778: -- [~ngangam] Daniel is away for few days, so I doubt if he is going to get back on this soon. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778_TestCliDriver.patch, > HIVE-18788_SparkNegative.patch, HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16441444#comment-16441444 ] Naveen Gangam commented on HIVE-18778: -- [~daijy] Any luck with the tests? Thanks > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778_TestCliDriver.patch, > HIVE-18788_SparkNegative.patch, HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16422907#comment-16422907 ] Daniel Dai commented on HIVE-18778: --- [~ngangam], sorry for delay, there're other issues blocking me, but I should be back on it in next 2-3 days. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778_TestCliDriver.patch, > HIVE-18788_SparkNegative.patch, HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16420586#comment-16420586 ] Naveen Gangam commented on HIVE-18778: -- [~daijy] I understand you are busy but do you have an update when we might be able to upload a full patch? Thanks > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778_TestCliDriver.patch, > HIVE-18788_SparkNegative.patch, HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16409997#comment-16409997 ] Daniel Dai commented on HIVE-18778: --- Yes, still in background. There are lots of tests have statistics difference in addition to PREHOOK/POSTHOOK changes. Need to get sometime to clear these out. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778_TestCliDriver.patch, > HIVE-18788_SparkNegative.patch, HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16409912#comment-16409912 ] Naveen Gangam commented on HIVE-18778: -- [~daijy] Have you had a chance to work on these lately? Thanks > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778_TestCliDriver.patch, > HIVE-18788_SparkNegative.patch, HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16400491#comment-16400491 ] Naveen Gangam commented on HIVE-18778: -- [~daijy] Looks like it ran with your patch instead of mine. I was hoping to see if the Spark issues have been resolved. But since it ran with your patch, there are failures with the Spark Driver in the latest run. However, I expected to not see any failures with the TestCliDriver. There are quite a few of them. Can you take a look to ensure that all of them have been addressed? Thanks > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778_TestCliDriver.patch, > HIVE-18788_SparkNegative.patch, HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16400313#comment-16400313 ] Hive QA commented on HIVE-18778: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || || || || || {color:brown} Prechecks {color} || | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 0s{color} | {color:blue} Findbugs executables are not available. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | || || || || {color:brown} master Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 6m 25s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 3s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 40s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 54s{color} | {color:green} master passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 19s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 2s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 1m 2s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} checkstyle {color} | {color:red} 0m 38s{color} | {color:red} ql: The patch generated 1 new + 3 unchanged - 0 fixed = 4 total (was 3) {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 52s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:red}-1{color} | {color:red} asflicense {color} | {color:red} 0m 13s{color} | {color:red} The patch generated 49 ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 15m 14s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Optional Tests | asflicense javac javadoc findbugs checkstyle compile | | uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux | | Build tool | maven | | Personality | /data/hiveptest/working/yetus_PreCommit-HIVE-Build-9642/dev-support/hive-personality.sh | | git revision | master / d5cb7f6 | | Default Java | 1.8.0_111 | | checkstyle | http://104.198.109.242/logs//PreCommit-HIVE-Build-9642/yetus/diff-checkstyle-ql.txt | | asflicense | http://104.198.109.242/logs//PreCommit-HIVE-Build-9642/yetus/patch-asflicense-problems.txt | | modules | C: ql U: ql | | Console output | http://104.198.109.242/logs//PreCommit-HIVE-Build-9642/yetus.txt | | Powered by | Apache Yetushttp://yetus.apache.org | This message was automatically generated. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778_TestCliDriver.patch, > HIVE-18788_SparkNegative.patch, HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16399595#comment-16399595 ] Daniel Dai commented on HIVE-18778: --- Finally begin to make process. TestCliDriver first, now Tez/LLAP/Druid secion. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18778_TestCliDriver.patch, > HIVE-18788_SparkNegative.patch, HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16398707#comment-16398707 ] Naveen Gangam commented on HIVE-18778: -- [~daijy] How is it going on ur end? I just added a patch for test purposes that includes the spark, encryptedHDFS tests. I also included the source fix so we know if there are additional failures in the areas I was assigned. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18778.3.patch, HIVE-18788_SparkNegative.patch, > HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16396291#comment-16396291 ] Hive QA commented on HIVE-18778: Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12914114/HIVE-18788_SparkPerf.patch {color:red}ERROR:{color} -1 due to no test(s) being added or modified. {color:red}ERROR:{color} -1 due to 198 failed/errored test(s), 13756 tests executed *Failed tests:* {noformat} org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[acid_table_stats] (batchId=53) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[ppd_join5] (batchId=36) org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[llap_smb] (batchId=152) org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[bucket_map_join_tez_empty] (batchId=158) org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[insert_values_orig_table_use_metadata] (batchId=168) org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[results_cache_1] (batchId=168) org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[sysdb] (batchId=161) org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[tez_smb_main] (batchId=159) org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[update_access_time_non_current_db] (batchId=170) org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[vectorization_div0] (batchId=169) org.apache.hadoop.hive.cli.TestNegativeCliDriver.org.apache.hadoop.hive.cli.TestNegativeCliDriver (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[mm_convert] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[mm_truncate_cols] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[smb_bucketmapjoin] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[smb_mapjoin_14] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[sortmerge_mapjoin_mismatch_1] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[split_sample_wrong_format] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[stats_aggregator_error_1] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[stats_aggregator_error_2] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[stats_publisher_error_1] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[strict_join_2] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[strict_orderby] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[strict_orderby_2] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[strict_pruning_2] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[subquery_corr_grandparent] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[subquery_corr_in_agg] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[subquery_in_groupby] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[subquery_multiple_cols_in_select] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[subquery_notin_implicit_gby] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[subquery_scalar_corr_multi_rows] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[subquery_scalar_multi_rows] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[subquery_select_aggregate] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[subquery_select_distinct] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[subquery_subquery_chain_exists] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[temp_table_rename] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[touch2] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[truncate_bucketed_column] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[truncate_column_list_bucketing] (batchId=94) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[truncate_column_seqfile] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[truncate_nonexistant_column] (batchId=94) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[truncate_table_failure3] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[truncate_table_failure5] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[udaf_collect_set_unsupported] (batchId=95) org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[udf_array_contains_wrong2] (batchId=95)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16396227#comment-16396227 ] Hive QA commented on HIVE-18778: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | || || || || {color:brown} master Compile Tests {color} || || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | || || || || {color:brown} Other Tests {color} || | {color:red}-1{color} | {color:red} asflicense {color} | {color:red} 0m 39s{color} | {color:red} The patch generated 49 ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 1m 59s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Optional Tests | asflicense | | uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux | | Build tool | maven | | Personality | /data/hiveptest/working/yetus_PreCommit-HIVE-Build-9606/dev-support/hive-personality.sh | | git revision | master / d717d38 | | asflicense | http://104.198.109.242/logs//PreCommit-HIVE-Build-9606/yetus/patch-asflicense-problems.txt | | modules | C: ql U: ql | | Console output | http://104.198.109.242/logs//PreCommit-HIVE-Build-9606/yetus.txt | | Powered by | Apache Yetushttp://yetus.apache.org | This message was automatically generated. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18788_SparkNegative.patch, HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16395823#comment-16395823 ] Naveen Gangam commented on HIVE-18778: -- how are things on your end? Thanks > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18788_SparkNegative.patch, HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16395821#comment-16395821 ] Naveen Gangam commented on HIVE-18778: -- [~daijy] I could only attach files if I were the owner of the jira. So I had to temporarily flip the ownership of this jira. I have attached 3 separate patches with the names suffixed by the Spark Driver names. Hopefully no one checks in any changes before we are able to commit this. I had to do them in small batches as the running them in single run kept failing. I will get to the EncryptedHDFS ones later today or tomorrow. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778-SparkPositive.patch, HIVE-18778.1.patch, > HIVE-18778.2.patch, HIVE-18788_SparkNegative.patch, HIVE-18788_SparkPerf.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16390299#comment-16390299 ] Daniel Dai commented on HIVE-18778: --- I am struggling to run tests locally, but hopefully I can get some sections done in 1 or 2 days. Will keep you updated. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778.1.patch, HIVE-18778.2.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16389734#comment-16389734 ] Naveen Gangam commented on HIVE-18778: -- [~daijy] I havent been able to get to it thus far this week. I will be starting on this soon. How is your side coming along? Thanks > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778.1.patch, HIVE-18778.2.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16388613#comment-16388613 ] Daniel Dai commented on HIVE-18778: --- [~ngangam], are you able working on this? > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778.1.patch, HIVE-18778.2.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16382658#comment-16382658 ] Daniel Dai commented on HIVE-18778: --- [~ngangam], I glance through the list, and summarize the number of failures here: * TestCliDriver: TestCliDriver 1247 * Spark/EncrypedHDFS TestSparkCliDriver 536 TestSparkPerfCliDriver 95 TestSparkNegativeCliDriver 2 TestLocalSparkCliDriver 1 TestMiniSparkOnYarnCliDriver 47 TestEncryptedHDFSCliDriver 2 * Tez/LLAP/Druid TestMiniLlapCliDriver/TestMiniLlapLocalCliDriver 515 TestTezPerfCliDriver 97 TestMiniTezCliDriver 14 TestMiniDruidCliDriver 3 * Others TestHBaseCliDriver 11 TestContribCliDriver 18 TestBeeLineDriver 11 TestMinimrCliDriver 4 TestNegativeCliDriver 4 TestBlobstoreCliDriver 7 TestAccumuloCliDriver 3 TestContribNegativeCliDriver 1 TestMTQueries 1 TestTriggersMoveWorkloadManager 1 I would like to start with (Tez/LLAP/Druid), and then (TestCliDriver), would you like to help with (Spark/EncrypedHDFS)? > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778.1.patch, HIVE-18778.2.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16382065#comment-16382065 ] Hive QA commented on HIVE-18778: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || || || || || {color:brown} Prechecks {color} || | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 0s{color} | {color:blue} Findbugs executables are not available. {color} | | {color:red}-1{color} | {color:red} @author {color} | {color:red} 0m 0s{color} | {color:red} The patch appears to contain 6 @author tags which the community has agreed to not allow in code contributions. {color} | || || || || {color:brown} master Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 7m 38s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 7s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 38s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 55s{color} | {color:green} master passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 17s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 2s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 1m 2s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} checkstyle {color} | {color:red} 0m 37s{color} | {color:red} ql: The patch generated 1 new + 3 unchanged - 0 fixed = 4 total (was 3) {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 3s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:red}-1{color} | {color:red} asflicense {color} | {color:red} 0m 13s{color} | {color:red} The patch generated 49 ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 14m 46s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Optional Tests | asflicense javac javadoc findbugs checkstyle compile | | uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux | | Build tool | maven | | Personality | /data/hiveptest/working/yetus_PreCommit-HIVE-Build-9422/dev-support/hive-personality.sh | | git revision | master / 1e6a182 | | Default Java | 1.8.0_111 | | @author | http://104.198.109.242/logs//PreCommit-HIVE-Build-9422/yetus/author-tags.txt | | checkstyle | http://104.198.109.242/logs//PreCommit-HIVE-Build-9422/yetus/diff-checkstyle-ql.txt | | asflicense | http://104.198.109.242/logs//PreCommit-HIVE-Build-9422/yetus/patch-asflicense-problems.txt | | modules | C: ql U: ql | | Console output | http://104.198.109.242/logs//PreCommit-HIVE-Build-9422/yetus.txt | | Powered by | Apache Yetushttp://yetus.apache.org | This message was automatically generated. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778.1.patch, HIVE-18778.2.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16381514#comment-16381514 ] Daniel Dai commented on HIVE-18778: --- There are around 2600 failures. I think we need to run all tests with the overwrite flag and check the diff. Let me kick off another round to get a list, so we can split. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778.1.patch, HIVE-18778.2.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16381498#comment-16381498 ] Naveen Gangam commented on HIVE-18778: -- [~daijy] Sure, I could help. Do you have a list? The pre-commit build above is already deleted. Do you want to divi it up and tackle them together? Thanks > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778.1.patch, HIVE-18778.2.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16381326#comment-16381326 ] Daniel Dai commented on HIVE-18778: --- Thanks [~ngangam]. However, there are a huge number of golden files need to update, it might take sometime. I wonder if you'd like to help? > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778.1.patch, HIVE-18778.2.patch > > > With Sentry enabled, commands like explain drop table foo fail with {{explain > drop table foo;}} > {code} > Error: Error while compiling statement: FAILED: SemanticException No valid > privileges > Required privilege( Table) not available in input privileges > The required privileges: (state=42000,code=4) > {code} > Sentry fails to authorize because the ExplainSemanticAnalyzer uses an > instance of DDLSemanticAnalyzer to analyze the explain query. > {code} > BaseSemanticAnalyzer sem = SemanticAnalyzerFactory.get(conf, input); > sem.analyze(input, ctx); > sem.validate() > {code} > The inputs/outputs entities for this query are set in the above code. > However, these are never set on the instance of ExplainSemanticAnalyzer > itself and thus is not propagated into the HookContext in the calling Driver > code. > {code} > sem.analyze(tree, ctx); --> this results in calling the above code that uses > DDLSA > hookCtx.update(sem); --> sem is an instance of ExplainSemanticAnalyzer, this > code attempts to update the HookContext with the input/output info from ESA > which is never set. > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16381260#comment-16381260 ] Naveen Gangam commented on HIVE-18778: -- Patch looks good to me. I have tested it on a sentry-enabled cluster. So +1 for me. I had opened a duplicate jira. I am copying the description from the other jira so this jira is more descriptive of the problem. Hope you dont mind. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778.1.patch, HIVE-18778.2.patch > > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=1633#comment-1633 ] Daniel Dai commented on HIVE-18778: --- Here is the UT result: [https://builds.apache.org/job/PreCommit-HIVE-Build/9357/] Need to update quite a few golden files. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778.1.patch, HIVE-18778.2.patch > > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16375854#comment-16375854 ] Hive QA commented on HIVE-18778: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || || || || || {color:brown} Prechecks {color} || | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 1s{color} | {color:blue} Findbugs executables are not available. {color} | | {color:red}-1{color} | {color:red} @author {color} | {color:red} 0m 0s{color} | {color:red} The patch appears to contain 6 @author tags which the community has agreed to not allow in code contributions. {color} | || || || || {color:brown} master Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 6m 47s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 55s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 35s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 52s{color} | {color:green} master passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 16s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 54s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 54s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} checkstyle {color} | {color:red} 0m 33s{color} | {color:red} ql: The patch generated 1 new + 3 unchanged - 0 fixed = 4 total (was 3) {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 50s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:red}-1{color} | {color:red} asflicense {color} | {color:red} 0m 12s{color} | {color:red} The patch generated 49 ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 13m 9s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Optional Tests | asflicense javac javadoc findbugs checkstyle compile | | uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux | | Build tool | maven | | Personality | /data/hiveptest/working/yetus_PreCommit-HIVE-Build-9357/dev-support/hive-personality.sh | | git revision | master / 53a590b | | Default Java | 1.8.0_111 | | @author | http://104.198.109.242/logs//PreCommit-HIVE-Build-9357/yetus/author-tags.txt | | checkstyle | http://104.198.109.242/logs//PreCommit-HIVE-Build-9357/yetus/diff-checkstyle-ql.txt | | asflicense | http://104.198.109.242/logs//PreCommit-HIVE-Build-9357/yetus/patch-asflicense-problems.txt | | modules | C: ql U: ql | | Console output | http://104.198.109.242/logs//PreCommit-HIVE-Build-9357/yetus.txt | | Powered by | Apache Yetushttp://yetus.apache.org | This message was automatically generated. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778.1.patch, HIVE-18778.2.patch > > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16375764#comment-16375764 ] Daniel Dai commented on HIVE-18778: --- To kick off ptest. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778.1.patch, HIVE-18778.2.patch > > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (HIVE-18778) Needs to capture input/output entities in explain
[ https://issues.apache.org/jira/browse/HIVE-18778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16373999#comment-16373999 ] Hive QA commented on HIVE-18778: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || || || || || {color:brown} Prechecks {color} || | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 1s{color} | {color:blue} Findbugs executables are not available. {color} | | {color:red}-1{color} | {color:red} @author {color} | {color:red} 0m 0s{color} | {color:red} The patch appears to contain 6 @author tags which the community has agreed to not allow in code contributions. {color} | || || || || {color:brown} master Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 7m 36s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 2s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 38s{color} | {color:green} master passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 58s{color} | {color:green} master passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 17s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 0s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 1m 0s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} checkstyle {color} | {color:red} 0m 34s{color} | {color:red} ql: The patch generated 1 new + 3 unchanged - 0 fixed = 4 total (was 3) {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 56s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:red}-1{color} | {color:red} asflicense {color} | {color:red} 0m 13s{color} | {color:red} The patch generated 49 ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 14m 29s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Optional Tests | asflicense javac javadoc findbugs checkstyle compile | | uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux | | Build tool | maven | | Personality | /data/hiveptest/working/yetus_PreCommit-HIVE-Build-9324/dev-support/hive-personality.sh | | git revision | master / 1593c10 | | Default Java | 1.8.0_111 | | @author | http://104.198.109.242/logs//PreCommit-HIVE-Build-9324/yetus/author-tags.txt | | checkstyle | http://104.198.109.242/logs//PreCommit-HIVE-Build-9324/yetus/diff-checkstyle-ql.txt | | asflicense | http://104.198.109.242/logs//PreCommit-HIVE-Build-9324/yetus/patch-asflicense-problems.txt | | modules | C: ql U: ql | | Console output | http://104.198.109.242/logs//PreCommit-HIVE-Build-9324/yetus.txt | | Powered by | Apache Yetushttp://yetus.apache.org | This message was automatically generated. > Needs to capture input/output entities in explain > - > > Key: HIVE-18778 > URL: https://issues.apache.org/jira/browse/HIVE-18778 > Project: Hive > Issue Type: Bug >Reporter: Daniel Dai >Assignee: Daniel Dai >Priority: Major > Attachments: HIVE-18778.1.patch > > -- This message was sent by Atlassian JIRA (v7.6.3#76005)