This is an automated email from the ASF dual-hosted git repository. prasanthj pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/hive.git
The following commit(s) were added to refs/heads/master by this push: new d8bc9dc HIVE-23972: Add external client ID to LLAP external client (#1350) d8bc9dc is described below commit d8bc9dc5a35cb20861be090ff171a77a81618625 Author: jdere <jd...@hortonworks.com> AuthorDate: Thu Aug 13 13:31:15 2020 -0700 HIVE-23972: Add external client ID to LLAP external client (#1350) Change-Id: Ica5fd933d26dc76c9f6fbbd4c8b578eaf20d5e69 Co-authored-by: Jason Dere <jd...@cloudera.com> --- common/src/java/org/apache/hadoop/hive/conf/HiveConf.java | 1 + .../apache/hadoop/hive/llap/daemon/impl/TaskExecutorService.java | 3 +++ .../apache/hadoop/hive/ql/udf/generic/GenericUDTFGetSplits.java | 9 ++++++++- 3 files changed, 12 insertions(+), 1 deletion(-) diff --git a/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java b/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java index d58326f..4b69576 100644 --- a/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java +++ b/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java @@ -6240,6 +6240,7 @@ public class HiveConf extends Configuration { "hive\\.parquet\\..*", "hive\\.ppd\\..*", "hive\\.prewarm\\..*", + "hive\\.query\\.name", "hive\\.server2\\.thrift\\.resultset\\.default\\.fetch\\.size", "hive\\.server2\\.proxy\\.user", "hive\\.skewjoin\\..*", diff --git a/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/TaskExecutorService.java b/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/TaskExecutorService.java index 3195ea7..e44dd2c 100644 --- a/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/TaskExecutorService.java +++ b/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/TaskExecutorService.java @@ -342,6 +342,9 @@ public class TaskExecutorService extends AbstractService SignableVertexSpec fs = c.getVertexSpec(); value.append(isFirst ? " (" : ", ").append(c.getQueryId()) .append("/").append(fs.getVertexName()).append(c.isGuaranteed() ? ", guaranteed" : ""); + if (fs.getDagName() != null) { + value.append(", dagName ").append(fs.getDagName()); + } isFirst = false; } value.append(isFirst ? " (" : ", "); diff --git a/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTFGetSplits.java b/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTFGetSplits.java index 01b2037..ebd041b 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTFGetSplits.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/udf/generic/GenericUDTFGetSplits.java @@ -233,7 +233,14 @@ public class GenericUDTFGetSplits extends GenericUDTF { + ConfVars.LLAP_HS2_ENABLE_COORDINATOR.varname + " enabled"); } ApplicationId applicationId = coordinator.createExtClientAppId(); - LOG.info("Generated appID {} for LLAP splits", applicationId.toString()); + String externalDagName = SessionState.get().getConf().getVar(ConfVars.HIVEQUERYNAME); + + StringBuilder sb = new StringBuilder(); + sb.append("Generated appID ").append(applicationId.toString()).append(" for LLAP splits"); + if (externalDagName != null) { + sb.append(", with externalID ").append(externalDagName); + } + LOG.info(sb.toString()); PlanFragment fragment = createPlanFragment(inputArgQuery, applicationId); TezWork tezWork = fragment.work;