hive git commit: HIVE-20868 : SMB Join fails intermittently when TezDummyOperator has child op in getFinalOp in MapRecordProcessor(Deepak Jaiswal, reviewed by Gopal V)
Repository: hive Updated Branches: refs/heads/master ad597b026 -> ffb78b521 HIVE-20868 : SMB Join fails intermittently when TezDummyOperator has child op in getFinalOp in MapRecordProcessor(Deepak Jaiswal, reviewed by Gopal V) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/ffb78b52 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/ffb78b52 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/ffb78b52 Branch: refs/heads/master Commit: ffb78b521cbd977d575374169392b0b3231024dc Parents: ad597b0 Author: Deepak Jaiswal Authored: Thu Nov 8 07:30:04 2018 -0800 Committer: Deepak Jaiswal Committed: Thu Nov 8 07:30:04 2018 -0800 -- .../apache/hadoop/hive/ql/exec/TezDummyStoreOperator.java| 8 +++- 1 file changed, 7 insertions(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/ffb78b52/ql/src/java/org/apache/hadoop/hive/ql/exec/TezDummyStoreOperator.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/TezDummyStoreOperator.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/TezDummyStoreOperator.java index d00057a..08c1cc4 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/TezDummyStoreOperator.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/TezDummyStoreOperator.java @@ -37,7 +37,7 @@ public class TezDummyStoreOperator extends DummyStoreOperator { super(ctx); } - private boolean fetchDone = false; + private transient boolean fetchDone = false; /** * Unlike the MR counterpoint, on Tez we want processOp to forward @@ -56,4 +56,10 @@ public class TezDummyStoreOperator extends DummyStoreOperator { public void setFetchDone(boolean fetchDone) { this.fetchDone = fetchDone; } + + @Override + public void closeOp(boolean abort) throws HiveException { +super.closeOp(abort); +fetchDone = false; + } }
hive git commit: HIVE-16976 : DPP: SyntheticJoinPredicate transitivity for < > and BETWEEN (Deepak Jaiswal, reviewed by Jesus)
Repository: hive Updated Branches: refs/heads/master cb9d5ccd8 -> 305e7887e HIVE-16976 : DPP: SyntheticJoinPredicate transitivity for < > and BETWEEN (Deepak Jaiswal, reviewed by Jesus) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/305e7887 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/305e7887 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/305e7887 Branch: refs/heads/master Commit: 305e7887e542b680ba41e32c49dab3a18a84e68a Parents: cb9d5cc Author: Deepak Jaiswal Authored: Fri Jan 11 13:51:31 2019 -0800 Committer: Deepak Jaiswal Committed: Fri Jan 11 13:51:31 2019 -0800 -- .../org/apache/hadoop/hive/conf/HiveConf.java | 2 +- .../hadoop/hive/ql/exec/FunctionRegistry.java | 20 +++ .../hive/ql/metadata/HiveStorageHandler.java| 12 ++ .../DynamicPartitionPruningOptimization.java| 23 ++- .../stats/annotation/StatsRulesProcFactory.java | 13 +- .../hadoop/hive/ql/parse/GenTezUtils.java | 9 ++ .../hive/ql/plan/ExprNodeDynamicListDesc.java | 20 ++- .../hive/ql/ppd/SyntheticJoinPredicate.java | 160 ++- .../clientpositive/llap/cross_prod_1.q.out | 4 +- .../llap/groupby_groupingset_bug.q.out | 74 - .../results/clientpositive/llap/semijoin.q.out | 4 +- .../clientpositive/llap/subquery_in.q.out | 4 +- .../clientpositive/llap/subquery_notin.q.out| 4 +- .../clientpositive/llap/subquery_scalar.q.out | 48 +++--- .../clientpositive/llap/subquery_select.q.out | 4 +- 15 files changed, 283 insertions(+), 118 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/305e7887/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java -- diff --git a/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java b/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java index b213609..a6bb9ef 100644 --- a/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java +++ b/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java @@ -2489,7 +2489,7 @@ public class HiveConf extends Configuration { "UDTFs change the number of rows of the output. A common UDTF is the explode() method that creates\n" + "multiple rows for each element in the input array. This factor is applied to the number of\n" + "output rows and output size."), - +HIVE_STATS_USE_NDV_ADJUSTMENT("hive.stats.use.ndv.adjustment", false, "Use nDV adjustment for other columns in join"), // Concurrency HIVE_SUPPORT_CONCURRENCY("hive.support.concurrency", false, "Whether Hive supports concurrency control or not. \n" + http://git-wip-us.apache.org/repos/asf/hive/blob/305e7887/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java index d7f069e..bbe7fb0 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java @@ -1851,4 +1851,24 @@ public final class FunctionRegistry { String blackListStr) { system.setupPermissionsForUDFs(whiteListStr, blackListStr); } + + /** + * Function to invert non-equi function texts + * @param funcText + */ + public static String invertFuncText(final String funcText) { +// Reverse the text +switch (funcText) { + case "<": +return ">"; + case "<=": +return ">="; + case ">": +return "<"; + case ">=": +return "<="; + default: +return null; // helps identify unsupported functions +} + } } http://git-wip-us.apache.org/repos/asf/hive/blob/305e7887/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java b/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java index 2ebb149..c010b18 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java @@ -26,6 +26,7 @@ import org.apache.hadoop.hive.metastore.api.LockType; import org.apache.hadoop.hive.metastore.api.MetaException; import org.apache.hadoop.hive.metastore.api.Table; import org.apache.hadoop.hive.ql.hooks.WriteEntity; +import org.apache.hadoop.hive.ql.plan.ExprNodeDesc; import org.apache.hadoop.hive.ql.plan.TableDesc; import org.apache.hadoop.hive.ql.security.authorization.HiveAuthorizationProvider; import org.apache.hadoop.hive.serde2.Abstrac
hive git commit: Revert "HIVE-16976 : DPP: SyntheticJoinPredicate transitivity for < > and BETWEEN (Deepak Jaiswal, reviewed by Jesus)"
Repository: hive Updated Branches: refs/heads/master 305e7887e -> a81e8de30 Revert "HIVE-16976 : DPP: SyntheticJoinPredicate transitivity for < > and BETWEEN (Deepak Jaiswal, reviewed by Jesus)" This reverts commit 305e7887e542b680ba41e32c49dab3a18a84e68a. Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/a81e8de3 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/a81e8de3 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/a81e8de3 Branch: refs/heads/master Commit: a81e8de30b742c26c8ebb21eeeb057d52fca9b8e Parents: 305e788 Author: Deepak Jaiswal Authored: Fri Jan 11 14:00:27 2019 -0800 Committer: Deepak Jaiswal Committed: Fri Jan 11 14:00:27 2019 -0800 -- .../org/apache/hadoop/hive/conf/HiveConf.java | 2 +- .../hadoop/hive/ql/exec/FunctionRegistry.java | 20 --- .../hive/ql/metadata/HiveStorageHandler.java| 12 -- .../DynamicPartitionPruningOptimization.java| 23 +-- .../stats/annotation/StatsRulesProcFactory.java | 13 +- .../hadoop/hive/ql/parse/GenTezUtils.java | 9 -- .../hive/ql/plan/ExprNodeDynamicListDesc.java | 20 +-- .../hive/ql/ppd/SyntheticJoinPredicate.java | 160 +-- .../clientpositive/llap/cross_prod_1.q.out | 4 +- .../llap/groupby_groupingset_bug.q.out | 74 - .../results/clientpositive/llap/semijoin.q.out | 4 +- .../clientpositive/llap/subquery_in.q.out | 4 +- .../clientpositive/llap/subquery_notin.q.out| 4 +- .../clientpositive/llap/subquery_scalar.q.out | 48 +++--- .../clientpositive/llap/subquery_select.q.out | 4 +- 15 files changed, 118 insertions(+), 283 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/a81e8de3/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java -- diff --git a/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java b/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java index a6bb9ef..b213609 100644 --- a/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java +++ b/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java @@ -2489,7 +2489,7 @@ public class HiveConf extends Configuration { "UDTFs change the number of rows of the output. A common UDTF is the explode() method that creates\n" + "multiple rows for each element in the input array. This factor is applied to the number of\n" + "output rows and output size."), -HIVE_STATS_USE_NDV_ADJUSTMENT("hive.stats.use.ndv.adjustment", false, "Use nDV adjustment for other columns in join"), + // Concurrency HIVE_SUPPORT_CONCURRENCY("hive.support.concurrency", false, "Whether Hive supports concurrency control or not. \n" + http://git-wip-us.apache.org/repos/asf/hive/blob/a81e8de3/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java index bbe7fb0..d7f069e 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java @@ -1851,24 +1851,4 @@ public final class FunctionRegistry { String blackListStr) { system.setupPermissionsForUDFs(whiteListStr, blackListStr); } - - /** - * Function to invert non-equi function texts - * @param funcText - */ - public static String invertFuncText(final String funcText) { -// Reverse the text -switch (funcText) { - case "<": -return ">"; - case "<=": -return ">="; - case ">": -return "<"; - case ">=": -return "<="; - default: -return null; // helps identify unsupported functions -} - } } http://git-wip-us.apache.org/repos/asf/hive/blob/a81e8de3/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java b/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java index c010b18..2ebb149 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java @@ -26,7 +26,6 @@ import org.apache.hadoop.hive.metastore.api.LockType; import org.apache.hadoop.hive.metastore.api.MetaException; import org.apache.hadoop.hive.metastore.api.Table; import org.apache.hadoop.hive.ql.hooks.WriteEntity; -import org.apache.hadoop.hive.ql.plan.ExprNodeDesc; import org.apache.hadoop.hive.ql.plan.TableDesc; import org.apache.hadoop.hive.ql.security.authorization.
hive git commit: HIVE-16976 : DPP: SyntheticJoinPredicate transitivity for < > and BETWEEN (Deepak Jaiswal, reviewed by Jesus)
Repository: hive Updated Branches: refs/heads/master a81e8de30 -> 28db173b9 HIVE-16976 : DPP: SyntheticJoinPredicate transitivity for < > and BETWEEN (Deepak Jaiswal, reviewed by Jesus) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/28db173b Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/28db173b Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/28db173b Branch: refs/heads/master Commit: 28db173b9c24991e184fdf69625d932294a15d15 Parents: a81e8de Author: Deepak Jaiswal Authored: Fri Jan 11 14:02:34 2019 -0800 Committer: Deepak Jaiswal Committed: Fri Jan 11 14:02:34 2019 -0800 -- .../hadoop/hive/ql/exec/FunctionRegistry.java | 20 +++ .../hive/ql/metadata/HiveStorageHandler.java| 12 ++ .../DynamicPartitionPruningOptimization.java| 23 ++- .../hadoop/hive/ql/parse/GenTezUtils.java | 9 ++ .../hive/ql/plan/ExprNodeDynamicListDesc.java | 20 ++- .../hive/ql/ppd/SyntheticJoinPredicate.java | 160 ++- .../clientpositive/llap/cross_prod_1.q.out | 4 +- .../llap/groupby_groupingset_bug.q.out | 74 - .../results/clientpositive/llap/semijoin.q.out | 4 +- .../clientpositive/llap/subquery_in.q.out | 4 +- .../clientpositive/llap/subquery_notin.q.out| 4 +- .../clientpositive/llap/subquery_scalar.q.out | 48 +++--- .../clientpositive/llap/subquery_select.q.out | 4 +- 13 files changed, 270 insertions(+), 116 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/28db173b/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java index d7f069e..bbe7fb0 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java @@ -1851,4 +1851,24 @@ public final class FunctionRegistry { String blackListStr) { system.setupPermissionsForUDFs(whiteListStr, blackListStr); } + + /** + * Function to invert non-equi function texts + * @param funcText + */ + public static String invertFuncText(final String funcText) { +// Reverse the text +switch (funcText) { + case "<": +return ">"; + case "<=": +return ">="; + case ">": +return "<"; + case ">=": +return "<="; + default: +return null; // helps identify unsupported functions +} + } } http://git-wip-us.apache.org/repos/asf/hive/blob/28db173b/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java b/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java index 2ebb149..c010b18 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java @@ -26,6 +26,7 @@ import org.apache.hadoop.hive.metastore.api.LockType; import org.apache.hadoop.hive.metastore.api.MetaException; import org.apache.hadoop.hive.metastore.api.Table; import org.apache.hadoop.hive.ql.hooks.WriteEntity; +import org.apache.hadoop.hive.ql.plan.ExprNodeDesc; import org.apache.hadoop.hive.ql.plan.TableDesc; import org.apache.hadoop.hive.ql.security.authorization.HiveAuthorizationProvider; import org.apache.hadoop.hive.serde2.AbstractSerDe; @@ -172,4 +173,15 @@ public interface HiveStorageHandler extends Configurable { default LockType getLockType(WriteEntity writeEntity){ return LockType.EXCLUSIVE; } + + /** + * Test if the storage handler allows the push-down of join filter predicate to prune further the splits. + * + * @param syntheticFilterPredicate Join filter predicate. + * @return true if supports dynamic split pruning for the given predicate. + */ + + default boolean addDynamicSplitPruningEdge(ExprNodeDesc syntheticFilterPredicate) { +return false; + } } http://git-wip-us.apache.org/repos/asf/hive/blob/28db173b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/DynamicPartitionPruningOptimization.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/DynamicPartitionPruningOptimization.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/DynamicPartitionPruningOptimization.java index a1401aa..439fb75 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/DynamicPartitionPruningOptimization.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/DynamicPartitionPruningOptimization.
[hive] branch master updated: HIVE-21214 : MoveTask : Use attemptId instead of file size for deduplication of files compareTempOrDuplicateFiles() (Deepak Jaiswal, reviewed by Jason Dere)
This is an automated email from the ASF dual-hosted git repository. djaiswal pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/hive.git The following commit(s) were added to refs/heads/master by this push: new fae6256 HIVE-21214 : MoveTask : Use attemptId instead of file size for deduplication of files compareTempOrDuplicateFiles() (Deepak Jaiswal, reviewed by Jason Dere) fae6256 is described below commit fae6256ace38d106b62d3bcade7b84b51bf4e1ec Author: Deepak Jaiswal AuthorDate: Tue Feb 5 14:06:19 2019 -0800 HIVE-21214 : MoveTask : Use attemptId instead of file size for deduplication of files compareTempOrDuplicateFiles() (Deepak Jaiswal, reviewed by Jason Dere) --- .../org/apache/hadoop/hive/ql/exec/Utilities.java | 71 ++ 1 file changed, 59 insertions(+), 12 deletions(-) diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java index 8937b43..b84b052 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java @@ -1308,7 +1308,7 @@ public final class Utilities { * filename to extract taskid from */ public static String getTaskIdFromFilename(String filename) { -return getIdFromFilename(filename, FILE_NAME_TO_TASK_ID_REGEX); +return getTaskIdFromFilename(filename, FILE_NAME_TO_TASK_ID_REGEX); } /** @@ -1319,10 +1319,19 @@ public final class Utilities { * filename to extract taskid from */ public static String getPrefixedTaskIdFromFilename(String filename) { -return getIdFromFilename(filename, FILE_NAME_PREFIXED_TASK_ID_REGEX); +return getTaskIdFromFilename(filename, FILE_NAME_PREFIXED_TASK_ID_REGEX); } - private static String getIdFromFilename(String filename, Pattern pattern) { + private static String getTaskIdFromFilename(String filename, Pattern pattern) { +return getIdFromFilename(filename, pattern, 1); + } + + public static int getAttemptIdFromFilename(String filename) { +String attemptStr = getIdFromFilename(filename, FILE_NAME_PREFIXED_TASK_ID_REGEX, 3); +return Integer.parseInt(attemptStr.substring(1)); + } + + private static String getIdFromFilename(String filename, Pattern pattern, int group) { String taskId = filename; int dirEnd = filename.lastIndexOf(Path.SEPARATOR); if (dirEnd != -1) { @@ -1334,7 +1343,7 @@ public final class Utilities { LOG.warn("Unable to get task id from file name: {}. Using last component {}" + " as task id.", filename, taskId); } else { - taskId = m.group(1); + taskId = m.group(group); } LOG.debug("TaskId for {} = {}", filename, taskId); return taskId; @@ -1823,10 +1832,10 @@ public final class Utilities { private static FileStatus compareTempOrDuplicateFiles(FileSystem fs, FileStatus file, FileStatus existingFile) throws IOException { -// Compare the file sizes of all the attempt files for the same task, the largest win -// any attempt files could contain partial results (due to task failures or -// speculative runs), but the largest should be the correct one since the result -// of a successful run should never be smaller than a failed/speculative run. +// Pick the one with mewest attempt ID. For sanity, check the file sizes too. +// If the file size of newest attempt is less than that for older one, +// Throw an exception as it maybe a correctness issue causing it. +// This breaks speculative execution if it ends prematurely. FileStatus toDelete = null, toRetain = null; // "LOAD .. INTO" and "INSERT INTO" commands will generate files with @@ -1847,12 +1856,26 @@ public final class Utilities { return existingFile; } -if (existingFile.getLen() >= file.getLen()) { - toDelete = file; +int existingFileAttemptId = getAttemptIdFromFilename(existingFile.getPath().getName()); +int fileAttemptId = getAttemptIdFromFilename(file.getPath().getName()); + +long existingFileSz = getFileSizeRecursively(fs, existingFile); +long fileSz = getFileSizeRecursively(fs, file); +// Files may come in any order irrespective of their attempt IDs +if (existingFileAttemptId > fileAttemptId && +existingFileSz >= fileSz) { + // keep existing toRetain = existingFile; -} else { - toDelete = existingFile; + toDelete = file; +} else if (existingFileAttemptId < fileAttemptId && +existingFileSz <= fileSz) { + // keep file toRetain = file; + toDelete = existingFile; +} else { + throw new IOException(" File " + filePath + +" with newer attempt ID " + fileAttemptId + " is smaller than the file " ++ existingFile.getPath() + " wit
[hive] branch master updated: HIVE-21221 : Make HS2 and LLAP consistent - Bring up LLAP WebUI in test mode if WebUI port is configured (Oliver Draese via Deepak Jaiswal)
This is an automated email from the ASF dual-hosted git repository. djaiswal pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/hive.git The following commit(s) were added to refs/heads/master by this push: new aaf01ae HIVE-21221 : Make HS2 and LLAP consistent - Bring up LLAP WebUI in test mode if WebUI port is configured (Oliver Draese via Deepak Jaiswal) aaf01ae is described below commit aaf01ae60bb987965a45adc243605e625fd39140 Author: Oliver Draese AuthorDate: Fri Feb 15 14:36:39 2019 -0800 HIVE-21221 : Make HS2 and LLAP consistent - Bring up LLAP WebUI in test mode if WebUI port is configured (Oliver Draese via Deepak Jaiswal) --- .../java/org/apache/hadoop/hive/llap/daemon/impl/LlapDaemon.java| 6 +- 1 file changed, 5 insertions(+), 1 deletion(-) diff --git a/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/LlapDaemon.java b/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/LlapDaemon.java index 41eca8f..bafa029 100644 --- a/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/LlapDaemon.java +++ b/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/LlapDaemon.java @@ -313,7 +313,11 @@ public class LlapDaemon extends CompositeService implements ContainerRunner, Lla // Not adding the registry as a service, since we need to control when it is initialized - conf used to pickup properties. this.registry = new LlapRegistryService(true); -if (HiveConf.getBoolVar(daemonConf, HiveConf.ConfVars.HIVE_IN_TEST)) { +// disable web UI in test mode until a specific port was configured +if (HiveConf.getBoolVar(daemonConf, HiveConf.ConfVars.HIVE_IN_TEST) +&& Integer.parseInt(ConfVars.LLAP_DAEMON_WEB_PORT.getDefaultValue()) == webPort) { + LOG.info("Web UI was disabled in test mode because hive.llap.daemon.web.port was not " + + "specified or has default value ({})", webPort); this.webServices = null; } else { this.webServices = new LlapWebServices(webPort, this, registry);
[hive] branch master updated: HIVE-21167 : Bucketing: Bucketing version 1 is incorrectly partitioning data (Deepak Jaiswal, reviewed by Jason Dere and Vineet Garg)
This is an automated email from the ASF dual-hosted git repository. djaiswal pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/hive.git The following commit(s) were added to refs/heads/master by this push: new 69a7fc5 HIVE-21167 : Bucketing: Bucketing version 1 is incorrectly partitioning data (Deepak Jaiswal, reviewed by Jason Dere and Vineet Garg) 69a7fc5 is described below commit 69a7fc5955ceac1245233c0d297929e730750929 Author: Deepak Jaiswal AuthorDate: Wed Feb 13 22:04:33 2019 -0800 HIVE-21167 : Bucketing: Bucketing version 1 is incorrectly partitioning data (Deepak Jaiswal, reviewed by Jason Dere and Vineet Garg) --- .../apache/hadoop/hive/ql/parse/TezCompiler.java | 47 ++- .../queries/clientpositive/murmur_hash_migration.q | 35 +++ .../llap/dynpart_sort_opt_vectorization.q.out | 4 +- .../llap/murmur_hash_migration.q.out | 332 + 4 files changed, 392 insertions(+), 26 deletions(-) diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/TezCompiler.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/TezCompiler.java index 4b10e89..a917617 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/TezCompiler.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/TezCompiler.java @@ -174,9 +174,6 @@ public class TezCompiler extends TaskCompiler { runStatsAnnotation(procCtx); perfLogger.PerfLogEnd(this.getClass().getName(), PerfLogger.TEZ_COMPILER, "Setup stats in the operator plan"); -// Update bucketing version of ReduceSinkOp if needed -updateBucketingVersionForUpgrade(procCtx); - // run Sorted dynamic partition optimization if(HiveConf.getBoolVar(procCtx.conf, HiveConf.ConfVars.DYNAMICPARTITIONING) && HiveConf.getVar(procCtx.conf, HiveConf.ConfVars.DYNAMICPARTITIONINGMODE).equals("nonstrict") && @@ -229,6 +226,15 @@ public class TezCompiler extends TaskCompiler { new ConstantPropagate(ConstantPropagateOption.SHORTCUT).transform(procCtx.parseContext); } +// ATTENTION : DO NOT, I REPEAT, DO NOT WRITE ANYTHING AFTER updateBucketingVersionForUpgrade() +// ANYTHING WHICH NEEDS TO BE ADDED MUST BE ADDED ABOVE +// This call updates the bucketing version of final ReduceSinkOp based on +// the bucketing version of FileSinkOp. This operation must happen at the +// end to ensure there is no further rewrite of plan which may end up +// removing/updating the ReduceSinkOp as was the case with SortedDynPartitionOptimizer +// Update bucketing version of ReduceSinkOp if needed +updateBucketingVersionForUpgrade(procCtx); + } private void runCycleAnalysisForPartitionPruning(OptimizeTezProcContext procCtx, @@ -1833,30 +1839,23 @@ public class TezCompiler extends TaskCompiler { for (FileSinkOperator fsOp : fsOpsAll) { - Operator parentOfFS = fsOp.getParentOperators().get(0); - if (parentOfFS instanceof GroupByOperator) { -GroupByOperator gbyOp = (GroupByOperator) parentOfFS; -List aggs = gbyOp.getConf().getAggregatorStrings(); -boolean compute_stats = false; -for (String agg : aggs) { - if (agg.equalsIgnoreCase("compute_stats")) { -compute_stats = true; -break; - } -} -if (compute_stats) { + if (!fsOp.getConf().getTableInfo().isSetBucketingVersion()) { +continue; + } + // Look for direct parent ReduceSinkOp + // If there are more than 1 parent, bail out. + Operator parent = fsOp; + List> parentOps = parent.getParentOperators(); + while (parentOps != null && parentOps.size() == 1) { +parent = parentOps.get(0); +if (!(parent instanceof ReduceSinkOperator)) { + parentOps = parent.getParentOperators(); continue; } - } - // Not compute_stats - Set rsOps = OperatorUtils.findOperatorsUpstream(parentOfFS, ReduceSinkOperator.class); - if (rsOps.isEmpty()) { -continue; - } - // Skip setting if the bucketing version is not set in FileSinkOp. - if (fsOp.getConf().getTableInfo().isSetBucketingVersion()) { - rsOps.iterator().next().setBucketingVersion(fsOp.getConf().getTableInfo().getBucketingVersion()); +// Found the target RSOp + parent.setBucketingVersion(fsOp.getConf().getTableInfo().getBucketingVersion()); +break; } } } diff --git a/ql/src/test/queries/clientpositive/murmur_hash_migration.q b/ql/src/test/queries/clientpositive/murmur_hash_migration.q index 2b8da9f..6bf3785 100644 --- a/ql/src/test/queries/clientpositive/murmur_hash_migration.q +++ b/ql/src/test/queries/clientpositive/murmur_hash_migration.q @@ -59,3 +59,38 @@ select t1.key, t1.value, t2.key, t2.value from srcbucket_mapjoin_n18 t1, srcbuck explain select t1.key, t1.value, t2.key, t2.value from tab_part_n11 t1, tab_n10 t2
[hive] branch master updated: HIVE-21408 : Disable synthetic join predicates for non-equi joins for unintended cases (Deepak Jaiswal, reviewed by Vineet Garg)
This is an automated email from the ASF dual-hosted git repository. djaiswal pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/hive.git The following commit(s) were added to refs/heads/master by this push: new cdd8fa5 HIVE-21408 : Disable synthetic join predicates for non-equi joins for unintended cases (Deepak Jaiswal, reviewed by Vineet Garg) cdd8fa5 is described below commit cdd8fa58f4b7f6aa2973fb4a114bb891ed397416 Author: Deepak Jaiswal AuthorDate: Fri Mar 8 11:14:10 2019 -0800 HIVE-21408 : Disable synthetic join predicates for non-equi joins for unintended cases (Deepak Jaiswal, reviewed by Vineet Garg) --- .../DynamicPartitionPruningOptimization.java | 24 +++-- .../clientpositive/dynamic_semijoin_reduction.q| 3 + .../llap/dynamic_semijoin_reduction.q.out | 111 + 3 files changed, 129 insertions(+), 9 deletions(-) diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/DynamicPartitionPruningOptimization.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/DynamicPartitionPruningOptimization.java index 439fb75..cc01034 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/DynamicPartitionPruningOptimization.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/DynamicPartitionPruningOptimization.java @@ -167,7 +167,8 @@ public class DynamicPartitionPruningOptimization implements NodeProcessor { Table table = ts.getConf().getTableMetadata(); -if (table != null && table.isPartitionKey(column)) { +boolean nonEquiJoin = isNonEquiJoin(ctx.parent); +if (table != null && table.isPartitionKey(column) && !nonEquiJoin) { String columnType = table.getPartColByName(column).getType(); String alias = ts.getConf().getAlias(); PrunedPartitionList plist = parseContext.getPrunedPartitions(alias, ts); @@ -197,7 +198,7 @@ public class DynamicPartitionPruningOptimization implements NodeProcessor { } else { // semijoin LOG.debug("Column " + column + " is not a partition column"); if (semiJoin && !disableSemiJoinOptDueToExternalTable(parseContext.getConf(), ts, ctx) - && ts.getConf().getFilterExpr() != null) { + && ts.getConf().getFilterExpr() != null && !nonEquiJoin) { LOG.debug("Initiate semijoin reduction for " + column + " (" + ts.getConf().getFilterExpr().getExprString()); @@ -434,6 +435,18 @@ public class DynamicPartitionPruningOptimization implements NodeProcessor { } } + private boolean isNonEquiJoin(ExprNodeDesc predicate) { +Preconditions.checkArgument(predicate instanceof ExprNodeGenericFuncDesc); + +ExprNodeGenericFuncDesc funcDesc = (ExprNodeGenericFuncDesc) predicate; +if (funcDesc.getGenericUDF() instanceof GenericUDFIn) { + return false; +} + +return true; + } + + private void generateEventOperatorPlan(DynamicListContext ctx, ParseContext parseContext, TableScanOperator ts, String column, String columnType) { @@ -530,13 +543,6 @@ public class DynamicPartitionPruningOptimization implements NodeProcessor { TableScanOperator ts, String keyBaseAlias, String internalColName, String colName, SemiJoinHint sjHint) throws SemanticException { -// Semijoin reduction for non-equi join not yet supported, check for it -ExprNodeGenericFuncDesc funcDesc = (ExprNodeGenericFuncDesc) ctx.parent; -if (!(funcDesc.getGenericUDF() instanceof GenericUDFIn)) { - LOG.info("Semijoin reduction for non-equi joins is currently disabled."); - return false; -} - // we will put a fork in the plan at the source of the reduce sink Operator parentOfRS = ctx.generator.getParentOperators().get(0); diff --git a/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction.q b/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction.q index 32c2854..f5a5824 100644 --- a/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction.q +++ b/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction.q @@ -141,6 +141,9 @@ explain select * from alltypesorc_int_n1 join select srcpart_small_n3.key1 as key from srcpart_small_n3) unionsrc on (alltypesorc_int_n1.cstring = unionsrc.key); +-- Non equi joins : Should NOT create a semijoin branch +EXPLAIN select count(*) from srcpart_date_n7 join srcpart_small_n3 on (srcpart_date_n7.key = srcpart_small_n3.key1 and srcpart_date_n7.value > srcpart_small_n3.value1); + drop table srcpart_date_n7; drop table srcpart_small_n3; diff --git a/ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction.q.out b/ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction.q.out index 74bb1a0..d5e4a2a 100644 --- a/ql/src/te
hive git commit: HIVE-19162 : SMB : Test tez_smb_1.q stops making SMB join for a query (Deepak Jaiswal, reviewed by Ashutosh Chauhan)
Repository: hive Updated Branches: refs/heads/master 5213abf02 -> 3915980fe HIVE-19162 : SMB : Test tez_smb_1.q stops making SMB join for a query (Deepak Jaiswal, reviewed by Ashutosh Chauhan) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/3915980f Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/3915980f Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/3915980f Branch: refs/heads/master Commit: 3915980fe686d0c3edc375dd89587be78bb285fe Parents: 5213abf Author: Deepak Jaiswal Authored: Fri Apr 13 11:42:34 2018 -0700 Committer: Deepak Jaiswal Committed: Fri Apr 13 11:43:24 2018 -0700 -- .../results/clientpositive/llap/tez_smb_1.q.out | 176 +-- 1 file changed, 82 insertions(+), 94 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/3915980f/ql/src/test/results/clientpositive/llap/tez_smb_1.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/tez_smb_1.q.out b/ql/src/test/results/clientpositive/llap/tez_smb_1.q.out index 57e5e1a..faa94862 100644 --- a/ql/src/test/results/clientpositive/llap/tez_smb_1.q.out +++ b/ql/src/test/results/clientpositive/llap/tez_smb_1.q.out @@ -123,61 +123,49 @@ STAGE PLANS: Tez A masked pattern was here Edges: -Map 1 <- Map 3 (CUSTOM_EDGE) Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) A masked pattern was here Vertices: Map 1 Map Operator Tree: TableScan + alias: s3 + Statistics: Num rows: 242 Data size: 3490 Basic stats: COMPLETE Column stats: NONE + Filter Operator +predicate: key is not null (type: boolean) +Statistics: Num rows: 230 Data size: 3316 Basic stats: COMPLETE Column stats: NONE +Select Operator + expressions: key (type: int) + outputColumnNames: _col0 + Statistics: Num rows: 230 Data size: 3316 Basic stats: COMPLETE Column stats: NONE +Map Operator Tree: +TableScan alias: s1 - Statistics: Num rows: 242 Data size: 968 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 242 Data size: 3490 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: key is not null (type: boolean) -Statistics: Num rows: 242 Data size: 968 Basic stats: COMPLETE Column stats: COMPLETE +Statistics: Num rows: 230 Data size: 3316 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: int) outputColumnNames: _col0 - Statistics: Num rows: 242 Data size: 968 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator + Statistics: Num rows: 230 Data size: 3316 Basic stats: COMPLETE Column stats: NONE + Merge Join Operator condition map: Inner Join 0 to 1 keys: 0 _col0 (type: int) 1 _col0 (type: int) -input vertices: - 1 Map 3 -Statistics: Num rows: 382 Data size: 3056 Basic stats: COMPLETE Column stats: COMPLETE +Statistics: Num rows: 253 Data size: 3647 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: -Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE +Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint) Execution mode: llap -LLAP IO: no inputs -Map 3 -Map Operator Tree: -TableScan - alias: s3 - Statistics: Num rows: 242 Data size: 968 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator -predicate: key is n
[09/11] hive git commit: HIVE-19195, HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run llap specific tests in llap only.
http://git-wip-us.apache.org/repos/asf/hive/blob/77da6997/ql/src/test/results/clientpositive/auto_sortmerge_join_11.q.out -- diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_11.q.out b/ql/src/test/results/clientpositive/auto_sortmerge_join_11.q.out deleted file mode 100644 index a065887..000 --- a/ql/src/test/results/clientpositive/auto_sortmerge_join_11.q.out +++ /dev/null @@ -1,1625 +0,0 @@ -PREHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (KEY) INTO 2 BUCKETS STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@bucket_small -POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (KEY) INTO 2 BUCKETS STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@bucket_small -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY(KEY) INTO 4 BUCKETS STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@bucket_big -POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY(KEY) INTO 4 BUCKETS STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@bucket_big -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/00_0' INTO TABLE bucket_big partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_big -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/00_0' INTO TABLE bucket_big partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_big -POSTHOOK: Output: default@bucket_big@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/01_0' INTO TABLE bucket_big partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_big@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/01_0' INTO TABLE bucket_big partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_big@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/02_0' INTO TABLE bucket_big partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_big@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/02_0' INTO TABLE bucket_big partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_big@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/03_0' INTO TABLE bucket_big partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_big@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/03_0' INTO TABLE bucket_big partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_big@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/00_0' INTO TABLE bucket_big partition(ds='2008-04-09') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_big -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join
[01/11] hive git commit: HIVE-19195, HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run llap specific tests in llap only.
Repository: hive Updated Branches: refs/heads/master 1e9f3d77c -> 77da6997d http://git-wip-us.apache.org/repos/asf/hive/blob/77da6997/ql/src/test/results/clientpositive/tez_join_hash.q.out -- diff --git a/ql/src/test/results/clientpositive/tez_join_hash.q.out b/ql/src/test/results/clientpositive/tez_join_hash.q.out deleted file mode 100644 index c9b8169..000 --- a/ql/src/test/results/clientpositive/tez_join_hash.q.out +++ /dev/null @@ -1,921 +0,0 @@ -PREHOOK: query: create table orc_src (key string, value string) STORED AS ORC -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@orc_src -POSTHOOK: query: create table orc_src (key string, value string) STORED AS ORC -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@orc_src -PREHOOK: query: insert into table orc_src select * from src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@orc_src -POSTHOOK: query: insert into table orc_src select * from src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@orc_src -POSTHOOK: Lineage: orc_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: orc_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: explain -SELECT count(*) FROM src, orc_src where src.key=orc_src.key -PREHOOK: type: QUERY -POSTHOOK: query: explain -SELECT count(*) FROM src, orc_src where src.key=orc_src.key -POSTHOOK: type: QUERY -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 -Map Reduce - Map Operator Tree: - TableScan -alias: src -Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE -Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE - Select Operator -expressions: key (type: string) -outputColumnNames: _col0 -Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE -Reduce Output Operator - key expressions: _col0 (type: string) - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE - TableScan -alias: orc_src -Statistics: Num rows: 500 Data size: 88000 Basic stats: COMPLETE Column stats: NONE -Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 88000 Basic stats: COMPLETE Column stats: NONE - Select Operator -expressions: key (type: string) -outputColumnNames: _col0 -Statistics: Num rows: 500 Data size: 88000 Basic stats: COMPLETE Column stats: NONE -Reduce Output Operator - key expressions: _col0 (type: string) - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 88000 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: -Join Operator - condition map: - Inner Join 0 to 1 - keys: -0 _col0 (type: string) -1 _col0 (type: string) - Statistics: Num rows: 550 Data size: 5843 Basic stats: COMPLETE Column stats: NONE - Group By Operator -aggregations: count() -mode: hash -outputColumnNames: _col0 -Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE -File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 -Map Reduce - Map Operator Tree: - TableScan -Reduce Output Operator - sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: -Group By Operator - aggregations: count(VALUE._col0) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - File Out
[05/11] hive git commit: HIVE-19195, HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run llap specific tests in llap only.
http://git-wip-us.apache.org/repos/asf/hive/blob/77da6997/ql/src/test/results/clientpositive/auto_sortmerge_join_3.q.out -- diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_3.q.out b/ql/src/test/results/clientpositive/auto_sortmerge_join_3.q.out deleted file mode 100644 index 34959b4..000 --- a/ql/src/test/results/clientpositive/auto_sortmerge_join_3.q.out +++ /dev/null @@ -1,1153 +0,0 @@ -PREHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@bucket_small -POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@bucket_small -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small partition(ds='2008-04-09') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small partition(ds='2008-04-09') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small -POSTHOOK: Output: default@bucket_small@ds=2008-04-09 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small partition(ds='2008-04-09') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small@ds=2008-04-09 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small partition(ds='2008-04-09') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small@ds=2008-04-09 -PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@bucket_big -POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@bucket_big -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/00_0' INTO TABLE bucket_big partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_big -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/00_0' INTO TABLE bucket_big partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_big -POSTHOOK: Output: default@bucket_big@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/01_0' INTO TABLE bucket_big partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_big@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/01_0' INTO TABLE bucket_big partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_big@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/02_0' INTO TABLE bucket_big partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_big@ds=2008-04-08 -POSTHOOK: query:
[11/11] hive git commit: HIVE-19195, HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run llap specific tests in llap only.
HIVE-19195,HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run llap specific tests in llap only. Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/77da6997 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/77da6997 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/77da6997 Branch: refs/heads/master Commit: 77da6997d4357c1d88f3a966fe549f1f41efeab5 Parents: 1e9f3d7 Author: Deepak Jaiswal Authored: Mon Apr 16 11:08:30 2018 -0700 Committer: Deepak Jaiswal Committed: Mon Apr 16 11:08:58 2018 -0700 -- .../test/resources/testconfiguration.properties | 60 +- .../clientpositive/auto_sortmerge_join_1.q.out | 1324 -- .../clientpositive/auto_sortmerge_join_10.q.out | 620 - .../clientpositive/auto_sortmerge_join_11.q.out | 1625 - .../clientpositive/auto_sortmerge_join_12.q.out | 2295 -- .../clientpositive/auto_sortmerge_join_13.q.out | 775 -- .../clientpositive/auto_sortmerge_join_14.q.out | 298 --- .../clientpositive/auto_sortmerge_join_15.q.out | 266 -- .../clientpositive/auto_sortmerge_join_2.q.out | 1109 - .../clientpositive/auto_sortmerge_join_3.q.out | 1153 - .../clientpositive/auto_sortmerge_join_4.q.out | 1169 - .../clientpositive/auto_sortmerge_join_5.q.out | 856 --- .../clientpositive/auto_sortmerge_join_7.q.out | 1489 .../results/clientpositive/llap/llap_smb.q.out | 10 +- .../clientpositive/llap/tez_smb_main.q.out | 142 +- .../test/results/clientpositive/llap_acid.q.out | 401 --- .../results/clientpositive/llap_acid_fast.q.out | 395 --- .../results/clientpositive/tez_join_hash.q.out | 921 --- 18 files changed, 85 insertions(+), 14823 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/77da6997/itests/src/test/resources/testconfiguration.properties -- diff --git a/itests/src/test/resources/testconfiguration.properties b/itests/src/test/resources/testconfiguration.properties index f99d6a1..2845ab6 100644 --- a/itests/src/test/resources/testconfiguration.properties +++ b/itests/src/test/resources/testconfiguration.properties @@ -99,30 +99,12 @@ minillaplocal.shared.query.files=alter_merge_2_orc.q,\ auto_join30.q,\ auto_join_filters.q,\ auto_join_nulls.q,\ - auto_sortmerge_join_1.q,\ - auto_sortmerge_join_10.q,\ - auto_sortmerge_join_11.q,\ - auto_sortmerge_join_12.q,\ - auto_sortmerge_join_13.q,\ - auto_sortmerge_join_14.q,\ - auto_sortmerge_join_15.q,\ - auto_sortmerge_join_16.q,\ - auto_sortmerge_join_2.q,\ - auto_sortmerge_join_3.q,\ - auto_sortmerge_join_4.q,\ - auto_sortmerge_join_5.q,\ - auto_sortmerge_join_6.q,\ - auto_sortmerge_join_7.q,\ - auto_sortmerge_join_8.q,\ - auto_sortmerge_join_9.q,\ autoColumnStats_1.q,\ autoColumnStats_10.q,\ autoColumnStats_2.q,\ bucket2.q,\ bucket3.q,\ bucket4.q,\ - bucket_map_join_tez1.q,\ - bucket_map_join_tez2.q,\ cbo_gby.q,\ cbo_gby_empty.q,\ cbo_join.q,\ @@ -169,8 +151,6 @@ minillaplocal.shared.query.files=alter_merge_2_orc.q,\ delete_where_partitioned.q,\ delete_whole_partition.q,\ disable_merge_for_bucketing.q,\ - dynamic_partition_pruning.q,\ - dynamic_semijoin_reduction.q,\ dynpart_sort_opt_vectorization.q,\ dynpart_sort_optimization.q,\ dynpart_sort_optimization2.q,\ @@ -200,8 +180,6 @@ minillaplocal.shared.query.files=alter_merge_2_orc.q,\ kill_query.q,\ leftsemijoin.q,\ limit_pushdown.q,\ - llap_acid.q,\ - llap_acid_fast.q,\ load_dyn_part1.q,\ load_dyn_part2.q,\ load_dyn_part3.q,\ @@ -253,29 +231,6 @@ minillaplocal.shared.query.files=alter_merge_2_orc.q,\ subquery_exists.q,\ subquery_in.q,\ temp_table.q,\ - tez_bmj_schema_evolution.q,\ - tez_dml.q,\ - tez_dynpart_hashjoin_1.q,\ - tez_dynpart_hashjoin_2.q,\ - tez_fsstat.q,\ - tez_insert_overwrite_local_directory_1.q,\ - tez_join.q,\ - tez_join_hash.q,\ - tez_join_result_complex.q,\ - tez_join_tests.q,\ - tez_joins_explain.q,\ - tez_multi_union.q,\ - tez_schema_evolution.q,\ - tez_self_join.q,\ - tez_smb_1.q,\ - tez_smb_main.q,\ - tez_union.q,\ - tez_union2.q,\ - tez_union_decimal.q,\ - tez_union_group_by.q,\ - tez_union_multiinsert.q,\ - tez_vector_dynpart_hashjoin_1.q,\ - tez_vector_dynpart_hashjoin_2.q,\ union2.q,\ union3.q,\ union4.q,\ @@ -448,7 +403,6 @@ minillap.query.files=acid_bucket_pruning.q,\ intersect_all.q,\ intersect_distinct.q,\ intersect_merge.q,\ - llap_smb.q,\ llap_udf.q,\ llapdecider.q,\ reduce_deduplicate.q,\ @@ -494,8 +448,20 @@ minillaplocal.query.files=\ auto_join30.q,\ auto_join_filters.q,\ auto_join_nulls.q,\ + auto_sortmerge_join_1.q,\ + auto_sortmerge_join_10.q,\ + auto_sortmerge_
[04/11] hive git commit: HIVE-19195, HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run llap specific tests in llap only.
http://git-wip-us.apache.org/repos/asf/hive/blob/77da6997/ql/src/test/results/clientpositive/auto_sortmerge_join_4.q.out -- diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_4.q.out b/ql/src/test/results/clientpositive/auto_sortmerge_join_4.q.out deleted file mode 100644 index 0d586fd..000 --- a/ql/src/test/results/clientpositive/auto_sortmerge_join_4.q.out +++ /dev/null @@ -1,1169 +0,0 @@ -PREHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@bucket_small -POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@bucket_small -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/02_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/02_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/03_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/03_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small partition(ds='2008-04-09') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small partition(ds='2008-04-09') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small -POSTHOOK: Output: default@bucket_small@ds=2008-04-09 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small partition(ds='2008-04-09') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small@ds=2008-04-09 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small partition(ds='2008-04-09') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small@ds=2008-04-09 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/02_0' INTO TABLE bucket_small partition(ds='2008-04-09') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small@ds=2008-04-09 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/02_0' INTO TABLE bucket_small partition(ds='2008-04-09') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small@ds=2008-04-09 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/03_0' INTO TABLE bucket_small partition(ds='2008-04-09') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small@ds=2008-04-09 -POSTHOOK: query: load data
[06/11] hive git commit: HIVE-19195, HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run llap specific tests in llap only.
http://git-wip-us.apache.org/repos/asf/hive/blob/77da6997/ql/src/test/results/clientpositive/auto_sortmerge_join_2.q.out -- diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_2.q.out b/ql/src/test/results/clientpositive/auto_sortmerge_join_2.q.out deleted file mode 100644 index 5cfc35a..000 --- a/ql/src/test/results/clientpositive/auto_sortmerge_join_2.q.out +++ /dev/null @@ -1,1109 +0,0 @@ -PREHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@bucket_small -POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@bucket_small -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/02_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/02_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/03_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/03_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@bucket_big -POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@bucket_big -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/00_0' INTO TABLE bucket_big partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_big -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/00_0' INTO TABLE bucket_big partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_big -POSTHOOK: Output: default@bucket_big@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/01_0' INTO TABLE bucket_big partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_big@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/01_0' INTO TABLE bucket_big partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_big@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/00_0' INTO TABLE bucket_big partition(ds='2008-04-09') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_big -POSTHOOK: query: load data local inpath '../../data/fil
[10/11] hive git commit: HIVE-19195, HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run llap specific tests in llap only.
http://git-wip-us.apache.org/repos/asf/hive/blob/77da6997/ql/src/test/results/clientpositive/auto_sortmerge_join_10.q.out -- diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_10.q.out b/ql/src/test/results/clientpositive/auto_sortmerge_join_10.q.out deleted file mode 100644 index 3a90b05..000 --- a/ql/src/test/results/clientpositive/auto_sortmerge_join_10.q.out +++ /dev/null @@ -1,620 +0,0 @@ -PREHOOK: query: CREATE TABLE tbl1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@tbl1 -POSTHOOK: query: CREATE TABLE tbl1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@tbl1 -PREHOOK: query: CREATE TABLE tbl2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@tbl2 -POSTHOOK: query: CREATE TABLE tbl2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@tbl2 -PREHOOK: query: insert overwrite table tbl1 -select * from src where key < 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@tbl1 -POSTHOOK: query: insert overwrite table tbl1 -select * from src where key < 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@tbl1 -POSTHOOK: Lineage: tbl1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: tbl1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: insert overwrite table tbl2 -select * from src where key < 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@tbl2 -POSTHOOK: query: insert overwrite table tbl2 -select * from src where key < 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@tbl2 -POSTHOOK: Lineage: tbl2.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: tbl2.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: explain -select count(*) from - ( - select * from - (select a.key as key, a.value as value from tbl1 a where key < 6 - union all - select a.key as key, a.value as value from tbl1 a where key < 6 - ) usubq1 ) subq1 -join - (select a.key as key, a.value as value from tbl2 a where key < 6) subq2 - on subq1.key = subq2.key -PREHOOK: type: QUERY -POSTHOOK: query: explain -select count(*) from - ( - select * from - (select a.key as key, a.value as value from tbl1 a where key < 6 - union all - select a.key as key, a.value as value from tbl1 a where key < 6 - ) usubq1 ) subq1 -join - (select a.key as key, a.value as value from tbl2 a where key < 6) subq2 - on subq1.key = subq2.key -POSTHOOK: type: QUERY -STAGE DEPENDENCIES: - Stage-7 is a root stage , consists of Stage-8, Stage-9, Stage-1 - Stage-8 has a backup stage: Stage-1 - Stage-5 depends on stages: Stage-8 - Stage-2 depends on stages: Stage-1, Stage-5, Stage-6 - Stage-9 has a backup stage: Stage-1 - Stage-6 depends on stages: Stage-9 - Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-7 -Conditional Operator - - Stage: Stage-8 -Map Reduce Local Work - Alias -> Map Local Tables: -$hdt$_1:a - Fetch Operator -limit: -1 - Alias -> Map Local Operator Tree: -$hdt$_1:a - TableScan -alias: a -Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE -Filter Operator - predicate: (key < 6) (type: boolean) - Statistics: Num rows: 3 Data size: 21 Basic stats: COMPLETE Column stats: NONE - Select Operator -expressions: key (type: int) -outputColumnNames: _col0 -Statistics: Num rows: 3 Data size: 21 Basic stats: COMPLETE Column stats: NONE -HashTable Sink Operator - keys: -0 _col0 (type: int) -1 _col0 (type: int) - - Stage: Stage-5 -Map Reduce - Map Operator Tree: - TableScan -alias: a -Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE -Filter Operator - predicate: (key < 6) (type: boolean) - Statistics: Num rows: 3 Data size: 21 Basic stats: COMPLETE Column stats: NONE - Select Operator -expressions: key (type: int) -outputColumnNames: _col0 -Statistics: Num rows: 3 Data size: 21 Basic stats: COM
[02/11] hive git commit: HIVE-19195, HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run llap specific tests in llap only.
http://git-wip-us.apache.org/repos/asf/hive/blob/77da6997/ql/src/test/results/clientpositive/llap/tez_smb_main.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/tez_smb_main.q.out b/ql/src/test/results/clientpositive/llap/tez_smb_main.q.out index 705bcc3..0dcca1f 100644 --- a/ql/src/test/results/clientpositive/llap/tez_smb_main.q.out +++ b/ql/src/test/results/clientpositive/llap/tez_smb_main.q.out @@ -263,33 +263,33 @@ STAGE PLANS: Map Operator Tree: TableScan alias: a - Statistics: Num rows: 242 Data size: 968 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 242 Data size: 3490 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: key is not null (type: boolean) -Statistics: Num rows: 242 Data size: 968 Basic stats: COMPLETE Column stats: COMPLETE +Statistics: Num rows: 230 Data size: 3316 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: int) outputColumnNames: _col0 - Statistics: Num rows: 242 Data size: 968 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 230 Data size: 3316 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + Map-reduce partition columns: _col0 (type: int) -Statistics: Num rows: 242 Data size: 968 Basic stats: COMPLETE Column stats: COMPLETE +Statistics: Num rows: 230 Data size: 3316 Basic stats: COMPLETE Column stats: NONE Execution mode: llap LLAP IO: no inputs Map 2 Map Operator Tree: TableScan alias: b - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 500 Data size: 7216 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: key is not null (type: boolean) -Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE +Statistics: Num rows: 475 Data size: 6855 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: int) outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 475 Data size: 6855 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -298,15 +298,15 @@ STAGE PLANS: 1 _col0 (type: int) input vertices: 0 Map 1 -Statistics: Num rows: 391 Data size: 3128 Basic stats: COMPLETE Column stats: COMPLETE +Statistics: Num rows: 522 Data size: 7540 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: -Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE +Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint) Execution mode: llap LLAP IO: no inputs @@ -317,10 +317,10 @@ STAGE PLANS: aggregations: count(VALUE._col0) mode: mergepartial outputColumnNames: _col0 -Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE +Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE table:
[08/11] hive git commit: HIVE-19195, HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run llap specific tests in llap only.
http://git-wip-us.apache.org/repos/asf/hive/blob/77da6997/ql/src/test/results/clientpositive/auto_sortmerge_join_12.q.out -- diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_12.q.out b/ql/src/test/results/clientpositive/auto_sortmerge_join_12.q.out deleted file mode 100644 index 96b0016..000 --- a/ql/src/test/results/clientpositive/auto_sortmerge_join_12.q.out +++ /dev/null @@ -1,2295 +0,0 @@ -PREHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) -CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@bucket_small -POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) -CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@bucket_small -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@bucket_big -POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@bucket_big -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/00_0' INTO TABLE bucket_big partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_big -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/00_0' INTO TABLE bucket_big partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_big -POSTHOOK: Output: default@bucket_big@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/01_0' INTO TABLE bucket_big partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_big@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/01_0' INTO TABLE bucket_big partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_big@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/02_0' INTO TABLE bucket_big partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_big@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/02_0' INTO TABLE bucket_big partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_big@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/03_0' INTO TABLE bucket_big partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_big@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/03_0' INTO TABLE bucket_big partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_big@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/big/00_0' INTO TABLE bucket_big partition(ds='2008-04-09') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_big -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_jo
[07/11] hive git commit: HIVE-19195, HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run llap specific tests in llap only.
http://git-wip-us.apache.org/repos/asf/hive/blob/77da6997/ql/src/test/results/clientpositive/auto_sortmerge_join_13.q.out -- diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_13.q.out b/ql/src/test/results/clientpositive/auto_sortmerge_join_13.q.out deleted file mode 100644 index ac143dd..000 --- a/ql/src/test/results/clientpositive/auto_sortmerge_join_13.q.out +++ /dev/null @@ -1,775 +0,0 @@ -PREHOOK: query: CREATE TABLE tbl1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@tbl1 -POSTHOOK: query: CREATE TABLE tbl1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@tbl1 -PREHOOK: query: CREATE TABLE tbl2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@tbl2 -POSTHOOK: query: CREATE TABLE tbl2(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@tbl2 -PREHOOK: query: insert overwrite table tbl1 select * from src where key < 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@tbl1 -POSTHOOK: query: insert overwrite table tbl1 select * from src where key < 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@tbl1 -POSTHOOK: Lineage: tbl1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: tbl1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: insert overwrite table tbl2 select * from src where key < 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@tbl2 -POSTHOOK: query: insert overwrite table tbl2 select * from src where key < 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@tbl2 -POSTHOOK: Lineage: tbl2.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: tbl2.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: CREATE TABLE dest1(k1 int, k2 int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest1 -POSTHOOK: query: CREATE TABLE dest1(k1 int, k2 int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest1 -PREHOOK: query: CREATE TABLE dest2(k1 string, k2 string) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest2 -POSTHOOK: query: CREATE TABLE dest2(k1 string, k2 string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest2 -PREHOOK: query: explain -from ( - SELECT a.key key1, a.value value1, b.key key2, b.value value2 - FROM tbl1 a JOIN tbl2 b - ON a.key = b.key ) subq -INSERT OVERWRITE TABLE dest1 select key1, key2 -INSERT OVERWRITE TABLE dest2 select value1, value2 -PREHOOK: type: QUERY -POSTHOOK: query: explain -from ( - SELECT a.key key1, a.value value1, b.key key2, b.value value2 - FROM tbl1 a JOIN tbl2 b - ON a.key = b.key ) subq -INSERT OVERWRITE TABLE dest1 select key1, key2 -INSERT OVERWRITE TABLE dest2 select value1, value2 -POSTHOOK: type: QUERY -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-5 - Stage-1 depends on stages: Stage-2 - Stage-4 depends on stages: Stage-1, Stage-5 - Stage-5 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-2 -Map Reduce - Map Operator Tree: - TableScan -alias: a -Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE -Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE - Select Operator -expressions: key (type: int), value (type: string) -outputColumnNames: _col0, _col1 -Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE -Sorted Merge Bucket Map Join Operator - condition map: - Inner Join 0 to 1 - keys: -0 _col0 (type: int) -1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3 - Select Operator -expressions: _col0 (type: int), _col2 (type: int) -outputColumnNames: _col0, _col1 -File Output Operator - compressed: false -
[03/11] hive git commit: HIVE-19195, HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run llap specific tests in llap only.
http://git-wip-us.apache.org/repos/asf/hive/blob/77da6997/ql/src/test/results/clientpositive/auto_sortmerge_join_7.q.out -- diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_7.q.out b/ql/src/test/results/clientpositive/auto_sortmerge_join_7.q.out deleted file mode 100644 index 1959075..000 --- a/ql/src/test/results/clientpositive/auto_sortmerge_join_7.q.out +++ /dev/null @@ -1,1489 +0,0 @@ -PREHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@bucket_small -POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@bucket_small -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/02_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/02_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/03_0' INTO TABLE bucket_small partition(ds='2008-04-08') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small@ds=2008-04-08 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/03_0' INTO TABLE bucket_small partition(ds='2008-04-08') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small@ds=2008-04-08 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small partition(ds='2008-04-09') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small partition(ds='2008-04-09') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small -POSTHOOK: Output: default@bucket_small@ds=2008-04-09 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small partition(ds='2008-04-09') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small@ds=2008-04-09 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small partition(ds='2008-04-09') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small@ds=2008-04-09 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/02_0' INTO TABLE bucket_small partition(ds='2008-04-09') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small@ds=2008-04-09 -POSTHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/02_0' INTO TABLE bucket_small partition(ds='2008-04-09') -POSTHOOK: type: LOAD - A masked pattern was here -POSTHOOK: Output: default@bucket_small@ds=2008-04-09 -PREHOOK: query: load data local inpath '../../data/files/auto_sortmerge_join/small/03_0' INTO TABLE bucket_small partition(ds='2008-04-09') -PREHOOK: type: LOAD - A masked pattern was here -PREHOOK: Output: default@bucket_small@ds=2008-04-09 -POSTHOOK: query: load data
hive git commit: HIVE-17055 : Flaky test: TestMiniLlapCliDriver.testCliDriver[llap_smb] (Deepak Jaiswal, reviewed by Vineet Garg)
Repository: hive Updated Branches: refs/heads/master a94f382a0 -> f6d2e6e08 HIVE-17055 : Flaky test: TestMiniLlapCliDriver.testCliDriver[llap_smb] (Deepak Jaiswal, reviewed by Vineet Garg) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/f6d2e6e0 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/f6d2e6e0 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/f6d2e6e0 Branch: refs/heads/master Commit: f6d2e6e0825ba1415882d0f7ddcc09ee6911cbd7 Parents: a94f382 Author: Deepak Jaiswal Authored: Fri Apr 20 17:00:32 2018 -0700 Committer: Deepak Jaiswal Committed: Fri Apr 20 17:00:32 2018 -0700 -- itests/src/test/resources/testconfiguration.properties | 7 +++ 1 file changed, 3 insertions(+), 4 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/f6d2e6e0/itests/src/test/resources/testconfiguration.properties -- diff --git a/itests/src/test/resources/testconfiguration.properties b/itests/src/test/resources/testconfiguration.properties index 4e7c519..3aaa68b 100644 --- a/itests/src/test/resources/testconfiguration.properties +++ b/itests/src/test/resources/testconfiguration.properties @@ -83,10 +83,7 @@ minillap.shared.query.files=insert_into1.q,\ orc_struct_type_vectorization.q,\ union_type_chk.q,\ cte_2.q,\ - cte_4.q,\ - llap_nullscan.q,\ - dynamic_partition_pruning_2.q,\ - tez_union_dynamic_partition.q + cte_4.q minillaplocal.shared.query.files=alter_merge_2_orc.q,\ alter_merge_orc.q,\ @@ -552,6 +549,7 @@ minillaplocal.query.files=\ llap_acid_fast.q,\ llap_acid2.q,\ llap_partitioned.q,\ + llap_smb.q,\ llap_vector_nohybridgrace.q,\ load_data_acid_rename.q,\ load_dyn_part5.q,\ @@ -687,6 +685,7 @@ minillaplocal.query.files=\ tez_fsstat.q,\ tez_insert_overwrite_local_directory_1.q,\ tez_join.q,\ + tez_join_hash.q,\ tez_join_result_complex.q,\ tez_join_tests.q,\ tez_joins_explain.q,\
hive git commit: Preparing for storage-api 2.7.0 release.
Repository: hive Updated Branches: refs/heads/storage-branch-2.7 [created] 13802a2cb Preparing for storage-api 2.7.0 release. Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/13802a2c Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/13802a2c Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/13802a2c Branch: refs/heads/storage-branch-2.7 Commit: 13802a2cbdfbd22532cb80ae00cb42303f036dea Parents: eae5225 Author: Deepak Jaiswal Authored: Fri Jul 6 14:02:38 2018 -0700 Committer: Deepak Jaiswal Committed: Fri Jul 6 14:02:38 2018 -0700 -- storage-api/pom.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/13802a2c/storage-api/pom.xml -- diff --git a/storage-api/pom.xml b/storage-api/pom.xml index 06c10d0..d7cd0fb 100644 --- a/storage-api/pom.xml +++ b/storage-api/pom.xml @@ -25,7 +25,7 @@ org.apache.hive hive-storage-api - 2.7.0-SNAPSHOT + 2.7.0 jar Hive Storage API
[hive] Git Push Summary
Repository: hive Updated Tags: refs/tags/storage-release-2.7.0-rc0 [created] df9693a66
hive git commit: HIVE-20100 : OpTraits : Select Optraits should stop when a mismatch is detected (Deepak Jaiswal, reviewed by Jason Dere)
Repository: hive Updated Branches: refs/heads/master 5016d6f05 -> ddf7e25d3 HIVE-20100 : OpTraits : Select Optraits should stop when a mismatch is detected (Deepak Jaiswal, reviewed by Jason Dere) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/ddf7e25d Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/ddf7e25d Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/ddf7e25d Branch: refs/heads/master Commit: ddf7e25d3ee94224c17e353e334728241515fffb Parents: 5016d6f Author: Deepak Jaiswal Authored: Mon Jul 9 15:31:13 2018 -0700 Committer: Deepak Jaiswal Committed: Mon Jul 9 15:31:13 2018 -0700 -- .../annotation/OpTraitsRulesProcFactory.java| 59 +- .../clientpositive/llap/subquery_notin.q.out| 118 --- .../results/clientpositive/llap/tez_join.q.out | 51 +--- 3 files changed, 138 insertions(+), 90 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/ddf7e25d/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java index 263770e..89db530 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java @@ -308,44 +308,40 @@ public class OpTraitsRulesProcFactory { public static class SelectRule implements NodeProcessor { -boolean processSortCols = false; - // For bucket columns // If all the columns match to the parent, put them in the bucket cols // else, add empty list. // For sort columns // Keep the subset of all the columns as long as order is maintained. public List> getConvertedColNames( -List> parentColNames, SelectOperator selOp) { +List> parentColNames, SelectOperator selOp, boolean processSortCols) { List> listBucketCols = new ArrayList<>(); - if (selOp.getColumnExprMap() != null) { -if (parentColNames != null) { - for (List colNames : parentColNames) { -List bucketColNames = new ArrayList<>(); -boolean found = false; -for (String colName : colNames) { - for (Entry entry : selOp.getColumnExprMap().entrySet()) { -if ((entry.getValue() instanceof ExprNodeColumnDesc) && -(((ExprNodeColumnDesc) (entry.getValue())).getColumn().equals(colName))) { - bucketColNames.add(entry.getKey()); - found = true; - break; -} - } - if (!found) { -// Bail out on first missed column. -break; - } -} -if (!processSortCols && !found) { - // While processing bucket columns, atleast one bucket column - // missed. This results in a different bucketing scheme. - // Add empty list - listBucketCols.add(new ArrayList<>()); -} else { - listBucketCols.add(bucketColNames); + for (List colNames : parentColNames) { +List bucketColNames = new ArrayList<>(); +boolean found = false; +for (String colName : colNames) { + // Reset found + found = false; + for (Entry entry : selOp.getColumnExprMap().entrySet()) { +if ((entry.getValue() instanceof ExprNodeColumnDesc) && +(((ExprNodeColumnDesc) (entry.getValue())).getColumn().equals(colName))) { + bucketColNames.add(entry.getKey()); + found = true; + break; } } + if (!found) { +// Bail out on first missed column. +break; + } +} +if (!processSortCols && !found) { + // While processing bucket columns, atleast one bucket column + // missed. This results in a different bucketing scheme. + // Add empty list + listBucketCols.add(new ArrayList<>()); +} else { + listBucketCols.add(bucketColNames); } } @@ -363,13 +359,12 @@ public class OpTraitsRulesProcFactory { List> listSortCols = null; if (selOp.getColumnExprMap() != null) { if (parentBucketColNames != null) { - listBucketCols = getConvertedColNames(parentBucketColNames, selOp); + listBucketCols = getConvertedColNames(parentBucketColNames, selOp, false); } List> parentSortC
hive git commit: HIVE-20100 : OpTraits : Select Optraits should stop when a mismatch is detected (Deepak Jaiswal, reviewed by Jason Dere)
Repository: hive Updated Branches: refs/heads/branch-3 c462214aa -> 6c55afb04 HIVE-20100 : OpTraits : Select Optraits should stop when a mismatch is detected (Deepak Jaiswal, reviewed by Jason Dere) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/6c55afb0 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/6c55afb0 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/6c55afb0 Branch: refs/heads/branch-3 Commit: 6c55afb046daa5c28c4fb1e2a573a4a05e3f5572 Parents: c462214 Author: Deepak Jaiswal Authored: Mon Jul 9 15:31:13 2018 -0700 Committer: Deepak Jaiswal Committed: Tue Jul 10 22:10:58 2018 -0700 -- .../annotation/OpTraitsRulesProcFactory.java| 59 +- .../clientpositive/llap/subquery_notin.q.out| 118 --- .../results/clientpositive/llap/tez_join.q.out | 51 +--- 3 files changed, 138 insertions(+), 90 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/6c55afb0/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java index 263770e..89db530 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java @@ -308,44 +308,40 @@ public class OpTraitsRulesProcFactory { public static class SelectRule implements NodeProcessor { -boolean processSortCols = false; - // For bucket columns // If all the columns match to the parent, put them in the bucket cols // else, add empty list. // For sort columns // Keep the subset of all the columns as long as order is maintained. public List> getConvertedColNames( -List> parentColNames, SelectOperator selOp) { +List> parentColNames, SelectOperator selOp, boolean processSortCols) { List> listBucketCols = new ArrayList<>(); - if (selOp.getColumnExprMap() != null) { -if (parentColNames != null) { - for (List colNames : parentColNames) { -List bucketColNames = new ArrayList<>(); -boolean found = false; -for (String colName : colNames) { - for (Entry entry : selOp.getColumnExprMap().entrySet()) { -if ((entry.getValue() instanceof ExprNodeColumnDesc) && -(((ExprNodeColumnDesc) (entry.getValue())).getColumn().equals(colName))) { - bucketColNames.add(entry.getKey()); - found = true; - break; -} - } - if (!found) { -// Bail out on first missed column. -break; - } -} -if (!processSortCols && !found) { - // While processing bucket columns, atleast one bucket column - // missed. This results in a different bucketing scheme. - // Add empty list - listBucketCols.add(new ArrayList<>()); -} else { - listBucketCols.add(bucketColNames); + for (List colNames : parentColNames) { +List bucketColNames = new ArrayList<>(); +boolean found = false; +for (String colName : colNames) { + // Reset found + found = false; + for (Entry entry : selOp.getColumnExprMap().entrySet()) { +if ((entry.getValue() instanceof ExprNodeColumnDesc) && +(((ExprNodeColumnDesc) (entry.getValue())).getColumn().equals(colName))) { + bucketColNames.add(entry.getKey()); + found = true; + break; } } + if (!found) { +// Bail out on first missed column. +break; + } +} +if (!processSortCols && !found) { + // While processing bucket columns, atleast one bucket column + // missed. This results in a different bucketing scheme. + // Add empty list + listBucketCols.add(new ArrayList<>()); +} else { + listBucketCols.add(bucketColNames); } } @@ -363,13 +359,12 @@ public class OpTraitsRulesProcFactory { List> listSortCols = null; if (selOp.getColumnExprMap() != null) { if (parentBucketColNames != null) { - listBucketCols = getConvertedColNames(parentBucketColNames, selOp); + listBucketCols = getConvertedColNames(parentBucketColNames, selOp, false); } List> parent
[hive] Git Push Summary
Repository: hive Updated Branches: refs/heads/storage-branch-2.7 [deleted] 13802a2cb
hive git commit: Preparing for storage-api 2.7.0 release
Repository: hive Updated Branches: refs/heads/storage-branch-2.7 [created] e59fdf986 Preparing for storage-api 2.7.0 release Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/e59fdf98 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/e59fdf98 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/e59fdf98 Branch: refs/heads/storage-branch-2.7 Commit: e59fdf9861d20747f1e69623a73782979258d61d Parents: 8fd90ea Author: Deepak Jaiswal Authored: Thu Jul 12 15:53:28 2018 -0700 Committer: Deepak Jaiswal Committed: Thu Jul 12 15:53:28 2018 -0700 -- storage-api/pom.xml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/e59fdf98/storage-api/pom.xml -- diff --git a/storage-api/pom.xml b/storage-api/pom.xml index a40feff..5d3c7d4 100644 --- a/storage-api/pom.xml +++ b/storage-api/pom.xml @@ -25,7 +25,7 @@ org.apache.hive hive-storage-api - 2.6.1-SNAPSHOT + 2.7.0 jar Hive Storage API
[hive] Git Push Summary
Repository: hive Updated Tags: refs/tags/storage-release-2.7.0-rc1 [created] 7b909acc5
[hive] Git Push Summary
Repository: hive Updated Tags: refs/tags/rel/storage-release-2.7.0 [created] 5d3241345
[02/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/union_remove_5.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/union_remove_5.q.out b/ql/src/test/results/clientpositive/spark/union_remove_5.q.out index 92273f8..95b4de3 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_5.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_5.q.out @@ -234,6 +234,7 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 numFiles4 totalSize 68 A masked pattern was here http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/union_remove_7.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/union_remove_7.q.out b/ql/src/test/results/clientpositive/spark/union_remove_7.q.out index b0f1818..ed4559b 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_7.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_7.q.out @@ -158,8 +158,9 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 numFiles4 - totalSize 332 + totalSize 336 A masked pattern was here # Storage Information http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/union_remove_8.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/union_remove_8.q.out b/ql/src/test/results/clientpositive/spark/union_remove_8.q.out index 63dd553..5d51136 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_8.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_8.q.out @@ -184,8 +184,9 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 numFiles4 - totalSize 348 + totalSize 350 A masked pattern was here # Storage Information http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/union_remove_9.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/union_remove_9.q.out b/ql/src/test/results/clientpositive/spark/union_remove_9.q.out index 4dc614f..2d8c21d 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_9.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_9.q.out @@ -242,8 +242,9 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 numFiles4 - totalSize 348 + totalSize 350 A masked pattern was here # Storage Information http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/vectorization_0.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/vectorization_0.q.out b/ql/src/test/results/clientpositive/spark/vectorization_0.q.out index c3201bf..8b73ef2 100644 --- a/ql/src/test/results/clientpositive/spark/vectorization_0.q.out +++ b/ql/src/test/results/clientpositive/spark/vectorization_0.q.out @@ -1817,6 +1817,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 columns.comments @@ -1838,6 +1839,7 @@ STAGE PLANS:
[15/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/vectorization_0.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/vectorization_0.q.out b/ql/src/test/results/clientpositive/llap/vectorization_0.q.out index b2db5a5..80da5a9 100644 --- a/ql/src/test/results/clientpositive/llap/vectorization_0.q.out +++ b/ql/src/test/results/clientpositive/llap/vectorization_0.q.out @@ -1839,6 +1839,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 columns.comments @@ -1860,6 +1861,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 columns.comments @@ -30660,6 +30662,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 columns.comments @@ -30681,6 +30684,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 columns.comments @@ -30775,6 +30779,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 columns.comments @@ -30796,6 +30801,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 columns.comments @@ -30890,6 +30896,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} bucket_count -1 +bucketing_version
[21/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/orc_create.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/orc_create.q.out b/ql/src/test/results/clientpositive/llap/orc_create.q.out index 72d94e5..df286b2 100644 --- a/ql/src/test/results/clientpositive/llap/orc_create.q.out +++ b/ql/src/test/results/clientpositive/llap/orc_create.q.out @@ -66,6 +66,7 @@ Retention:0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"lst\":\"true\",\"mp\":\"true\",\"str\":\"true\",\"strct\":\"true\"}} + bucketing_version 2 numFiles0 numRows 0 rawDataSize 0 @@ -119,6 +120,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + bucketing_version 2 numFiles0 numPartitions 0 numRows 0 @@ -176,6 +178,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + bucketing_version 2 numFiles0 numPartitions 0 numRows 0 @@ -223,6 +226,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + bucketing_version 2 A masked pattern was here numFiles0 numPartitions 0 @@ -281,6 +285,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + bucketing_version 2 numFiles0 numPartitions 0 numRows 0 @@ -336,6 +341,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"lst\":\"true\",\"mp\":\"true\",\"str\":\"true\",\"strct\":\"true\"}} + bucketing_version 2 numFiles0 numRows 0 rawDataSize 0 http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/orc_llap_counters.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/orc_llap_counters.q.out b/ql/src/test/results/clientpositive/llap/orc_llap_counters.q.out index af237ea..98d9e06 100644 --- a/ql/src/test/results/clientpositive/llap/orc_llap_counters.q.out +++ b/ql/src/test/results/clientpositive/llap/orc_llap_counters.q.out @@ -227,6 +227,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + bucketing_version 2 numFiles1 numRows 2100 orc.bloom.filter.columns* http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/orc_llap_counters1.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/orc_llap_counters1.q.out b/ql/src/test/results/clientpositive/llap/orc_llap_counters1.q.out index 01fb4af..c078326 100644 --- a/ql/src/test/results/clientpositive/llap/orc_llap_counters1.q.out +++ b/ql/src/test/results/clientpositive/llap/orc_llap_counters1.q.out @@ -227,6 +227,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + bucketing_version 2 numFiles1 numRows 2100 orc.bloom.filter.columns* http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/ll
[24/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/cte_1.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/cte_1.q.out b/ql/src/test/results/clientpositive/llap/cte_1.q.out index d7bc062..d98862c 100644 --- a/ql/src/test/results/clientpositive/llap/cte_1.q.out +++ b/ql/src/test/results/clientpositive/llap/cte_1.q.out @@ -159,178 +159,217 @@ POSTHOOK: Input: default@src 10 val_10 val_10 100val_100 val_100 100val_100 val_100 +103val_103 val_103 +103val_103 val_103 104val_104 val_104 104val_104 val_104 -111val_111 val_111 -114val_114 val_114 -116val_116 val_116 +105val_105 val_105 +11 val_11 val_11 126val_126 val_126 128val_128 val_128 128val_128 val_128 128val_128 val_128 -131val_131 val_131 +129val_129 val_129 +129val_129 val_129 +133val_133 val_133 134val_134 val_134 134val_134 val_134 +136val_136 val_136 +143val_143 val_143 +149val_149 val_149 +149val_149 val_149 152val_152 val_152 152val_152 val_152 -155val_155 val_155 +153val_153 val_153 +157val_157 val_157 +162val_162 val_162 +166val_166 val_166 169val_169 val_169 169val_169 val_169 169val_169 val_169 169val_169 val_169 -17 val_17 val_17 170val_170 val_170 -172val_172 val_172 -172val_172 val_172 -178val_178 val_178 -179val_179 val_179 -179val_179 val_179 -189val_189 val_189 -19 val_19 val_19 -192val_192 val_192 -194val_194 val_194 -195val_195 val_195 -195val_195 val_195 -199val_199 val_199 -199val_199 val_199 -199val_199 val_199 -20 val_20 val_20 -207val_207 val_207 -207val_207 val_207 +174val_174 val_174 +174val_174 val_174 +176val_176 val_176 +176val_176 val_176 +193val_193 val_193 +193val_193 val_193 +193val_193 val_193 +196val_196 val_196 +201val_201 val_201 +205val_205 val_205 +205val_205 val_205 +209val_209 val_209 +209val_209 val_209 213val_213 val_213 213val_213 val_213 218val_218 val_218 +221val_221 val_221 +221val_221 val_221 223val_223 val_223 223val_223 val_223 -229val_229 val_229 -229val_229 val_229 233val_233 val_233 233val_233 val_233 -235val_235 val_235 +239val_239 val_239 +239val_239 val_239 24 val_24 val_24 24 val_24 val_24 -256val_256 val_256 -256val_256 val_256 -257val_257 val_257 +241val_241 val_241 +242val_242 val_242 +242val_242 val_242 +247val_247 val_247 +248val_248 val_248 26 val_26 val_26 26 val_26 val_26 -265val_265 val_265 -265val_265 val_265 -266val_266 val_266 -27 val_27 val_27 -278val_278 val_278 -278val_278 val_278 -280val_280 val_280 -280val_280 val_280 -284val_284 val_284 -291val_291 val_291 -30 val_30 val_30 +263val_263 val_263 +272val_272 val_272 +272val_272 val_272 +275val_275 val_275 +285val_285 val_285 +296val_296 val_296 +298val_298 val_298 +298val_298 val_298 +298val_298 val_298 302val_302 val_302 -305val_305 val_305 -316val_316 val_316 -316val_316 val_316 -316val_316 val_316 -321val_321 val_321 -321val_321 val_321 +307val_307 val_307 +307val_307 val_307 +310val_310 val_310 323val_323 val_323 -33 val_33 val_33 +325val_325 val_325 +325val_325 val_325 333val_333 val_333 333val_333 val_333 +335val_335 val_335 338val_338 val_338 -344val_344 val_344 -344val_344 val_344 -353val_353 val_353 -353val_353 val_353 -356val_356 val_356 -364val_364 val_364 -368val_368 val_368 -369val_369 val_369 -369val_369 val_369 -369val_369 val_369 -37 val_37 val_37 -37 val_37 val_37 +360val_360 val_360 +362val_362 val_362 +367val_367 val_367 +367val_367 val_367 +375val_375 val_375 377val_377 val_377 -378val_378 val_378 -386val_386 val_386 +384val_384 val_384 +384val_384 val_384 +384val_384 val_384 392val_392 val_392 -397val_397 val_397 -397val_397 val_397 -404val_404 val_404 -404val_404 val_404 -407val_407 val_407 -411val_411 val_411 -418val_418 val_418 +399val_399 val_399 +399val_399 val_399 +401val_401 val_401 +401val_401 val_401 +401val_401 val_401 +401val_401 val_401 +401val_401 val_401 +402val_402 val_402 +403val_403 val_403 +403val_403 val_403 +403val_403 val_403 +413val_413 val_413 +413val_413 val_413 +417val_417 val_417 +417val_417 val_417 +417val_417 val_417 419val_419 val_419 -424val_424 val_424 -424val_424 val_424 -430val_430 val_430 -430val_430 val_430 -430val_430 val_430 -432val_432 val_432 -436val_436 val_436 -
[13/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/sample1.q.out -- diff --git a/ql/src/test/results/clientpositive/sample1.q.out b/ql/src/test/results/clientpositive/sample1.q.out index 6d58658..09b9ac7 100644 --- a/ql/src/test/results/clientpositive/sample1.q.out +++ b/ql/src/test/results/clientpositive/sample1.q.out @@ -55,6 +55,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value,dt,hr columns.comments @@ -126,6 +127,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -193,6 +195,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value,dt,hr columns.comments @@ -236,6 +239,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value,dt,hr columns.comments @@ -266,6 +270,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value,dt,hr columns.comments @@ -287,6 +292,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value,dt,hr columns.comments @@ -323,6 +329,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value,dt,hr columns.comments @@ -353,6 +360,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value,dt,hr columns.comments @@ -374,6 +382,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value,dt,hr columns.comments http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/sample2.q.out -- diff --git a/ql/src/test/results/clientpositive/sample2.q.out b/ql/src/test/results/clientpositive/sample2.q.out index 164ce43..53c3df2 100644 --- a/ql/src/test/results/clientpositive/sample2.q.out +++ b/ql/src/test/results/clientpositive/sample2.q.out @@ -54,6 +54,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -101,6 +102,7 @@ STAGE PLANS: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key + bucketing_version 2 column.name
[30/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out -- diff --git a/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out b/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out index 4d8dd74..11ef8b1 100644 --- a/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out +++ b/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out @@ -87,6 +87,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -135,6 +136,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -158,6 +160,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -227,6 +230,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -270,6 +274,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -300,6 +305,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -321,6 +327,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -357,6 +364,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -387,6 +395,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -408,6 +417,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -522,6 +532,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -545,6 +556,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -589,6 +601,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key1,key2,cnt columns.comments @@ -647,6 +660,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","
[17/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/vector_ptf_part_simple.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/vector_ptf_part_simple.q.out b/ql/src/test/results/clientpositive/llap/vector_ptf_part_simple.q.out index 9aeb650..ec8611d 100644 --- a/ql/src/test/results/clientpositive/llap/vector_ptf_part_simple.q.out +++ b/ql/src/test/results/clientpositive/llap/vector_ptf_part_simple.q.out @@ -309,46 +309,46 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@vector_ptf_part_simple_orc A masked pattern was here p_mfgr p_name p_retailprice rn r dr fv lv c cs -Manufacturer#4 almond azure aquamarine papaya violet 1290.35 1 1 1 1290.35 1206.26 4 6 -Manufacturer#4 almond antique violet mint lemon1375.42 2 1 1 1290.35 1206.26 4 6 -Manufacturer#4 almond aquamarine floral ivory bisque NULL3 1 1 1290.35 1206.26 4 6 -Manufacturer#4 almond antique gainsboro frosted violet NULL4 1 1 1290.35 1206.26 4 6 -Manufacturer#4 almond aquamarine yellow dodger mint1844.92 5 1 1 1290.35 1206.26 4 6 -Manufacturer#4 almond aquamarine floral ivory bisque 1206.26 6 1 1 1290.35 1206.26 4 6 -Manufacturer#5 almond azure blanched chiffon midnight 1464.48 1 1 1 1464.48 1788.73 6 6 -Manufacturer#5 almond aquamarine dodger light gainsboro1018.1 2 1 1 1464.48 1788.73 6 6 -Manufacturer#5 almond antique medium spring khaki 1611.66 3 1 1 1464.48 1788.73 6 6 -Manufacturer#5 almond antique blue firebrick mint 1789.69 4 1 1 1464.48 1788.73 6 6 -Manufacturer#5 almond antique medium spring khaki 1611.66 5 1 1 1464.48 1788.73 6 6 -Manufacturer#5 almond antique sky peru orange 1788.73 6 1 1 1464.48 1788.73 6 6 -Manufacturer#2 almond aquamarine rose maroon antique 900.66 1 1 1 900.66 1800.7 8 8 -Manufacturer#2 almond aquamarine rose maroon antique 1698.66 2 1 1 900.66 1800.7 8 8 -Manufacturer#2 almond antique violet turquoise frosted 1800.7 3 1 1 900.66 1800.7 8 8 -Manufacturer#2 almond antique violet chocolate turquoise 1690.68 4 1 1 900.66 1800.7 8 8 -Manufacturer#2 almond antique violet turquoise frosted 1800.7 5 1 1 900.66 1800.7 8 8 -Manufacturer#2 almond aquamarine sandy cyan gainsboro 1000.6 6 1 1 900.66 1800.7 8 8 -Manufacturer#2 almond aquamarine midnight light salmon 2031.98 7 1 1 900.66 1800.7 8 8 -Manufacturer#2 almond antique violet turquoise frosted 1800.7 8 1 1 900.66 1800.7 8 8 -Manufacturer#1 almond antique chartreuse lavender yellow 1753.76 1 1 1 1753.76 1632.66 11 12 -Manufacturer#1 almond aquamarine pink moccasin thistle 1632.66 2 1 1 1753.76 1632.66 11 12 -Manufacturer#1 almond aquamarine pink moccasin thistle 1632.66 3 1 1 1753.76 1632.66 11 12 -Manufacturer#1 almond antique chartreuse lavender yellow 1753.76 4 1 1 1753.76 1632.66 11 12 -Manufacturer#1 almond aquamarine burnished black steel 1414.42 5 1 1 1753.76 1632.66 11 12 -Manufacturer#1 almond antique burnished rose metallic 1173.15 6 1 1 1753.76 1632.66 11 12 -Manufacturer#1 almond antique salmon chartreuse burlywood 1602.59 7 1 1 1753.76 1632.66 11 12 -Manufacturer#1 almond antique chartreuse lavender yellow 1753.76 8 1 1 1753.76 1632.66 11 12 -Manufacturer#1 almond antique burnished rose metallic 1173.15 9 1 1 1753.76 1632.66 11 12 -Manufacturer#1 almond antique chartreuse lavender yellow 1753.76 10 1 1 1753.76 1632.66 11 12 -Manufacturer#1 almond aquamarine pink moccasin thistle NULL11 1 1 1753.76 1632.66 11 12 -Manufacturer#1 almond aquamarine pink moccasin thistle 1632.66 12 1 1 1753.76 1632.66 11 12 -Manufacturer#3 almond antique forest lavender goldenrod590.27 1 1 1 590.27 99.68 7 8 -Manufacturer#3 almond antique metallic orange dim 55.39 2 1 1 590.27 99.68 7 8 -Manufacturer#3 almond antique forest lavender goldenrod1190.27 3 1 1 590.27 99.68 7 8 -Manufacturer#3 almond antique olive coral navajo 1337.29 4 1 1 590.27 99.68 7 8 -Manufacturer#3 almond antique m
[27/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out index 885fd18..8fa1b35 100644 --- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out +++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out @@ -147,6 +147,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 4 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -234,6 +235,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -282,6 +284,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -421,6 +424,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 4 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -508,6 +512,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -556,6 +561,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out index ff98ee5..bd7f60a 100644 --- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out +++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out @@ -147,6 +147,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -195,6 +196,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -283,6 +285,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 4 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -421,6 +424,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -469,6 +473,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -557,6 +562,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 4 bucket_field_name key + bucketing_version 2 colu
[11/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/smb_mapjoin_46.q.out -- diff --git a/ql/src/test/results/clientpositive/smb_mapjoin_46.q.out b/ql/src/test/results/clientpositive/smb_mapjoin_46.q.out index 383ab31..f259c06 100644 --- a/ql/src/test/results/clientpositive/smb_mapjoin_46.q.out +++ b/ql/src/test/results/clientpositive/smb_mapjoin_46.q.out @@ -143,11 +143,11 @@ POSTHOOK: Input: default@test2 A masked pattern was here 98 NULLNoneNULLNULLNULL NULL NULLNoneNULLNULLNULL -99 0 Alice NULLNULLNULL 1012 Car 103 2 Ema 1012 Car 102 2 Del 99 2 Mat 103 2 Ema 99 2 Mat 102 2 Del +99 0 Alice NULLNULLNULL 1001 Bob NULLNULLNULL PREHOOK: query: EXPLAIN SELECT * @@ -221,9 +221,9 @@ POSTHOOK: Input: default@test2 A masked pattern was here 98 NULLNoneNULLNULLNULL NULL NULLNoneNULLNULLNULL -99 0 Alice NULLNULLNULL 1012 Car 102 2 Del 99 2 Mat NULLNULLNULL +99 0 Alice NULLNULLNULL 1001 Bob NULLNULLNULL Warning: Map Join MAPJOIN[11][bigTable=?] in task 'Stage-3:MAPRED' is a cross product PREHOOK: query: EXPLAIN @@ -327,9 +327,9 @@ POSTHOOK: Input: default@test2 A masked pattern was here 98 NULLNoneNULLNULLNULL NULL NULLNoneNULLNULLNULL -99 0 Alice NULLNULLNULL 1012 Car 102 2 Del 99 2 Mat NULLNULLNULL +99 0 Alice NULLNULLNULL 1001 Bob 102 2 Del PREHOOK: query: EXPLAIN SELECT * @@ -491,12 +491,12 @@ POSTHOOK: Input: default@test2 A masked pattern was here 98 NULLNoneNULLNULLNULL NULL NULLNoneNULLNULLNULL -99 0 Alice NULLNULLNULL 1012 Car 105 NULLNone 1012 Car 103 2 Ema 1012 Car 102 2 Del 1012 Car 104 3 Fli 99 2 Mat NULLNULLNULL +99 0 Alice NULLNULLNULL 1001 Bob 105 NULLNone 1001 Bob 103 2 Ema 1001 Bob 102 2 Del @@ -598,13 +598,13 @@ POSTHOOK: Input: default@test2 A masked pattern was here 98 NULLNone102 2 Del NULL NULLNone102 2 Del -99 0 Alice 102 2 Del 1012 Car 105 NULLNone 1012 Car 103 2 Ema 1012 Car 102 2 Del 1012 Car 104 3 Fli 99 2 Mat 103 2 Ema 99 2 Mat 102 2 Del +99 0 Alice 102 2 Del 1001 Bob 105 NULLNone 1001 Bob 103 2 Ema 1001 Bob 102 2 Del @@ -702,13 +702,13 @@ POSTHOOK: Input: default@test2 A masked pattern was here 98 NULLNoneNULLNULLNULL NULL NULLNoneNULLNULLNULL -99 0 Alice NULLNULLNULL 1012 Car 105 NULLNone 1012 Car 103 2 Ema 1012 Car 102 2 Del 1012 Car 104 3 Fli 99 2 Mat 103 2 Ema 99 2 Mat 102 2 Del +99 0 Alice NULLNULLNULL 1001 Bob 105 NULLNone 1001 Bob 103 2 Ema 1001 Bob 102 2 Del @@ -806,11 +806,11 @@ POSTHOOK: Input: default@test2 A masked pattern was here 98 NULLNone102 2 Del NULL NULLNone102 2 Del -99 0 Alice 102 2 Del 1012 Car 103 2 Ema 1012 Car 102 2 Del 99 2 Mat 103 2 Ema 99 2 Mat 102 2 Del +99 0 Alice 102 2 Del 1001 Bob 102 2 Del PREHOOK: query: EXPLAIN SELECT * @@ -882,10 +882,10 @@ POSTHOOK: Input: default@test2 A masked pattern was here 98 NULLNoneNULLNULLNULL NULL NULLNoneNULLNULLNULL -99 0 Alice NULLNULLNULL 1012 Car 103 2 Ema 1012 Car 102 2 Del 99 2 Mat 102 2 Del +99 0 Alice NULLNULLNULL 1001 Bob NULLNULLNULL Warning: Map Join MAPJOIN[9][bigTable=?] in task 'Stage-3:MAPRED' is a cross product PREHOOK: query: EXPLAIN @@ -989,9 +989,9 @
[31/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/druid_basic1.q.out -- diff --git a/ql/src/test/results/clientpositive/druid_basic1.q.out b/ql/src/test/results/clientpositive/druid_basic1.q.out index 25ec0ca..a5c0687 100644 --- a/ql/src/test/results/clientpositive/druid_basic1.q.out +++ b/ql/src/test/results/clientpositive/druid_basic1.q.out @@ -41,6 +41,7 @@ Table Type: EXTERNAL_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"__time\":\"true\",\"added\":\"true\",\"anonymous\":\"true\",\"count\":\"true\",\"deleted\":\"true\",\"delta\":\"true\",\"language\":\"true\",\"namespace\":\"true\",\"newpage\":\"true\",\"page\":\"true\",\"robot\":\"true\",\"unpatrolled\":\"true\",\"user\":\"true\",\"variation\":\"true\"}} EXTERNALTRUE + bucketing_version 2 druid.datasourcewikipedia numFiles0 numRows 0 @@ -102,6 +103,7 @@ Table Type: EXTERNAL_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"__time\":\"true\",\"added\":\"true\",\"anonymous\":\"true\",\"count\":\"true\",\"deleted\":\"true\",\"delta\":\"true\",\"language\":\"true\",\"namespace\":\"true\",\"newpage\":\"true\",\"page\":\"true\",\"robot\":\"true\",\"unpatrolled\":\"true\",\"user\":\"true\",\"variation\":\"true\"}} EXTERNALTRUE + bucketing_version 2 druid.datasourcewikipedia numFiles0 numRows 0 http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/druid_basic2.q.out -- diff --git a/ql/src/test/results/clientpositive/druid_basic2.q.out b/ql/src/test/results/clientpositive/druid_basic2.q.out index 5ed039b..0e5e957 100644 --- a/ql/src/test/results/clientpositive/druid_basic2.q.out +++ b/ql/src/test/results/clientpositive/druid_basic2.q.out @@ -41,6 +41,7 @@ Table Type: EXTERNAL_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"__time\":\"true\",\"added\":\"true\",\"anonymous\":\"true\",\"count\":\"true\",\"deleted\":\"true\",\"delta\":\"true\",\"language\":\"true\",\"namespace\":\"true\",\"newpage\":\"true\",\"page\":\"true\",\"robot\":\"true\",\"unpatrolled\":\"true\",\"user\":\"true\",\"variation\":\"true\"}} EXTERNALTRUE + bucketing_version 2 druid.datasourcewikipedia numFiles0 numRows 0 @@ -266,6 +267,7 @@ STAGE PLANS: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"__time":"true","added":"true","anonymous":"true","count":"true","deleted":"true","delta":"true","language":"true","namespace":"true","newpage":"true","page":"true","robot":"true","unpatrolled":"true","user":"true","variation":"true"}} EXTERNAL TRUE bucket_count -1 + bucketing_version 2 column.name.delimiter , columns __time,robot,namespace,anonymous,unpatrolled,page,language,newpage,user,count,added,delta,variation,deleted columns.comments 'from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer' @@ -294,6 +296,7 @@ STAGE PLANS: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"__time":"true","added":"true","anonymous":"true","count":"true","deleted":"true","delta":"true","language":"true","namespace":"true","newpage":"true","page":"true","robot":"true","unpatrolled":"true","user":"true","variation":"true"}} EXTERNAL TRUE bucket_count -1 +bucketing_version 2 column.name.delimiter , columns __time,robot,namespace,anonymous,unpatrolled,page,language,newpage,user,count,added,delta,variation,deleted columns.comments 'from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer' @@ -444,6 +447,7 @@ STAGE PLANS:
[36/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/ac6b2a3f Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/ac6b2a3f Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/ac6b2a3f Branch: refs/heads/master Commit: ac6b2a3fb195916e22b2e5f465add2ffbcdc7430 Parents: a095c3d Author: Deepak Jaiswal Authored: Sun Apr 29 22:24:15 2018 -0700 Committer: Deepak Jaiswal Committed: Sun Apr 29 22:24:15 2018 -0700 -- .../results/positive/external_table_ppd.q.out | 1 + .../positive/hbase_binary_storage_queries.q.out | 2 + .../src/test/results/positive/hbase_ddl.q.out | 2 + .../src/test/results/positive/hbasestats.q.out | 5 + .../streaming/AbstractRecordWriter.java | 9 +- .../hive/hcatalog/streaming/TestStreaming.java |10 +- .../mutate/worker/TestBucketIdResolverImpl.java | 2 +- .../org/apache/hive/hcatalog/api/HCatTable.java | 4 + .../hive/hcatalog/api/TestHCatClient.java | 2 + .../insert_into_dynamic_partitions.q.out| 4 + .../clientpositive/insert_into_table.q.out |11 + .../insert_overwrite_directory.q.out| 4 + .../insert_overwrite_dynamic_partitions.q.out | 4 + .../clientpositive/insert_overwrite_table.q.out |11 + .../write_final_output_blobstore.q.out | 8 + .../hive/ql/txn/compactor/TestCompactor.java| 2 +- .../test/resources/testconfiguration.properties | 6 +- .../hadoop/hive/ql/exec/FileSinkOperator.java |16 +- .../hadoop/hive/ql/exec/FunctionRegistry.java | 1 + .../apache/hadoop/hive/ql/exec/Operator.java| 9 + .../hadoop/hive/ql/exec/OperatorFactory.java| 3 + .../hadoop/hive/ql/exec/ReduceSinkOperator.java |44 +- .../apache/hadoop/hive/ql/exec/Utilities.java |13 + .../bootstrap/load/table/LoadPartitions.java| 3 +- .../VectorKeySeriesSerializedImpl.java | 4 +- .../VectorReduceSinkObjectHashOperator.java | 135 +- .../hadoop/hive/ql/io/HiveFileFormatUtils.java |11 +- .../hadoop/hive/ql/io/orc/OrcInputFormat.java | 8 +- .../apache/hadoop/hive/ql/metadata/Table.java |11 + .../hive/ql/optimizer/ConvertJoinMapJoin.java |63 +- .../optimizer/FixedBucketPruningOptimizer.java |18 +- .../ql/optimizer/PrunerOperatorFactory.java | 3 +- .../optimizer/SortedDynPartitionOptimizer.java | 6 +- ...tedDynPartitionTimeGranularityOptimizer.java | 9 +- .../annotation/OpTraitsRulesProcFactory.java|37 +- .../hive/ql/optimizer/physical/Vectorizer.java |16 +- .../optimizer/spark/SparkMapJoinOptimizer.java | 3 +- .../hadoop/hive/ql/parse/SemanticAnalyzer.java |33 +- .../apache/hadoop/hive/ql/plan/OpTraits.java|24 +- .../apache/hadoop/hive/ql/plan/PlanUtils.java | 2 +- .../hadoop/hive/ql/plan/ReduceSinkDesc.java |10 +- .../apache/hadoop/hive/ql/plan/TableDesc.java | 5 + .../hive/ql/plan/VectorReduceSinkDesc.java | 9 + .../hive/ql/udf/generic/GenericUDFHash.java | 3 +- .../ql/udf/generic/GenericUDFMurmurHash.java|64 + .../hadoop/hive/ql/TestTxnAddPartition.java |11 +- .../apache/hadoop/hive/ql/TestTxnCommands.java |20 +- .../apache/hadoop/hive/ql/TestTxnCommands2.java |63 +- .../apache/hadoop/hive/ql/TestTxnNoBuckets.java |14 +- .../parse/authorization/TestPrivilegesV1.java | 2 +- .../clientpositive/archive_excludeHadoop20.q| 4 +- .../test/queries/clientpositive/bucket_many.q | 4 +- .../clientpositive/bucket_map_join_tez1.q | 2 +- .../clientpositive/bucket_num_reducers.q| 5 +- .../clientpositive/bucket_num_reducers2.q | 6 +- .../clientpositive/bucket_num_reducers_acid2.q | 3 - .../clientpositive/insert_update_delete.q | 1 - ql/src/test/queries/clientpositive/sample10.q | 4 +- .../clientpositive/tez_dynpart_hashjoin_3.q | 4 +- ql/src/test/queries/clientpositive/tez_smb_1.q | 2 +- .../clientnegative/alter_file_format.q.out | 1 + .../alter_view_as_select_with_partition.q.out | 1 + .../clientnegative/unset_table_property.q.out | 1 + .../clientnegative/unset_view_property.q.out| 1 + .../results/clientpositive/acid_mapjoin.q.out |20 +- .../results/clientpositive/acid_nullscan.q.out | 2 + .../clientpositive/acid_table_stats.q.out | 2 + .../clientpositive/alterColumnStats.q.out | 3 + .../clientpositive/alterColumnStatsPart.q.out | 1 + .../clientpositive/alter_file_format.q.out | 7 + ...lter_numbuckets_partitioned_table2_h23.q.out | 9 + ...alter_numbuckets_partitioned_table_h23.q.out | 8 + .../alter_partition_clusterby_sortby.q.out |
[03/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/subquery_notin.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/subquery_notin.q.out b/ql/src/test/results/clientpositive/spark/subquery_notin.q.out index ffa524e..0f2f717 100644 --- a/ql/src/test/results/clientpositive/spark/subquery_notin.q.out +++ b/ql/src/test/results/clientpositive/spark/subquery_notin.q.out @@ -1213,25 +1213,25 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@part A masked pattern was here Manufacturer#1 almond aquamarine burnished black steel 28 -Manufacturer#1 almond antique chartreuse lavender yellow 34 Manufacturer#1 almond aquamarine pink moccasin thistle 42 -Manufacturer#2 almond aquamarine rose maroon antique 25 +Manufacturer#2 almond antique violet chocolate turquoise 14 +Manufacturer#2 almond aquamarine sandy cyan gainsboro 18 Manufacturer#2 almond antique violet turquoise frosted 40 -Manufacturer#3 almond antique chartreuse khaki white 17 Manufacturer#3 almond antique metallic orange dim 19 +Manufacturer#4 almond aquamarine floral ivory bisque 27 +Manufacturer#4 almond antique violet mint lemon39 Manufacturer#5 almond antique medium spring khaki 6 Manufacturer#5 almond azure blanched chiffon midnight 23 Manufacturer#5 almond antique blue firebrick mint 31 Manufacturer#5 almond aquamarine dodger light gainsboro46 Manufacturer#1 almond antique salmon chartreuse burlywood 6 -Manufacturer#2 almond antique violet chocolate turquoise 14 -Manufacturer#2 almond aquamarine sandy cyan gainsboro 18 +Manufacturer#1 almond antique chartreuse lavender yellow 34 +Manufacturer#2 almond aquamarine rose maroon antique 25 Manufacturer#3 almond antique forest lavender goldenrod14 +Manufacturer#3 almond antique chartreuse khaki white 17 Manufacturer#3 almond antique olive coral navajo 45 Manufacturer#4 almond antique gainsboro frosted violet 10 Manufacturer#4 almond azure aquamarine papaya violet 12 -Manufacturer#4 almond aquamarine floral ivory bisque 27 -Manufacturer#4 almond antique violet mint lemon39 Warning: Shuffle Join JOIN[19][tables = [$hdt$_0, $hdt$_1]] in Work 'Reducer 2' is a cross product PREHOOK: query: select li.l_partkey, count(*) from lineitem li @@ -1250,21 +1250,21 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@lineitem A masked pattern was here 22630 1 -7068 1 -88362 1 -123076 1 -139636 1 -106170 1 -182052 1 -59694 1 175839 1 -21636 1 119767 1 +139636 1 +21636 1 119477 1 -88035 1 -4501 -61931 1 +88362 1 +106170 1 +59694 1 +7068 1 +123076 1 85951 1 +61931 1 +4501 +182052 1 +88035 1 Warning: Shuffle Join JOIN[18][tables = [$hdt$_0, $hdt$_1]] in Work 'Reducer 2' is a cross product PREHOOK: query: select * from src @@ -1789,29 +1789,29 @@ POSTHOOK: query: select * from part where p_brand <> 'Brand#14' AND p_size NOT I POSTHOOK: type: QUERY POSTHOOK: Input: default@part A masked pattern was here -17927 almond aquamarine yellow dodger mintManufacturer#4 Brand#41 ECONOMY BRUSHED COPPER 7 SM PKG 1844.92 ites. eve 191709 almond antique violet turquoise frosted Manufacturer#2 Brand#22 ECONOMY POLISHED STEEL 40 MED BOX 1800.7 haggle -78486 almond azure blanched chiffon midnight Manufacturer#5 Brand#52 LARGE BRUSHED BRASS 23 MED BAG 1464.48 hely blith +85768 almond antique chartreuse lavender yellow Manufacturer#1 Brand#12LARGE BRUSHED STEEL 34 SM BAG 1753.76 refull +65667 almond aquamarine pink moccasin thistle Manufacturer#1 Brand#12 LARGE BURNISHED STEEL 42 JUMBO CASE 1632.66 e across the expr 105685 almond antique violet chocolate turquoise Manufacturer#2 Brand#22MEDIUM ANODIZED COPPER 14 MED CAN 1690.68 ly pending requ 112398 almond antique metallic orange dim Manufacturer#3 Brand#32 MEDIUM BURNISHED BRASS 19 JUMBO JAR 1410.39 ole car 146985 almond aquamarine midnight light salmon Manufacturer#2 Brand#23 MEDIUM BURNISHED COPPER 2 SM CASE 2031.98 s cajole caref -90681 almond antique chartreuse khaki white Manufacturer#3 Brand#31 MEDIUM BURNISHED TIN17 SM CASE 1671.68 are slyly after the sl 110592 almond antique salmon chartreuse burlywood Manufacturer#1 Brand#15PROMO BURNISHED NICKEL 6 JUMBO PKG 1602.59 to the furiously +48427 almond antique violet mint lemonManufacturer#4 Brand#42 PROMO POLISHED STEEL39 SM CASE 1375.42 hely ironic i 49671 almond antique gainsboro frosted violet Manufacturer#4 Brand#41 SMALL BRUSHED BRASS 10 SM BOX 1620.67 ccounts run quick -33357 almond azure aquamarine papaya violet Manufacturer#4 Brand#41 STANDAR
[33/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/bucket1.q.out -- diff --git a/ql/src/test/results/clientpositive/bucket1.q.out b/ql/src/test/results/clientpositive/bucket1.q.out index 3060df3..07b5b9d 100644 --- a/ql/src/test/results/clientpositive/bucket1.q.out +++ b/ql/src/test/results/clientpositive/bucket1.q.out @@ -52,6 +52,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -73,6 +74,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -112,6 +114,7 @@ STAGE PLANS: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 100 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -171,6 +174,7 @@ STAGE PLANS: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 100 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/bucket2.q.out -- diff --git a/ql/src/test/results/clientpositive/bucket2.q.out b/ql/src/test/results/clientpositive/bucket2.q.out index 16218f9..e8a1dd7 100644 --- a/ql/src/test/results/clientpositive/bucket2.q.out +++ b/ql/src/test/results/clientpositive/bucket2.q.out @@ -51,6 +51,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -72,6 +73,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -111,6 +113,7 @@ STAGE PLANS: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -178,6 +181,7 @@ STAGE PLANS: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -264,66 +268,59 @@ POSTHOOK: query: select * from bucket2_1 tablesample (bucket 1 out of 2) s POSTHOOK: type: QUERY POSTHOOK: Input: default@bucket2_1 A masked pattern was here -0 val_0 -0 val_0 -0 val_0 10 val_10 -100val_100 -100val_100 -104val_104 -104val_104 -114val_114 +105val_105 +113val_113 +113val_113 116val_116 -118val_118 -118val_118 -12 val_12 -12 val_12 -120val_120 -120val_120 +119val_119 +119val_119 +119val_119 126val_126 -128val_128 -128val_128 -128val_128 +133val_133 134val_134 134val_134 -136val_136 -138val_138 -138val_138 -138val_138 -138val_138 +137val_137 +137val_137 146val_146 146val_146 -150val_150 -152val_152 -152val_152 +149val_149 +149val_149 +153val_153 156val_156 +157val_157 158val_158 -160val_160 162val_162 +163val_163 164val_164 164val_164 -166val_166 -168val_168 -170val_170 -172val_172 -172val_172 -174val_174 -174val_174 -176val_176 -176val_176 +165val_165 +165val_165 +169val_169 +169val_169 +
[32/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out -- diff --git a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out index b59c4bc..49598d1 100644 --- a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out +++ b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out @@ -128,7 +128,7 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table2 POSTHOOK: Input: default@test_table2@ds=1 A masked pattern was here -247 +243 PREHOOK: query: select count(*) from test_table2 tablesample (bucket 2 out of 2) s where ds = '1' PREHOOK: type: QUERY PREHOOK: Input: default@test_table2 @@ -139,7 +139,7 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table2 POSTHOOK: Input: default@test_table2@ds=1 A masked pattern was here -253 +257 PREHOOK: query: EXPLAIN INSERT OVERWRITE TABLE test_table2 PARTITION (ds = '1') SELECT * from @@ -238,7 +238,7 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table2 POSTHOOK: Input: default@test_table2@ds=1 A masked pattern was here -247 +243 PREHOOK: query: select count(*) from test_table2 tablesample (bucket 2 out of 2) s where ds = '1' PREHOOK: type: QUERY PREHOOK: Input: default@test_table2 @@ -249,7 +249,7 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table2 POSTHOOK: Input: default@test_table2@ds=1 A masked pattern was here -253 +257 PREHOOK: query: EXPLAIN INSERT OVERWRITE TABLE test_table2 PARTITION (ds = '1') SELECT x.key, concat(x.value, x.value) from http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out -- diff --git a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out index 52ef3db..d1c11f4 100644 --- a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out +++ b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out @@ -120,7 +120,7 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table2 POSTHOOK: Input: default@test_table2@ds=1 A masked pattern was here -247 +243 PREHOOK: query: select count(*) from test_table2 tablesample (bucket 2 out of 2) s where ds = '1' PREHOOK: type: QUERY PREHOOK: Input: default@test_table2 @@ -131,7 +131,7 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table2 POSTHOOK: Input: default@test_table2@ds=1 A masked pattern was here -253 +257 PREHOOK: query: CREATE TABLE test_table3 (key INT, value STRING) PARTITIONED BY (ds STRING) CLUSTERED BY (value) SORTED BY (value) INTO 2 BUCKETS PREHOOK: type: CREATETABLE http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out -- diff --git a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out index eaf85c3..72e918f 100644 --- a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out +++ b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out @@ -199,18 +199,7 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table3 POSTHOOK: Input: default@test_table3@ds=1 A masked pattern was here -0 0 val_0val_0 1 -0 0 val_0val_0 1 -0 0 val_0val_0 1 -0 0 val_0val_0 1 -0 0 val_0val_0 1 -0 0 val_0val_0 1 -0 0 val_0val_0 1 -0 0 val_0val_0 1 -0 0 val_0val_0 1 2 2 val_2val_2 1 -4 4 val_4val_4 1 -8 8 val_8val_8 1 PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s where ds = '1' PREHOOK: type: QUERY PREHOOK: Input: default@test_table3 @@ -221,6 +210,16 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table3 POSTHOOK: Input: default@test_table3@ds=1 A masked pattern was here +0 0 val_0val_0 1 +0 0 val_0val_0 1 +0 0 val_0val_0 1 +0 0 val_0val_0 1 +0 0 val_0val_0 1 +0 0 val_0val_0 1 +0 0 val_0val_0 1 +0 0 val_0val_0 1 +0 0 val_0val_0 1 +4 4 val_4val_4 1 5 5 val_5val_5 1 5 5 val_5val_5 1 5 5 val_5val_5 1 @@ -230,6 +229,7 @@ POSTHOOK: Input: default@test_table3@ds=1 5 5 val_5val_5 1 5 5 val_5val_5 1 5 5 val_5val_5 1 +8 8
[09/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/bucketsortoptimize_insert_2.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/bucketsortoptimize_insert_2.q.out b/ql/src/test/results/clientpositive/spark/bucketsortoptimize_insert_2.q.out index 3e8ea6f..9b14b92 100644 --- a/ql/src/test/results/clientpositive/spark/bucketsortoptimize_insert_2.q.out +++ b/ql/src/test/results/clientpositive/spark/bucketsortoptimize_insert_2.q.out @@ -193,18 +193,7 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table3 POSTHOOK: Input: default@test_table3@ds=1 A masked pattern was here -0 val_0val_0 1 -0 val_0val_0 1 -0 val_0val_0 1 -0 val_0val_0 1 -0 val_0val_0 1 -0 val_0val_0 1 -0 val_0val_0 1 -0 val_0val_0 1 -0 val_0val_0 1 2 val_2val_2 1 -4 val_4val_4 1 -8 val_8val_8 1 PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s where ds = '1' PREHOOK: type: QUERY PREHOOK: Input: default@test_table3 @@ -215,6 +204,16 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table3 POSTHOOK: Input: default@test_table3@ds=1 A masked pattern was here +0 val_0val_0 1 +0 val_0val_0 1 +0 val_0val_0 1 +0 val_0val_0 1 +0 val_0val_0 1 +0 val_0val_0 1 +0 val_0val_0 1 +0 val_0val_0 1 +0 val_0val_0 1 +4 val_4val_4 1 5 val_5val_5 1 5 val_5val_5 1 5 val_5val_5 1 @@ -224,6 +223,7 @@ POSTHOOK: Input: default@test_table3@ds=1 5 val_5val_5 1 5 val_5val_5 1 5 val_5val_5 1 +8 val_8val_8 1 9 val_9val_9 1 PREHOOK: query: EXPLAIN INSERT OVERWRITE TABLE test_table3 PARTITION (ds = '1') @@ -344,6 +344,18 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table3 POSTHOOK: Input: default@test_table3@ds=1 A masked pattern was here +2 val_2val_2 1 +2 val_2val_2 1 +PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s where ds = '1' +PREHOOK: type: QUERY +PREHOOK: Input: default@test_table3 +PREHOOK: Input: default@test_table3@ds=1 + A masked pattern was here +POSTHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s where ds = '1' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@test_table3 +POSTHOOK: Input: default@test_table3@ds=1 + A masked pattern was here 0 val_0val_0 1 0 val_0val_0 1 0 val_0val_0 1 @@ -362,22 +374,8 @@ POSTHOOK: Input: default@test_table3@ds=1 0 val_0val_0 1 0 val_0val_0 1 0 val_0val_0 1 -2 val_2val_2 1 -2 val_2val_2 1 4 val_4val_4 1 4 val_4val_4 1 -8 val_8val_8 1 -8 val_8val_8 1 -PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s where ds = '1' -PREHOOK: type: QUERY -PREHOOK: Input: default@test_table3 -PREHOOK: Input: default@test_table3@ds=1 - A masked pattern was here -POSTHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s where ds = '1' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@test_table3 -POSTHOOK: Input: default@test_table3@ds=1 - A masked pattern was here 5 val_5val_5 1 5 val_5val_5 1 5 val_5val_5 1 @@ -396,6 +394,8 @@ POSTHOOK: Input: default@test_table3@ds=1 5 val_5val_5 1 5 val_5val_5 1 5 val_5val_5 1 +8 val_8val_8 1 +8 val_8val_8 1 9 val_9val_9 1 9 val_9val_9 1 PREHOOK: query: EXPLAIN @@ -517,6 +517,18 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table3 POSTHOOK: Input: default@test_table3@ds=1 A masked pattern was here +2 val_2val_2 1 +2 val_2val_2 1 +PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s where ds = '1' +PREHOOK: type: QUERY +PREHOOK: Input: default@test_table3 +PREHOOK: Input: default@test_table3@ds=1 + A masked pattern was here +POSTHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s where ds = '1' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@test_table3 +POSTHOOK: Input: default@test_table3@ds=1 + A masked pattern was here 0 val_0val_0 1 0 val_0val_0 1 0 val_0val_0 1 @@ -535,22 +547,8 @@ POSTHOOK: Input: default@test_table3@ds=1 0 val_0val_0 1 0 val_0val_0 1 0 val_0val_0 1 -2 val_2val_2 1 -2 val_2val_2 1 4 val_4val_4 1 4 val_4val_4 1 -8 val_8val_8 1 -8 val_8val_8 1 -PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s where ds = '1' -PREHOOK: type: QUERY -P
[16/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/vector_windowing_range_multiorder.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/vector_windowing_range_multiorder.q.out b/ql/src/test/results/clientpositive/llap/vector_windowing_range_multiorder.q.out index 256b80d..9acbf2e 100644 --- a/ql/src/test/results/clientpositive/llap/vector_windowing_range_multiorder.q.out +++ b/ql/src/test/results/clientpositive/llap/vector_windowing_range_multiorder.q.out @@ -900,10005 +900,10005 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@over10k A masked pattern was here s si i avg_window_0 -alice falkner 323 65669 65695.76470588235 alice falkner 477 65722 65695.76470588235 -alice falkner 455 65718 65695.76470588235 -alice falkner 481 65709 65695.76470588235 -alice falkner 345 65773 65695.76470588235 -alice falkner 280 65597 65695.76470588235 -alice falkner 500 65775 65695.76470588235 alice falkner 339 65785 65695.76470588235 -alice falkner 452 65596 65695.76470588235 -alice falkner 382 65690 65695.76470588235 alice falkner 382 65622 65695.76470588235 alice falkner 393 65611 65695.76470588235 +alice falkner 389 65699 65695.76470588235 +alice falkner 500 65775 65695.76470588235 +alice falkner 371 65710 65695.76470588235 +alice falkner 481 65709 65695.76470588235 +alice falkner 452 65596 65695.76470588235 +alice falkner 323 65669 65695.76470588235 +alice falkner 382 65690 65695.76470588235 +alice falkner 455 65718 65695.76470588235 alice falkner 393 65685 65695.76470588235 -alice falkner 342 65752 65695.76470588235 alice falkner 311 65715 65695.76470588235 -alice falkner 371 65710 65695.76470588235 -alice falkner 389 65699 65695.76470588235 -alice ichabod 366 65590 65654.95454545454 +alice falkner 280 65597 65695.76470588235 +alice falkner 342 65752 65695.76470588235 +alice falkner 345 65773 65695.76470588235 +alice ichabod 292 65788 65654.95454545454 alice ichabod 458 65550 65654.95454545454 +alice ichabod 412 65718 65654.95454545454 +alice ichabod 303 65692 65654.95454545454 +alice ichabod 300 65704 65654.95454545454 +alice ichabod 366 65590 65654.95454545454 +alice ichabod 301 65693 65654.95454545454 +alice ichabod 398 65659 65654.95454545454 +alice ichabod 292 65585 65654.95454545454 +alice ichabod 305 65617 65654.95454545454 alice ichabod 436 65738 65654.95454545454 -alice ichabod 315 65772 65654.95454545454 +alice ichabod 440 65725 65654.95454545454 alice ichabod 453 65780 65654.95454545454 +alice ichabod 398 65785 65654.95454545454 alice ichabod 347 65547 65654.95454545454 -alice ichabod 398 65659 65654.95454545454 -alice ichabod 338 65538 65654.95454545454 -alice ichabod 440 65725 65654.95454545454 alice ichabod 320 65622 65654.95454545454 -alice ichabod 412 65718 65654.95454545454 -alice ichabod 305 65617 65654.95454545454 -alice ichabod 292 65585 65654.95454545454 -alice ichabod 303 65692 65654.95454545454 -alice ichabod 338 65545 65654.95454545454 alice ichabod 398 65680 65654.95454545454 +alice ichabod 315 65772 65654.95454545454 alice ichabod 416 65536 65654.95454545454 +alice ichabod 338 65545 65654.95454545454 alice ichabod 344 65545 65654.95454545454 -alice ichabod 300 65704 65654.95454545454 -alice ichabod 292 65788 65654.95454545454 -alice ichabod 398 65785 65654.95454545454 -alice ichabod 301 65693 65654.95454545454 -alice polk 443 65734 65661.57142857143 -alice polk 444 65564 65661.57142857143 -alice polk 357 65550 65661.57142857143 -alice polk 321 65744 65661.57142857143 -alice polk 273 65548 65661.57142857143 -alice polk 366 65595 65661.57142857143 -alice polk 285 65761 65661.57142857143 -alice polk 466 65561 65661.57142857143 -alice polk 324 65749 65661.57142857143 -alice polk 487 65746 65661.57142857143 -alice polk 378 65598 65661.57142857143 -alice polk 395 65751 65661.57142857143 -alice polk 407 65617 65661.57142857143 -alice polk 507 65744 65661.57142857143 -alice young468 65649 65706.63636363637 -alice young489 65646 65706.63636363637 -alice young286 65705 65706.63636363637 -alice young447 65789 65706.63636363637 -alice young425 65677 65706.63636363637 -alice young282 65671 65706.63636363637 -alice young351 65776 65706.63636363637 -alice young308 65776 65706.63636363637
[04/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/smb_mapjoin_12.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/smb_mapjoin_12.q.out b/ql/src/test/results/clientpositive/spark/smb_mapjoin_12.q.out index ec8b046..9b421d5 100644 --- a/ql/src/test/results/clientpositive/spark/smb_mapjoin_12.q.out +++ b/ql/src/test/results/clientpositive/spark/smb_mapjoin_12.q.out @@ -141,6 +141,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -181,6 +182,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -213,6 +215,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -267,7 +270,7 @@ POSTHOOK: Input: default@test_table1@ds=1 POSTHOOK: Input: default@test_table3 POSTHOOK: Input: default@test_table3@ds=1 A masked pattern was here -879 +420 PREHOOK: query: explain extended INSERT OVERWRITE TABLE test_table3 PARTITION (ds = '2') SELECT /*+mapjoin(b)*/ a.key, concat(a.value, b.value) FROM test_table3 a JOIN test_table1 b ON a.key = b.key AND a.ds = '1' AND b.ds='1' @@ -359,6 +362,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -399,6 +403,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -431,6 +436,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -479,4 +485,4 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table3 POSTHOOK: Input: default@test_table3@ds=2 A masked pattern was here -879 +420 http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out b/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out index f8d0fb2..a364374 100644 --- a/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out +++ b/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out @@ -124,6 +124,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -147,6 +148,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -273,6 +275,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -296,6 +299,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -373,6 +377,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 16
[07/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/list_bucket_dml_2.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/list_bucket_dml_2.q.out b/ql/src/test/results/clientpositive/spark/list_bucket_dml_2.q.out index 30dc8b5..0ee050f 100644 --- a/ql/src/test/results/clientpositive/spark/list_bucket_dml_2.q.out +++ b/ql/src/test/results/clientpositive/spark/list_bucket_dml_2.q.out @@ -55,6 +55,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -109,6 +110,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -157,6 +159,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -189,6 +192,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -338,6 +342,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out b/ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out index 10d12a6..71996fe 100644 --- a/ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out +++ b/ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out @@ -173,6 +173,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -221,6 +222,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -269,6 +271,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -317,6 +320,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out b/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out index 972a7a8..7fbd873 100644 --- a/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out +++ b/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out @@ -62,6 +62,7 @@ STAGE PLANS: properties: COLUMN_STAT
[35/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/keyseries/VectorKeySeriesSerializedImpl.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/keyseries/VectorKeySeriesSerializedImpl.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/keyseries/VectorKeySeriesSerializedImpl.java index 86f466f..77c9ecc 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/keyseries/VectorKeySeriesSerializedImpl.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/keyseries/VectorKeySeriesSerializedImpl.java @@ -21,9 +21,9 @@ package org.apache.hadoop.hive.ql.exec.vector.keyseries; import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch; import org.apache.hadoop.hive.serde2.ByteStream.Output; import org.apache.hadoop.hive.serde2.fast.SerializeWrite; -import org.apache.hive.common.util.HashCodeUtil; import com.google.common.base.Preconditions; +import org.apache.hive.common.util.Murmur3; /** * Implementation of base serialization interface. @@ -103,7 +103,7 @@ public abstract class VectorKeySeriesSerializedImpl byte[] bytes = output.getData(); for (int i = 0; i < nonNullKeyCount; i++) { keyLength = serializedKeyLengths[i]; - hashCodes[i] = HashCodeUtil.murmurHash(bytes, offset, keyLength); + hashCodes[i] = Murmur3.hash32(bytes, offset, keyLength, 0); offset += keyLength; } } http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java index 1bc3fda..42b7784 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java @@ -32,6 +32,7 @@ import org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression; import org.apache.hadoop.hive.ql.metadata.HiveException; import org.apache.hadoop.hive.ql.plan.OperatorDesc; import org.apache.hadoop.hive.ql.plan.VectorDesc; +import org.apache.hadoop.hive.serde2.ByteStream; import org.apache.hadoop.hive.serde2.ByteStream.Output; import org.apache.hadoop.hive.serde2.binarysortable.fast.BinarySortableSerializeWrite; import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector; @@ -40,6 +41,7 @@ import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo; import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoUtils; import com.google.common.base.Preconditions; +import org.apache.hive.common.util.Murmur3; /** * This class is the object hash (not Uniform Hash) operator class for native vectorized reduce sink. @@ -226,61 +228,110 @@ public class VectorReduceSinkObjectHashOperator extends VectorReduceSinkCommonOp int[] selected = batch.selected; final int size = batch.size; - for (int logical = 0; logical < size; logical++) { -final int batchIndex = (selectedInUse ? selected[logical] : logical); - -final int hashCode; -if (isEmptyBuckets) { - if (isEmptyPartitions) { -hashCode = nonPartitionRandom.nextInt(); - } else { + + // EmptyBuckets = true + if (isEmptyBuckets) { +if (isEmptyPartitions) { + for (int logical = 0; logical< size; logical++) { +final int batchIndex = (selectedInUse ? selected[logical] : logical); +final int hashCode = nonPartitionRandom.nextInt(); +postProcess(batch, batchIndex, tag, hashCode); + } +} else { // isEmptyPartition = false + for (int logical = 0; logical< size; logical++) { +final int batchIndex = (selectedInUse ? selected[logical] : logical); partitionVectorExtractRow.extractRow(batch, batchIndex, partitionFieldValues); -hashCode = +final int hashCode = bucketingVersion == 2 && !vectorDesc.getIsAcidChange() ? ObjectInspectorUtils.getBucketHashCode( +partitionFieldValues, partitionObjectInspectors) : +ObjectInspectorUtils.getBucketHashCodeOld( partitionFieldValues, partitionObjectInspectors); +postProcess(batch, batchIndex, tag, hashCode); } -} else { - bucketVectorExtractRow.extractRow(batch, batchIndex, bucketFieldValues); - final int bucketNum = - ObjectInspectorUtils.getBucketNumber( +} + } else { // EmptyBuckets = false +if (isEmptyPartitions) { + for (int logical = 0; logical< size; logical
[28/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/join9.q.out -- diff --git a/ql/src/test/results/clientpositive/join9.q.out b/ql/src/test/results/clientpositive/join9.q.out index 50ca52c..0a86a6e 100644 --- a/ql/src/test/results/clientpositive/join9.q.out +++ b/ql/src/test/results/clientpositive/join9.q.out @@ -76,6 +76,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -97,6 +98,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -147,6 +149,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -192,6 +195,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -250,6 +254,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/join_filters_overlap.q.out -- diff --git a/ql/src/test/results/clientpositive/join_filters_overlap.q.out b/ql/src/test/results/clientpositive/join_filters_overlap.q.out index db0b26d..7d4855a 100644 --- a/ql/src/test/results/clientpositive/join_filters_overlap.q.out +++ b/ql/src/test/results/clientpositive/join_filters_overlap.q.out @@ -92,6 +92,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -113,6 +114,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -283,6 +285,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -304,6 +307,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -474,6 +478,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -495,6 +500,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -682,6 +688,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -703,6 +710,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 +bucketing_version 2 column.name.d
[19/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out b/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out index 6a654d0..42d5528 100644 --- a/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out +++ b/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out @@ -651,314 +651,314 @@ POSTHOOK: query: select unionsrc.key, unionsrc.value FROM (select s1.key as key, POSTHOOK: type: QUERY POSTHOOK: Input: default@src POSTHOOK: Output: hdfs://### HDFS PATH ### -0 val_0 -103val_103 -105val_105 116val_116 +118val_118 120val_120 -129val_129 131val_131 -136val_136 +133val_133 +15 val_15 +150val_150 +152val_152 +153val_153 155val_155 -157val_157 160val_160 -180val_180 +169val_169 +175val_175 +176val_176 +178val_178 183val_183 -190val_190 -191val_191 -193val_193 -205val_205 -207val_207 +186val_186 +194val_194 +202val_202 +203val_203 213val_213 223val_223 -235val_235 -238val_238 +224val_224 +226val_226 +228val_228 239val_239 -24 val_24 -248val_248 -260val_260 +244val_244 +247val_247 +252val_252 +255val_255 +256val_256 +257val_257 +266val_266 +272val_272 +274val_274 +275val_275 28 val_28 +284val_284 286val_286 -298val_298 -305val_305 +289val_289 +296val_296 +302val_302 306val_306 307val_307 -308val_308 -315val_315 317val_317 321val_321 +323val_323 33 val_33 -331val_331 -356val_356 +338val_338 +34 val_34 +35 val_35 360val_360 -366val_366 -367val_367 -368val_368 -37 val_37 -378val_378 -379val_379 -382val_382 -389val_389 -395val_395 -400val_400 -402val_402 +365val_365 +374val_374 +377val_377 +393val_393 +399val_399 +404val_404 406val_406 -407val_407 +409val_409 +41 val_41 +417val_417 +418val_418 42 val_42 -430val_430 -436val_436 +431val_431 +439val_439 44 val_44 -452val_452 -466val_466 -467val_467 -472val_472 -480val_480 -484val_484 -485val_485 -487val_487 -495val_495 -496val_496 -498val_498 -51 val_51 -57 val_57 -58 val_58 -65 val_65 +443val_443 +448val_448 +455val_455 +459val_459 +462val_462 +463val_463 +470val_470 +490val_490 +491val_491 +494val_494 +5 val_5 +53 val_53 66 val_66 -69 val_69 -70 val_70 -78 val_78 -8 val_8 -92 val_92 -96 val_96 -104val_104 +72 val_72 +76 val_76 +77 val_77 +80 val_80 +84 val_84 +90 val_90 +10 val_10 +105val_105 111val_111 -114val_114 -118val_118 -12 val_12 -133val_133 -138val_138 -153val_153 -156val_156 -162val_162 -168val_168 +113val_113 +119val_119 +129val_129 +136val_136 +143val_143 +149val_149 +163val_163 +167val_167 174val_174 -176val_176 -178val_178 -181val_181 -186val_186 187val_187 -192val_192 197val_197 -199val_199 2 val_2 -201val_201 -203val_203 +200val_200 208val_208 -214val_214 +209val_209 +216val_216 +217val_217 218val_218 -219val_219 +221val_221 230val_230 -237val_237 -249val_249 -256val_256 -257val_257 +258val_258 +26 val_26 +263val_263 +273val_273 277val_277 -278val_278 -284val_284 -285val_285 -310val_310 -322val_322 -327val_327 -333val_333 -336val_336 -338val_338 +281val_281 +292val_292 +30 val_30 +316val_316 +335val_335 341val_341 +342val_342 344val_344 -348val_348 -351val_351 +345val_345 362val_362 -364val_364 -374val_374 +367val_367 +368val_368 +37 val_37 392val_392 -397val_397 +394val_394 4 val_4 -403val_403 -404val_404 -419val_419 -421val_421 -427val_427 -429val_429 +413val_413 437val_437 -438val_438 -439val_439 -454val_454 -455val_455 +444val_444 +446val_446 +452val_452 +453val_453 +457val_457 460val_460 -463val_463 -470val_470 -478val_478 +466val_466 +468val_468 +477val_477 479val_479 -489val_489 -491val_491 +483val_483 +487val_487 492val_492 -53 val_53 +493val_493 +498val_498 +51 val_51 +54 val_54 +64 val_64 +65 val_65 67 val_67 82 val_82 -90 val_90 -97 val_97 -10 val_10 -125val_125 -134val_134 -143val_143 +85 val_85
[18/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out b/ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out index 687b4af..8bec959 100644 --- a/ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out +++ b/ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out @@ -232,17 +232,17 @@ POSTHOOK: query: select fl_date, count(*) from flights_tiny_orc group by fl_date POSTHOOK: type: QUERY POSTHOOK: Input: default@flights_tiny_orc A masked pattern was here -2010-10-20 11 2010-10-21 12 -2010-10-26 13 -2010-10-30 11 -2010-10-23 12 2010-10-24 12 -2010-10-25 12 -2010-10-22 11 +2010-10-26 13 2010-10-28 12 +2010-10-22 11 2010-10-29 12 +2010-10-20 11 +2010-10-23 12 +2010-10-25 12 2010-10-27 11 +2010-10-30 11 2010-10-31 8 PREHOOK: query: explain vectorization expression select * from flights_tiny_orc sort by fl_num, fl_date limit 25 @@ -529,17 +529,17 @@ POSTHOOK: query: select fl_date, count(*) from flights_tiny_orc group by fl_date POSTHOOK: type: QUERY POSTHOOK: Input: default@flights_tiny_orc A masked pattern was here -2010-10-20 11 2010-10-21 12 -2010-10-26 13 -2010-10-30 11 -2010-10-23 12 2010-10-24 12 -2010-10-25 12 -2010-10-22 11 +2010-10-26 13 2010-10-28 12 +2010-10-22 11 2010-10-29 12 +2010-10-20 11 +2010-10-23 12 +2010-10-25 12 2010-10-27 11 +2010-10-30 11 2010-10-31 8 PREHOOK: query: CREATE TABLE flights_tiny_orc_partitioned_date ( origin_city_name STRING, @@ -908,17 +908,17 @@ POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-29 POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-30 POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-31 A masked pattern was here -2010-10-20 11 2010-10-21 12 -2010-10-26 13 -2010-10-30 11 -2010-10-23 12 2010-10-24 12 -2010-10-25 12 -2010-10-22 11 +2010-10-26 13 2010-10-28 12 +2010-10-22 11 2010-10-29 12 +2010-10-20 11 +2010-10-23 12 +2010-10-25 12 2010-10-27 11 +2010-10-30 11 2010-10-31 8 PREHOOK: query: explain vectorization expression select * from flights_tiny_orc_partitioned_date @@ -1484,17 +1484,17 @@ POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-29 POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-30 POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-31 A masked pattern was here -2010-10-20 11 2010-10-21 12 -2010-10-26 13 -2010-10-30 11 -2010-10-23 12 2010-10-24 12 -2010-10-25 12 -2010-10-22 11 +2010-10-26 13 2010-10-28 12 +2010-10-22 11 2010-10-29 12 +2010-10-20 11 +2010-10-23 12 +2010-10-25 12 2010-10-27 11 +2010-10-30 11 2010-10-31 8 PREHOOK: query: CREATE TABLE flights_tiny_orc_partitioned_timestamp ( origin_city_name STRING, @@ -1863,18 +1863,18 @@ POSTHOOK: Input: default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10- POSTHOOK: Input: default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-30 07%3A00%3A00 POSTHOOK: Input: default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-31 07%3A00%3A00 A masked pattern was here -2010-10-20 07:00:0011 2010-10-23 07:00:0012 2010-10-25 07:00:0012 2010-10-22 07:00:0011 +2010-10-24 07:00:0012 2010-10-26 07:00:0013 +2010-10-20 07:00:0011 +2010-10-28 07:00:0012 2010-10-29 07:00:0012 +2010-10-30 07:00:0011 2010-10-31 07:00:008 2010-10-21 07:00:0012 -2010-10-24 07:00:0012 2010-10-27 07:00:0011 -2010-10-28 07:00:0012 -2010-10-30 07:00:0011 PREHOOK: query: explain vectorization expression select * from flights_tiny_orc_partitioned_timestamp PREHOOK: type: QUERY @@ -2439,18 +2439,18 @@ POSTHOOK: Input: default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10- POSTHOOK: Input: default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-30 07%3A00%3A00 POSTHOOK: Input: default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-31 07%3A00%3A00 A masked pattern was here -2010-10-20 07:00:0011 2010-10-23 07:00:0012 2010-10-25 07:00:0012 2010-10-22 07:00:0011 +2010-10-24 07:00:0012 2010-10-26 07:00:0013 +2010-10-20 07:00:0011 +2010-10-28 07:00:0012 2010-10-29 07:00:0012 +2010-10-30 07:00:0011 2010-10-31 07:00:008 2010-10-21 07:00:0012 -2010-10-24 07:00:0012 2010-10-27 07:00:0011 -2010-10-28 07:00:0012 -2010-10-30 07:00:
[12/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/sample7.q.out -- diff --git a/ql/src/test/results/clientpositive/sample7.q.out b/ql/src/test/results/clientpositive/sample7.q.out index f24f297..07b6f0e 100644 --- a/ql/src/test/results/clientpositive/sample7.q.out +++ b/ql/src/test/results/clientpositive/sample7.q.out @@ -55,6 +55,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -102,6 +103,7 @@ STAGE PLANS: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -124,6 +126,7 @@ STAGE PLANS: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -193,6 +196,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -236,6 +240,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -266,6 +271,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -287,6 +293,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -323,6 +330,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -353,6 +361,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -374,6 +383,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -424,107 +434,48 @@ order by key, value POSTHOOK: type: QUERY POSTHOOK: Input: default@dest1 A masked pattern was here -104val_104 -104val_104 -104val_105 -104val_105 -104val_105 +110val_111 116val_116 116val_117 -120val_120 -120val_120 -120val_121 -128val_128 -128val_128 -128val_128 -128val_129 -128val_129 -132val_133 -132val_133 -136val_136 -136val_137 -140val_141 -144val_145 -152val_152 -152val_152 -152val_153 -152val_153 -152val_153 +126val_126 +126val_127 +126val_127 +134val_134 +134val_134 +134val_135 156val_156 156val_157 156val_157 -160val_160 -160val_161 -164val_164 -164val_164 -164val_165 -164val_165 -168val_168 -168val_169 -172val_172 -172val_172 -172val_173 -176val_176 -176val_176 -180val_180 +158val_158 +178val_178 +178val_179 +178
[20/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/sysdb.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/sysdb.q.out b/ql/src/test/results/clientpositive/llap/sysdb.q.out index 7d977d3..de20f2d 100644 --- a/ql/src/test/results/clientpositive/llap/sysdb.q.out +++ b/ql/src/test/results/clientpositive/llap/sysdb.q.out @@ -58,34 +58,10 @@ default alltypesorc hive_test_user USERDELETE true-1 hive_test_user defaultalltypesorc hive_test_user USERINSERT true-1 hive_test_user defaultalltypesorc hive_test_user USERSELECT true-1 hive_test_user defaultalltypesorc hive_test_user USERUPDATE true-1 hive_test_user -defaultalltypesparquet hive_test_user USERDELETE true-1 hive_test_user -defaultalltypesparquet hive_test_user USERINSERT true-1 hive_test_user -defaultalltypesparquet hive_test_user USERSELECT true-1 hive_test_user -defaultalltypesparquet hive_test_user USERUPDATE true-1 hive_test_user -defaultcbo_t1 hive_test_user USERDELETE true -1 hive_test_user -defaultcbo_t1 hive_test_user USERINSERT true -1 hive_test_user -defaultcbo_t1 hive_test_user USERSELECT true -1 hive_test_user -defaultcbo_t1 hive_test_user USERUPDATE true -1 hive_test_user -defaultcbo_t2 hive_test_user USERDELETE true -1 hive_test_user -defaultcbo_t2 hive_test_user USERINSERT true -1 hive_test_user -defaultcbo_t2 hive_test_user USERSELECT true -1 hive_test_user -defaultcbo_t2 hive_test_user USERUPDATE true -1 hive_test_user -defaultcbo_t3 hive_test_user USERDELETE true -1 hive_test_user -defaultcbo_t3 hive_test_user USERINSERT true -1 hive_test_user -defaultcbo_t3 hive_test_user USERSELECT true -1 hive_test_user -defaultcbo_t3 hive_test_user USERUPDATE true -1 hive_test_user -defaultlineitemhive_test_user USERDELETE true-1 hive_test_user -defaultlineitemhive_test_user USERINSERT true-1 hive_test_user -defaultlineitemhive_test_user USERSELECT true-1 hive_test_user -defaultlineitemhive_test_user USERUPDATE true-1 hive_test_user defaultmoretypes hive_test_user USERDELETE true-1 hive_test_user defaultmoretypes hive_test_user USERINSERT true-1 hive_test_user defaultmoretypes hive_test_user USERSELECT true-1 hive_test_user defaultmoretypes hive_test_user USERUPDATE true-1 hive_test_user -defaultparthive_test_user USERDELETE true -1 hive_test_user -defaultparthive_test_user USERINSERT true -1 hive_test_user -defaultparthive_test_user USERSELECT true -1 hive_test_user -defaultparthive_test_user USERUPDATE true -1 hive_test_user defaultscr_txn hive_test_user USERDELETE true -1 hive_test_user defaultscr_txn hive_test_user USERINSERT true -1 hive_test_user defaultscr_txn hive_test_user USERSELECT true -1 hive_test_user @@ -94,46 +70,14 @@ default src hive_test_user USER DELETE true-1 hive_test_user defaultsrc hive_test_user USERINSERT true -1 hive_test_user defaultsrc hive_test_user USERSELECT true -1 hive_test_user defaultsrc hive_test_user USERUPDATE true -1 hive_test_user -defaultsrc1hive_test_user USERDELETE true -1 hive_test_user -defaultsrc1hive_test_user USERINSERT true -1 hive_test_user -defaultsrc1hive_test_user USERSELECT true -1 hive_test_user -defaultsrc1
[34/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientnegative/alter_file_format.q.out -- diff --git a/ql/src/test/results/clientnegative/alter_file_format.q.out b/ql/src/test/results/clientnegative/alter_file_format.q.out index 96259e5..3033cbe 100644 --- a/ql/src/test/results/clientnegative/alter_file_format.q.out +++ b/ql/src/test/results/clientnegative/alter_file_format.q.out @@ -24,6 +24,7 @@ Retention:0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} + bucketing_version 2 numFiles0 numRows 0 rawDataSize 0 http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out -- diff --git a/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out b/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out index 8e240ea..90c3d02 100644 --- a/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out +++ b/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out @@ -51,6 +51,7 @@ Retention:0 Table Type:VIRTUAL_VIEW Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + bucketing_version 2 numFiles0 numPartitions 2 numRows 0 http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientnegative/unset_table_property.q.out -- diff --git a/ql/src/test/results/clientnegative/unset_table_property.q.out b/ql/src/test/results/clientnegative/unset_table_property.q.out index 2dd047a..eb308eb 100644 --- a/ql/src/test/results/clientnegative/unset_table_property.q.out +++ b/ql/src/test/results/clientnegative/unset_table_property.q.out @@ -20,6 +20,7 @@ POSTHOOK: query: SHOW TBLPROPERTIES testTable POSTHOOK: type: SHOW_TBLPROPERTIES COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"col1":"true","col2":"true"}} a 1 +bucketing_version 2 c 3 A masked pattern was here numFiles 0 http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientnegative/unset_view_property.q.out -- diff --git a/ql/src/test/results/clientnegative/unset_view_property.q.out b/ql/src/test/results/clientnegative/unset_view_property.q.out index ae7f4f6..b5038d9 100644 --- a/ql/src/test/results/clientnegative/unset_view_property.q.out +++ b/ql/src/test/results/clientnegative/unset_view_property.q.out @@ -21,6 +21,7 @@ PREHOOK: query: SHOW TBLPROPERTIES testView PREHOOK: type: SHOW_TBLPROPERTIES POSTHOOK: query: SHOW TBLPROPERTIES testView POSTHOOK: type: SHOW_TBLPROPERTIES +bucketing_version 2 A masked pattern was here propA 100 propB 200 http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/acid_mapjoin.q.out -- diff --git a/ql/src/test/results/clientpositive/acid_mapjoin.q.out b/ql/src/test/results/clientpositive/acid_mapjoin.q.out index 5569a03..76a781e 100644 --- a/ql/src/test/results/clientpositive/acid_mapjoin.q.out +++ b/ql/src/test/results/clientpositive/acid_mapjoin.q.out @@ -73,21 +73,21 @@ STAGE PLANS: Stage: Stage-5 Map Reduce Local Work Alias -> Map Local Tables: -$hdt$_1:acid2 +$hdt$_0:acid1 Fetch Operator limit: -1 Alias -> Map Local Operator Tree: -$hdt$_1:acid2 +$hdt$_0:acid1 TableScan -alias: acid2 -Statistics: Num rows: 210 Data size: 840 Basic stats: COMPLETE Column stats: NONE +alias: acid1 +Statistics: Num rows: 316 Data size: 1265 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: key is not null (type: boolean) - Statistics: Num rows: 210 Data size: 840 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 316 Data size: 1265 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: int) outputColumnNames: _col0 -Statistics: Num rows: 210 Data size: 840 Basic stats: COMPLETE Column stats: NONE +
[23/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/default_constraint.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/default_constraint.q.out b/ql/src/test/results/clientpositive/llap/default_constraint.q.out index 15cbd84..d987ef2 100644 --- a/ql/src/test/results/clientpositive/llap/default_constraint.q.out +++ b/ql/src/test/results/clientpositive/llap/default_constraint.q.out @@ -31,6 +31,7 @@ Retention:0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 transactional true transactional_propertiesdefault A masked pattern was here @@ -301,6 +302,7 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 A masked pattern was here # Storage Information @@ -530,6 +532,7 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 transactional true transactional_propertiesdefault A masked pattern was here @@ -788,6 +791,7 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 transactional true transactional_propertiesdefault A masked pattern was here @@ -1059,6 +1063,7 @@ Retention:0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 A masked pattern was here # Storage Information @@ -1336,6 +1341,7 @@ Retention:0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 transactional true transactional_propertiesdefault A masked pattern was here @@ -1490,8 +1496,9 @@ Retention:0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 numFiles1 - totalSize 1062 + totalSize 1070 transactional true transactional_propertiesdefault A masked pattern was here @@ -1658,9 +1665,10 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 A masked pattern was here numFiles2 - totalSize 2124 + totalSize 2140 transactional true transactional_propertiesdefault A masked pattern was here @@ -1737,9 +1745,10 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 A masked pattern was here numFiles2 - totalSize 2124 + totalSize 2140 transactional true transactional_propertiesdefault A masked pattern was here @@ -1913,9 +1922,10 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 A masked pattern was here numFiles3 - totalSize 3188 + totalSize 3199 transactional true transactional_propertiesdefault A masked pattern was here @@ -1991,9 +2001,10 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TAB
[22/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/intersect_distinct.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/intersect_distinct.q.out b/ql/src/test/results/clientpositive/llap/intersect_distinct.q.out index 44362be..fb59f10 100644 --- a/ql/src/test/results/clientpositive/llap/intersect_distinct.q.out +++ b/ql/src/test/results/clientpositive/llap/intersect_distinct.q.out @@ -281,315 +281,315 @@ POSTHOOK: query: select * from src intersect distinct select * from src POSTHOOK: type: QUERY POSTHOOK: Input: default@src POSTHOOK: Output: hdfs://### HDFS PATH ### -0 val_0 -103val_103 +104val_104 105val_105 -116val_116 +113val_113 +118val_118 +12 val_12 120val_120 -129val_129 -131val_131 +133val_133 136val_136 155val_155 -157val_157 +158val_158 160val_160 +162val_162 +168val_168 +17 val_17 +170val_170 +175val_175 +18 val_18 180val_180 -183val_183 -190val_190 -191val_191 -193val_193 -205val_205 -207val_207 -213val_213 -223val_223 +186val_186 +19 val_19 +197val_197 +200val_200 +216val_216 +218val_218 +222val_222 +224val_224 +228val_228 235val_235 -238val_238 +237val_237 239val_239 -24 val_24 -248val_248 -260val_260 +244val_244 +247val_247 +256val_256 +263val_263 +273val_273 28 val_28 +283val_283 286val_286 -298val_298 -305val_305 +292val_292 306val_306 -307val_307 308val_308 +327val_327 +33 val_33 +348val_348 +353val_353 +362val_362 +366val_366 +396val_396 +397val_397 +401val_401 +409val_409 +411val_411 +419val_419 +427val_427 +43 val_43 +432val_432 +436val_436 +439val_439 +443val_443 +453val_453 +460val_460 +462val_462 +47 val_47 +472val_472 +485val_485 +496val_496 +54 val_54 +64 val_64 +70 val_70 +8 val_8 +83 val_83 +84 val_84 +85 val_85 +90 val_90 +0 val_0 +103val_103 +114val_114 +125val_125 +138val_138 +146val_146 +150val_150 +152val_152 +153val_153 +156val_156 +157val_157 +165val_165 +172val_172 +177val_177 +179val_179 +187val_187 +195val_195 +196val_196 +217val_217 +242val_242 +248val_248 +252val_252 +265val_265 +27 val_27 +272val_272 +280val_280 +291val_291 +305val_305 +309val_309 +311val_311 315val_315 317val_317 -321val_321 -33 val_33 -331val_331 +322val_322 +333val_333 +34 val_34 +345val_345 +35 val_35 356val_356 -360val_360 -366val_366 -367val_367 +364val_364 368val_368 +369val_369 37 val_37 -378val_378 -379val_379 -382val_382 -389val_389 -395val_395 -400val_400 +373val_373 +377val_377 +4 val_4 402val_402 -406val_406 -407val_407 +404val_404 +413val_413 42 val_42 430val_430 -436val_436 -44 val_44 +431val_431 +444val_444 +449val_449 452val_452 +454val_454 +457val_457 +463val_463 466val_466 -467val_467 -472val_472 -480val_480 -484val_484 -485val_485 -487val_487 -495val_495 -496val_496 -498val_498 -51 val_51 +470val_470 +475val_475 +481val_481 +489val_489 +491val_491 57 val_57 -58 val_58 65 val_65 66 val_66 -69 val_69 -70 val_70 +74 val_74 +76 val_76 78 val_78 -8 val_8 +9 val_9 92 val_92 -96 val_96 -104val_104 -111val_111 -114val_114 -118val_118 -12 val_12 -133val_133 -138val_138 -153val_153 -156val_156 -162val_162 -168val_168 +95 val_95 +100val_100 +119val_119 +128val_128 +129val_129 +145val_145 +167val_167 174val_174 -176val_176 178val_178 181val_181 -186val_186 -187val_187 -192val_192 -197val_197 +193val_193 199val_199 -2 val_2 +20 val_20 201val_201 -203val_203 -208val_208 +213val_213 214val_214 -218val_218 219val_219 -230val_230 -237val_237 +221val_221 +223val_223 +226val_226 +233val_233 +241val_241 249val_249 -256val_256 -257val_257 +260val_260 +262val_262 277val_277 -278val_278 -284val_284 -285val_285 +281val_281 +287val_287 +288val_288 +298val_298 +302val_302 310val_310 -322val_322 -327val_327 -333val_333 +323val_323 336val_336 -338val_338 341val_341 344val_344 -348val_348 351val_351 -362val_362 -364val_364 -374val_374 -392val_392 -397val_397 -4 v
[01/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
Repository: hive Updated Branches: refs/heads/master a095c3d93 -> ac6b2a3fb http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/union_remove_7.q.out -- diff --git a/ql/src/test/results/clientpositive/union_remove_7.q.out b/ql/src/test/results/clientpositive/union_remove_7.q.out index b60f7bb..437b08a 100644 --- a/ql/src/test/results/clientpositive/union_remove_7.q.out +++ b/ql/src/test/results/clientpositive/union_remove_7.q.out @@ -173,6 +173,7 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 numFiles2 totalSize 178 A masked pattern was here http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/union_remove_8.q.out -- diff --git a/ql/src/test/results/clientpositive/union_remove_8.q.out b/ql/src/test/results/clientpositive/union_remove_8.q.out index 0fbaa3c..4971a74 100644 --- a/ql/src/test/results/clientpositive/union_remove_8.q.out +++ b/ql/src/test/results/clientpositive/union_remove_8.q.out @@ -183,6 +183,7 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 numFiles3 totalSize 271 A masked pattern was here http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/union_remove_9.q.out -- diff --git a/ql/src/test/results/clientpositive/union_remove_9.q.out b/ql/src/test/results/clientpositive/union_remove_9.q.out index 2dadf0f..786a81f 100644 --- a/ql/src/test/results/clientpositive/union_remove_9.q.out +++ b/ql/src/test/results/clientpositive/union_remove_9.q.out @@ -230,6 +230,7 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 numFiles2 totalSize 192 A masked pattern was here http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/union_stats.q.out -- diff --git a/ql/src/test/results/clientpositive/union_stats.q.out b/ql/src/test/results/clientpositive/union_stats.q.out index 5b6ada7..8bd3f44 100644 --- a/ql/src/test/results/clientpositive/union_stats.q.out +++ b/ql/src/test/results/clientpositive/union_stats.q.out @@ -90,6 +90,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -111,6 +112,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -313,6 +315,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + bucketing_version 2 numFiles1 numRows 1000 rawDataSize 10624 @@ -361,6 +364,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + bucketing_version 2 numFiles1 numRows 1500 rawDataSize 15936 @@ -417,6 +421,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + bucketing_version 2 numFiles1 numRows 1500 rawDataSize 15936 http://git-wip-us.apache.o
[14/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/pcr.q.out -- diff --git a/ql/src/test/results/clientpositive/pcr.q.out b/ql/src/test/results/clientpositive/pcr.q.out index 0adee06..e30f6d1 100644 --- a/ql/src/test/results/clientpositive/pcr.q.out +++ b/ql/src/test/results/clientpositive/pcr.q.out @@ -116,6 +116,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -163,6 +164,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -306,6 +308,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -353,6 +356,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -400,6 +404,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -581,6 +586,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -628,6 +634,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -773,6 +780,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -820,6 +828,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -967,6 +976,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -1014,6 +1024,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -1061,6 +1072,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -1219,6 +1231,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -1266,6 +1279,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter ,
[05/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/sample6.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/sample6.q.out b/ql/src/test/results/clientpositive/spark/sample6.q.out index 34aee1e..cf172ea 100644 --- a/ql/src/test/results/clientpositive/spark/sample6.q.out +++ b/ql/src/test/results/clientpositive/spark/sample6.q.out @@ -52,6 +52,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -84,6 +85,7 @@ STAGE PLANS: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -106,6 +108,7 @@ STAGE PLANS: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -137,6 +140,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -181,172 +185,175 @@ order by key, value POSTHOOK: type: QUERY POSTHOOK: Input: default@dest1 A masked pattern was here -0 val_0 -0 val_0 -0 val_0 -0 val_1 -0 val_1 -4 val_4 -4 val_5 -8 val_8 -8 val_9 -12 val_12 -12 val_12 -12 val_13 -16 val_17 -16 val_17 -20 val_20 -20 val_21 -20 val_21 -24 val_24 -24 val_24 -28 val_28 -32 val_33 +2 val_2 +2 val_3 +6 val_7 +6 val_7 +10 val_10 +10 val_11 +17 val_17 +21 val_22 +21 val_22 +21 val_22 +21 val_22 +27 val_27 +30 val_30 +30 val_31 +31 val_32 40 val_41 40 val_41 -44 val_44 -48 val_49 -48 val_49 -52 val_53 -52 val_53 -52 val_53 -52 val_53 -56 val_57 -60 val_61 +51 val_51 +51 val_51 +51 val_52 +57 val_57 +58 val_58 +58 val_58 +58 val_59 +58 val_59 64 val_64 -68 val_69 -72 val_72 -72 val_72 -76 val_76 -76 val_76 -76 val_77 -76 val_77 -76 val_77 +65 val_65 +65 val_66 +65 val_66 +70 val_70 +70 val_70 +70 val_70 +70 val_71 80 val_80 80 val_81 -84 val_84 -84 val_84 -92 val_92 -96 val_96 -100val_100 -100val_100 -100val_101 -100val_101 -104val_104 -104val_104 -104val_105 -104val_105 -104val_105 +83 val_83 +83 val_83 +86 val_86 +86 val_87 +90 val_90 +90 val_90 +90 val_90 +91 val_92 +98 val_98 +98 val_98 +105val_105 +105val_106 +105val_106 +110val_111 +113val_113 +113val_113 116val_116 116val_117 -120val_120 -120val_120 -120val_121 -128val_128 -128val_128 -128val_128 -128val_129 -128val_129 -132val_133 -132val_133 -136val_136 -136val_137 -140val_141 -144val_145 -152val_152 -152val_152 -152val_153 -152val_153 -152val_153 +117val_118 +117val_118 +119val_119 +119val_119 +119val_119 +119val_120 +119val_120 +119val_120 +121val_122 +121val_122 +123val_124 +123val_124 +126val_126 +126val_127 +126val_127 +134val_134 +134val_134 +134val_135 +137val_137 +137val_137 +137val_138 +153val_153 +153val_154 +153val_154 156val_156 156val_157 156val_157 -160val_160 -160val_161 -164val_164 -164val_164 -164val_165 -164val_165 -168val_168 -168val_169 -172val_172 -172val_172 -172val_173 -176val_176 -176val_176 -180val_180 +157val_157 +157val_158 +157val_158 +158val_158 +163val_163 +169val_169 +169val_169 +169val_169 +169val_169 +177val_177 +177val_178 +177val_178 +178val_178 +178val_179 +178val_179 184val_185 -192val_192 -192val_193 -196
[29/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out -- diff --git a/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out b/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out index cb897d1..5a00d10 100644 --- a/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out +++ b/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out @@ -87,6 +87,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -135,6 +136,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -158,6 +160,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -227,6 +230,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -270,6 +274,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -300,6 +305,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -321,6 +327,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -357,6 +364,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -387,6 +395,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -408,6 +417,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -523,6 +533,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -546,6 +557,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -659,6 +671,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key1,key2,cnt columns.comments @@ -717,6 +730,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE
[26/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/bucket_num_reducers2.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/bucket_num_reducers2.q.out b/ql/src/test/results/clientpositive/llap/bucket_num_reducers2.q.out new file mode 100644 index 000..79d9755 --- /dev/null +++ b/ql/src/test/results/clientpositive/llap/bucket_num_reducers2.q.out @@ -0,0 +1,242 @@ +PREHOOK: query: CREATE TABLE test_table(key int, value string) CLUSTERED BY (key) INTO 3 BUCKETS +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@test_table +POSTHOOK: query: CREATE TABLE test_table(key int, value string) CLUSTERED BY (key) INTO 3 BUCKETS +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@test_table +PREHOOK: query: explain extended insert overwrite table test_table + select * from src +PREHOOK: type: QUERY +POSTHOOK: query: explain extended insert overwrite table test_table + select * from src +POSTHOOK: type: QUERY +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 +Tez + A masked pattern was here + Edges: +Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) + A masked pattern was here + Vertices: +Map 1 +Map Operator Tree: +TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator +expressions: key (type: string), value (type: string) +outputColumnNames: _col0, _col1 +Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE +Reduce Output Operator + null sort order: + sort order: + Map-reduce partition columns: UDFToInteger(_col0) (type: int) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: string), _col1 (type: string) + auto parallelism: false +Execution mode: llap +LLAP IO: no inputs +Path -> Alias: + A masked pattern was here +Path -> Partition: + A masked pattern was here +Partition + base file name: src + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: +COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} +bucket_count -1 +bucketing_version 2 +column.name.delimiter , +columns key,value +columns.comments 'default','default' +columns.types string:string + A masked pattern was here +name default.src +numFiles 1 +numRows 500 +rawDataSize 5312 +serialization.ddl struct src { string key, string value} +serialization.format 1 +serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +totalSize 5812 + A masked pattern was here + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + +input format: org.apache.hadoop.mapred.TextInputFormat +output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat +properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string + A masked pattern was here + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 + A masked pattern was here +serde:
[10/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/bucket_map_join_spark4.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/bucket_map_join_spark4.q.out b/ql/src/test/results/clientpositive/spark/bucket_map_join_spark4.q.out index 68889d3..6ac9dd4 100644 --- a/ql/src/test/results/clientpositive/spark/bucket_map_join_spark4.q.out +++ b/ql/src/test/results/clientpositive/spark/bucket_map_join_spark4.q.out @@ -110,6 +110,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -133,6 +134,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -186,6 +188,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -209,6 +212,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -310,6 +314,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -333,6 +338,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -483,6 +489,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -506,6 +513,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -559,6 +567,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -582,6 +591,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -683,6 +693,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -706,6 +717,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/bucket_map_join_tez1.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/bucket_map_join_tez1.q.out b/ql/src/test/results/clientpositive/spark/bucket_map_join_tez1.q.out index 158d1c3..987a98c 100644 --- a/ql/src/test/results/clientpositive/spark/bucket_map_join_tez1.q.out +++ b/ql/src/test/results/clientpositive/spark
[25/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/column_table_stats.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/column_table_stats.q.out b/ql/src/test/results/clientpositive/llap/column_table_stats.q.out index 7f3cc02..1457932 100644 --- a/ql/src/test/results/clientpositive/llap/column_table_stats.q.out +++ b/ql/src/test/results/clientpositive/llap/column_table_stats.q.out @@ -35,6 +35,7 @@ Retention:0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 numFiles1 numRows 0 rawDataSize 0 @@ -102,6 +103,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -122,6 +124,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -211,6 +214,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} + bucketing_version 2 numFiles1 numRows 500 rawDataSize 5312 @@ -287,6 +291,7 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 numFiles2 numPartitions 2 numRows 0 @@ -383,6 +388,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -430,6 +436,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -536,6 +543,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + bucketing_version 2 numFiles2 numPartitions 2 numRows 1000 @@ -693,6 +701,7 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 numFiles2 numPartitions 2 numRows 0 @@ -789,6 +798,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -836,6 +846,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -942,6 +953,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters:
[08/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/groupby_sort_skew_1_23.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/groupby_sort_skew_1_23.q.out b/ql/src/test/results/clientpositive/spark/groupby_sort_skew_1_23.q.out index 9b068ed..d8b9dca 100644 --- a/ql/src/test/results/clientpositive/spark/groupby_sort_skew_1_23.q.out +++ b/ql/src/test/results/clientpositive/spark/groupby_sort_skew_1_23.q.out @@ -85,6 +85,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -118,6 +119,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -141,6 +143,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -172,6 +175,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -287,6 +291,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -310,6 +315,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -375,6 +381,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key1,key2,cnt columns.comments @@ -406,6 +413,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key1,key2,cnt columns.comments @@ -506,6 +514,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -539,6 +548,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -562,6 +572,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -593,6 +604,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -691,6 +703,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
[06/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/sample1.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/sample1.q.out b/ql/src/test/results/clientpositive/spark/sample1.q.out index b711292..164378b 100644 --- a/ql/src/test/results/clientpositive/spark/sample1.q.out +++ b/ql/src/test/results/clientpositive/spark/sample1.q.out @@ -53,6 +53,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value,dt,hr columns.comments @@ -109,6 +110,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -138,6 +140,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value,dt,hr columns.comments http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/sample10.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/sample10.q.out b/ql/src/test/results/clientpositive/spark/sample10.q.out index 418f63a..ac28779 100644 --- a/ql/src/test/results/clientpositive/spark/sample10.q.out +++ b/ql/src/test/results/clientpositive/spark/sample10.q.out @@ -34,10 +34,10 @@ POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-09,hr=11).value SIMPLE [(s POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-09,hr=12).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-09,hr=12).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] PREHOOK: query: explain extended -select ds, count(1) from srcpartbucket tablesample (bucket 1 out of 4 on key) where ds is not null group by ds ORDER BY ds ASC +select ds, count(1) from srcpartbucket tablesample (bucket 2 out of 4 on key) where ds is not null group by ds ORDER BY ds ASC PREHOOK: type: QUERY POSTHOOK: query: explain extended -select ds, count(1) from srcpartbucket tablesample (bucket 1 out of 4 on key) where ds is not null group by ds ORDER BY ds ASC +select ds, count(1) from srcpartbucket tablesample (bucket 2 out of 4 on key) where ds is not null group by ds ORDER BY ds ASC POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage @@ -59,8 +59,8 @@ STAGE PLANS: GatherStats: false Filter Operator isSamplingPred: true -predicate: (((hash(key) & 2147483647) % 4) = 0) (type: boolean) -sampleDesc: BUCKET 1 OUT OF 4 +predicate: (((hash(key) & 2147483647) % 4) = 1) (type: boolean) +sampleDesc: BUCKET 2 OUT OF 4 Statistics: Num rows: 20 Data size: 120 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: ds (type: string) @@ -87,7 +87,7 @@ STAGE PLANS: Path -> Partition: A masked pattern was here Partition - base file name: 00_0 + base file name: 01_0 input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat partition values: @@ -120,6 +120,7 @@ STAGE PLANS: properties: bucket_count 4 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -137,7 +138,7 @@ STAGE PLANS: name: default.srcpartbucket A masked pattern was here Partition - base file name: 00_0 + base file name: 01_0 input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputF
[2/2] hive git commit: HIVE-19311 : Partition and bucketing support for “load data” statement (Deepak Jaiswal, reviewed by Prasanth Jayachandran)
HIVE-19311 : Partition and bucketing support for âload dataâ statement (Deepak Jaiswal, reviewed by Prasanth Jayachandran) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/46c5580b Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/46c5580b Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/46c5580b Branch: refs/heads/master Commit: 46c5580b735f9b3b4f3e482d8f7a524dd88d4771 Parents: e8651cb Author: Deepak Jaiswal Authored: Wed May 2 01:27:02 2018 -0700 Committer: Deepak Jaiswal Committed: Wed May 2 01:27:02 2018 -0700 -- data/files/load_data_job/bucketing.txt | 118 + .../load_data_job/load_data_1_partition.txt | 118 + .../partitions/load_data_1_partition.txt| 118 + .../partitions/load_data_2_partitions.txt | 118 + .../partitions/subdir/load_data_1_partition.txt | 118 + .../subdir/load_data_2_partitions.txt | 118 + .../test/resources/testconfiguration.properties |1 + .../java/org/apache/hadoop/hive/ql/Context.java | 12 +- .../org/apache/hadoop/hive/ql/ErrorMsg.java |2 + .../apache/hadoop/hive/ql/metadata/Table.java |6 +- .../hive/ql/parse/LoadSemanticAnalyzer.java | 172 +- .../hadoop/hive/ql/parse/SemanticAnalyzer.java | 43 +- .../ql/parse/UpdateDeleteSemanticAnalyzer.java | 43 +- .../apache/hadoop/hive/ql/TestTxnLoadData.java |5 +- .../queries/clientnegative/load_part_nospec.q |2 - .../test/queries/clientnegative/nopart_load.q |5 - .../clientpositive/load_data_using_job.q| 83 + .../clientnegative/load_part_nospec.q.out |9 - .../results/clientnegative/nopart_load.q.out|9 - .../llap/load_data_using_job.q.out | 2769 ++ 20 files changed, 3747 insertions(+), 122 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/46c5580b/data/files/load_data_job/bucketing.txt -- diff --git a/data/files/load_data_job/bucketing.txt b/data/files/load_data_job/bucketing.txt new file mode 100644 index 000..8334415 --- /dev/null +++ b/data/files/load_data_job/bucketing.txt @@ -0,0 +1,118 @@ +165val_165 +484val_484 +150val_150 +224val_224 +66val_66 +213val_213 +374val_374 +495val_495 +37val_37 +327val_327 +15val_15 +338val_338 +459val_459 +466val_466 +396val_396 +309val_309 +367val_367 +0val_0 +455val_455 +316val_316 +345val_345 +129val_129 +378val_378 +4val_4 +356val_356 +169val_169 +125val_125 +437val_437 +286val_286 +187val_187 +176val_176 +459val_459 +51val_51 +103val_103 +239val_239 +213val_213 +176val_176 +275val_275 +260val_260 +404val_404 +217val_217 +84val_84 +466val_466 +8val_8 +411val_411 +172val_172 +129val_129 +158val_158 +0val_0 +26val_26 +165val_165 +327val_327 +51val_51 +404val_404 +95val_95 +282val_282 +187val_187 +316val_316 +169val_169 +77val_77 +0val_0 +118val_118 +282val_282 +419val_419 +15val_15 +118val_118 +19val_19 +224val_224 +309val_309 +389val_389 +327val_327 +242val_242 +392val_392 +242val_242 +396val_396 +95val_95 +11val_11 +143val_143 +228val_228 +33val_33 +103val_103 +367val_367 +239val_239 +480val_480 +202val_202 +316val_316 +235val_235 +80val_80 +44val_44 +466val_466 +257val_257 +190val_190 +114val_114 +396val_396 +217val_217 +125val_125 +187val_187 +480val_480 +491val_491 +305val_305 +444val_444 +169val_169 +323val_323 +480val_480 +136val_136 +172val_172 +462val_462 +26val_26 +462val_462 +341val_341 +183val_183 +84val_84 +37val_37 +448val_448 +194val_194 +477val_477 +169val_169 +400val_400 http://git-wip-us.apache.org/repos/asf/hive/blob/46c5580b/data/files/load_data_job/load_data_1_partition.txt -- diff --git a/data/files/load_data_job/load_data_1_partition.txt b/data/files/load_data_job/load_data_1_partition.txt new file mode 100644 index 000..9f34604 --- /dev/null +++ b/data/files/load_data_job/load_data_1_partition.txt @@ -0,0 +1,118 @@ +165val_1652008-04-08 +484val_4842008-04-08 +150val_1502008-04-08 +224val_2242008-04-08 +66val_662008-04-08 +213val_2132008-04-08 +374val_3742008-04-08 +495val_4952008-04-08 +37val_372008-04-08 +327val_3272008-04-08 +15val_152008-04-08 +338val_3382008-04-08 +459val_4592008-04-08 +466val_4662008-04-08 +396val_3962008-04-08 +309val_3092008-04-08 +367val_3672008-04-08 +0val_02008-04-08 +455val_4552008-04-08 +316val_3162008-04-08 +345val_3452008-04-08 +129val_1292008-04-08 +378val_3782008-04-08 +4val_42008-04-08 +356val_3562008-04-08 +169val_1692008-04-08 +125val_1252008-04-08 +437val_4372008-04-08 +286val_2862008-04-08 +187val_1872008-04-08 +176val_
[1/2] hive git commit: HIVE-19311 : Partition and bucketing support for “load data” statement (Deepak Jaiswal, reviewed by Prasanth Jayachandran)
Repository: hive Updated Branches: refs/heads/master e8651cb97 -> 46c5580b7 http://git-wip-us.apache.org/repos/asf/hive/blob/46c5580b/ql/src/test/results/clientpositive/llap/load_data_using_job.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/load_data_using_job.q.out b/ql/src/test/results/clientpositive/llap/load_data_using_job.q.out new file mode 100644 index 000..8077310 --- /dev/null +++ b/ql/src/test/results/clientpositive/llap/load_data_using_job.q.out @@ -0,0 +1,2769 @@ +PREHOOK: query: CREATE TABLE srcbucket_mapjoin(key int, value string) partitioned by (ds string) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@srcbucket_mapjoin +POSTHOOK: query: CREATE TABLE srcbucket_mapjoin(key int, value string) partitioned by (ds string) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@srcbucket_mapjoin +PREHOOK: query: explain load data local inpath '../../data/files/bmj/00_0' INTO TABLE srcbucket_mapjoin partition(ds='2008-04-08') +PREHOOK: type: LOAD +POSTHOOK: query: explain load data local inpath '../../data/files/bmj/00_0' INTO TABLE srcbucket_mapjoin partition(ds='2008-04-08') +POSTHOOK: type: LOAD +STAGE DEPENDENCIES: + Stage-0 is a root stage + Stage-1 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-0 +Move Operator + tables: + partition: +ds 2008-04-08 + replace: false + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcbucket_mapjoin + + Stage: Stage-1 +Stats Work + Basic Stats Work: + +PREHOOK: query: load data local inpath '../../data/files/bmj/00_0' INTO TABLE srcbucket_mapjoin partition(ds='2008-04-08') +PREHOOK: type: LOAD + A masked pattern was here +PREHOOK: Output: default@srcbucket_mapjoin +POSTHOOK: query: load data local inpath '../../data/files/bmj/00_0' INTO TABLE srcbucket_mapjoin partition(ds='2008-04-08') +POSTHOOK: type: LOAD + A masked pattern was here +POSTHOOK: Output: default@srcbucket_mapjoin +POSTHOOK: Output: default@srcbucket_mapjoin@ds=2008-04-08 +PREHOOK: query: select * from srcbucket_mapjoin +PREHOOK: type: QUERY +PREHOOK: Input: default@srcbucket_mapjoin +PREHOOK: Input: default@srcbucket_mapjoin@ds=2008-04-08 + A masked pattern was here +POSTHOOK: query: select * from srcbucket_mapjoin +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcbucket_mapjoin +POSTHOOK: Input: default@srcbucket_mapjoin@ds=2008-04-08 + A masked pattern was here +165val_165 2008-04-08 +484val_484 2008-04-08 +150val_150 2008-04-08 +224val_224 2008-04-08 +66 val_66 2008-04-08 +213val_213 2008-04-08 +374val_374 2008-04-08 +495val_495 2008-04-08 +37 val_37 2008-04-08 +327val_327 2008-04-08 +15 val_15 2008-04-08 +338val_338 2008-04-08 +459val_459 2008-04-08 +466val_466 2008-04-08 +396val_396 2008-04-08 +309val_309 2008-04-08 +367val_367 2008-04-08 +0 val_0 2008-04-08 +455val_455 2008-04-08 +316val_316 2008-04-08 +345val_345 2008-04-08 +129val_129 2008-04-08 +378val_378 2008-04-08 +4 val_4 2008-04-08 +356val_356 2008-04-08 +169val_169 2008-04-08 +125val_125 2008-04-08 +437val_437 2008-04-08 +286val_286 2008-04-08 +187val_187 2008-04-08 +176val_176 2008-04-08 +459val_459 2008-04-08 +51 val_51 2008-04-08 +103val_103 2008-04-08 +239val_239 2008-04-08 +213val_213 2008-04-08 +176val_176 2008-04-08 +275val_275 2008-04-08 +260val_260 2008-04-08 +404val_404 2008-04-08 +217val_217 2008-04-08 +84 val_84 2008-04-08 +466val_466 2008-04-08 +8 val_8 2008-04-08 +411val_411 2008-04-08 +172val_172 2008-04-08 +129val_129 2008-04-08 +158val_158 2008-04-08 +0 val_0 2008-04-08 +26 val_26 2008-04-08 +165val_165 2008-04-08 +327val_327 2008-04-08 +51 val_51 2008-04-08 +404val_404 2008-04-08 +95 val_95 2008-04-08 +282val_282 2008-04-08 +187val_187 2008-04-08 +316val_316 2008-04-08 +169val_169 2008-04-08 +77 val_77 2008-04-08 +0 val_0 2008-04-08 +118val_118 2008-04-08 +282val_282 2008-04-08 +419val_419 2008-04-08 +15 val_15 2008-04-08 +118val_118 2008-04-08 +19 val_19 2008-04-08 +224val_224 2008-04-08 +309val_309 2008-04-08 +389val_389 2008-04-08 +327val_327 2008-04-08 +242val_242 2008-04-08 +392val_392 2008-04-08 +242val_242 2008-04-08 +396val_396 2008-04-08 +95 val_95 2008-04-08 +11 val_11 2008-04-08 +143val_143 2008-04-08 +228val_22
[25/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/column_table_stats.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/column_table_stats.q.out b/ql/src/test/results/clientpositive/llap/column_table_stats.q.out index 7f3cc02..1457932 100644 --- a/ql/src/test/results/clientpositive/llap/column_table_stats.q.out +++ b/ql/src/test/results/clientpositive/llap/column_table_stats.q.out @@ -35,6 +35,7 @@ Retention:0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 numFiles1 numRows 0 rawDataSize 0 @@ -102,6 +103,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -122,6 +124,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -211,6 +214,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} + bucketing_version 2 numFiles1 numRows 500 rawDataSize 5312 @@ -287,6 +291,7 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 numFiles2 numPartitions 2 numRows 0 @@ -383,6 +388,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -430,6 +436,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -536,6 +543,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + bucketing_version 2 numFiles2 numPartitions 2 numRows 1000 @@ -693,6 +701,7 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 numFiles2 numPartitions 2 numRows 0 @@ -789,6 +798,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -836,6 +846,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -942,6 +953,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters:
[29/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out -- diff --git a/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out b/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out index cb897d1..5a00d10 100644 --- a/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out +++ b/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out @@ -87,6 +87,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -135,6 +136,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -158,6 +160,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -227,6 +230,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -270,6 +274,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -300,6 +305,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -321,6 +327,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -357,6 +364,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -387,6 +395,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -408,6 +417,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -523,6 +533,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -546,6 +557,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -659,6 +671,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key1,key2,cnt columns.comments @@ -717,6 +730,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE
[22/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/intersect_distinct.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/intersect_distinct.q.out b/ql/src/test/results/clientpositive/llap/intersect_distinct.q.out index 44362be..fb59f10 100644 --- a/ql/src/test/results/clientpositive/llap/intersect_distinct.q.out +++ b/ql/src/test/results/clientpositive/llap/intersect_distinct.q.out @@ -281,315 +281,315 @@ POSTHOOK: query: select * from src intersect distinct select * from src POSTHOOK: type: QUERY POSTHOOK: Input: default@src POSTHOOK: Output: hdfs://### HDFS PATH ### -0 val_0 -103val_103 +104val_104 105val_105 -116val_116 +113val_113 +118val_118 +12 val_12 120val_120 -129val_129 -131val_131 +133val_133 136val_136 155val_155 -157val_157 +158val_158 160val_160 +162val_162 +168val_168 +17 val_17 +170val_170 +175val_175 +18 val_18 180val_180 -183val_183 -190val_190 -191val_191 -193val_193 -205val_205 -207val_207 -213val_213 -223val_223 +186val_186 +19 val_19 +197val_197 +200val_200 +216val_216 +218val_218 +222val_222 +224val_224 +228val_228 235val_235 -238val_238 +237val_237 239val_239 -24 val_24 -248val_248 -260val_260 +244val_244 +247val_247 +256val_256 +263val_263 +273val_273 28 val_28 +283val_283 286val_286 -298val_298 -305val_305 +292val_292 306val_306 -307val_307 308val_308 +327val_327 +33 val_33 +348val_348 +353val_353 +362val_362 +366val_366 +396val_396 +397val_397 +401val_401 +409val_409 +411val_411 +419val_419 +427val_427 +43 val_43 +432val_432 +436val_436 +439val_439 +443val_443 +453val_453 +460val_460 +462val_462 +47 val_47 +472val_472 +485val_485 +496val_496 +54 val_54 +64 val_64 +70 val_70 +8 val_8 +83 val_83 +84 val_84 +85 val_85 +90 val_90 +0 val_0 +103val_103 +114val_114 +125val_125 +138val_138 +146val_146 +150val_150 +152val_152 +153val_153 +156val_156 +157val_157 +165val_165 +172val_172 +177val_177 +179val_179 +187val_187 +195val_195 +196val_196 +217val_217 +242val_242 +248val_248 +252val_252 +265val_265 +27 val_27 +272val_272 +280val_280 +291val_291 +305val_305 +309val_309 +311val_311 315val_315 317val_317 -321val_321 -33 val_33 -331val_331 +322val_322 +333val_333 +34 val_34 +345val_345 +35 val_35 356val_356 -360val_360 -366val_366 -367val_367 +364val_364 368val_368 +369val_369 37 val_37 -378val_378 -379val_379 -382val_382 -389val_389 -395val_395 -400val_400 +373val_373 +377val_377 +4 val_4 402val_402 -406val_406 -407val_407 +404val_404 +413val_413 42 val_42 430val_430 -436val_436 -44 val_44 +431val_431 +444val_444 +449val_449 452val_452 +454val_454 +457val_457 +463val_463 466val_466 -467val_467 -472val_472 -480val_480 -484val_484 -485val_485 -487val_487 -495val_495 -496val_496 -498val_498 -51 val_51 +470val_470 +475val_475 +481val_481 +489val_489 +491val_491 57 val_57 -58 val_58 65 val_65 66 val_66 -69 val_69 -70 val_70 +74 val_74 +76 val_76 78 val_78 -8 val_8 +9 val_9 92 val_92 -96 val_96 -104val_104 -111val_111 -114val_114 -118val_118 -12 val_12 -133val_133 -138val_138 -153val_153 -156val_156 -162val_162 -168val_168 +95 val_95 +100val_100 +119val_119 +128val_128 +129val_129 +145val_145 +167val_167 174val_174 -176val_176 178val_178 181val_181 -186val_186 -187val_187 -192val_192 -197val_197 +193val_193 199val_199 -2 val_2 +20 val_20 201val_201 -203val_203 -208val_208 +213val_213 214val_214 -218val_218 219val_219 -230val_230 -237val_237 +221val_221 +223val_223 +226val_226 +233val_233 +241val_241 249val_249 -256val_256 -257val_257 +260val_260 +262val_262 277val_277 -278val_278 -284val_284 -285val_285 +281val_281 +287val_287 +288val_288 +298val_298 +302val_302 310val_310 -322val_322 -327val_327 -333val_333 +323val_323 336val_336 -338val_338 341val_341 344val_344 -348val_348 351val_351 -362val_362 -364val_364 -374val_374 -392val_392 -397val_397 -4 v
[31/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/druid_basic1.q.out -- diff --git a/ql/src/test/results/clientpositive/druid_basic1.q.out b/ql/src/test/results/clientpositive/druid_basic1.q.out index 25ec0ca..a5c0687 100644 --- a/ql/src/test/results/clientpositive/druid_basic1.q.out +++ b/ql/src/test/results/clientpositive/druid_basic1.q.out @@ -41,6 +41,7 @@ Table Type: EXTERNAL_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"__time\":\"true\",\"added\":\"true\",\"anonymous\":\"true\",\"count\":\"true\",\"deleted\":\"true\",\"delta\":\"true\",\"language\":\"true\",\"namespace\":\"true\",\"newpage\":\"true\",\"page\":\"true\",\"robot\":\"true\",\"unpatrolled\":\"true\",\"user\":\"true\",\"variation\":\"true\"}} EXTERNALTRUE + bucketing_version 2 druid.datasourcewikipedia numFiles0 numRows 0 @@ -102,6 +103,7 @@ Table Type: EXTERNAL_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"__time\":\"true\",\"added\":\"true\",\"anonymous\":\"true\",\"count\":\"true\",\"deleted\":\"true\",\"delta\":\"true\",\"language\":\"true\",\"namespace\":\"true\",\"newpage\":\"true\",\"page\":\"true\",\"robot\":\"true\",\"unpatrolled\":\"true\",\"user\":\"true\",\"variation\":\"true\"}} EXTERNALTRUE + bucketing_version 2 druid.datasourcewikipedia numFiles0 numRows 0 http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/druid_basic2.q.out -- diff --git a/ql/src/test/results/clientpositive/druid_basic2.q.out b/ql/src/test/results/clientpositive/druid_basic2.q.out index 5ed039b..0e5e957 100644 --- a/ql/src/test/results/clientpositive/druid_basic2.q.out +++ b/ql/src/test/results/clientpositive/druid_basic2.q.out @@ -41,6 +41,7 @@ Table Type: EXTERNAL_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"__time\":\"true\",\"added\":\"true\",\"anonymous\":\"true\",\"count\":\"true\",\"deleted\":\"true\",\"delta\":\"true\",\"language\":\"true\",\"namespace\":\"true\",\"newpage\":\"true\",\"page\":\"true\",\"robot\":\"true\",\"unpatrolled\":\"true\",\"user\":\"true\",\"variation\":\"true\"}} EXTERNALTRUE + bucketing_version 2 druid.datasourcewikipedia numFiles0 numRows 0 @@ -266,6 +267,7 @@ STAGE PLANS: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"__time":"true","added":"true","anonymous":"true","count":"true","deleted":"true","delta":"true","language":"true","namespace":"true","newpage":"true","page":"true","robot":"true","unpatrolled":"true","user":"true","variation":"true"}} EXTERNAL TRUE bucket_count -1 + bucketing_version 2 column.name.delimiter , columns __time,robot,namespace,anonymous,unpatrolled,page,language,newpage,user,count,added,delta,variation,deleted columns.comments 'from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer' @@ -294,6 +296,7 @@ STAGE PLANS: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"__time":"true","added":"true","anonymous":"true","count":"true","deleted":"true","delta":"true","language":"true","namespace":"true","newpage":"true","page":"true","robot":"true","unpatrolled":"true","user":"true","variation":"true"}} EXTERNAL TRUE bucket_count -1 +bucketing_version 2 column.name.delimiter , columns __time,robot,namespace,anonymous,unpatrolled,page,language,newpage,user,count,added,delta,variation,deleted columns.comments 'from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer' @@ -444,6 +447,7 @@ STAGE PLANS:
[26/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/bucket_num_reducers2.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/bucket_num_reducers2.q.out b/ql/src/test/results/clientpositive/llap/bucket_num_reducers2.q.out new file mode 100644 index 000..d0ca23c --- /dev/null +++ b/ql/src/test/results/clientpositive/llap/bucket_num_reducers2.q.out @@ -0,0 +1,242 @@ +PREHOOK: query: CREATE TABLE test_table(key int, value string) CLUSTERED BY (key) INTO 3 BUCKETS +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@test_table +POSTHOOK: query: CREATE TABLE test_table(key int, value string) CLUSTERED BY (key) INTO 3 BUCKETS +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@test_table +PREHOOK: query: explain extended insert overwrite table test_table + select * from src +PREHOOK: type: QUERY +POSTHOOK: query: explain extended insert overwrite table test_table + select * from src +POSTHOOK: type: QUERY +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 +Tez + A masked pattern was here + Edges: +Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) + A masked pattern was here + Vertices: +Map 1 +Map Operator Tree: +TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator +expressions: key (type: string), value (type: string) +outputColumnNames: _col0, _col1 +Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE +Reduce Output Operator + null sort order: + sort order: + Map-reduce partition columns: UDFToInteger(_col0) (type: int) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: string), _col1 (type: string) + auto parallelism: false +Execution mode: vectorized, llap +LLAP IO: no inputs +Path -> Alias: + A masked pattern was here +Path -> Partition: + A masked pattern was here +Partition + base file name: src + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: +COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} +bucket_count -1 +bucketing_version 2 +column.name.delimiter , +columns key,value +columns.comments 'default','default' +columns.types string:string + A masked pattern was here +name default.src +numFiles 1 +numRows 500 +rawDataSize 5312 +serialization.ddl struct src { string key, string value} +serialization.format 1 +serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +totalSize 5812 + A masked pattern was here + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + +input format: org.apache.hadoop.mapred.TextInputFormat +output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat +properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string + A masked pattern was here + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 + A masked pattern was here +
[30/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out -- diff --git a/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out b/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out index 4d8dd74..11ef8b1 100644 --- a/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out +++ b/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out @@ -87,6 +87,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -135,6 +136,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -158,6 +160,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -227,6 +230,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -270,6 +274,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -300,6 +305,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -321,6 +327,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -357,6 +364,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -387,6 +395,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -408,6 +417,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -522,6 +532,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -545,6 +556,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -589,6 +601,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key1,key2,cnt columns.comments @@ -647,6 +660,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","
[35/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/keyseries/VectorKeySeriesSerializedImpl.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/keyseries/VectorKeySeriesSerializedImpl.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/keyseries/VectorKeySeriesSerializedImpl.java index 86f466f..77c9ecc 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/keyseries/VectorKeySeriesSerializedImpl.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/keyseries/VectorKeySeriesSerializedImpl.java @@ -21,9 +21,9 @@ package org.apache.hadoop.hive.ql.exec.vector.keyseries; import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch; import org.apache.hadoop.hive.serde2.ByteStream.Output; import org.apache.hadoop.hive.serde2.fast.SerializeWrite; -import org.apache.hive.common.util.HashCodeUtil; import com.google.common.base.Preconditions; +import org.apache.hive.common.util.Murmur3; /** * Implementation of base serialization interface. @@ -103,7 +103,7 @@ public abstract class VectorKeySeriesSerializedImpl byte[] bytes = output.getData(); for (int i = 0; i < nonNullKeyCount; i++) { keyLength = serializedKeyLengths[i]; - hashCodes[i] = HashCodeUtil.murmurHash(bytes, offset, keyLength); + hashCodes[i] = Murmur3.hash32(bytes, offset, keyLength, 0); offset += keyLength; } } http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java index 1bc3fda..42b7784 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java @@ -32,6 +32,7 @@ import org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression; import org.apache.hadoop.hive.ql.metadata.HiveException; import org.apache.hadoop.hive.ql.plan.OperatorDesc; import org.apache.hadoop.hive.ql.plan.VectorDesc; +import org.apache.hadoop.hive.serde2.ByteStream; import org.apache.hadoop.hive.serde2.ByteStream.Output; import org.apache.hadoop.hive.serde2.binarysortable.fast.BinarySortableSerializeWrite; import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector; @@ -40,6 +41,7 @@ import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo; import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoUtils; import com.google.common.base.Preconditions; +import org.apache.hive.common.util.Murmur3; /** * This class is the object hash (not Uniform Hash) operator class for native vectorized reduce sink. @@ -226,61 +228,110 @@ public class VectorReduceSinkObjectHashOperator extends VectorReduceSinkCommonOp int[] selected = batch.selected; final int size = batch.size; - for (int logical = 0; logical < size; logical++) { -final int batchIndex = (selectedInUse ? selected[logical] : logical); - -final int hashCode; -if (isEmptyBuckets) { - if (isEmptyPartitions) { -hashCode = nonPartitionRandom.nextInt(); - } else { + + // EmptyBuckets = true + if (isEmptyBuckets) { +if (isEmptyPartitions) { + for (int logical = 0; logical< size; logical++) { +final int batchIndex = (selectedInUse ? selected[logical] : logical); +final int hashCode = nonPartitionRandom.nextInt(); +postProcess(batch, batchIndex, tag, hashCode); + } +} else { // isEmptyPartition = false + for (int logical = 0; logical< size; logical++) { +final int batchIndex = (selectedInUse ? selected[logical] : logical); partitionVectorExtractRow.extractRow(batch, batchIndex, partitionFieldValues); -hashCode = +final int hashCode = bucketingVersion == 2 && !vectorDesc.getIsAcidChange() ? ObjectInspectorUtils.getBucketHashCode( +partitionFieldValues, partitionObjectInspectors) : +ObjectInspectorUtils.getBucketHashCodeOld( partitionFieldValues, partitionObjectInspectors); +postProcess(batch, batchIndex, tag, hashCode); } -} else { - bucketVectorExtractRow.extractRow(batch, batchIndex, bucketFieldValues); - final int bucketNum = - ObjectInspectorUtils.getBucketNumber( +} + } else { // EmptyBuckets = false +if (isEmptyPartitions) { + for (int logical = 0; logical< size; logical
[08/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/groupby_sort_skew_1_23.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/groupby_sort_skew_1_23.q.out b/ql/src/test/results/clientpositive/spark/groupby_sort_skew_1_23.q.out index 9b068ed..d8b9dca 100644 --- a/ql/src/test/results/clientpositive/spark/groupby_sort_skew_1_23.q.out +++ b/ql/src/test/results/clientpositive/spark/groupby_sort_skew_1_23.q.out @@ -85,6 +85,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -118,6 +119,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -141,6 +143,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -172,6 +175,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -287,6 +291,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -310,6 +315,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -375,6 +381,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key1,key2,cnt columns.comments @@ -406,6 +413,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key1,key2,cnt columns.comments @@ -506,6 +514,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -539,6 +548,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -562,6 +572,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,val columns.comments @@ -593,6 +604,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,cnt columns.comments @@ -691,6 +703,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
[05/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/sample6.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/sample6.q.out b/ql/src/test/results/clientpositive/spark/sample6.q.out index 34aee1e..cf172ea 100644 --- a/ql/src/test/results/clientpositive/spark/sample6.q.out +++ b/ql/src/test/results/clientpositive/spark/sample6.q.out @@ -52,6 +52,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -84,6 +85,7 @@ STAGE PLANS: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -106,6 +108,7 @@ STAGE PLANS: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -137,6 +140,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -181,172 +185,175 @@ order by key, value POSTHOOK: type: QUERY POSTHOOK: Input: default@dest1 A masked pattern was here -0 val_0 -0 val_0 -0 val_0 -0 val_1 -0 val_1 -4 val_4 -4 val_5 -8 val_8 -8 val_9 -12 val_12 -12 val_12 -12 val_13 -16 val_17 -16 val_17 -20 val_20 -20 val_21 -20 val_21 -24 val_24 -24 val_24 -28 val_28 -32 val_33 +2 val_2 +2 val_3 +6 val_7 +6 val_7 +10 val_10 +10 val_11 +17 val_17 +21 val_22 +21 val_22 +21 val_22 +21 val_22 +27 val_27 +30 val_30 +30 val_31 +31 val_32 40 val_41 40 val_41 -44 val_44 -48 val_49 -48 val_49 -52 val_53 -52 val_53 -52 val_53 -52 val_53 -56 val_57 -60 val_61 +51 val_51 +51 val_51 +51 val_52 +57 val_57 +58 val_58 +58 val_58 +58 val_59 +58 val_59 64 val_64 -68 val_69 -72 val_72 -72 val_72 -76 val_76 -76 val_76 -76 val_77 -76 val_77 -76 val_77 +65 val_65 +65 val_66 +65 val_66 +70 val_70 +70 val_70 +70 val_70 +70 val_71 80 val_80 80 val_81 -84 val_84 -84 val_84 -92 val_92 -96 val_96 -100val_100 -100val_100 -100val_101 -100val_101 -104val_104 -104val_104 -104val_105 -104val_105 -104val_105 +83 val_83 +83 val_83 +86 val_86 +86 val_87 +90 val_90 +90 val_90 +90 val_90 +91 val_92 +98 val_98 +98 val_98 +105val_105 +105val_106 +105val_106 +110val_111 +113val_113 +113val_113 116val_116 116val_117 -120val_120 -120val_120 -120val_121 -128val_128 -128val_128 -128val_128 -128val_129 -128val_129 -132val_133 -132val_133 -136val_136 -136val_137 -140val_141 -144val_145 -152val_152 -152val_152 -152val_153 -152val_153 -152val_153 +117val_118 +117val_118 +119val_119 +119val_119 +119val_119 +119val_120 +119val_120 +119val_120 +121val_122 +121val_122 +123val_124 +123val_124 +126val_126 +126val_127 +126val_127 +134val_134 +134val_134 +134val_135 +137val_137 +137val_137 +137val_138 +153val_153 +153val_154 +153val_154 156val_156 156val_157 156val_157 -160val_160 -160val_161 -164val_164 -164val_164 -164val_165 -164val_165 -168val_168 -168val_169 -172val_172 -172val_172 -172val_173 -176val_176 -176val_176 -180val_180 +157val_157 +157val_158 +157val_158 +158val_158 +163val_163 +169val_169 +169val_169 +169val_169 +169val_169 +177val_177 +177val_178 +177val_178 +178val_178 +178val_179 +178val_179 184val_185 -192val_192 -192val_193 -196
[28/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/join9.q.out -- diff --git a/ql/src/test/results/clientpositive/join9.q.out b/ql/src/test/results/clientpositive/join9.q.out index 50ca52c..0a86a6e 100644 --- a/ql/src/test/results/clientpositive/join9.q.out +++ b/ql/src/test/results/clientpositive/join9.q.out @@ -76,6 +76,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -97,6 +98,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -147,6 +149,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -192,6 +195,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -250,6 +254,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/join_filters_overlap.q.out -- diff --git a/ql/src/test/results/clientpositive/join_filters_overlap.q.out b/ql/src/test/results/clientpositive/join_filters_overlap.q.out index db0b26d..7d4855a 100644 --- a/ql/src/test/results/clientpositive/join_filters_overlap.q.out +++ b/ql/src/test/results/clientpositive/join_filters_overlap.q.out @@ -92,6 +92,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -113,6 +114,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -283,6 +285,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -304,6 +307,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -474,6 +478,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -495,6 +500,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -682,6 +688,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -703,6 +710,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 +bucketing_version 2 column.name.d
[18/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out b/ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out index 687b4af..8bec959 100644 --- a/ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out +++ b/ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out @@ -232,17 +232,17 @@ POSTHOOK: query: select fl_date, count(*) from flights_tiny_orc group by fl_date POSTHOOK: type: QUERY POSTHOOK: Input: default@flights_tiny_orc A masked pattern was here -2010-10-20 11 2010-10-21 12 -2010-10-26 13 -2010-10-30 11 -2010-10-23 12 2010-10-24 12 -2010-10-25 12 -2010-10-22 11 +2010-10-26 13 2010-10-28 12 +2010-10-22 11 2010-10-29 12 +2010-10-20 11 +2010-10-23 12 +2010-10-25 12 2010-10-27 11 +2010-10-30 11 2010-10-31 8 PREHOOK: query: explain vectorization expression select * from flights_tiny_orc sort by fl_num, fl_date limit 25 @@ -529,17 +529,17 @@ POSTHOOK: query: select fl_date, count(*) from flights_tiny_orc group by fl_date POSTHOOK: type: QUERY POSTHOOK: Input: default@flights_tiny_orc A masked pattern was here -2010-10-20 11 2010-10-21 12 -2010-10-26 13 -2010-10-30 11 -2010-10-23 12 2010-10-24 12 -2010-10-25 12 -2010-10-22 11 +2010-10-26 13 2010-10-28 12 +2010-10-22 11 2010-10-29 12 +2010-10-20 11 +2010-10-23 12 +2010-10-25 12 2010-10-27 11 +2010-10-30 11 2010-10-31 8 PREHOOK: query: CREATE TABLE flights_tiny_orc_partitioned_date ( origin_city_name STRING, @@ -908,17 +908,17 @@ POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-29 POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-30 POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-31 A masked pattern was here -2010-10-20 11 2010-10-21 12 -2010-10-26 13 -2010-10-30 11 -2010-10-23 12 2010-10-24 12 -2010-10-25 12 -2010-10-22 11 +2010-10-26 13 2010-10-28 12 +2010-10-22 11 2010-10-29 12 +2010-10-20 11 +2010-10-23 12 +2010-10-25 12 2010-10-27 11 +2010-10-30 11 2010-10-31 8 PREHOOK: query: explain vectorization expression select * from flights_tiny_orc_partitioned_date @@ -1484,17 +1484,17 @@ POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-29 POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-30 POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-31 A masked pattern was here -2010-10-20 11 2010-10-21 12 -2010-10-26 13 -2010-10-30 11 -2010-10-23 12 2010-10-24 12 -2010-10-25 12 -2010-10-22 11 +2010-10-26 13 2010-10-28 12 +2010-10-22 11 2010-10-29 12 +2010-10-20 11 +2010-10-23 12 +2010-10-25 12 2010-10-27 11 +2010-10-30 11 2010-10-31 8 PREHOOK: query: CREATE TABLE flights_tiny_orc_partitioned_timestamp ( origin_city_name STRING, @@ -1863,18 +1863,18 @@ POSTHOOK: Input: default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10- POSTHOOK: Input: default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-30 07%3A00%3A00 POSTHOOK: Input: default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-31 07%3A00%3A00 A masked pattern was here -2010-10-20 07:00:0011 2010-10-23 07:00:0012 2010-10-25 07:00:0012 2010-10-22 07:00:0011 +2010-10-24 07:00:0012 2010-10-26 07:00:0013 +2010-10-20 07:00:0011 +2010-10-28 07:00:0012 2010-10-29 07:00:0012 +2010-10-30 07:00:0011 2010-10-31 07:00:008 2010-10-21 07:00:0012 -2010-10-24 07:00:0012 2010-10-27 07:00:0011 -2010-10-28 07:00:0012 -2010-10-30 07:00:0011 PREHOOK: query: explain vectorization expression select * from flights_tiny_orc_partitioned_timestamp PREHOOK: type: QUERY @@ -2439,18 +2439,18 @@ POSTHOOK: Input: default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10- POSTHOOK: Input: default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-30 07%3A00%3A00 POSTHOOK: Input: default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-31 07%3A00%3A00 A masked pattern was here -2010-10-20 07:00:0011 2010-10-23 07:00:0012 2010-10-25 07:00:0012 2010-10-22 07:00:0011 +2010-10-24 07:00:0012 2010-10-26 07:00:0013 +2010-10-20 07:00:0011 +2010-10-28 07:00:0012 2010-10-29 07:00:0012 +2010-10-30 07:00:0011 2010-10-31 07:00:008 2010-10-21 07:00:0012 -2010-10-24 07:00:0012 2010-10-27 07:00:0011 -2010-10-28 07:00:0012 -2010-10-30 07:00:
[13/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/sample1.q.out -- diff --git a/ql/src/test/results/clientpositive/sample1.q.out b/ql/src/test/results/clientpositive/sample1.q.out index 6d58658..09b9ac7 100644 --- a/ql/src/test/results/clientpositive/sample1.q.out +++ b/ql/src/test/results/clientpositive/sample1.q.out @@ -55,6 +55,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value,dt,hr columns.comments @@ -126,6 +127,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -193,6 +195,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value,dt,hr columns.comments @@ -236,6 +239,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value,dt,hr columns.comments @@ -266,6 +270,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value,dt,hr columns.comments @@ -287,6 +292,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value,dt,hr columns.comments @@ -323,6 +329,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value,dt,hr columns.comments @@ -353,6 +360,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value,dt,hr columns.comments @@ -374,6 +382,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value,dt,hr columns.comments http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/sample2.q.out -- diff --git a/ql/src/test/results/clientpositive/sample2.q.out b/ql/src/test/results/clientpositive/sample2.q.out index 164ce43..17e5491 100644 --- a/ql/src/test/results/clientpositive/sample2.q.out +++ b/ql/src/test/results/clientpositive/sample2.q.out @@ -54,6 +54,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -101,6 +102,7 @@ STAGE PLANS: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key + bucketing_version 2 column.name
[33/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/bucket1.q.out -- diff --git a/ql/src/test/results/clientpositive/bucket1.q.out b/ql/src/test/results/clientpositive/bucket1.q.out index 3060df3..07b5b9d 100644 --- a/ql/src/test/results/clientpositive/bucket1.q.out +++ b/ql/src/test/results/clientpositive/bucket1.q.out @@ -52,6 +52,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -73,6 +74,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -112,6 +114,7 @@ STAGE PLANS: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 100 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -171,6 +174,7 @@ STAGE PLANS: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 100 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/bucket2.q.out -- diff --git a/ql/src/test/results/clientpositive/bucket2.q.out b/ql/src/test/results/clientpositive/bucket2.q.out index 16218f9..e8a1dd7 100644 --- a/ql/src/test/results/clientpositive/bucket2.q.out +++ b/ql/src/test/results/clientpositive/bucket2.q.out @@ -51,6 +51,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -72,6 +73,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -111,6 +113,7 @@ STAGE PLANS: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -178,6 +181,7 @@ STAGE PLANS: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -264,66 +268,59 @@ POSTHOOK: query: select * from bucket2_1 tablesample (bucket 1 out of 2) s POSTHOOK: type: QUERY POSTHOOK: Input: default@bucket2_1 A masked pattern was here -0 val_0 -0 val_0 -0 val_0 10 val_10 -100val_100 -100val_100 -104val_104 -104val_104 -114val_114 +105val_105 +113val_113 +113val_113 116val_116 -118val_118 -118val_118 -12 val_12 -12 val_12 -120val_120 -120val_120 +119val_119 +119val_119 +119val_119 126val_126 -128val_128 -128val_128 -128val_128 +133val_133 134val_134 134val_134 -136val_136 -138val_138 -138val_138 -138val_138 -138val_138 +137val_137 +137val_137 146val_146 146val_146 -150val_150 -152val_152 -152val_152 +149val_149 +149val_149 +153val_153 156val_156 +157val_157 158val_158 -160val_160 162val_162 +163val_163 164val_164 164val_164 -166val_166 -168val_168 -170val_170 -172val_172 -172val_172 -174val_174 -174val_174 -176val_176 -176val_176 +165val_165 +165val_165 +169val_169 +169val_169 +
[10/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/bucket_map_join_tez1.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/bucket_map_join_tez1.q.out b/ql/src/test/results/clientpositive/spark/bucket_map_join_tez1.q.out index 158d1c3..987a98c 100644 --- a/ql/src/test/results/clientpositive/spark/bucket_map_join_tez1.q.out +++ b/ql/src/test/results/clientpositive/spark/bucket_map_join_tez1.q.out @@ -4323,110 +4323,104 @@ PREHOOK: type: QUERY POSTHOOK: query: explain select a.key, b.key from tab_part a join tab_part c on a.key = c.key join tab_part b on a.value = b.value POSTHOOK: type: QUERY STAGE DEPENDENCIES: - Stage-1 is a root stage + Stage-2 is a root stage + Stage-1 depends on stages: Stage-2 Stage-0 depends on stages: Stage-1 STAGE PLANS: - Stage: Stage-1 + Stage: Stage-2 Spark - Edges: -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 2), Map 4 (PARTITION-LEVEL SORT, 2) -Reducer 3 <- Map 5 (PARTITION-LEVEL SORT, 2), Reducer 2 (PARTITION-LEVEL SORT, 2) A masked pattern was here Vertices: -Map 1 +Map 2 Map Operator Tree: TableScan - alias: a + alias: c Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE Filter Operator -predicate: (key is not null and value is not null) (type: boolean) +predicate: key is not null (type: boolean) Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: key (type: int), value (type: string) - outputColumnNames: _col0, _col1 + expressions: key (type: int) + outputColumnNames: _col0 Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator -key expressions: _col0 (type: int) -sort order: + -Map-reduce partition columns: _col0 (type: int) -Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE -value expressions: _col1 (type: string) + Spark HashTable Sink Operator +keys: + 0 _col0 (type: int) + 1 _col0 (type: int) Execution mode: vectorized -Map 4 +Local Work: + Map Reduce Local Work +Map 3 Map Operator Tree: TableScan - alias: c + alias: b Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE Filter Operator -predicate: key is not null (type: boolean) +predicate: value is not null (type: boolean) Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: key (type: int) - outputColumnNames: _col0 + expressions: key (type: int), value (type: string) + outputColumnNames: _col0, _col1 Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator -key expressions: _col0 (type: int) -sort order: + -Map-reduce partition columns: _col0 (type: int) -Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE + Spark HashTable Sink Operator +keys: + 0 _col1 (type: string) + 1 _col1 (type: string) Execution mode: vectorized -Map 5 +Local Work: + Map Reduce Local Work + + Stage: Stage-1 +Spark + A masked pattern was here + Vertices: +Map 1 Map Operator Tree: TableScan - alias: b + alias: a Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE Filter Operator -predicate: value is not null (type: boolean) +predicate: (key is not null and value is not null) (type: boolean) Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type
[27/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out index 885fd18..8fa1b35 100644 --- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out +++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out @@ -147,6 +147,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 4 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -234,6 +235,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -282,6 +284,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -421,6 +424,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 4 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -508,6 +512,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -556,6 +561,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out index ff98ee5..bd7f60a 100644 --- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out +++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out @@ -147,6 +147,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -195,6 +196,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -283,6 +285,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 4 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -421,6 +424,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -469,6 +473,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -557,6 +562,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 4 bucket_field_name key + bucketing_version 2 colu
[04/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/smb_mapjoin_12.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/smb_mapjoin_12.q.out b/ql/src/test/results/clientpositive/spark/smb_mapjoin_12.q.out index ec8b046..9b421d5 100644 --- a/ql/src/test/results/clientpositive/spark/smb_mapjoin_12.q.out +++ b/ql/src/test/results/clientpositive/spark/smb_mapjoin_12.q.out @@ -141,6 +141,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -181,6 +182,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -213,6 +215,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -267,7 +270,7 @@ POSTHOOK: Input: default@test_table1@ds=1 POSTHOOK: Input: default@test_table3 POSTHOOK: Input: default@test_table3@ds=1 A masked pattern was here -879 +420 PREHOOK: query: explain extended INSERT OVERWRITE TABLE test_table3 PARTITION (ds = '2') SELECT /*+mapjoin(b)*/ a.key, concat(a.value, b.value) FROM test_table3 a JOIN test_table1 b ON a.key = b.key AND a.ds = '1' AND b.ds='1' @@ -359,6 +362,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -399,6 +403,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -431,6 +436,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -479,4 +485,4 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table3 POSTHOOK: Input: default@test_table3@ds=2 A masked pattern was here -879 +420 http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out b/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out index f8d0fb2..a364374 100644 --- a/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out +++ b/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out @@ -124,6 +124,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -147,6 +148,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -273,6 +275,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -296,6 +299,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -373,6 +377,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 16
[06/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/sample1.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/sample1.q.out b/ql/src/test/results/clientpositive/spark/sample1.q.out index b711292..164378b 100644 --- a/ql/src/test/results/clientpositive/spark/sample1.q.out +++ b/ql/src/test/results/clientpositive/spark/sample1.q.out @@ -53,6 +53,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}} bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value,dt,hr columns.comments @@ -109,6 +110,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -138,6 +140,7 @@ STAGE PLANS: properties: COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}} bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value,dt,hr columns.comments http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/sample10.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/sample10.q.out b/ql/src/test/results/clientpositive/spark/sample10.q.out index 418f63a..ac28779 100644 --- a/ql/src/test/results/clientpositive/spark/sample10.q.out +++ b/ql/src/test/results/clientpositive/spark/sample10.q.out @@ -34,10 +34,10 @@ POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-09,hr=11).value SIMPLE [(s POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-09,hr=12).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-09,hr=12).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] PREHOOK: query: explain extended -select ds, count(1) from srcpartbucket tablesample (bucket 1 out of 4 on key) where ds is not null group by ds ORDER BY ds ASC +select ds, count(1) from srcpartbucket tablesample (bucket 2 out of 4 on key) where ds is not null group by ds ORDER BY ds ASC PREHOOK: type: QUERY POSTHOOK: query: explain extended -select ds, count(1) from srcpartbucket tablesample (bucket 1 out of 4 on key) where ds is not null group by ds ORDER BY ds ASC +select ds, count(1) from srcpartbucket tablesample (bucket 2 out of 4 on key) where ds is not null group by ds ORDER BY ds ASC POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage @@ -59,8 +59,8 @@ STAGE PLANS: GatherStats: false Filter Operator isSamplingPred: true -predicate: (((hash(key) & 2147483647) % 4) = 0) (type: boolean) -sampleDesc: BUCKET 1 OUT OF 4 +predicate: (((hash(key) & 2147483647) % 4) = 1) (type: boolean) +sampleDesc: BUCKET 2 OUT OF 4 Statistics: Num rows: 20 Data size: 120 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: ds (type: string) @@ -87,7 +87,7 @@ STAGE PLANS: Path -> Partition: A masked pattern was here Partition - base file name: 00_0 + base file name: 01_0 input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat partition values: @@ -120,6 +120,7 @@ STAGE PLANS: properties: bucket_count 4 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -137,7 +138,7 @@ STAGE PLANS: name: default.srcpartbucket A masked pattern was here Partition - base file name: 00_0 + base file name: 01_0 input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputF
[02/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/union_remove_2.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/union_remove_2.q.out b/ql/src/test/results/clientpositive/spark/union_remove_2.q.out index f681966..18563c7 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_2.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_2.q.out @@ -184,6 +184,7 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 numFiles4 totalSize 68 A masked pattern was here http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/union_remove_20.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/union_remove_20.q.out b/ql/src/test/results/clientpositive/spark/union_remove_20.q.out index 402d2b6..9f2bf0e 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_20.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_20.q.out @@ -166,6 +166,7 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 numFiles4 totalSize 40 A masked pattern was here http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/union_remove_21.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/union_remove_21.q.out b/ql/src/test/results/clientpositive/spark/union_remove_21.q.out index 0df7318..6ef6d4e 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_21.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_21.q.out @@ -152,6 +152,7 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 numFiles4 totalSize 20 A masked pattern was here http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/union_remove_22.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/union_remove_22.q.out b/ql/src/test/results/clientpositive/spark/union_remove_22.q.out index 967d6fe..16d52bb 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_22.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_22.q.out @@ -168,6 +168,7 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 numFiles4 totalSize 60 A masked pattern was here http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/union_remove_23.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/union_remove_23.q.out b/ql/src/test/results/clientpositive/spark/union_remove_23.q.out index 19a3262..0e25378 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_23.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_23.q.out @@ -221,6 +221,7 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 numFiles4 totalSize 40 A masked pattern was here http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/union_remove_24.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/union_remove_24.q.out b/ql/src/test/results/clientpositive/spark/union_remove_24.q.out index 998ffa9..16b9272 100644 --- a/ql/src/test/results/clientpositive/spark/union_remove_24.q.out +++ b/ql/src/test/results/clientpositive/spark/union_remove_24.q.out @@ -166,6 +166,7 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Paramete
[19/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out b/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out index 6a654d0..42d5528 100644 --- a/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out +++ b/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out @@ -651,314 +651,314 @@ POSTHOOK: query: select unionsrc.key, unionsrc.value FROM (select s1.key as key, POSTHOOK: type: QUERY POSTHOOK: Input: default@src POSTHOOK: Output: hdfs://### HDFS PATH ### -0 val_0 -103val_103 -105val_105 116val_116 +118val_118 120val_120 -129val_129 131val_131 -136val_136 +133val_133 +15 val_15 +150val_150 +152val_152 +153val_153 155val_155 -157val_157 160val_160 -180val_180 +169val_169 +175val_175 +176val_176 +178val_178 183val_183 -190val_190 -191val_191 -193val_193 -205val_205 -207val_207 +186val_186 +194val_194 +202val_202 +203val_203 213val_213 223val_223 -235val_235 -238val_238 +224val_224 +226val_226 +228val_228 239val_239 -24 val_24 -248val_248 -260val_260 +244val_244 +247val_247 +252val_252 +255val_255 +256val_256 +257val_257 +266val_266 +272val_272 +274val_274 +275val_275 28 val_28 +284val_284 286val_286 -298val_298 -305val_305 +289val_289 +296val_296 +302val_302 306val_306 307val_307 -308val_308 -315val_315 317val_317 321val_321 +323val_323 33 val_33 -331val_331 -356val_356 +338val_338 +34 val_34 +35 val_35 360val_360 -366val_366 -367val_367 -368val_368 -37 val_37 -378val_378 -379val_379 -382val_382 -389val_389 -395val_395 -400val_400 -402val_402 +365val_365 +374val_374 +377val_377 +393val_393 +399val_399 +404val_404 406val_406 -407val_407 +409val_409 +41 val_41 +417val_417 +418val_418 42 val_42 -430val_430 -436val_436 +431val_431 +439val_439 44 val_44 -452val_452 -466val_466 -467val_467 -472val_472 -480val_480 -484val_484 -485val_485 -487val_487 -495val_495 -496val_496 -498val_498 -51 val_51 -57 val_57 -58 val_58 -65 val_65 +443val_443 +448val_448 +455val_455 +459val_459 +462val_462 +463val_463 +470val_470 +490val_490 +491val_491 +494val_494 +5 val_5 +53 val_53 66 val_66 -69 val_69 -70 val_70 -78 val_78 -8 val_8 -92 val_92 -96 val_96 -104val_104 +72 val_72 +76 val_76 +77 val_77 +80 val_80 +84 val_84 +90 val_90 +10 val_10 +105val_105 111val_111 -114val_114 -118val_118 -12 val_12 -133val_133 -138val_138 -153val_153 -156val_156 -162val_162 -168val_168 +113val_113 +119val_119 +129val_129 +136val_136 +143val_143 +149val_149 +163val_163 +167val_167 174val_174 -176val_176 -178val_178 -181val_181 -186val_186 187val_187 -192val_192 197val_197 -199val_199 2 val_2 -201val_201 -203val_203 +200val_200 208val_208 -214val_214 +209val_209 +216val_216 +217val_217 218val_218 -219val_219 +221val_221 230val_230 -237val_237 -249val_249 -256val_256 -257val_257 +258val_258 +26 val_26 +263val_263 +273val_273 277val_277 -278val_278 -284val_284 -285val_285 -310val_310 -322val_322 -327val_327 -333val_333 -336val_336 -338val_338 +281val_281 +292val_292 +30 val_30 +316val_316 +335val_335 341val_341 +342val_342 344val_344 -348val_348 -351val_351 +345val_345 362val_362 -364val_364 -374val_374 +367val_367 +368val_368 +37 val_37 392val_392 -397val_397 +394val_394 4 val_4 -403val_403 -404val_404 -419val_419 -421val_421 -427val_427 -429val_429 +413val_413 437val_437 -438val_438 -439val_439 -454val_454 -455val_455 +444val_444 +446val_446 +452val_452 +453val_453 +457val_457 460val_460 -463val_463 -470val_470 -478val_478 +466val_466 +468val_468 +477val_477 479val_479 -489val_489 -491val_491 +483val_483 +487val_487 492val_492 -53 val_53 +493val_493 +498val_498 +51 val_51 +54 val_54 +64 val_64 +65 val_65 67 val_67 82 val_82 -90 val_90 -97 val_97 -10 val_10 -125val_125 -134val_134 -143val_143 +85 val_85
[16/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/vector_windowing_range_multiorder.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/vector_windowing_range_multiorder.q.out b/ql/src/test/results/clientpositive/llap/vector_windowing_range_multiorder.q.out index 256b80d..9acbf2e 100644 --- a/ql/src/test/results/clientpositive/llap/vector_windowing_range_multiorder.q.out +++ b/ql/src/test/results/clientpositive/llap/vector_windowing_range_multiorder.q.out @@ -900,10005 +900,10005 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@over10k A masked pattern was here s si i avg_window_0 -alice falkner 323 65669 65695.76470588235 alice falkner 477 65722 65695.76470588235 -alice falkner 455 65718 65695.76470588235 -alice falkner 481 65709 65695.76470588235 -alice falkner 345 65773 65695.76470588235 -alice falkner 280 65597 65695.76470588235 -alice falkner 500 65775 65695.76470588235 alice falkner 339 65785 65695.76470588235 -alice falkner 452 65596 65695.76470588235 -alice falkner 382 65690 65695.76470588235 alice falkner 382 65622 65695.76470588235 alice falkner 393 65611 65695.76470588235 +alice falkner 389 65699 65695.76470588235 +alice falkner 500 65775 65695.76470588235 +alice falkner 371 65710 65695.76470588235 +alice falkner 481 65709 65695.76470588235 +alice falkner 452 65596 65695.76470588235 +alice falkner 323 65669 65695.76470588235 +alice falkner 382 65690 65695.76470588235 +alice falkner 455 65718 65695.76470588235 alice falkner 393 65685 65695.76470588235 -alice falkner 342 65752 65695.76470588235 alice falkner 311 65715 65695.76470588235 -alice falkner 371 65710 65695.76470588235 -alice falkner 389 65699 65695.76470588235 -alice ichabod 366 65590 65654.95454545454 +alice falkner 280 65597 65695.76470588235 +alice falkner 342 65752 65695.76470588235 +alice falkner 345 65773 65695.76470588235 +alice ichabod 292 65788 65654.95454545454 alice ichabod 458 65550 65654.95454545454 +alice ichabod 412 65718 65654.95454545454 +alice ichabod 303 65692 65654.95454545454 +alice ichabod 300 65704 65654.95454545454 +alice ichabod 366 65590 65654.95454545454 +alice ichabod 301 65693 65654.95454545454 +alice ichabod 398 65659 65654.95454545454 +alice ichabod 292 65585 65654.95454545454 +alice ichabod 305 65617 65654.95454545454 alice ichabod 436 65738 65654.95454545454 -alice ichabod 315 65772 65654.95454545454 +alice ichabod 440 65725 65654.95454545454 alice ichabod 453 65780 65654.95454545454 +alice ichabod 398 65785 65654.95454545454 alice ichabod 347 65547 65654.95454545454 -alice ichabod 398 65659 65654.95454545454 -alice ichabod 338 65538 65654.95454545454 -alice ichabod 440 65725 65654.95454545454 alice ichabod 320 65622 65654.95454545454 -alice ichabod 412 65718 65654.95454545454 -alice ichabod 305 65617 65654.95454545454 -alice ichabod 292 65585 65654.95454545454 -alice ichabod 303 65692 65654.95454545454 -alice ichabod 338 65545 65654.95454545454 alice ichabod 398 65680 65654.95454545454 +alice ichabod 315 65772 65654.95454545454 alice ichabod 416 65536 65654.95454545454 +alice ichabod 338 65545 65654.95454545454 alice ichabod 344 65545 65654.95454545454 -alice ichabod 300 65704 65654.95454545454 -alice ichabod 292 65788 65654.95454545454 -alice ichabod 398 65785 65654.95454545454 -alice ichabod 301 65693 65654.95454545454 -alice polk 443 65734 65661.57142857143 -alice polk 444 65564 65661.57142857143 -alice polk 357 65550 65661.57142857143 -alice polk 321 65744 65661.57142857143 -alice polk 273 65548 65661.57142857143 -alice polk 366 65595 65661.57142857143 -alice polk 285 65761 65661.57142857143 -alice polk 466 65561 65661.57142857143 -alice polk 324 65749 65661.57142857143 -alice polk 487 65746 65661.57142857143 -alice polk 378 65598 65661.57142857143 -alice polk 395 65751 65661.57142857143 -alice polk 407 65617 65661.57142857143 -alice polk 507 65744 65661.57142857143 -alice young468 65649 65706.63636363637 -alice young489 65646 65706.63636363637 -alice young286 65705 65706.63636363637 -alice young447 65789 65706.63636363637 -alice young425 65677 65706.63636363637 -alice young282 65671 65706.63636363637 -alice young351 65776 65706.63636363637 -alice young308 65776 65706.63636363637
[20/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/sysdb.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/sysdb.q.out b/ql/src/test/results/clientpositive/llap/sysdb.q.out index 7d977d3..de20f2d 100644 --- a/ql/src/test/results/clientpositive/llap/sysdb.q.out +++ b/ql/src/test/results/clientpositive/llap/sysdb.q.out @@ -58,34 +58,10 @@ default alltypesorc hive_test_user USERDELETE true-1 hive_test_user defaultalltypesorc hive_test_user USERINSERT true-1 hive_test_user defaultalltypesorc hive_test_user USERSELECT true-1 hive_test_user defaultalltypesorc hive_test_user USERUPDATE true-1 hive_test_user -defaultalltypesparquet hive_test_user USERDELETE true-1 hive_test_user -defaultalltypesparquet hive_test_user USERINSERT true-1 hive_test_user -defaultalltypesparquet hive_test_user USERSELECT true-1 hive_test_user -defaultalltypesparquet hive_test_user USERUPDATE true-1 hive_test_user -defaultcbo_t1 hive_test_user USERDELETE true -1 hive_test_user -defaultcbo_t1 hive_test_user USERINSERT true -1 hive_test_user -defaultcbo_t1 hive_test_user USERSELECT true -1 hive_test_user -defaultcbo_t1 hive_test_user USERUPDATE true -1 hive_test_user -defaultcbo_t2 hive_test_user USERDELETE true -1 hive_test_user -defaultcbo_t2 hive_test_user USERINSERT true -1 hive_test_user -defaultcbo_t2 hive_test_user USERSELECT true -1 hive_test_user -defaultcbo_t2 hive_test_user USERUPDATE true -1 hive_test_user -defaultcbo_t3 hive_test_user USERDELETE true -1 hive_test_user -defaultcbo_t3 hive_test_user USERINSERT true -1 hive_test_user -defaultcbo_t3 hive_test_user USERSELECT true -1 hive_test_user -defaultcbo_t3 hive_test_user USERUPDATE true -1 hive_test_user -defaultlineitemhive_test_user USERDELETE true-1 hive_test_user -defaultlineitemhive_test_user USERINSERT true-1 hive_test_user -defaultlineitemhive_test_user USERSELECT true-1 hive_test_user -defaultlineitemhive_test_user USERUPDATE true-1 hive_test_user defaultmoretypes hive_test_user USERDELETE true-1 hive_test_user defaultmoretypes hive_test_user USERINSERT true-1 hive_test_user defaultmoretypes hive_test_user USERSELECT true-1 hive_test_user defaultmoretypes hive_test_user USERUPDATE true-1 hive_test_user -defaultparthive_test_user USERDELETE true -1 hive_test_user -defaultparthive_test_user USERINSERT true -1 hive_test_user -defaultparthive_test_user USERSELECT true -1 hive_test_user -defaultparthive_test_user USERUPDATE true -1 hive_test_user defaultscr_txn hive_test_user USERDELETE true -1 hive_test_user defaultscr_txn hive_test_user USERINSERT true -1 hive_test_user defaultscr_txn hive_test_user USERSELECT true -1 hive_test_user @@ -94,46 +70,14 @@ default src hive_test_user USER DELETE true-1 hive_test_user defaultsrc hive_test_user USERINSERT true -1 hive_test_user defaultsrc hive_test_user USERSELECT true -1 hive_test_user defaultsrc hive_test_user USERUPDATE true -1 hive_test_user -defaultsrc1hive_test_user USERDELETE true -1 hive_test_user -defaultsrc1hive_test_user USERINSERT true -1 hive_test_user -defaultsrc1hive_test_user USERSELECT true -1 hive_test_user -defaultsrc1
[03/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/statsfs.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/statsfs.q.out b/ql/src/test/results/clientpositive/spark/statsfs.q.out index fef2040..1af1495 100644 --- a/ql/src/test/results/clientpositive/spark/statsfs.q.out +++ b/ql/src/test/results/clientpositive/spark/statsfs.q.out @@ -274,6 +274,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + bucketing_version 2 numFiles1 numRows 500 rawDataSize 5312 @@ -334,6 +335,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + bucketing_version 2 numFiles1 numRows 500 rawDataSize 5312 http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/subquery_nested_subquery.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/subquery_nested_subquery.q.out b/ql/src/test/results/clientpositive/spark/subquery_nested_subquery.q.out index 15f33f0..0d894e3 100644 --- a/ql/src/test/results/clientpositive/spark/subquery_nested_subquery.q.out +++ b/ql/src/test/results/clientpositive/spark/subquery_nested_subquery.q.out @@ -11,28 +11,28 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@part A masked pattern was here 192697 almond antique blue firebrick mint Manufacturer#5 Brand#52 MEDIUM BURNISHED TIN31 LG DRUM 1789.69 ickly ir +121152 almond antique burnished rose metallic Manufacturer#1 Brand#14 PROMO PLATED TIN2 JUMBO BOX 1173.15 e pinto beans h +121152 almond antique burnished rose metallic Manufacturer#1 Brand#14 PROMO PLATED TIN2 JUMBO BOX 1173.15 e pinto beans h 90681 almond antique chartreuse khaki white Manufacturer#3 Brand#31 MEDIUM BURNISHED TIN17 SM CASE 1671.68 are slyly after the sl -85768 almond antique chartreuse lavender yellow Manufacturer#1 Brand#12LARGE BRUSHED STEEL 34 SM BAG 1753.76 refull +49671 almond antique gainsboro frosted violet Manufacturer#4 Brand#41 SMALL BRUSHED BRASS 10 SM BOX 1620.67 ccounts run quick 42669 almond antique medium spring khaki Manufacturer#5 Brand#51 STANDARD BURNISHED TIN 6 MED CAN 1611.66 sits haggl -105685 almond antique violet chocolate turquoise Manufacturer#2 Brand#22MEDIUM ANODIZED COPPER 14 MED CAN 1690.68 ly pending requ +112398 almond antique metallic orange dim Manufacturer#3 Brand#32 MEDIUM BURNISHED BRASS 19 JUMBO JAR 1410.39 ole car +110592 almond antique salmon chartreuse burlywood Manufacturer#1 Brand#15PROMO BURNISHED NICKEL 6 JUMBO PKG 1602.59 to the furiously +155733 almond antique sky peru orange Manufacturer#5 Brand#53SMALL PLATED BRASS 2 WRAP DRUM 1788.73 furiously. bra 48427 almond antique violet mint lemonManufacturer#4 Brand#42 PROMO POLISHED STEEL39 SM CASE 1375.42 hely ironic i 86428 almond aquamarine burnished black steel Manufacturer#1 Brand#12 STANDARD ANODIZED STEEL 28 WRAP BAG1414.42 arefully 15103 almond aquamarine dodger light gainsboroManufacturer#5 Brand#53ECONOMY BURNISHED STEEL 46 LG PACK 1018.1 packages hinder carefu -45261 almond aquamarine floral ivory bisque Manufacturer#4 Brand#42 SMALL PLATED STEEL 27 WRAP CASE 1206.26 careful -65667 almond aquamarine pink moccasin thistle Manufacturer#1 Brand#12 LARGE BURNISHED STEEL 42 JUMBO CASE 1632.66 e across the expr -132666 almond aquamarine rose maroon antique Manufacturer#2 Brand#24 SMALL POLISHED NICKEL 25 MED BOX 1698.66 even 195606 almond aquamarine sandy cyan gainsboro Manufacturer#2 Brand#25 STANDARD PLATED TIN 18 SM PKG 1701.6 ic de -17927 almond aquamarine yellow dodger mintManufacturer#4 Brand#41 ECONOMY BRUSHED COPPER 7 SM PKG 1844.92 ites. eve 33357 almond azure aquamarine papaya violet Manufacturer#4 Brand#41 STANDARD ANODIZED TIN 12 WRAP CASE 1290.35 reful 78486 almond azure blanched chiffon midnight Manufacturer#5 Brand#52 LARGE BRUSHED BRASS 23 MED BAG 1464.48 hel
[23/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/default_constraint.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/default_constraint.q.out b/ql/src/test/results/clientpositive/llap/default_constraint.q.out index 15cbd84..d987ef2 100644 --- a/ql/src/test/results/clientpositive/llap/default_constraint.q.out +++ b/ql/src/test/results/clientpositive/llap/default_constraint.q.out @@ -31,6 +31,7 @@ Retention:0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 transactional true transactional_propertiesdefault A masked pattern was here @@ -301,6 +302,7 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 A masked pattern was here # Storage Information @@ -530,6 +532,7 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 transactional true transactional_propertiesdefault A masked pattern was here @@ -788,6 +791,7 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 transactional true transactional_propertiesdefault A masked pattern was here @@ -1059,6 +1063,7 @@ Retention:0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 A masked pattern was here # Storage Information @@ -1336,6 +1341,7 @@ Retention:0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 transactional true transactional_propertiesdefault A masked pattern was here @@ -1490,8 +1496,9 @@ Retention:0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 numFiles1 - totalSize 1062 + totalSize 1070 transactional true transactional_propertiesdefault A masked pattern was here @@ -1658,9 +1665,10 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 A masked pattern was here numFiles2 - totalSize 2124 + totalSize 2140 transactional true transactional_propertiesdefault A masked pattern was here @@ -1737,9 +1745,10 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 A masked pattern was here numFiles2 - totalSize 2124 + totalSize 2140 transactional true transactional_propertiesdefault A masked pattern was here @@ -1913,9 +1922,10 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TABLE Table Parameters: + bucketing_version 2 A masked pattern was here numFiles3 - totalSize 3188 + totalSize 3199 transactional true transactional_propertiesdefault A masked pattern was here @@ -1991,9 +2001,10 @@ Retention: 0 A masked pattern was here Table Type:MANAGED_TAB
[07/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/list_bucket_dml_2.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/list_bucket_dml_2.q.out b/ql/src/test/results/clientpositive/spark/list_bucket_dml_2.q.out index 30dc8b5..0ee050f 100644 --- a/ql/src/test/results/clientpositive/spark/list_bucket_dml_2.q.out +++ b/ql/src/test/results/clientpositive/spark/list_bucket_dml_2.q.out @@ -55,6 +55,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -109,6 +110,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -157,6 +159,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -189,6 +192,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -338,6 +342,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out b/ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out index 10d12a6..71996fe 100644 --- a/ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out +++ b/ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out @@ -173,6 +173,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -221,6 +222,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -269,6 +271,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -317,6 +320,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 + bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out b/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out index 972a7a8..7fbd873 100644 --- a/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out +++ b/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out @@ -62,6 +62,7 @@ STAGE PLANS: properties: COLUMN_STAT
[12/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/sample8.q.out -- diff --git a/ql/src/test/results/clientpositive/sample8.q.out b/ql/src/test/results/clientpositive/sample8.q.out index 365b7cf..3280cb5 100644 --- a/ql/src/test/results/clientpositive/sample8.q.out +++ b/ql/src/test/results/clientpositive/sample8.q.out @@ -86,6 +86,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -134,6 +135,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -182,6 +184,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -230,6 +233,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments 'default','default' @@ -319,358 +323,334 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 A masked pattern was here -105val_105 -105val_105 -105val_105 -105val_105 -114val_114 -114val_114 -114val_114 -114val_114 -150val_150 -150val_150 -150val_150 -150val_150 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -169val_169 -178val_178 -178val_178 -178val_178 -178val_178 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 -187val_187 +149val_149 +149val_149 +149val_149 +149val_149 +149val_149 +149val_149 +149val_149 +149val_149 +149val_149 +149val_149 +149val_149 +149val_149 +149val_149 +149val_149 +149val_149 +149val_149 196val_196 196val_196 196val_196 196val_196 -2 val_2 -2 val_2 -2 val_2 -2 val_2 -213val_213 -213val_213 -213val_213 -213val_213 -213val_213 -213val_213 -213val_213 -213val_213 -213val_213 -213val_213 -213val_213 -213val_213 -213val_213 -213val_213 -213val_213 -213val_213 -222val_222 -222val_222 -222val_222 -222val_222 -277val_277 -277val_277 -277val_277 -277val_277 -277val_277 -277val_277 -277val_277 -277val_277 -277val_277 -277val_277 -277val_277 -277val_277 -277val_277 -277val_277 -277val_277 -277val_277 -277val_277 -277val_277 -277val_277 -277val_277 -277val_277 -277val_277 -277val_277 -277val_277 -277val
[24/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/cte_1.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/cte_1.q.out b/ql/src/test/results/clientpositive/llap/cte_1.q.out index d7bc062..d98862c 100644 --- a/ql/src/test/results/clientpositive/llap/cte_1.q.out +++ b/ql/src/test/results/clientpositive/llap/cte_1.q.out @@ -159,178 +159,217 @@ POSTHOOK: Input: default@src 10 val_10 val_10 100val_100 val_100 100val_100 val_100 +103val_103 val_103 +103val_103 val_103 104val_104 val_104 104val_104 val_104 -111val_111 val_111 -114val_114 val_114 -116val_116 val_116 +105val_105 val_105 +11 val_11 val_11 126val_126 val_126 128val_128 val_128 128val_128 val_128 128val_128 val_128 -131val_131 val_131 +129val_129 val_129 +129val_129 val_129 +133val_133 val_133 134val_134 val_134 134val_134 val_134 +136val_136 val_136 +143val_143 val_143 +149val_149 val_149 +149val_149 val_149 152val_152 val_152 152val_152 val_152 -155val_155 val_155 +153val_153 val_153 +157val_157 val_157 +162val_162 val_162 +166val_166 val_166 169val_169 val_169 169val_169 val_169 169val_169 val_169 169val_169 val_169 -17 val_17 val_17 170val_170 val_170 -172val_172 val_172 -172val_172 val_172 -178val_178 val_178 -179val_179 val_179 -179val_179 val_179 -189val_189 val_189 -19 val_19 val_19 -192val_192 val_192 -194val_194 val_194 -195val_195 val_195 -195val_195 val_195 -199val_199 val_199 -199val_199 val_199 -199val_199 val_199 -20 val_20 val_20 -207val_207 val_207 -207val_207 val_207 +174val_174 val_174 +174val_174 val_174 +176val_176 val_176 +176val_176 val_176 +193val_193 val_193 +193val_193 val_193 +193val_193 val_193 +196val_196 val_196 +201val_201 val_201 +205val_205 val_205 +205val_205 val_205 +209val_209 val_209 +209val_209 val_209 213val_213 val_213 213val_213 val_213 218val_218 val_218 +221val_221 val_221 +221val_221 val_221 223val_223 val_223 223val_223 val_223 -229val_229 val_229 -229val_229 val_229 233val_233 val_233 233val_233 val_233 -235val_235 val_235 +239val_239 val_239 +239val_239 val_239 24 val_24 val_24 24 val_24 val_24 -256val_256 val_256 -256val_256 val_256 -257val_257 val_257 +241val_241 val_241 +242val_242 val_242 +242val_242 val_242 +247val_247 val_247 +248val_248 val_248 26 val_26 val_26 26 val_26 val_26 -265val_265 val_265 -265val_265 val_265 -266val_266 val_266 -27 val_27 val_27 -278val_278 val_278 -278val_278 val_278 -280val_280 val_280 -280val_280 val_280 -284val_284 val_284 -291val_291 val_291 -30 val_30 val_30 +263val_263 val_263 +272val_272 val_272 +272val_272 val_272 +275val_275 val_275 +285val_285 val_285 +296val_296 val_296 +298val_298 val_298 +298val_298 val_298 +298val_298 val_298 302val_302 val_302 -305val_305 val_305 -316val_316 val_316 -316val_316 val_316 -316val_316 val_316 -321val_321 val_321 -321val_321 val_321 +307val_307 val_307 +307val_307 val_307 +310val_310 val_310 323val_323 val_323 -33 val_33 val_33 +325val_325 val_325 +325val_325 val_325 333val_333 val_333 333val_333 val_333 +335val_335 val_335 338val_338 val_338 -344val_344 val_344 -344val_344 val_344 -353val_353 val_353 -353val_353 val_353 -356val_356 val_356 -364val_364 val_364 -368val_368 val_368 -369val_369 val_369 -369val_369 val_369 -369val_369 val_369 -37 val_37 val_37 -37 val_37 val_37 +360val_360 val_360 +362val_362 val_362 +367val_367 val_367 +367val_367 val_367 +375val_375 val_375 377val_377 val_377 -378val_378 val_378 -386val_386 val_386 +384val_384 val_384 +384val_384 val_384 +384val_384 val_384 392val_392 val_392 -397val_397 val_397 -397val_397 val_397 -404val_404 val_404 -404val_404 val_404 -407val_407 val_407 -411val_411 val_411 -418val_418 val_418 +399val_399 val_399 +399val_399 val_399 +401val_401 val_401 +401val_401 val_401 +401val_401 val_401 +401val_401 val_401 +401val_401 val_401 +402val_402 val_402 +403val_403 val_403 +403val_403 val_403 +403val_403 val_403 +413val_413 val_413 +413val_413 val_413 +417val_417 val_417 +417val_417 val_417 +417val_417 val_417 419val_419 val_419 -424val_424 val_424 -424val_424 val_424 -430val_430 val_430 -430val_430 val_430 -430val_430 val_430 -432val_432 val_432 -436val_436 val_436 -
[11/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/sort_merge_join_desc_5.q.out -- diff --git a/ql/src/test/results/clientpositive/sort_merge_join_desc_5.q.out b/ql/src/test/results/clientpositive/sort_merge_join_desc_5.q.out index 5754a74..38d2b96 100644 --- a/ql/src/test/results/clientpositive/sort_merge_join_desc_5.q.out +++ b/ql/src/test/results/clientpositive/sort_merge_join_desc_5.q.out @@ -128,6 +128,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 1 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/sort_merge_join_desc_6.q.out -- diff --git a/ql/src/test/results/clientpositive/sort_merge_join_desc_6.q.out b/ql/src/test/results/clientpositive/sort_merge_join_desc_6.q.out index 23cb97a..64bf678 100644 --- a/ql/src/test/results/clientpositive/sort_merge_join_desc_6.q.out +++ b/ql/src/test/results/clientpositive/sort_merge_join_desc_6.q.out @@ -101,6 +101,7 @@ STAGE PLANS: properties: bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -212,6 +213,7 @@ STAGE PLANS: SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/sort_merge_join_desc_7.q.out -- diff --git a/ql/src/test/results/clientpositive/sort_merge_join_desc_7.q.out b/ql/src/test/results/clientpositive/sort_merge_join_desc_7.q.out index 426f63f..77f7792 100644 --- a/ql/src/test/results/clientpositive/sort_merge_join_desc_7.q.out +++ b/ql/src/test/results/clientpositive/sort_merge_join_desc_7.q.out @@ -137,6 +137,7 @@ STAGE PLANS: properties: bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -184,6 +185,7 @@ STAGE PLANS: properties: bucket_count 2 bucket_field_name key + bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -286,6 +288,7 @@ STAGE PLANS: properties: bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -335,6 +338,7 @@ STAGE PLANS: properties: bucket_count 2 bucket_field_name key +bucketing_version 2 column.name.delimiter , columns key,value columns.comments http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/alter_merge_stats_orc.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/alter_merge_stats_orc.q.out b/ql/src/test/results/clientpositive/spark/alter_merge_stats_orc.q.out index 2b069ec..9337387 100644 --- a/ql/src/test/results/clientpositive/spark/alter_merge_stats_orc.q.out +++ b/ql/src/test/results/clientpositive/spark/alter_merge_stats_orc.q.out @@ -89,6 +89,7 @@ Retention:0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + bucketing_version 2 numFiles3 numRows 1500 rawDataSize 141000 @@ -139,6 +140,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + bucketing_version 2 numFiles1 numRows 1500 rawDataSize 141000 http://git-wip-us.apache.org
[32/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out -- diff --git a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out index b59c4bc..49598d1 100644 --- a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out +++ b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out @@ -128,7 +128,7 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table2 POSTHOOK: Input: default@test_table2@ds=1 A masked pattern was here -247 +243 PREHOOK: query: select count(*) from test_table2 tablesample (bucket 2 out of 2) s where ds = '1' PREHOOK: type: QUERY PREHOOK: Input: default@test_table2 @@ -139,7 +139,7 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table2 POSTHOOK: Input: default@test_table2@ds=1 A masked pattern was here -253 +257 PREHOOK: query: EXPLAIN INSERT OVERWRITE TABLE test_table2 PARTITION (ds = '1') SELECT * from @@ -238,7 +238,7 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table2 POSTHOOK: Input: default@test_table2@ds=1 A masked pattern was here -247 +243 PREHOOK: query: select count(*) from test_table2 tablesample (bucket 2 out of 2) s where ds = '1' PREHOOK: type: QUERY PREHOOK: Input: default@test_table2 @@ -249,7 +249,7 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table2 POSTHOOK: Input: default@test_table2@ds=1 A masked pattern was here -253 +257 PREHOOK: query: EXPLAIN INSERT OVERWRITE TABLE test_table2 PARTITION (ds = '1') SELECT x.key, concat(x.value, x.value) from http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out -- diff --git a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out index 52ef3db..d1c11f4 100644 --- a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out +++ b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out @@ -120,7 +120,7 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table2 POSTHOOK: Input: default@test_table2@ds=1 A masked pattern was here -247 +243 PREHOOK: query: select count(*) from test_table2 tablesample (bucket 2 out of 2) s where ds = '1' PREHOOK: type: QUERY PREHOOK: Input: default@test_table2 @@ -131,7 +131,7 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table2 POSTHOOK: Input: default@test_table2@ds=1 A masked pattern was here -253 +257 PREHOOK: query: CREATE TABLE test_table3 (key INT, value STRING) PARTITIONED BY (ds STRING) CLUSTERED BY (value) SORTED BY (value) INTO 2 BUCKETS PREHOOK: type: CREATETABLE http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out -- diff --git a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out index eaf85c3..72e918f 100644 --- a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out +++ b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out @@ -199,18 +199,7 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table3 POSTHOOK: Input: default@test_table3@ds=1 A masked pattern was here -0 0 val_0val_0 1 -0 0 val_0val_0 1 -0 0 val_0val_0 1 -0 0 val_0val_0 1 -0 0 val_0val_0 1 -0 0 val_0val_0 1 -0 0 val_0val_0 1 -0 0 val_0val_0 1 -0 0 val_0val_0 1 2 2 val_2val_2 1 -4 4 val_4val_4 1 -8 8 val_8val_8 1 PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s where ds = '1' PREHOOK: type: QUERY PREHOOK: Input: default@test_table3 @@ -221,6 +210,16 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table3 POSTHOOK: Input: default@test_table3@ds=1 A masked pattern was here +0 0 val_0val_0 1 +0 0 val_0val_0 1 +0 0 val_0val_0 1 +0 0 val_0val_0 1 +0 0 val_0val_0 1 +0 0 val_0val_0 1 +0 0 val_0val_0 1 +0 0 val_0val_0 1 +0 0 val_0val_0 1 +4 4 val_4val_4 1 5 5 val_5val_5 1 5 5 val_5val_5 1 5 5 val_5val_5 1 @@ -230,6 +229,7 @@ POSTHOOK: Input: default@test_table3@ds=1 5 5 val_5val_5 1 5 5 val_5val_5 1 5 5 val_5val_5 1 +8 8
[14/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/pcr.q.out -- diff --git a/ql/src/test/results/clientpositive/pcr.q.out b/ql/src/test/results/clientpositive/pcr.q.out index 0adee06..e30f6d1 100644 --- a/ql/src/test/results/clientpositive/pcr.q.out +++ b/ql/src/test/results/clientpositive/pcr.q.out @@ -116,6 +116,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -163,6 +164,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -306,6 +308,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -353,6 +356,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -400,6 +404,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -581,6 +586,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -628,6 +634,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -773,6 +780,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -820,6 +828,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -967,6 +976,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -1014,6 +1024,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -1061,6 +1072,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -1219,6 +1231,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter , columns key,value columns.comments @@ -1266,6 +1279,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 +bucketing_version 2 column.name.delimiter ,
[21/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/orc_create.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/orc_create.q.out b/ql/src/test/results/clientpositive/llap/orc_create.q.out index 72d94e5..df286b2 100644 --- a/ql/src/test/results/clientpositive/llap/orc_create.q.out +++ b/ql/src/test/results/clientpositive/llap/orc_create.q.out @@ -66,6 +66,7 @@ Retention:0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"lst\":\"true\",\"mp\":\"true\",\"str\":\"true\",\"strct\":\"true\"}} + bucketing_version 2 numFiles0 numRows 0 rawDataSize 0 @@ -119,6 +120,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + bucketing_version 2 numFiles0 numPartitions 0 numRows 0 @@ -176,6 +178,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + bucketing_version 2 numFiles0 numPartitions 0 numRows 0 @@ -223,6 +226,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + bucketing_version 2 A masked pattern was here numFiles0 numPartitions 0 @@ -281,6 +285,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + bucketing_version 2 numFiles0 numPartitions 0 numRows 0 @@ -336,6 +341,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"lst\":\"true\",\"mp\":\"true\",\"str\":\"true\",\"strct\":\"true\"}} + bucketing_version 2 numFiles0 numRows 0 rawDataSize 0 http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/orc_llap_counters.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/orc_llap_counters.q.out b/ql/src/test/results/clientpositive/llap/orc_llap_counters.q.out index af237ea..98d9e06 100644 --- a/ql/src/test/results/clientpositive/llap/orc_llap_counters.q.out +++ b/ql/src/test/results/clientpositive/llap/orc_llap_counters.q.out @@ -227,6 +227,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + bucketing_version 2 numFiles1 numRows 2100 orc.bloom.filter.columns* http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/orc_llap_counters1.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/orc_llap_counters1.q.out b/ql/src/test/results/clientpositive/llap/orc_llap_counters1.q.out index 01fb4af..c078326 100644 --- a/ql/src/test/results/clientpositive/llap/orc_llap_counters1.q.out +++ b/ql/src/test/results/clientpositive/llap/orc_llap_counters1.q.out @@ -227,6 +227,7 @@ Retention: 0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + bucketing_version 2 numFiles1 numRows 2100 orc.bloom.filter.columns* http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/ll
[09/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/bucketsortoptimize_insert_2.q.out -- diff --git a/ql/src/test/results/clientpositive/spark/bucketsortoptimize_insert_2.q.out b/ql/src/test/results/clientpositive/spark/bucketsortoptimize_insert_2.q.out index 3e8ea6f..9b14b92 100644 --- a/ql/src/test/results/clientpositive/spark/bucketsortoptimize_insert_2.q.out +++ b/ql/src/test/results/clientpositive/spark/bucketsortoptimize_insert_2.q.out @@ -193,18 +193,7 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table3 POSTHOOK: Input: default@test_table3@ds=1 A masked pattern was here -0 val_0val_0 1 -0 val_0val_0 1 -0 val_0val_0 1 -0 val_0val_0 1 -0 val_0val_0 1 -0 val_0val_0 1 -0 val_0val_0 1 -0 val_0val_0 1 -0 val_0val_0 1 2 val_2val_2 1 -4 val_4val_4 1 -8 val_8val_8 1 PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s where ds = '1' PREHOOK: type: QUERY PREHOOK: Input: default@test_table3 @@ -215,6 +204,16 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table3 POSTHOOK: Input: default@test_table3@ds=1 A masked pattern was here +0 val_0val_0 1 +0 val_0val_0 1 +0 val_0val_0 1 +0 val_0val_0 1 +0 val_0val_0 1 +0 val_0val_0 1 +0 val_0val_0 1 +0 val_0val_0 1 +0 val_0val_0 1 +4 val_4val_4 1 5 val_5val_5 1 5 val_5val_5 1 5 val_5val_5 1 @@ -224,6 +223,7 @@ POSTHOOK: Input: default@test_table3@ds=1 5 val_5val_5 1 5 val_5val_5 1 5 val_5val_5 1 +8 val_8val_8 1 9 val_9val_9 1 PREHOOK: query: EXPLAIN INSERT OVERWRITE TABLE test_table3 PARTITION (ds = '1') @@ -344,6 +344,18 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table3 POSTHOOK: Input: default@test_table3@ds=1 A masked pattern was here +2 val_2val_2 1 +2 val_2val_2 1 +PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s where ds = '1' +PREHOOK: type: QUERY +PREHOOK: Input: default@test_table3 +PREHOOK: Input: default@test_table3@ds=1 + A masked pattern was here +POSTHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s where ds = '1' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@test_table3 +POSTHOOK: Input: default@test_table3@ds=1 + A masked pattern was here 0 val_0val_0 1 0 val_0val_0 1 0 val_0val_0 1 @@ -362,22 +374,8 @@ POSTHOOK: Input: default@test_table3@ds=1 0 val_0val_0 1 0 val_0val_0 1 0 val_0val_0 1 -2 val_2val_2 1 -2 val_2val_2 1 4 val_4val_4 1 4 val_4val_4 1 -8 val_8val_8 1 -8 val_8val_8 1 -PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s where ds = '1' -PREHOOK: type: QUERY -PREHOOK: Input: default@test_table3 -PREHOOK: Input: default@test_table3@ds=1 - A masked pattern was here -POSTHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s where ds = '1' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@test_table3 -POSTHOOK: Input: default@test_table3@ds=1 - A masked pattern was here 5 val_5val_5 1 5 val_5val_5 1 5 val_5val_5 1 @@ -396,6 +394,8 @@ POSTHOOK: Input: default@test_table3@ds=1 5 val_5val_5 1 5 val_5val_5 1 5 val_5val_5 1 +8 val_8val_8 1 +8 val_8val_8 1 9 val_9val_9 1 9 val_9val_9 1 PREHOOK: query: EXPLAIN @@ -517,6 +517,18 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@test_table3 POSTHOOK: Input: default@test_table3@ds=1 A masked pattern was here +2 val_2val_2 1 +2 val_2val_2 1 +PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s where ds = '1' +PREHOOK: type: QUERY +PREHOOK: Input: default@test_table3 +PREHOOK: Input: default@test_table3@ds=1 + A masked pattern was here +POSTHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s where ds = '1' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@test_table3 +POSTHOOK: Input: default@test_table3@ds=1 + A masked pattern was here 0 val_0val_0 1 0 val_0val_0 1 0 val_0val_0 1 @@ -535,22 +547,8 @@ POSTHOOK: Input: default@test_table3@ds=1 0 val_0val_0 1 0 val_0val_0 1 0 val_0val_0 1 -2 val_2val_2 1 -2 val_2val_2 1 4 val_4val_4 1 4 val_4val_4 1 -8 val_8val_8 1 -8 val_8val_8 1 -PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s where ds = '1' -PREHOOK: type: QUERY -P
[17/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/vector_ptf_part_simple.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/vector_ptf_part_simple.q.out b/ql/src/test/results/clientpositive/llap/vector_ptf_part_simple.q.out index 9aeb650..ec8611d 100644 --- a/ql/src/test/results/clientpositive/llap/vector_ptf_part_simple.q.out +++ b/ql/src/test/results/clientpositive/llap/vector_ptf_part_simple.q.out @@ -309,46 +309,46 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@vector_ptf_part_simple_orc A masked pattern was here p_mfgr p_name p_retailprice rn r dr fv lv c cs -Manufacturer#4 almond azure aquamarine papaya violet 1290.35 1 1 1 1290.35 1206.26 4 6 -Manufacturer#4 almond antique violet mint lemon1375.42 2 1 1 1290.35 1206.26 4 6 -Manufacturer#4 almond aquamarine floral ivory bisque NULL3 1 1 1290.35 1206.26 4 6 -Manufacturer#4 almond antique gainsboro frosted violet NULL4 1 1 1290.35 1206.26 4 6 -Manufacturer#4 almond aquamarine yellow dodger mint1844.92 5 1 1 1290.35 1206.26 4 6 -Manufacturer#4 almond aquamarine floral ivory bisque 1206.26 6 1 1 1290.35 1206.26 4 6 -Manufacturer#5 almond azure blanched chiffon midnight 1464.48 1 1 1 1464.48 1788.73 6 6 -Manufacturer#5 almond aquamarine dodger light gainsboro1018.1 2 1 1 1464.48 1788.73 6 6 -Manufacturer#5 almond antique medium spring khaki 1611.66 3 1 1 1464.48 1788.73 6 6 -Manufacturer#5 almond antique blue firebrick mint 1789.69 4 1 1 1464.48 1788.73 6 6 -Manufacturer#5 almond antique medium spring khaki 1611.66 5 1 1 1464.48 1788.73 6 6 -Manufacturer#5 almond antique sky peru orange 1788.73 6 1 1 1464.48 1788.73 6 6 -Manufacturer#2 almond aquamarine rose maroon antique 900.66 1 1 1 900.66 1800.7 8 8 -Manufacturer#2 almond aquamarine rose maroon antique 1698.66 2 1 1 900.66 1800.7 8 8 -Manufacturer#2 almond antique violet turquoise frosted 1800.7 3 1 1 900.66 1800.7 8 8 -Manufacturer#2 almond antique violet chocolate turquoise 1690.68 4 1 1 900.66 1800.7 8 8 -Manufacturer#2 almond antique violet turquoise frosted 1800.7 5 1 1 900.66 1800.7 8 8 -Manufacturer#2 almond aquamarine sandy cyan gainsboro 1000.6 6 1 1 900.66 1800.7 8 8 -Manufacturer#2 almond aquamarine midnight light salmon 2031.98 7 1 1 900.66 1800.7 8 8 -Manufacturer#2 almond antique violet turquoise frosted 1800.7 8 1 1 900.66 1800.7 8 8 -Manufacturer#1 almond antique chartreuse lavender yellow 1753.76 1 1 1 1753.76 1632.66 11 12 -Manufacturer#1 almond aquamarine pink moccasin thistle 1632.66 2 1 1 1753.76 1632.66 11 12 -Manufacturer#1 almond aquamarine pink moccasin thistle 1632.66 3 1 1 1753.76 1632.66 11 12 -Manufacturer#1 almond antique chartreuse lavender yellow 1753.76 4 1 1 1753.76 1632.66 11 12 -Manufacturer#1 almond aquamarine burnished black steel 1414.42 5 1 1 1753.76 1632.66 11 12 -Manufacturer#1 almond antique burnished rose metallic 1173.15 6 1 1 1753.76 1632.66 11 12 -Manufacturer#1 almond antique salmon chartreuse burlywood 1602.59 7 1 1 1753.76 1632.66 11 12 -Manufacturer#1 almond antique chartreuse lavender yellow 1753.76 8 1 1 1753.76 1632.66 11 12 -Manufacturer#1 almond antique burnished rose metallic 1173.15 9 1 1 1753.76 1632.66 11 12 -Manufacturer#1 almond antique chartreuse lavender yellow 1753.76 10 1 1 1753.76 1632.66 11 12 -Manufacturer#1 almond aquamarine pink moccasin thistle NULL11 1 1 1753.76 1632.66 11 12 -Manufacturer#1 almond aquamarine pink moccasin thistle 1632.66 12 1 1 1753.76 1632.66 11 12 -Manufacturer#3 almond antique forest lavender goldenrod590.27 1 1 1 590.27 99.68 7 8 -Manufacturer#3 almond antique metallic orange dim 55.39 2 1 1 590.27 99.68 7 8 -Manufacturer#3 almond antique forest lavender goldenrod1190.27 3 1 1 590.27 99.68 7 8 -Manufacturer#3 almond antique olive coral navajo 1337.29 4 1 1 590.27 99.68 7 8 -Manufacturer#3 almond antique m
[34/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientnegative/alter_file_format.q.out -- diff --git a/ql/src/test/results/clientnegative/alter_file_format.q.out b/ql/src/test/results/clientnegative/alter_file_format.q.out index 96259e5..3033cbe 100644 --- a/ql/src/test/results/clientnegative/alter_file_format.q.out +++ b/ql/src/test/results/clientnegative/alter_file_format.q.out @@ -24,6 +24,7 @@ Retention:0 Table Type:MANAGED_TABLE Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} + bucketing_version 2 numFiles0 numRows 0 rawDataSize 0 http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out -- diff --git a/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out b/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out index 8e240ea..90c3d02 100644 --- a/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out +++ b/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out @@ -51,6 +51,7 @@ Retention:0 Table Type:VIRTUAL_VIEW Table Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + bucketing_version 2 numFiles0 numPartitions 2 numRows 0 http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientnegative/unset_table_property.q.out -- diff --git a/ql/src/test/results/clientnegative/unset_table_property.q.out b/ql/src/test/results/clientnegative/unset_table_property.q.out index 2dd047a..eb308eb 100644 --- a/ql/src/test/results/clientnegative/unset_table_property.q.out +++ b/ql/src/test/results/clientnegative/unset_table_property.q.out @@ -20,6 +20,7 @@ POSTHOOK: query: SHOW TBLPROPERTIES testTable POSTHOOK: type: SHOW_TBLPROPERTIES COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"col1":"true","col2":"true"}} a 1 +bucketing_version 2 c 3 A masked pattern was here numFiles 0 http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientnegative/unset_view_property.q.out -- diff --git a/ql/src/test/results/clientnegative/unset_view_property.q.out b/ql/src/test/results/clientnegative/unset_view_property.q.out index ae7f4f6..b5038d9 100644 --- a/ql/src/test/results/clientnegative/unset_view_property.q.out +++ b/ql/src/test/results/clientnegative/unset_view_property.q.out @@ -21,6 +21,7 @@ PREHOOK: query: SHOW TBLPROPERTIES testView PREHOOK: type: SHOW_TBLPROPERTIES POSTHOOK: query: SHOW TBLPROPERTIES testView POSTHOOK: type: SHOW_TBLPROPERTIES +bucketing_version 2 A masked pattern was here propA 100 propB 200 http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/acid_mapjoin.q.out -- diff --git a/ql/src/test/results/clientpositive/acid_mapjoin.q.out b/ql/src/test/results/clientpositive/acid_mapjoin.q.out index 5569a03..76a781e 100644 --- a/ql/src/test/results/clientpositive/acid_mapjoin.q.out +++ b/ql/src/test/results/clientpositive/acid_mapjoin.q.out @@ -73,21 +73,21 @@ STAGE PLANS: Stage: Stage-5 Map Reduce Local Work Alias -> Map Local Tables: -$hdt$_1:acid2 +$hdt$_0:acid1 Fetch Operator limit: -1 Alias -> Map Local Operator Tree: -$hdt$_1:acid2 +$hdt$_0:acid1 TableScan -alias: acid2 -Statistics: Num rows: 210 Data size: 840 Basic stats: COMPLETE Column stats: NONE +alias: acid1 +Statistics: Num rows: 316 Data size: 1265 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: key is not null (type: boolean) - Statistics: Num rows: 210 Data size: 840 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 316 Data size: 1265 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: int) outputColumnNames: _col0 -Statistics: Num rows: 210 Data size: 840 Basic stats: COMPLETE Column stats: NONE +