hive git commit: HIVE-20868 : SMB Join fails intermittently when TezDummyOperator has child op in getFinalOp in MapRecordProcessor(Deepak Jaiswal, reviewed by Gopal V)

2018-11-08 Thread djaiswal
Repository: hive
Updated Branches:
  refs/heads/master ad597b026 -> ffb78b521


HIVE-20868 : SMB Join fails intermittently when TezDummyOperator has child op 
in getFinalOp in MapRecordProcessor(Deepak Jaiswal, reviewed by Gopal V)


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/ffb78b52
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/ffb78b52
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/ffb78b52

Branch: refs/heads/master
Commit: ffb78b521cbd977d575374169392b0b3231024dc
Parents: ad597b0
Author: Deepak Jaiswal 
Authored: Thu Nov 8 07:30:04 2018 -0800
Committer: Deepak Jaiswal 
Committed: Thu Nov 8 07:30:04 2018 -0800

--
 .../apache/hadoop/hive/ql/exec/TezDummyStoreOperator.java| 8 +++-
 1 file changed, 7 insertions(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/hive/blob/ffb78b52/ql/src/java/org/apache/hadoop/hive/ql/exec/TezDummyStoreOperator.java
--
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/exec/TezDummyStoreOperator.java 
b/ql/src/java/org/apache/hadoop/hive/ql/exec/TezDummyStoreOperator.java
index d00057a..08c1cc4 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/exec/TezDummyStoreOperator.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/TezDummyStoreOperator.java
@@ -37,7 +37,7 @@ public class TezDummyStoreOperator extends DummyStoreOperator 
{
 super(ctx);
   }
 
-  private boolean fetchDone = false;
+  private transient boolean fetchDone = false;
 
   /**
* Unlike the MR counterpoint, on Tez we want processOp to forward
@@ -56,4 +56,10 @@ public class TezDummyStoreOperator extends 
DummyStoreOperator {
   public void setFetchDone(boolean fetchDone) {
 this.fetchDone = fetchDone;
   }
+
+  @Override
+  public void closeOp(boolean abort) throws HiveException {
+super.closeOp(abort);
+fetchDone = false;
+  }
 }



hive git commit: HIVE-16976 : DPP: SyntheticJoinPredicate transitivity for < > and BETWEEN (Deepak Jaiswal, reviewed by Jesus)

2019-01-11 Thread djaiswal
Repository: hive
Updated Branches:
  refs/heads/master cb9d5ccd8 -> 305e7887e


HIVE-16976 : DPP: SyntheticJoinPredicate transitivity for < > and BETWEEN 
(Deepak Jaiswal, reviewed by Jesus)


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/305e7887
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/305e7887
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/305e7887

Branch: refs/heads/master
Commit: 305e7887e542b680ba41e32c49dab3a18a84e68a
Parents: cb9d5cc
Author: Deepak Jaiswal 
Authored: Fri Jan 11 13:51:31 2019 -0800
Committer: Deepak Jaiswal 
Committed: Fri Jan 11 13:51:31 2019 -0800

--
 .../org/apache/hadoop/hive/conf/HiveConf.java   |   2 +-
 .../hadoop/hive/ql/exec/FunctionRegistry.java   |  20 +++
 .../hive/ql/metadata/HiveStorageHandler.java|  12 ++
 .../DynamicPartitionPruningOptimization.java|  23 ++-
 .../stats/annotation/StatsRulesProcFactory.java |  13 +-
 .../hadoop/hive/ql/parse/GenTezUtils.java   |   9 ++
 .../hive/ql/plan/ExprNodeDynamicListDesc.java   |  20 ++-
 .../hive/ql/ppd/SyntheticJoinPredicate.java | 160 ++-
 .../clientpositive/llap/cross_prod_1.q.out  |   4 +-
 .../llap/groupby_groupingset_bug.q.out  |  74 -
 .../results/clientpositive/llap/semijoin.q.out  |   4 +-
 .../clientpositive/llap/subquery_in.q.out   |   4 +-
 .../clientpositive/llap/subquery_notin.q.out|   4 +-
 .../clientpositive/llap/subquery_scalar.q.out   |  48 +++---
 .../clientpositive/llap/subquery_select.q.out   |   4 +-
 15 files changed, 283 insertions(+), 118 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hive/blob/305e7887/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java
--
diff --git a/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java 
b/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java
index b213609..a6bb9ef 100644
--- a/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java
+++ b/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java
@@ -2489,7 +2489,7 @@ public class HiveConf extends Configuration {
 "UDTFs change the number of rows of the output. A common UDTF is the 
explode() method that creates\n" +
 "multiple rows for each element in the input array. This factor is 
applied to the number of\n" +
 "output rows and output size."),
-
+HIVE_STATS_USE_NDV_ADJUSTMENT("hive.stats.use.ndv.adjustment", false, "Use 
nDV adjustment for other columns in join"),
 // Concurrency
 HIVE_SUPPORT_CONCURRENCY("hive.support.concurrency", false,
 "Whether Hive supports concurrency control or not. \n" +

http://git-wip-us.apache.org/repos/asf/hive/blob/305e7887/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java
--
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java 
b/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java
index d7f069e..bbe7fb0 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java
@@ -1851,4 +1851,24 @@ public final class FunctionRegistry {
   String blackListStr) {
 system.setupPermissionsForUDFs(whiteListStr, blackListStr);
   }
+
+  /**
+   * Function to invert non-equi function texts
+   * @param funcText
+   */
+  public static String invertFuncText(final String funcText) {
+// Reverse the text
+switch (funcText) {
+  case "<":
+return ">";
+  case "<=":
+return ">=";
+  case ">":
+return "<";
+  case ">=":
+return "<=";
+  default:
+return null; // helps identify unsupported functions
+}
+  }
 }

http://git-wip-us.apache.org/repos/asf/hive/blob/305e7887/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java
--
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java 
b/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java
index 2ebb149..c010b18 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java
@@ -26,6 +26,7 @@ import org.apache.hadoop.hive.metastore.api.LockType;
 import org.apache.hadoop.hive.metastore.api.MetaException;
 import org.apache.hadoop.hive.metastore.api.Table;
 import org.apache.hadoop.hive.ql.hooks.WriteEntity;
+import org.apache.hadoop.hive.ql.plan.ExprNodeDesc;
 import org.apache.hadoop.hive.ql.plan.TableDesc;
 import 
org.apache.hadoop.hive.ql.security.authorization.HiveAuthorizationProvider;
 import org.apache.hadoop.hive.serde2.Abstrac

hive git commit: Revert "HIVE-16976 : DPP: SyntheticJoinPredicate transitivity for < > and BETWEEN (Deepak Jaiswal, reviewed by Jesus)"

2019-01-11 Thread djaiswal
Repository: hive
Updated Branches:
  refs/heads/master 305e7887e -> a81e8de30


Revert "HIVE-16976 : DPP: SyntheticJoinPredicate transitivity for < > and 
BETWEEN (Deepak Jaiswal, reviewed by Jesus)"

This reverts commit 305e7887e542b680ba41e32c49dab3a18a84e68a.


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/a81e8de3
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/a81e8de3
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/a81e8de3

Branch: refs/heads/master
Commit: a81e8de30b742c26c8ebb21eeeb057d52fca9b8e
Parents: 305e788
Author: Deepak Jaiswal 
Authored: Fri Jan 11 14:00:27 2019 -0800
Committer: Deepak Jaiswal 
Committed: Fri Jan 11 14:00:27 2019 -0800

--
 .../org/apache/hadoop/hive/conf/HiveConf.java   |   2 +-
 .../hadoop/hive/ql/exec/FunctionRegistry.java   |  20 ---
 .../hive/ql/metadata/HiveStorageHandler.java|  12 --
 .../DynamicPartitionPruningOptimization.java|  23 +--
 .../stats/annotation/StatsRulesProcFactory.java |  13 +-
 .../hadoop/hive/ql/parse/GenTezUtils.java   |   9 --
 .../hive/ql/plan/ExprNodeDynamicListDesc.java   |  20 +--
 .../hive/ql/ppd/SyntheticJoinPredicate.java | 160 +--
 .../clientpositive/llap/cross_prod_1.q.out  |   4 +-
 .../llap/groupby_groupingset_bug.q.out  |  74 -
 .../results/clientpositive/llap/semijoin.q.out  |   4 +-
 .../clientpositive/llap/subquery_in.q.out   |   4 +-
 .../clientpositive/llap/subquery_notin.q.out|   4 +-
 .../clientpositive/llap/subquery_scalar.q.out   |  48 +++---
 .../clientpositive/llap/subquery_select.q.out   |   4 +-
 15 files changed, 118 insertions(+), 283 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hive/blob/a81e8de3/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java
--
diff --git a/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java 
b/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java
index a6bb9ef..b213609 100644
--- a/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java
+++ b/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java
@@ -2489,7 +2489,7 @@ public class HiveConf extends Configuration {
 "UDTFs change the number of rows of the output. A common UDTF is the 
explode() method that creates\n" +
 "multiple rows for each element in the input array. This factor is 
applied to the number of\n" +
 "output rows and output size."),
-HIVE_STATS_USE_NDV_ADJUSTMENT("hive.stats.use.ndv.adjustment", false, "Use 
nDV adjustment for other columns in join"),
+
 // Concurrency
 HIVE_SUPPORT_CONCURRENCY("hive.support.concurrency", false,
 "Whether Hive supports concurrency control or not. \n" +

http://git-wip-us.apache.org/repos/asf/hive/blob/a81e8de3/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java
--
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java 
b/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java
index bbe7fb0..d7f069e 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java
@@ -1851,24 +1851,4 @@ public final class FunctionRegistry {
   String blackListStr) {
 system.setupPermissionsForUDFs(whiteListStr, blackListStr);
   }
-
-  /**
-   * Function to invert non-equi function texts
-   * @param funcText
-   */
-  public static String invertFuncText(final String funcText) {
-// Reverse the text
-switch (funcText) {
-  case "<":
-return ">";
-  case "<=":
-return ">=";
-  case ">":
-return "<";
-  case ">=":
-return "<=";
-  default:
-return null; // helps identify unsupported functions
-}
-  }
 }

http://git-wip-us.apache.org/repos/asf/hive/blob/a81e8de3/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java
--
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java 
b/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java
index c010b18..2ebb149 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java
@@ -26,7 +26,6 @@ import org.apache.hadoop.hive.metastore.api.LockType;
 import org.apache.hadoop.hive.metastore.api.MetaException;
 import org.apache.hadoop.hive.metastore.api.Table;
 import org.apache.hadoop.hive.ql.hooks.WriteEntity;
-import org.apache.hadoop.hive.ql.plan.ExprNodeDesc;
 import org.apache.hadoop.hive.ql.plan.TableDesc;
 import 
org.apache.hadoop.hive.ql.security.authorization.

hive git commit: HIVE-16976 : DPP: SyntheticJoinPredicate transitivity for < > and BETWEEN (Deepak Jaiswal, reviewed by Jesus)

2019-01-11 Thread djaiswal
Repository: hive
Updated Branches:
  refs/heads/master a81e8de30 -> 28db173b9


HIVE-16976 : DPP: SyntheticJoinPredicate transitivity for < > and BETWEEN 
(Deepak Jaiswal, reviewed by Jesus)


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/28db173b
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/28db173b
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/28db173b

Branch: refs/heads/master
Commit: 28db173b9c24991e184fdf69625d932294a15d15
Parents: a81e8de
Author: Deepak Jaiswal 
Authored: Fri Jan 11 14:02:34 2019 -0800
Committer: Deepak Jaiswal 
Committed: Fri Jan 11 14:02:34 2019 -0800

--
 .../hadoop/hive/ql/exec/FunctionRegistry.java   |  20 +++
 .../hive/ql/metadata/HiveStorageHandler.java|  12 ++
 .../DynamicPartitionPruningOptimization.java|  23 ++-
 .../hadoop/hive/ql/parse/GenTezUtils.java   |   9 ++
 .../hive/ql/plan/ExprNodeDynamicListDesc.java   |  20 ++-
 .../hive/ql/ppd/SyntheticJoinPredicate.java | 160 ++-
 .../clientpositive/llap/cross_prod_1.q.out  |   4 +-
 .../llap/groupby_groupingset_bug.q.out  |  74 -
 .../results/clientpositive/llap/semijoin.q.out  |   4 +-
 .../clientpositive/llap/subquery_in.q.out   |   4 +-
 .../clientpositive/llap/subquery_notin.q.out|   4 +-
 .../clientpositive/llap/subquery_scalar.q.out   |  48 +++---
 .../clientpositive/llap/subquery_select.q.out   |   4 +-
 13 files changed, 270 insertions(+), 116 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hive/blob/28db173b/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java
--
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java 
b/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java
index d7f069e..bbe7fb0 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/FunctionRegistry.java
@@ -1851,4 +1851,24 @@ public final class FunctionRegistry {
   String blackListStr) {
 system.setupPermissionsForUDFs(whiteListStr, blackListStr);
   }
+
+  /**
+   * Function to invert non-equi function texts
+   * @param funcText
+   */
+  public static String invertFuncText(final String funcText) {
+// Reverse the text
+switch (funcText) {
+  case "<":
+return ">";
+  case "<=":
+return ">=";
+  case ">":
+return "<";
+  case ">=":
+return "<=";
+  default:
+return null; // helps identify unsupported functions
+}
+  }
 }

http://git-wip-us.apache.org/repos/asf/hive/blob/28db173b/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java
--
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java 
b/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java
index 2ebb149..c010b18 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/metadata/HiveStorageHandler.java
@@ -26,6 +26,7 @@ import org.apache.hadoop.hive.metastore.api.LockType;
 import org.apache.hadoop.hive.metastore.api.MetaException;
 import org.apache.hadoop.hive.metastore.api.Table;
 import org.apache.hadoop.hive.ql.hooks.WriteEntity;
+import org.apache.hadoop.hive.ql.plan.ExprNodeDesc;
 import org.apache.hadoop.hive.ql.plan.TableDesc;
 import 
org.apache.hadoop.hive.ql.security.authorization.HiveAuthorizationProvider;
 import org.apache.hadoop.hive.serde2.AbstractSerDe;
@@ -172,4 +173,15 @@ public interface HiveStorageHandler extends Configurable {
   default LockType getLockType(WriteEntity writeEntity){
 return LockType.EXCLUSIVE;
   }
+
+  /**
+   * Test if the storage handler allows the push-down of join filter predicate 
to prune further the splits.
+   *
+   * @param syntheticFilterPredicate Join filter predicate.
+   * @return true if supports dynamic split pruning for the given predicate.
+   */
+
+  default boolean addDynamicSplitPruningEdge(ExprNodeDesc 
syntheticFilterPredicate) {
+return false;
+  }
 }

http://git-wip-us.apache.org/repos/asf/hive/blob/28db173b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/DynamicPartitionPruningOptimization.java
--
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/DynamicPartitionPruningOptimization.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/DynamicPartitionPruningOptimization.java
index a1401aa..439fb75 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/DynamicPartitionPruningOptimization.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/DynamicPartitionPruningOptimization.

[hive] branch master updated: HIVE-21214 : MoveTask : Use attemptId instead of file size for deduplication of files compareTempOrDuplicateFiles() (Deepak Jaiswal, reviewed by Jason Dere)

2019-02-06 Thread djaiswal
This is an automated email from the ASF dual-hosted git repository.

djaiswal pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hive.git


The following commit(s) were added to refs/heads/master by this push:
 new fae6256  HIVE-21214 : MoveTask : Use attemptId instead of file size 
for deduplication of files compareTempOrDuplicateFiles() (Deepak Jaiswal, 
reviewed by Jason Dere)
fae6256 is described below

commit fae6256ace38d106b62d3bcade7b84b51bf4e1ec
Author: Deepak Jaiswal 
AuthorDate: Tue Feb 5 14:06:19 2019 -0800

HIVE-21214 : MoveTask : Use attemptId instead of file size for 
deduplication of files compareTempOrDuplicateFiles() (Deepak Jaiswal, reviewed 
by Jason Dere)
---
 .../org/apache/hadoop/hive/ql/exec/Utilities.java  | 71 ++
 1 file changed, 59 insertions(+), 12 deletions(-)

diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java 
b/ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java
index 8937b43..b84b052 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/Utilities.java
@@ -1308,7 +1308,7 @@ public final class Utilities {
*  filename to extract taskid from
*/
   public static String getTaskIdFromFilename(String filename) {
-return getIdFromFilename(filename, FILE_NAME_TO_TASK_ID_REGEX);
+return getTaskIdFromFilename(filename, FILE_NAME_TO_TASK_ID_REGEX);
   }
 
   /**
@@ -1319,10 +1319,19 @@ public final class Utilities {
*  filename to extract taskid from
*/
   public static String getPrefixedTaskIdFromFilename(String filename) {
-return getIdFromFilename(filename, FILE_NAME_PREFIXED_TASK_ID_REGEX);
+return getTaskIdFromFilename(filename, FILE_NAME_PREFIXED_TASK_ID_REGEX);
   }
 
-  private static String getIdFromFilename(String filename, Pattern pattern) {
+  private static String getTaskIdFromFilename(String filename, Pattern 
pattern) {
+return getIdFromFilename(filename, pattern, 1);
+  }
+
+  public static int getAttemptIdFromFilename(String filename) {
+String attemptStr = getIdFromFilename(filename, 
FILE_NAME_PREFIXED_TASK_ID_REGEX, 3);
+return Integer.parseInt(attemptStr.substring(1));
+  }
+
+  private static String getIdFromFilename(String filename, Pattern pattern, 
int group) {
 String taskId = filename;
 int dirEnd = filename.lastIndexOf(Path.SEPARATOR);
 if (dirEnd != -1) {
@@ -1334,7 +1343,7 @@ public final class Utilities {
   LOG.warn("Unable to get task id from file name: {}. Using last component 
{}"
   + " as task id.", filename, taskId);
 } else {
-  taskId = m.group(1);
+  taskId = m.group(group);
 }
 LOG.debug("TaskId for {} = {}", filename, taskId);
 return taskId;
@@ -1823,10 +1832,10 @@ public final class Utilities {
 
   private static FileStatus compareTempOrDuplicateFiles(FileSystem fs,
   FileStatus file, FileStatus existingFile) throws IOException {
-// Compare the file sizes of all the attempt files for the same task, the 
largest win
-// any attempt files could contain partial results (due to task failures or
-// speculative runs), but the largest should be the correct one since the 
result
-// of a successful run should never be smaller than a failed/speculative 
run.
+// Pick the one with mewest attempt ID. For sanity, check the file sizes 
too.
+// If the file size of newest attempt is less than that for older one,
+// Throw an exception as it maybe a correctness issue causing it.
+// This breaks speculative execution if it ends prematurely.
 FileStatus toDelete = null, toRetain = null;
 
 // "LOAD .. INTO" and "INSERT INTO" commands will generate files with
@@ -1847,12 +1856,26 @@ public final class Utilities {
   return existingFile;
 }
 
-if (existingFile.getLen() >= file.getLen()) {
-  toDelete = file;
+int existingFileAttemptId = 
getAttemptIdFromFilename(existingFile.getPath().getName());
+int fileAttemptId = getAttemptIdFromFilename(file.getPath().getName());
+
+long existingFileSz = getFileSizeRecursively(fs, existingFile);
+long fileSz = getFileSizeRecursively(fs, file);
+// Files may come in any order irrespective of their attempt IDs
+if (existingFileAttemptId > fileAttemptId &&
+existingFileSz >= fileSz) {
+  // keep existing
   toRetain = existingFile;
-} else {
-  toDelete = existingFile;
+  toDelete = file;
+} else if (existingFileAttemptId < fileAttemptId &&
+existingFileSz <= fileSz) {
+  // keep file
   toRetain = file;
+  toDelete = existingFile;
+} else {
+  throw new IOException(" File " + filePath +
+" with newer attempt ID " + fileAttemptId + " is smaller than the file 
"
++ existingFile.getPath() + " wit

[hive] branch master updated: HIVE-21221 : Make HS2 and LLAP consistent - Bring up LLAP WebUI in test mode if WebUI port is configured (Oliver Draese via Deepak Jaiswal)

2019-02-15 Thread djaiswal
This is an automated email from the ASF dual-hosted git repository.

djaiswal pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hive.git


The following commit(s) were added to refs/heads/master by this push:
 new aaf01ae  HIVE-21221 : Make HS2 and LLAP consistent - Bring up LLAP 
WebUI in test mode if WebUI port is configured (Oliver Draese via Deepak 
Jaiswal)
aaf01ae is described below

commit aaf01ae60bb987965a45adc243605e625fd39140
Author: Oliver Draese 
AuthorDate: Fri Feb 15 14:36:39 2019 -0800

HIVE-21221 : Make HS2 and LLAP consistent - Bring up LLAP WebUI in test 
mode if WebUI port is configured (Oliver Draese via Deepak Jaiswal)
---
 .../java/org/apache/hadoop/hive/llap/daemon/impl/LlapDaemon.java| 6 +-
 1 file changed, 5 insertions(+), 1 deletion(-)

diff --git 
a/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/LlapDaemon.java 
b/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/LlapDaemon.java
index 41eca8f..bafa029 100644
--- 
a/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/LlapDaemon.java
+++ 
b/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/impl/LlapDaemon.java
@@ -313,7 +313,11 @@ public class LlapDaemon extends CompositeService 
implements ContainerRunner, Lla
 // Not adding the registry as a service, since we need to control when it 
is initialized - conf used to pickup properties.
 this.registry = new LlapRegistryService(true);
 
-if (HiveConf.getBoolVar(daemonConf, HiveConf.ConfVars.HIVE_IN_TEST)) {
+// disable web UI in test mode until a specific port was configured
+if (HiveConf.getBoolVar(daemonConf, HiveConf.ConfVars.HIVE_IN_TEST)
+&& Integer.parseInt(ConfVars.LLAP_DAEMON_WEB_PORT.getDefaultValue()) 
== webPort) {
+  LOG.info("Web UI was disabled in test mode because 
hive.llap.daemon.web.port was not "
+   + "specified or has default value ({})", webPort);
   this.webServices = null;
 } else {
   this.webServices = new LlapWebServices(webPort, this, registry);



[hive] branch master updated: HIVE-21167 : Bucketing: Bucketing version 1 is incorrectly partitioning data (Deepak Jaiswal, reviewed by Jason Dere and Vineet Garg)

2019-02-22 Thread djaiswal
This is an automated email from the ASF dual-hosted git repository.

djaiswal pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hive.git


The following commit(s) were added to refs/heads/master by this push:
 new 69a7fc5  HIVE-21167 : Bucketing: Bucketing version 1 is incorrectly 
partitioning data (Deepak Jaiswal, reviewed by Jason Dere and Vineet Garg)
69a7fc5 is described below

commit 69a7fc5955ceac1245233c0d297929e730750929
Author: Deepak Jaiswal 
AuthorDate: Wed Feb 13 22:04:33 2019 -0800

HIVE-21167 : Bucketing: Bucketing version 1 is incorrectly partitioning 
data (Deepak Jaiswal, reviewed by Jason Dere and Vineet Garg)
---
 .../apache/hadoop/hive/ql/parse/TezCompiler.java   |  47 ++-
 .../queries/clientpositive/murmur_hash_migration.q |  35 +++
 .../llap/dynpart_sort_opt_vectorization.q.out  |   4 +-
 .../llap/murmur_hash_migration.q.out   | 332 +
 4 files changed, 392 insertions(+), 26 deletions(-)

diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/TezCompiler.java 
b/ql/src/java/org/apache/hadoop/hive/ql/parse/TezCompiler.java
index 4b10e89..a917617 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/parse/TezCompiler.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/TezCompiler.java
@@ -174,9 +174,6 @@ public class TezCompiler extends TaskCompiler {
 runStatsAnnotation(procCtx);
 perfLogger.PerfLogEnd(this.getClass().getName(), PerfLogger.TEZ_COMPILER, 
"Setup stats in the operator plan");
 
-// Update bucketing version of ReduceSinkOp if needed
-updateBucketingVersionForUpgrade(procCtx);
-
 // run Sorted dynamic partition optimization
 if(HiveConf.getBoolVar(procCtx.conf, 
HiveConf.ConfVars.DYNAMICPARTITIONING) &&
 HiveConf.getVar(procCtx.conf, 
HiveConf.ConfVars.DYNAMICPARTITIONINGMODE).equals("nonstrict") &&
@@ -229,6 +226,15 @@ public class TezCompiler extends TaskCompiler {
   new 
ConstantPropagate(ConstantPropagateOption.SHORTCUT).transform(procCtx.parseContext);
 }
 
+// ATTENTION : DO NOT, I REPEAT, DO NOT WRITE ANYTHING AFTER 
updateBucketingVersionForUpgrade()
+// ANYTHING WHICH NEEDS TO BE ADDED MUST BE ADDED ABOVE
+// This call updates the bucketing version of final ReduceSinkOp based on
+// the bucketing version of FileSinkOp. This operation must happen at the
+// end to ensure there is no further rewrite of plan which may end up
+// removing/updating the ReduceSinkOp as was the case with 
SortedDynPartitionOptimizer
+// Update bucketing version of ReduceSinkOp if needed
+updateBucketingVersionForUpgrade(procCtx);
+
   }
 
   private void runCycleAnalysisForPartitionPruning(OptimizeTezProcContext 
procCtx,
@@ -1833,30 +1839,23 @@ public class TezCompiler extends TaskCompiler {
 
 
 for (FileSinkOperator fsOp : fsOpsAll) {
-  Operator parentOfFS = fsOp.getParentOperators().get(0);
-  if (parentOfFS instanceof GroupByOperator) {
-GroupByOperator gbyOp = (GroupByOperator) parentOfFS;
-List aggs = gbyOp.getConf().getAggregatorStrings();
-boolean compute_stats = false;
-for (String agg : aggs) {
-  if (agg.equalsIgnoreCase("compute_stats")) {
-compute_stats = true;
-break;
-  }
-}
-if (compute_stats) {
+  if (!fsOp.getConf().getTableInfo().isSetBucketingVersion()) {
+continue;
+  }
+  // Look for direct parent ReduceSinkOp
+  // If there are more than 1 parent, bail out.
+  Operator parent = fsOp;
+  List> parentOps = parent.getParentOperators();
+  while (parentOps != null && parentOps.size() == 1) {
+parent = parentOps.get(0);
+if (!(parent instanceof ReduceSinkOperator)) {
+  parentOps = parent.getParentOperators();
   continue;
 }
-  }
 
-  // Not compute_stats
-  Set rsOps = 
OperatorUtils.findOperatorsUpstream(parentOfFS, ReduceSinkOperator.class);
-  if (rsOps.isEmpty()) {
-continue;
-  }
-  // Skip setting if the bucketing version is not set in FileSinkOp.
-  if (fsOp.getConf().getTableInfo().isSetBucketingVersion()) {
-
rsOps.iterator().next().setBucketingVersion(fsOp.getConf().getTableInfo().getBucketingVersion());
+// Found the target RSOp
+
parent.setBucketingVersion(fsOp.getConf().getTableInfo().getBucketingVersion());
+break;
   }
 }
   }
diff --git a/ql/src/test/queries/clientpositive/murmur_hash_migration.q 
b/ql/src/test/queries/clientpositive/murmur_hash_migration.q
index 2b8da9f..6bf3785 100644
--- a/ql/src/test/queries/clientpositive/murmur_hash_migration.q
+++ b/ql/src/test/queries/clientpositive/murmur_hash_migration.q
@@ -59,3 +59,38 @@ select t1.key, t1.value, t2.key, t2.value from 
srcbucket_mapjoin_n18 t1, srcbuck
 explain
 select t1.key, t1.value, t2.key, t2.value from tab_part_n11 t1, tab_n10 t2

[hive] branch master updated: HIVE-21408 : Disable synthetic join predicates for non-equi joins for unintended cases (Deepak Jaiswal, reviewed by Vineet Garg)

2019-03-08 Thread djaiswal
This is an automated email from the ASF dual-hosted git repository.

djaiswal pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hive.git


The following commit(s) were added to refs/heads/master by this push:
 new cdd8fa5  HIVE-21408 : Disable synthetic join predicates for non-equi 
joins for unintended cases (Deepak Jaiswal, reviewed by Vineet Garg)
cdd8fa5 is described below

commit cdd8fa58f4b7f6aa2973fb4a114bb891ed397416
Author: Deepak Jaiswal 
AuthorDate: Fri Mar 8 11:14:10 2019 -0800

HIVE-21408 : Disable synthetic join predicates for non-equi joins for 
unintended cases (Deepak Jaiswal, reviewed by Vineet Garg)
---
 .../DynamicPartitionPruningOptimization.java   |  24 +++--
 .../clientpositive/dynamic_semijoin_reduction.q|   3 +
 .../llap/dynamic_semijoin_reduction.q.out  | 111 +
 3 files changed, 129 insertions(+), 9 deletions(-)

diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/DynamicPartitionPruningOptimization.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/DynamicPartitionPruningOptimization.java
index 439fb75..cc01034 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/DynamicPartitionPruningOptimization.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/DynamicPartitionPruningOptimization.java
@@ -167,7 +167,8 @@ public class DynamicPartitionPruningOptimization implements 
NodeProcessor {
 
 Table table = ts.getConf().getTableMetadata();
 
-if (table != null && table.isPartitionKey(column)) {
+boolean nonEquiJoin = isNonEquiJoin(ctx.parent);
+if (table != null && table.isPartitionKey(column) && !nonEquiJoin) {
   String columnType = table.getPartColByName(column).getType();
   String alias = ts.getConf().getAlias();
   PrunedPartitionList plist = parseContext.getPrunedPartitions(alias, 
ts);
@@ -197,7 +198,7 @@ public class DynamicPartitionPruningOptimization implements 
NodeProcessor {
 } else { // semijoin
   LOG.debug("Column " + column + " is not a partition column");
   if (semiJoin && 
!disableSemiJoinOptDueToExternalTable(parseContext.getConf(), ts, ctx)
-  && ts.getConf().getFilterExpr() != null) {
+  && ts.getConf().getFilterExpr() != null && !nonEquiJoin) {
 LOG.debug("Initiate semijoin reduction for " + column + " ("
 + ts.getConf().getFilterExpr().getExprString());
 
@@ -434,6 +435,18 @@ public class DynamicPartitionPruningOptimization 
implements NodeProcessor {
 }
   }
 
+  private boolean isNonEquiJoin(ExprNodeDesc predicate)  {
+Preconditions.checkArgument(predicate instanceof ExprNodeGenericFuncDesc);
+
+ExprNodeGenericFuncDesc funcDesc = (ExprNodeGenericFuncDesc) predicate;
+if (funcDesc.getGenericUDF() instanceof GenericUDFIn) {
+  return false;
+}
+
+return true;
+  }
+
+
   private void generateEventOperatorPlan(DynamicListContext ctx, ParseContext 
parseContext,
   TableScanOperator ts, String column, String columnType) {
 
@@ -530,13 +543,6 @@ public class DynamicPartitionPruningOptimization 
implements NodeProcessor {
   TableScanOperator ts, String keyBaseAlias, String internalColName,
   String colName, SemiJoinHint sjHint) throws SemanticException {
 
-// Semijoin reduction for non-equi join not yet supported, check for it
-ExprNodeGenericFuncDesc funcDesc = (ExprNodeGenericFuncDesc) ctx.parent;
-if (!(funcDesc.getGenericUDF() instanceof GenericUDFIn)) {
-  LOG.info("Semijoin reduction for non-equi joins is currently disabled.");
-  return false;
-}
-
 // we will put a fork in the plan at the source of the reduce sink
 Operator parentOfRS = 
ctx.generator.getParentOperators().get(0);
 
diff --git a/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction.q 
b/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction.q
index 32c2854..f5a5824 100644
--- a/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction.q
+++ b/ql/src/test/queries/clientpositive/dynamic_semijoin_reduction.q
@@ -141,6 +141,9 @@ explain select * from alltypesorc_int_n1 join
select srcpart_small_n3.key1 as key 
from srcpart_small_n3) unionsrc on (alltypesorc_int_n1.cstring = unionsrc.key);
 
 
+-- Non equi joins : Should NOT create a semijoin branch
+EXPLAIN select count(*) from srcpart_date_n7 join srcpart_small_n3 on 
(srcpart_date_n7.key = srcpart_small_n3.key1 and srcpart_date_n7.value > 
srcpart_small_n3.value1);
+
 
 drop table srcpart_date_n7;
 drop table srcpart_small_n3;
diff --git 
a/ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction.q.out 
b/ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction.q.out
index 74bb1a0..d5e4a2a 100644
--- a/ql/src/te

hive git commit: HIVE-19162 : SMB : Test tez_smb_1.q stops making SMB join for a query (Deepak Jaiswal, reviewed by Ashutosh Chauhan)

2018-04-13 Thread djaiswal
Repository: hive
Updated Branches:
  refs/heads/master 5213abf02 -> 3915980fe


HIVE-19162 : SMB : Test tez_smb_1.q stops making SMB join for a query (Deepak 
Jaiswal, reviewed by Ashutosh Chauhan)


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/3915980f
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/3915980f
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/3915980f

Branch: refs/heads/master
Commit: 3915980fe686d0c3edc375dd89587be78bb285fe
Parents: 5213abf
Author: Deepak Jaiswal 
Authored: Fri Apr 13 11:42:34 2018 -0700
Committer: Deepak Jaiswal 
Committed: Fri Apr 13 11:43:24 2018 -0700

--
 .../results/clientpositive/llap/tez_smb_1.q.out | 176 +--
 1 file changed, 82 insertions(+), 94 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hive/blob/3915980f/ql/src/test/results/clientpositive/llap/tez_smb_1.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/tez_smb_1.q.out 
b/ql/src/test/results/clientpositive/llap/tez_smb_1.q.out
index 57e5e1a..faa94862 100644
--- a/ql/src/test/results/clientpositive/llap/tez_smb_1.q.out
+++ b/ql/src/test/results/clientpositive/llap/tez_smb_1.q.out
@@ -123,61 +123,49 @@ STAGE PLANS:
 Tez
  A masked pattern was here 
   Edges:
-Map 1 <- Map 3 (CUSTOM_EDGE)
 Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE)
  A masked pattern was here 
   Vertices:
 Map 1 
 Map Operator Tree:
 TableScan
+  alias: s3
+  Statistics: Num rows: 242 Data size: 3490 Basic stats: 
COMPLETE Column stats: NONE
+  Filter Operator
+predicate: key is not null (type: boolean)
+Statistics: Num rows: 230 Data size: 3316 Basic stats: 
COMPLETE Column stats: NONE
+Select Operator
+  expressions: key (type: int)
+  outputColumnNames: _col0
+  Statistics: Num rows: 230 Data size: 3316 Basic stats: 
COMPLETE Column stats: NONE
+Map Operator Tree:
+TableScan
   alias: s1
-  Statistics: Num rows: 242 Data size: 968 Basic stats: 
COMPLETE Column stats: COMPLETE
+  Statistics: Num rows: 242 Data size: 3490 Basic stats: 
COMPLETE Column stats: NONE
   Filter Operator
 predicate: key is not null (type: boolean)
-Statistics: Num rows: 242 Data size: 968 Basic stats: 
COMPLETE Column stats: COMPLETE
+Statistics: Num rows: 230 Data size: 3316 Basic stats: 
COMPLETE Column stats: NONE
 Select Operator
   expressions: key (type: int)
   outputColumnNames: _col0
-  Statistics: Num rows: 242 Data size: 968 Basic stats: 
COMPLETE Column stats: COMPLETE
-  Map Join Operator
+  Statistics: Num rows: 230 Data size: 3316 Basic stats: 
COMPLETE Column stats: NONE
+  Merge Join Operator
 condition map:
  Inner Join 0 to 1
 keys:
   0 _col0 (type: int)
   1 _col0 (type: int)
-input vertices:
-  1 Map 3
-Statistics: Num rows: 382 Data size: 3056 Basic stats: 
COMPLETE Column stats: COMPLETE
+Statistics: Num rows: 253 Data size: 3647 Basic stats: 
COMPLETE Column stats: NONE
 Group By Operator
   aggregations: count()
   mode: hash
   outputColumnNames: _col0
-  Statistics: Num rows: 1 Data size: 8 Basic stats: 
COMPLETE Column stats: COMPLETE
+  Statistics: Num rows: 1 Data size: 8 Basic stats: 
COMPLETE Column stats: NONE
   Reduce Output Operator
 sort order: 
-Statistics: Num rows: 1 Data size: 8 Basic stats: 
COMPLETE Column stats: COMPLETE
+Statistics: Num rows: 1 Data size: 8 Basic stats: 
COMPLETE Column stats: NONE
 value expressions: _col0 (type: bigint)
 Execution mode: llap
-LLAP IO: no inputs
-Map 3 
-Map Operator Tree:
-TableScan
-  alias: s3
-  Statistics: Num rows: 242 Data size: 968 Basic stats: 
COMPLETE Column stats: COMPLETE
-  Filter Operator
-predicate: key is n

[09/11] hive git commit: HIVE-19195, HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run llap specific tests in llap only.

2018-04-16 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/77da6997/ql/src/test/results/clientpositive/auto_sortmerge_join_11.q.out
--
diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_11.q.out 
b/ql/src/test/results/clientpositive/auto_sortmerge_join_11.q.out
deleted file mode 100644
index a065887..000
--- a/ql/src/test/results/clientpositive/auto_sortmerge_join_11.q.out
+++ /dev/null
@@ -1,1625 +0,0 @@
-PREHOOK: query: CREATE TABLE bucket_small (key string, value string) 
partitioned by (ds string) CLUSTERED BY (key) SORTED BY (KEY) INTO 2 BUCKETS 
STORED AS TEXTFILE
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@bucket_small
-POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) 
partitioned by (ds string) CLUSTERED BY (key) SORTED BY (KEY) INTO 2 BUCKETS 
STORED AS TEXTFILE
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@bucket_small
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small
-POSTHOOK: Output: default@bucket_small@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small@ds=2008-04-08
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small@ds=2008-04-08
-PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned 
by (ds string) CLUSTERED BY (key) SORTED BY(KEY) INTO 4 BUCKETS STORED AS 
TEXTFILE
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@bucket_big
-POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) 
partitioned by (ds string) CLUSTERED BY (key) SORTED BY(KEY) INTO 4 BUCKETS 
STORED AS TEXTFILE
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@bucket_big
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/00_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_big
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/00_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_big
-POSTHOOK: Output: default@bucket_big@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/01_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_big@ds=2008-04-08
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/01_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_big@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/02_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_big@ds=2008-04-08
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/02_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_big@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/03_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_big@ds=2008-04-08
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/03_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_big@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/00_0' INTO TABLE bucket_big 
partition(ds='2008-04-09')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_big
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join

[01/11] hive git commit: HIVE-19195, HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run llap specific tests in llap only.

2018-04-16 Thread djaiswal
Repository: hive
Updated Branches:
  refs/heads/master 1e9f3d77c -> 77da6997d


http://git-wip-us.apache.org/repos/asf/hive/blob/77da6997/ql/src/test/results/clientpositive/tez_join_hash.q.out
--
diff --git a/ql/src/test/results/clientpositive/tez_join_hash.q.out 
b/ql/src/test/results/clientpositive/tez_join_hash.q.out
deleted file mode 100644
index c9b8169..000
--- a/ql/src/test/results/clientpositive/tez_join_hash.q.out
+++ /dev/null
@@ -1,921 +0,0 @@
-PREHOOK: query: create table orc_src (key string, value string) STORED AS ORC
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@orc_src
-POSTHOOK: query: create table orc_src (key string, value string) STORED AS ORC
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@orc_src
-PREHOOK: query: insert into table orc_src select * from src
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src
-PREHOOK: Output: default@orc_src
-POSTHOOK: query: insert into table orc_src select * from src
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src
-POSTHOOK: Output: default@orc_src
-POSTHOOK: Lineage: orc_src.key SIMPLE [(src)src.FieldSchema(name:key, 
type:string, comment:default), ]
-POSTHOOK: Lineage: orc_src.value SIMPLE [(src)src.FieldSchema(name:value, 
type:string, comment:default), ]
-PREHOOK: query: explain
-SELECT count(*) FROM src, orc_src where src.key=orc_src.key
-PREHOOK: type: QUERY
-POSTHOOK: query: explain
-SELECT count(*) FROM src, orc_src where src.key=orc_src.key
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-2 depends on stages: Stage-1
-  Stage-0 depends on stages: Stage-2
-
-STAGE PLANS:
-  Stage: Stage-1
-Map Reduce
-  Map Operator Tree:
-  TableScan
-alias: src
-Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE 
Column stats: NONE
-Filter Operator
-  predicate: key is not null (type: boolean)
-  Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE 
Column stats: NONE
-  Select Operator
-expressions: key (type: string)
-outputColumnNames: _col0
-Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
-Reduce Output Operator
-  key expressions: _col0 (type: string)
-  sort order: +
-  Map-reduce partition columns: _col0 (type: string)
-  Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
-  TableScan
-alias: orc_src
-Statistics: Num rows: 500 Data size: 88000 Basic stats: COMPLETE 
Column stats: NONE
-Filter Operator
-  predicate: key is not null (type: boolean)
-  Statistics: Num rows: 500 Data size: 88000 Basic stats: COMPLETE 
Column stats: NONE
-  Select Operator
-expressions: key (type: string)
-outputColumnNames: _col0
-Statistics: Num rows: 500 Data size: 88000 Basic stats: 
COMPLETE Column stats: NONE
-Reduce Output Operator
-  key expressions: _col0 (type: string)
-  sort order: +
-  Map-reduce partition columns: _col0 (type: string)
-  Statistics: Num rows: 500 Data size: 88000 Basic stats: 
COMPLETE Column stats: NONE
-  Reduce Operator Tree:
-Join Operator
-  condition map:
-   Inner Join 0 to 1
-  keys:
-0 _col0 (type: string)
-1 _col0 (type: string)
-  Statistics: Num rows: 550 Data size: 5843 Basic stats: COMPLETE 
Column stats: NONE
-  Group By Operator
-aggregations: count()
-mode: hash
-outputColumnNames: _col0
-Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column 
stats: NONE
-File Output Operator
-  compressed: false
-  table:
-  input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-  output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-  serde: 
org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe
-
-  Stage: Stage-2
-Map Reduce
-  Map Operator Tree:
-  TableScan
-Reduce Output Operator
-  sort order: 
-  Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
-  value expressions: _col0 (type: bigint)
-  Execution mode: vectorized
-  Reduce Operator Tree:
-Group By Operator
-  aggregations: count(VALUE._col0)
-  mode: mergepartial
-  outputColumnNames: _col0
-  Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column 
stats: NONE
-  File Out

[05/11] hive git commit: HIVE-19195, HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run llap specific tests in llap only.

2018-04-16 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/77da6997/ql/src/test/results/clientpositive/auto_sortmerge_join_3.q.out
--
diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_3.q.out 
b/ql/src/test/results/clientpositive/auto_sortmerge_join_3.q.out
deleted file mode 100644
index 34959b4..000
--- a/ql/src/test/results/clientpositive/auto_sortmerge_join_3.q.out
+++ /dev/null
@@ -1,1153 +0,0 @@
-PREHOOK: query: CREATE TABLE bucket_small (key string, value string) 
partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS 
STORED AS TEXTFILE
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@bucket_small
-POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) 
partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS 
STORED AS TEXTFILE
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@bucket_small
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small
-POSTHOOK: Output: default@bucket_small@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small@ds=2008-04-08
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small 
partition(ds='2008-04-09')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small 
partition(ds='2008-04-09')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small
-POSTHOOK: Output: default@bucket_small@ds=2008-04-09
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small 
partition(ds='2008-04-09')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small@ds=2008-04-09
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small 
partition(ds='2008-04-09')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small@ds=2008-04-09
-PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned 
by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS 
TEXTFILE
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@bucket_big
-POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) 
partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS 
STORED AS TEXTFILE
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@bucket_big
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/00_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_big
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/00_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_big
-POSTHOOK: Output: default@bucket_big@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/01_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_big@ds=2008-04-08
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/01_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_big@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/02_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_big@ds=2008-04-08
-POSTHOOK: query:

[11/11] hive git commit: HIVE-19195, HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run llap specific tests in llap only.

2018-04-16 Thread djaiswal
HIVE-19195,HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run 
llap specific tests in llap only.


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/77da6997
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/77da6997
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/77da6997

Branch: refs/heads/master
Commit: 77da6997d4357c1d88f3a966fe549f1f41efeab5
Parents: 1e9f3d7
Author: Deepak Jaiswal 
Authored: Mon Apr 16 11:08:30 2018 -0700
Committer: Deepak Jaiswal 
Committed: Mon Apr 16 11:08:58 2018 -0700

--
 .../test/resources/testconfiguration.properties |   60 +-
 .../clientpositive/auto_sortmerge_join_1.q.out  | 1324 --
 .../clientpositive/auto_sortmerge_join_10.q.out |  620 -
 .../clientpositive/auto_sortmerge_join_11.q.out | 1625 -
 .../clientpositive/auto_sortmerge_join_12.q.out | 2295 --
 .../clientpositive/auto_sortmerge_join_13.q.out |  775 --
 .../clientpositive/auto_sortmerge_join_14.q.out |  298 ---
 .../clientpositive/auto_sortmerge_join_15.q.out |  266 --
 .../clientpositive/auto_sortmerge_join_2.q.out  | 1109 -
 .../clientpositive/auto_sortmerge_join_3.q.out  | 1153 -
 .../clientpositive/auto_sortmerge_join_4.q.out  | 1169 -
 .../clientpositive/auto_sortmerge_join_5.q.out  |  856 ---
 .../clientpositive/auto_sortmerge_join_7.q.out  | 1489 
 .../results/clientpositive/llap/llap_smb.q.out  |   10 +-
 .../clientpositive/llap/tez_smb_main.q.out  |  142 +-
 .../test/results/clientpositive/llap_acid.q.out |  401 ---
 .../results/clientpositive/llap_acid_fast.q.out |  395 ---
 .../results/clientpositive/tez_join_hash.q.out  |  921 ---
 18 files changed, 85 insertions(+), 14823 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hive/blob/77da6997/itests/src/test/resources/testconfiguration.properties
--
diff --git a/itests/src/test/resources/testconfiguration.properties 
b/itests/src/test/resources/testconfiguration.properties
index f99d6a1..2845ab6 100644
--- a/itests/src/test/resources/testconfiguration.properties
+++ b/itests/src/test/resources/testconfiguration.properties
@@ -99,30 +99,12 @@ minillaplocal.shared.query.files=alter_merge_2_orc.q,\
   auto_join30.q,\
   auto_join_filters.q,\
   auto_join_nulls.q,\
-  auto_sortmerge_join_1.q,\
-  auto_sortmerge_join_10.q,\
-  auto_sortmerge_join_11.q,\
-  auto_sortmerge_join_12.q,\
-  auto_sortmerge_join_13.q,\
-  auto_sortmerge_join_14.q,\
-  auto_sortmerge_join_15.q,\
-  auto_sortmerge_join_16.q,\
-  auto_sortmerge_join_2.q,\
-  auto_sortmerge_join_3.q,\
-  auto_sortmerge_join_4.q,\
-  auto_sortmerge_join_5.q,\
-  auto_sortmerge_join_6.q,\
-  auto_sortmerge_join_7.q,\
-  auto_sortmerge_join_8.q,\
-  auto_sortmerge_join_9.q,\
   autoColumnStats_1.q,\
   autoColumnStats_10.q,\
   autoColumnStats_2.q,\
   bucket2.q,\
   bucket3.q,\
   bucket4.q,\
-  bucket_map_join_tez1.q,\
-  bucket_map_join_tez2.q,\
   cbo_gby.q,\
   cbo_gby_empty.q,\
   cbo_join.q,\
@@ -169,8 +151,6 @@ minillaplocal.shared.query.files=alter_merge_2_orc.q,\
   delete_where_partitioned.q,\
   delete_whole_partition.q,\
   disable_merge_for_bucketing.q,\
-  dynamic_partition_pruning.q,\
-  dynamic_semijoin_reduction.q,\
   dynpart_sort_opt_vectorization.q,\
   dynpart_sort_optimization.q,\
   dynpart_sort_optimization2.q,\
@@ -200,8 +180,6 @@ minillaplocal.shared.query.files=alter_merge_2_orc.q,\
   kill_query.q,\
   leftsemijoin.q,\
   limit_pushdown.q,\
-  llap_acid.q,\
-  llap_acid_fast.q,\
   load_dyn_part1.q,\
   load_dyn_part2.q,\
   load_dyn_part3.q,\
@@ -253,29 +231,6 @@ minillaplocal.shared.query.files=alter_merge_2_orc.q,\
   subquery_exists.q,\
   subquery_in.q,\
   temp_table.q,\
-  tez_bmj_schema_evolution.q,\
-  tez_dml.q,\
-  tez_dynpart_hashjoin_1.q,\
-  tez_dynpart_hashjoin_2.q,\
-  tez_fsstat.q,\
-  tez_insert_overwrite_local_directory_1.q,\
-  tez_join.q,\
-  tez_join_hash.q,\
-  tez_join_result_complex.q,\
-  tez_join_tests.q,\
-  tez_joins_explain.q,\
-  tez_multi_union.q,\
-  tez_schema_evolution.q,\
-  tez_self_join.q,\
-  tez_smb_1.q,\
-  tez_smb_main.q,\
-  tez_union.q,\
-  tez_union2.q,\
-  tez_union_decimal.q,\
-  tez_union_group_by.q,\
-  tez_union_multiinsert.q,\
-  tez_vector_dynpart_hashjoin_1.q,\
-  tez_vector_dynpart_hashjoin_2.q,\
   union2.q,\
   union3.q,\
   union4.q,\
@@ -448,7 +403,6 @@ minillap.query.files=acid_bucket_pruning.q,\
   intersect_all.q,\
   intersect_distinct.q,\
   intersect_merge.q,\
-  llap_smb.q,\
   llap_udf.q,\
   llapdecider.q,\
   reduce_deduplicate.q,\
@@ -494,8 +448,20 @@ minillaplocal.query.files=\
   auto_join30.q,\
   auto_join_filters.q,\
   auto_join_nulls.q,\
+  auto_sortmerge_join_1.q,\
+  auto_sortmerge_join_10.q,\
+  auto_sortmerge_

[04/11] hive git commit: HIVE-19195, HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run llap specific tests in llap only.

2018-04-16 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/77da6997/ql/src/test/results/clientpositive/auto_sortmerge_join_4.q.out
--
diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_4.q.out 
b/ql/src/test/results/clientpositive/auto_sortmerge_join_4.q.out
deleted file mode 100644
index 0d586fd..000
--- a/ql/src/test/results/clientpositive/auto_sortmerge_join_4.q.out
+++ /dev/null
@@ -1,1169 +0,0 @@
-PREHOOK: query: CREATE TABLE bucket_small (key string, value string) 
partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS 
STORED AS TEXTFILE
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@bucket_small
-POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) 
partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS 
STORED AS TEXTFILE
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@bucket_small
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small
-POSTHOOK: Output: default@bucket_small@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small@ds=2008-04-08
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/02_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small@ds=2008-04-08
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/02_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/03_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small@ds=2008-04-08
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/03_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small 
partition(ds='2008-04-09')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small 
partition(ds='2008-04-09')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small
-POSTHOOK: Output: default@bucket_small@ds=2008-04-09
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small 
partition(ds='2008-04-09')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small@ds=2008-04-09
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small 
partition(ds='2008-04-09')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small@ds=2008-04-09
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/02_0' INTO TABLE bucket_small 
partition(ds='2008-04-09')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small@ds=2008-04-09
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/02_0' INTO TABLE bucket_small 
partition(ds='2008-04-09')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small@ds=2008-04-09
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/03_0' INTO TABLE bucket_small 
partition(ds='2008-04-09')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small@ds=2008-04-09
-POSTHOOK: query: load data 

[06/11] hive git commit: HIVE-19195, HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run llap specific tests in llap only.

2018-04-16 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/77da6997/ql/src/test/results/clientpositive/auto_sortmerge_join_2.q.out
--
diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_2.q.out 
b/ql/src/test/results/clientpositive/auto_sortmerge_join_2.q.out
deleted file mode 100644
index 5cfc35a..000
--- a/ql/src/test/results/clientpositive/auto_sortmerge_join_2.q.out
+++ /dev/null
@@ -1,1109 +0,0 @@
-PREHOOK: query: CREATE TABLE bucket_small (key string, value string) 
partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS 
STORED AS TEXTFILE
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@bucket_small
-POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) 
partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS 
STORED AS TEXTFILE
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@bucket_small
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small
-POSTHOOK: Output: default@bucket_small@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small@ds=2008-04-08
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/02_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small@ds=2008-04-08
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/02_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/03_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small@ds=2008-04-08
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/03_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small@ds=2008-04-08
-PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned 
by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS 
TEXTFILE
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@bucket_big
-POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) 
partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS 
STORED AS TEXTFILE
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@bucket_big
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/00_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_big
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/00_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_big
-POSTHOOK: Output: default@bucket_big@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/01_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_big@ds=2008-04-08
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/01_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_big@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/00_0' INTO TABLE bucket_big 
partition(ds='2008-04-09')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_big
-POSTHOOK: query: load data local inpath 
'../../data/fil

[10/11] hive git commit: HIVE-19195, HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run llap specific tests in llap only.

2018-04-16 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/77da6997/ql/src/test/results/clientpositive/auto_sortmerge_join_10.q.out
--
diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_10.q.out 
b/ql/src/test/results/clientpositive/auto_sortmerge_join_10.q.out
deleted file mode 100644
index 3a90b05..000
--- a/ql/src/test/results/clientpositive/auto_sortmerge_join_10.q.out
+++ /dev/null
@@ -1,620 +0,0 @@
-PREHOOK: query: CREATE TABLE tbl1(key int, value string) CLUSTERED BY (key) 
SORTED BY (key) INTO 2 BUCKETS
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@tbl1
-POSTHOOK: query: CREATE TABLE tbl1(key int, value string) CLUSTERED BY (key) 
SORTED BY (key) INTO 2 BUCKETS
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@tbl1
-PREHOOK: query: CREATE TABLE tbl2(key int, value string) CLUSTERED BY (key) 
SORTED BY (key) INTO 2 BUCKETS
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@tbl2
-POSTHOOK: query: CREATE TABLE tbl2(key int, value string) CLUSTERED BY (key) 
SORTED BY (key) INTO 2 BUCKETS
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@tbl2
-PREHOOK: query: insert overwrite table tbl1
-select * from src where key < 10
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src
-PREHOOK: Output: default@tbl1
-POSTHOOK: query: insert overwrite table tbl1
-select * from src where key < 10
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src
-POSTHOOK: Output: default@tbl1
-POSTHOOK: Lineage: tbl1.key EXPRESSION [(src)src.FieldSchema(name:key, 
type:string, comment:default), ]
-POSTHOOK: Lineage: tbl1.value SIMPLE [(src)src.FieldSchema(name:value, 
type:string, comment:default), ]
-PREHOOK: query: insert overwrite table tbl2
-select * from src where key < 10
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src
-PREHOOK: Output: default@tbl2
-POSTHOOK: query: insert overwrite table tbl2
-select * from src where key < 10
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src
-POSTHOOK: Output: default@tbl2
-POSTHOOK: Lineage: tbl2.key EXPRESSION [(src)src.FieldSchema(name:key, 
type:string, comment:default), ]
-POSTHOOK: Lineage: tbl2.value SIMPLE [(src)src.FieldSchema(name:value, 
type:string, comment:default), ]
-PREHOOK: query: explain
-select count(*) from 
-  (
-  select * from
-  (select a.key as key, a.value as value from tbl1 a where key < 6
- union all
-   select a.key as key, a.value as value from tbl1 a where key < 6
-  ) usubq1 ) subq1
-join
-  (select a.key as key, a.value as value from tbl2 a where key < 6) subq2
-  on subq1.key = subq2.key
-PREHOOK: type: QUERY
-POSTHOOK: query: explain
-select count(*) from 
-  (
-  select * from
-  (select a.key as key, a.value as value from tbl1 a where key < 6
- union all
-   select a.key as key, a.value as value from tbl1 a where key < 6
-  ) usubq1 ) subq1
-join
-  (select a.key as key, a.value as value from tbl2 a where key < 6) subq2
-  on subq1.key = subq2.key
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
-  Stage-7 is a root stage , consists of Stage-8, Stage-9, Stage-1
-  Stage-8 has a backup stage: Stage-1
-  Stage-5 depends on stages: Stage-8
-  Stage-2 depends on stages: Stage-1, Stage-5, Stage-6
-  Stage-9 has a backup stage: Stage-1
-  Stage-6 depends on stages: Stage-9
-  Stage-1
-  Stage-0 depends on stages: Stage-2
-
-STAGE PLANS:
-  Stage: Stage-7
-Conditional Operator
-
-  Stage: Stage-8
-Map Reduce Local Work
-  Alias -> Map Local Tables:
-$hdt$_1:a 
-  Fetch Operator
-limit: -1
-  Alias -> Map Local Operator Tree:
-$hdt$_1:a 
-  TableScan
-alias: a
-Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE 
Column stats: NONE
-Filter Operator
-  predicate: (key < 6) (type: boolean)
-  Statistics: Num rows: 3 Data size: 21 Basic stats: COMPLETE 
Column stats: NONE
-  Select Operator
-expressions: key (type: int)
-outputColumnNames: _col0
-Statistics: Num rows: 3 Data size: 21 Basic stats: COMPLETE 
Column stats: NONE
-HashTable Sink Operator
-  keys:
-0 _col0 (type: int)
-1 _col0 (type: int)
-
-  Stage: Stage-5
-Map Reduce
-  Map Operator Tree:
-  TableScan
-alias: a
-Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE 
Column stats: NONE
-Filter Operator
-  predicate: (key < 6) (type: boolean)
-  Statistics: Num rows: 3 Data size: 21 Basic stats: COMPLETE 
Column stats: NONE
-  Select Operator
-expressions: key (type: int)
-outputColumnNames: _col0
-Statistics: Num rows: 3 Data size: 21 Basic stats: COM

[02/11] hive git commit: HIVE-19195, HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run llap specific tests in llap only.

2018-04-16 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/77da6997/ql/src/test/results/clientpositive/llap/tez_smb_main.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/tez_smb_main.q.out 
b/ql/src/test/results/clientpositive/llap/tez_smb_main.q.out
index 705bcc3..0dcca1f 100644
--- a/ql/src/test/results/clientpositive/llap/tez_smb_main.q.out
+++ b/ql/src/test/results/clientpositive/llap/tez_smb_main.q.out
@@ -263,33 +263,33 @@ STAGE PLANS:
 Map Operator Tree:
 TableScan
   alias: a
-  Statistics: Num rows: 242 Data size: 968 Basic stats: 
COMPLETE Column stats: COMPLETE
+  Statistics: Num rows: 242 Data size: 3490 Basic stats: 
COMPLETE Column stats: NONE
   Filter Operator
 predicate: key is not null (type: boolean)
-Statistics: Num rows: 242 Data size: 968 Basic stats: 
COMPLETE Column stats: COMPLETE
+Statistics: Num rows: 230 Data size: 3316 Basic stats: 
COMPLETE Column stats: NONE
 Select Operator
   expressions: key (type: int)
   outputColumnNames: _col0
-  Statistics: Num rows: 242 Data size: 968 Basic stats: 
COMPLETE Column stats: COMPLETE
+  Statistics: Num rows: 230 Data size: 3316 Basic stats: 
COMPLETE Column stats: NONE
   Reduce Output Operator
 key expressions: _col0 (type: int)
 sort order: +
 Map-reduce partition columns: _col0 (type: int)
-Statistics: Num rows: 242 Data size: 968 Basic stats: 
COMPLETE Column stats: COMPLETE
+Statistics: Num rows: 230 Data size: 3316 Basic stats: 
COMPLETE Column stats: NONE
 Execution mode: llap
 LLAP IO: no inputs
 Map 2 
 Map Operator Tree:
 TableScan
   alias: b
-  Statistics: Num rows: 500 Data size: 2000 Basic stats: 
COMPLETE Column stats: COMPLETE
+  Statistics: Num rows: 500 Data size: 7216 Basic stats: 
COMPLETE Column stats: NONE
   Filter Operator
 predicate: key is not null (type: boolean)
-Statistics: Num rows: 500 Data size: 2000 Basic stats: 
COMPLETE Column stats: COMPLETE
+Statistics: Num rows: 475 Data size: 6855 Basic stats: 
COMPLETE Column stats: NONE
 Select Operator
   expressions: key (type: int)
   outputColumnNames: _col0
-  Statistics: Num rows: 500 Data size: 2000 Basic stats: 
COMPLETE Column stats: COMPLETE
+  Statistics: Num rows: 475 Data size: 6855 Basic stats: 
COMPLETE Column stats: NONE
   Map Join Operator
 condition map:
  Inner Join 0 to 1
@@ -298,15 +298,15 @@ STAGE PLANS:
   1 _col0 (type: int)
 input vertices:
   0 Map 1
-Statistics: Num rows: 391 Data size: 3128 Basic stats: 
COMPLETE Column stats: COMPLETE
+Statistics: Num rows: 522 Data size: 7540 Basic stats: 
COMPLETE Column stats: NONE
 Group By Operator
   aggregations: count()
   mode: hash
   outputColumnNames: _col0
-  Statistics: Num rows: 1 Data size: 8 Basic stats: 
COMPLETE Column stats: COMPLETE
+  Statistics: Num rows: 1 Data size: 8 Basic stats: 
COMPLETE Column stats: NONE
   Reduce Output Operator
 sort order: 
-Statistics: Num rows: 1 Data size: 8 Basic stats: 
COMPLETE Column stats: COMPLETE
+Statistics: Num rows: 1 Data size: 8 Basic stats: 
COMPLETE Column stats: NONE
 value expressions: _col0 (type: bigint)
 Execution mode: llap
 LLAP IO: no inputs
@@ -317,10 +317,10 @@ STAGE PLANS:
 aggregations: count(VALUE._col0)
 mode: mergepartial
 outputColumnNames: _col0
-Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: COMPLETE
+Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
 File Output Operator
   compressed: false
-  Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: COMPLETE
+  Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
   table:

[08/11] hive git commit: HIVE-19195, HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run llap specific tests in llap only.

2018-04-16 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/77da6997/ql/src/test/results/clientpositive/auto_sortmerge_join_12.q.out
--
diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_12.q.out 
b/ql/src/test/results/clientpositive/auto_sortmerge_join_12.q.out
deleted file mode 100644
index 96b0016..000
--- a/ql/src/test/results/clientpositive/auto_sortmerge_join_12.q.out
+++ /dev/null
@@ -1,2295 +0,0 @@
-PREHOOK: query: CREATE TABLE bucket_small (key string, value string) 
partitioned by (ds string)
-CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@bucket_small
-POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) 
partitioned by (ds string)
-CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@bucket_small
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small
-POSTHOOK: Output: default@bucket_small@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small@ds=2008-04-08
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small@ds=2008-04-08
-PREHOOK: query: CREATE TABLE bucket_big (key string, value string) partitioned 
by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS STORED AS 
TEXTFILE
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@bucket_big
-POSTHOOK: query: CREATE TABLE bucket_big (key string, value string) 
partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS 
STORED AS TEXTFILE
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@bucket_big
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/00_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_big
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/00_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_big
-POSTHOOK: Output: default@bucket_big@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/01_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_big@ds=2008-04-08
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/01_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_big@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/02_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_big@ds=2008-04-08
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/02_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_big@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/03_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_big@ds=2008-04-08
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/03_0' INTO TABLE bucket_big 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_big@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/big/00_0' INTO TABLE bucket_big 
partition(ds='2008-04-09')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_big
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_jo

[07/11] hive git commit: HIVE-19195, HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run llap specific tests in llap only.

2018-04-16 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/77da6997/ql/src/test/results/clientpositive/auto_sortmerge_join_13.q.out
--
diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_13.q.out 
b/ql/src/test/results/clientpositive/auto_sortmerge_join_13.q.out
deleted file mode 100644
index ac143dd..000
--- a/ql/src/test/results/clientpositive/auto_sortmerge_join_13.q.out
+++ /dev/null
@@ -1,775 +0,0 @@
-PREHOOK: query: CREATE TABLE tbl1(key int, value string) CLUSTERED BY (key) 
SORTED BY (key) INTO 2 BUCKETS
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@tbl1
-POSTHOOK: query: CREATE TABLE tbl1(key int, value string) CLUSTERED BY (key) 
SORTED BY (key) INTO 2 BUCKETS
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@tbl1
-PREHOOK: query: CREATE TABLE tbl2(key int, value string) CLUSTERED BY (key) 
SORTED BY (key) INTO 2 BUCKETS
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@tbl2
-POSTHOOK: query: CREATE TABLE tbl2(key int, value string) CLUSTERED BY (key) 
SORTED BY (key) INTO 2 BUCKETS
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@tbl2
-PREHOOK: query: insert overwrite table tbl1 select * from src where key < 10
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src
-PREHOOK: Output: default@tbl1
-POSTHOOK: query: insert overwrite table tbl1 select * from src where key < 10
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src
-POSTHOOK: Output: default@tbl1
-POSTHOOK: Lineage: tbl1.key EXPRESSION [(src)src.FieldSchema(name:key, 
type:string, comment:default), ]
-POSTHOOK: Lineage: tbl1.value SIMPLE [(src)src.FieldSchema(name:value, 
type:string, comment:default), ]
-PREHOOK: query: insert overwrite table tbl2 select * from src where key < 10
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src
-PREHOOK: Output: default@tbl2
-POSTHOOK: query: insert overwrite table tbl2 select * from src where key < 10
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src
-POSTHOOK: Output: default@tbl2
-POSTHOOK: Lineage: tbl2.key EXPRESSION [(src)src.FieldSchema(name:key, 
type:string, comment:default), ]
-POSTHOOK: Lineage: tbl2.value SIMPLE [(src)src.FieldSchema(name:value, 
type:string, comment:default), ]
-PREHOOK: query: CREATE TABLE dest1(k1 int, k2 int)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@dest1
-POSTHOOK: query: CREATE TABLE dest1(k1 int, k2 int)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@dest1
-PREHOOK: query: CREATE TABLE dest2(k1 string, k2 string)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@dest2
-POSTHOOK: query: CREATE TABLE dest2(k1 string, k2 string)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@dest2
-PREHOOK: query: explain 
-from (
-  SELECT a.key key1, a.value value1, b.key key2, b.value value2 
-  FROM tbl1 a JOIN tbl2 b 
-  ON a.key = b.key ) subq
-INSERT OVERWRITE TABLE dest1 select key1, key2
-INSERT OVERWRITE TABLE dest2 select value1, value2
-PREHOOK: type: QUERY
-POSTHOOK: query: explain 
-from (
-  SELECT a.key key1, a.value value1, b.key key2, b.value value2 
-  FROM tbl1 a JOIN tbl2 b 
-  ON a.key = b.key ) subq
-INSERT OVERWRITE TABLE dest1 select key1, key2
-INSERT OVERWRITE TABLE dest2 select value1, value2
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
-  Stage-2 is a root stage
-  Stage-0 depends on stages: Stage-2
-  Stage-3 depends on stages: Stage-0, Stage-5
-  Stage-1 depends on stages: Stage-2
-  Stage-4 depends on stages: Stage-1, Stage-5
-  Stage-5 depends on stages: Stage-2
-
-STAGE PLANS:
-  Stage: Stage-2
-Map Reduce
-  Map Operator Tree:
-  TableScan
-alias: a
-Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE 
Column stats: NONE
-Filter Operator
-  predicate: key is not null (type: boolean)
-  Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE 
Column stats: NONE
-  Select Operator
-expressions: key (type: int), value (type: string)
-outputColumnNames: _col0, _col1
-Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE 
Column stats: NONE
-Sorted Merge Bucket Map Join Operator
-  condition map:
-   Inner Join 0 to 1
-  keys:
-0 _col0 (type: int)
-1 _col0 (type: int)
-  outputColumnNames: _col0, _col1, _col2, _col3
-  Select Operator
-expressions: _col0 (type: int), _col2 (type: int)
-outputColumnNames: _col0, _col1
-File Output Operator
-  compressed: false
-

[03/11] hive git commit: HIVE-19195, HIVE-17055 : Fix flaky tests and cleanup testconfiguration to run llap specific tests in llap only.

2018-04-16 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/77da6997/ql/src/test/results/clientpositive/auto_sortmerge_join_7.q.out
--
diff --git a/ql/src/test/results/clientpositive/auto_sortmerge_join_7.q.out 
b/ql/src/test/results/clientpositive/auto_sortmerge_join_7.q.out
deleted file mode 100644
index 1959075..000
--- a/ql/src/test/results/clientpositive/auto_sortmerge_join_7.q.out
+++ /dev/null
@@ -1,1489 +0,0 @@
-PREHOOK: query: CREATE TABLE bucket_small (key string, value string) 
partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS 
STORED AS TEXTFILE
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@bucket_small
-POSTHOOK: query: CREATE TABLE bucket_small (key string, value string) 
partitioned by (ds string) CLUSTERED BY (key) SORTED BY (key) INTO 4 BUCKETS 
STORED AS TEXTFILE
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@bucket_small
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small
-POSTHOOK: Output: default@bucket_small@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small@ds=2008-04-08
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/02_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small@ds=2008-04-08
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/02_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/03_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small@ds=2008-04-08
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/03_0' INTO TABLE bucket_small 
partition(ds='2008-04-08')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small@ds=2008-04-08
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small 
partition(ds='2008-04-09')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/00_0' INTO TABLE bucket_small 
partition(ds='2008-04-09')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small
-POSTHOOK: Output: default@bucket_small@ds=2008-04-09
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small 
partition(ds='2008-04-09')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small@ds=2008-04-09
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/01_0' INTO TABLE bucket_small 
partition(ds='2008-04-09')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small@ds=2008-04-09
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/02_0' INTO TABLE bucket_small 
partition(ds='2008-04-09')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small@ds=2008-04-09
-POSTHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/02_0' INTO TABLE bucket_small 
partition(ds='2008-04-09')
-POSTHOOK: type: LOAD
- A masked pattern was here 
-POSTHOOK: Output: default@bucket_small@ds=2008-04-09
-PREHOOK: query: load data local inpath 
'../../data/files/auto_sortmerge_join/small/03_0' INTO TABLE bucket_small 
partition(ds='2008-04-09')
-PREHOOK: type: LOAD
- A masked pattern was here 
-PREHOOK: Output: default@bucket_small@ds=2008-04-09
-POSTHOOK: query: load data 

hive git commit: HIVE-17055 : Flaky test: TestMiniLlapCliDriver.testCliDriver[llap_smb] (Deepak Jaiswal, reviewed by Vineet Garg)

2018-04-20 Thread djaiswal
Repository: hive
Updated Branches:
  refs/heads/master a94f382a0 -> f6d2e6e08


HIVE-17055 : Flaky test: TestMiniLlapCliDriver.testCliDriver[llap_smb] (Deepak 
Jaiswal, reviewed by Vineet Garg)


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/f6d2e6e0
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/f6d2e6e0
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/f6d2e6e0

Branch: refs/heads/master
Commit: f6d2e6e0825ba1415882d0f7ddcc09ee6911cbd7
Parents: a94f382
Author: Deepak Jaiswal 
Authored: Fri Apr 20 17:00:32 2018 -0700
Committer: Deepak Jaiswal 
Committed: Fri Apr 20 17:00:32 2018 -0700

--
 itests/src/test/resources/testconfiguration.properties | 7 +++
 1 file changed, 3 insertions(+), 4 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hive/blob/f6d2e6e0/itests/src/test/resources/testconfiguration.properties
--
diff --git a/itests/src/test/resources/testconfiguration.properties 
b/itests/src/test/resources/testconfiguration.properties
index 4e7c519..3aaa68b 100644
--- a/itests/src/test/resources/testconfiguration.properties
+++ b/itests/src/test/resources/testconfiguration.properties
@@ -83,10 +83,7 @@ minillap.shared.query.files=insert_into1.q,\
   orc_struct_type_vectorization.q,\
   union_type_chk.q,\
   cte_2.q,\
-  cte_4.q,\
-  llap_nullscan.q,\
-  dynamic_partition_pruning_2.q,\
-  tez_union_dynamic_partition.q
+  cte_4.q
 
 minillaplocal.shared.query.files=alter_merge_2_orc.q,\
   alter_merge_orc.q,\
@@ -552,6 +549,7 @@ minillaplocal.query.files=\
   llap_acid_fast.q,\
   llap_acid2.q,\
   llap_partitioned.q,\
+  llap_smb.q,\
   llap_vector_nohybridgrace.q,\
   load_data_acid_rename.q,\
   load_dyn_part5.q,\
@@ -687,6 +685,7 @@ minillaplocal.query.files=\
   tez_fsstat.q,\
   tez_insert_overwrite_local_directory_1.q,\
   tez_join.q,\
+  tez_join_hash.q,\
   tez_join_result_complex.q,\
   tez_join_tests.q,\
   tez_joins_explain.q,\



hive git commit: Preparing for storage-api 2.7.0 release.

2018-07-06 Thread djaiswal
Repository: hive
Updated Branches:
  refs/heads/storage-branch-2.7 [created] 13802a2cb


Preparing for storage-api 2.7.0 release.


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/13802a2c
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/13802a2c
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/13802a2c

Branch: refs/heads/storage-branch-2.7
Commit: 13802a2cbdfbd22532cb80ae00cb42303f036dea
Parents: eae5225
Author: Deepak Jaiswal 
Authored: Fri Jul 6 14:02:38 2018 -0700
Committer: Deepak Jaiswal 
Committed: Fri Jul 6 14:02:38 2018 -0700

--
 storage-api/pom.xml | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/hive/blob/13802a2c/storage-api/pom.xml
--
diff --git a/storage-api/pom.xml b/storage-api/pom.xml
index 06c10d0..d7cd0fb 100644
--- a/storage-api/pom.xml
+++ b/storage-api/pom.xml
@@ -25,7 +25,7 @@
 
   org.apache.hive
   hive-storage-api
-  2.7.0-SNAPSHOT
+  2.7.0
   jar
   Hive Storage API
 



[hive] Git Push Summary

2018-07-06 Thread djaiswal
Repository: hive
Updated Tags:  refs/tags/storage-release-2.7.0-rc0 [created] df9693a66


hive git commit: HIVE-20100 : OpTraits : Select Optraits should stop when a mismatch is detected (Deepak Jaiswal, reviewed by Jason Dere)

2018-07-09 Thread djaiswal
Repository: hive
Updated Branches:
  refs/heads/master 5016d6f05 -> ddf7e25d3


HIVE-20100 : OpTraits : Select Optraits should stop when a mismatch is detected 
(Deepak Jaiswal, reviewed by Jason Dere)


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/ddf7e25d
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/ddf7e25d
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/ddf7e25d

Branch: refs/heads/master
Commit: ddf7e25d3ee94224c17e353e334728241515fffb
Parents: 5016d6f
Author: Deepak Jaiswal 
Authored: Mon Jul 9 15:31:13 2018 -0700
Committer: Deepak Jaiswal 
Committed: Mon Jul 9 15:31:13 2018 -0700

--
 .../annotation/OpTraitsRulesProcFactory.java|  59 +-
 .../clientpositive/llap/subquery_notin.q.out| 118 ---
 .../results/clientpositive/llap/tez_join.q.out  |  51 +---
 3 files changed, 138 insertions(+), 90 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hive/blob/ddf7e25d/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java
--
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java
index 263770e..89db530 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java
@@ -308,44 +308,40 @@ public class OpTraitsRulesProcFactory {
 
   public static class SelectRule implements NodeProcessor {
 
-boolean processSortCols = false;
-
 // For bucket columns
 // If all the columns match to the parent, put them in the bucket cols
 // else, add empty list.
 // For sort columns
 // Keep the subset of all the columns as long as order is maintained.
 public List> getConvertedColNames(
-List> parentColNames, SelectOperator selOp) {
+List> parentColNames, SelectOperator selOp, boolean 
processSortCols) {
   List> listBucketCols = new ArrayList<>();
-  if (selOp.getColumnExprMap() != null) {
-if (parentColNames != null) {
-  for (List colNames : parentColNames) {
-List bucketColNames = new ArrayList<>();
-boolean found = false;
-for (String colName : colNames) {
-  for (Entry entry : 
selOp.getColumnExprMap().entrySet()) {
-if ((entry.getValue() instanceof ExprNodeColumnDesc) &&
-(((ExprNodeColumnDesc) 
(entry.getValue())).getColumn().equals(colName))) {
-  bucketColNames.add(entry.getKey());
-  found = true;
-  break;
-}
-  }
-  if (!found) {
-// Bail out on first missed column.
-break;
-  }
-}
-if (!processSortCols && !found) {
-  // While processing bucket columns, atleast one bucket column
-  // missed. This results in a different bucketing scheme.
-  // Add empty list
-  listBucketCols.add(new ArrayList<>());
-} else  {
-  listBucketCols.add(bucketColNames);
+  for (List colNames : parentColNames) {
+List bucketColNames = new ArrayList<>();
+boolean found = false;
+for (String colName : colNames) {
+  // Reset found
+  found = false;
+  for (Entry entry : 
selOp.getColumnExprMap().entrySet()) {
+if ((entry.getValue() instanceof ExprNodeColumnDesc) &&
+(((ExprNodeColumnDesc) 
(entry.getValue())).getColumn().equals(colName))) {
+  bucketColNames.add(entry.getKey());
+  found = true;
+  break;
 }
   }
+  if (!found) {
+// Bail out on first missed column.
+break;
+  }
+}
+if (!processSortCols && !found) {
+  // While processing bucket columns, atleast one bucket column
+  // missed. This results in a different bucketing scheme.
+  // Add empty list
+  listBucketCols.add(new ArrayList<>());
+} else  {
+  listBucketCols.add(bucketColNames);
 }
   }
 
@@ -363,13 +359,12 @@ public class OpTraitsRulesProcFactory {
   List> listSortCols = null;
   if (selOp.getColumnExprMap() != null) {
 if (parentBucketColNames != null) {
-  listBucketCols = getConvertedColNames(parentBucketColNames, selOp);
+  listBucketCols = getConvertedColNames(parentBucketColNames, selOp, 
false);
 }
 List> parentSortC

hive git commit: HIVE-20100 : OpTraits : Select Optraits should stop when a mismatch is detected (Deepak Jaiswal, reviewed by Jason Dere)

2018-07-10 Thread djaiswal
Repository: hive
Updated Branches:
  refs/heads/branch-3 c462214aa -> 6c55afb04


HIVE-20100 : OpTraits : Select Optraits should stop when a mismatch is detected 
(Deepak Jaiswal, reviewed by Jason Dere)


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/6c55afb0
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/6c55afb0
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/6c55afb0

Branch: refs/heads/branch-3
Commit: 6c55afb046daa5c28c4fb1e2a573a4a05e3f5572
Parents: c462214
Author: Deepak Jaiswal 
Authored: Mon Jul 9 15:31:13 2018 -0700
Committer: Deepak Jaiswal 
Committed: Tue Jul 10 22:10:58 2018 -0700

--
 .../annotation/OpTraitsRulesProcFactory.java|  59 +-
 .../clientpositive/llap/subquery_notin.q.out| 118 ---
 .../results/clientpositive/llap/tez_join.q.out  |  51 +---
 3 files changed, 138 insertions(+), 90 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hive/blob/6c55afb0/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java
--
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java
index 263770e..89db530 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java
@@ -308,44 +308,40 @@ public class OpTraitsRulesProcFactory {
 
   public static class SelectRule implements NodeProcessor {
 
-boolean processSortCols = false;
-
 // For bucket columns
 // If all the columns match to the parent, put them in the bucket cols
 // else, add empty list.
 // For sort columns
 // Keep the subset of all the columns as long as order is maintained.
 public List> getConvertedColNames(
-List> parentColNames, SelectOperator selOp) {
+List> parentColNames, SelectOperator selOp, boolean 
processSortCols) {
   List> listBucketCols = new ArrayList<>();
-  if (selOp.getColumnExprMap() != null) {
-if (parentColNames != null) {
-  for (List colNames : parentColNames) {
-List bucketColNames = new ArrayList<>();
-boolean found = false;
-for (String colName : colNames) {
-  for (Entry entry : 
selOp.getColumnExprMap().entrySet()) {
-if ((entry.getValue() instanceof ExprNodeColumnDesc) &&
-(((ExprNodeColumnDesc) 
(entry.getValue())).getColumn().equals(colName))) {
-  bucketColNames.add(entry.getKey());
-  found = true;
-  break;
-}
-  }
-  if (!found) {
-// Bail out on first missed column.
-break;
-  }
-}
-if (!processSortCols && !found) {
-  // While processing bucket columns, atleast one bucket column
-  // missed. This results in a different bucketing scheme.
-  // Add empty list
-  listBucketCols.add(new ArrayList<>());
-} else  {
-  listBucketCols.add(bucketColNames);
+  for (List colNames : parentColNames) {
+List bucketColNames = new ArrayList<>();
+boolean found = false;
+for (String colName : colNames) {
+  // Reset found
+  found = false;
+  for (Entry entry : 
selOp.getColumnExprMap().entrySet()) {
+if ((entry.getValue() instanceof ExprNodeColumnDesc) &&
+(((ExprNodeColumnDesc) 
(entry.getValue())).getColumn().equals(colName))) {
+  bucketColNames.add(entry.getKey());
+  found = true;
+  break;
 }
   }
+  if (!found) {
+// Bail out on first missed column.
+break;
+  }
+}
+if (!processSortCols && !found) {
+  // While processing bucket columns, atleast one bucket column
+  // missed. This results in a different bucketing scheme.
+  // Add empty list
+  listBucketCols.add(new ArrayList<>());
+} else  {
+  listBucketCols.add(bucketColNames);
 }
   }
 
@@ -363,13 +359,12 @@ public class OpTraitsRulesProcFactory {
   List> listSortCols = null;
   if (selOp.getColumnExprMap() != null) {
 if (parentBucketColNames != null) {
-  listBucketCols = getConvertedColNames(parentBucketColNames, selOp);
+  listBucketCols = getConvertedColNames(parentBucketColNames, selOp, 
false);
 }
 List> parent

[hive] Git Push Summary

2018-07-12 Thread djaiswal
Repository: hive
Updated Branches:
  refs/heads/storage-branch-2.7 [deleted] 13802a2cb


hive git commit: Preparing for storage-api 2.7.0 release

2018-07-12 Thread djaiswal
Repository: hive
Updated Branches:
  refs/heads/storage-branch-2.7 [created] e59fdf986


Preparing for storage-api 2.7.0 release


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/e59fdf98
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/e59fdf98
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/e59fdf98

Branch: refs/heads/storage-branch-2.7
Commit: e59fdf9861d20747f1e69623a73782979258d61d
Parents: 8fd90ea
Author: Deepak Jaiswal 
Authored: Thu Jul 12 15:53:28 2018 -0700
Committer: Deepak Jaiswal 
Committed: Thu Jul 12 15:53:28 2018 -0700

--
 storage-api/pom.xml | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/hive/blob/e59fdf98/storage-api/pom.xml
--
diff --git a/storage-api/pom.xml b/storage-api/pom.xml
index a40feff..5d3c7d4 100644
--- a/storage-api/pom.xml
+++ b/storage-api/pom.xml
@@ -25,7 +25,7 @@
 
   org.apache.hive
   hive-storage-api
-  2.6.1-SNAPSHOT
+  2.7.0
   jar
   Hive Storage API
 



[hive] Git Push Summary

2018-07-12 Thread djaiswal
Repository: hive
Updated Tags:  refs/tags/storage-release-2.7.0-rc1 [created] 7b909acc5


[hive] Git Push Summary

2018-07-18 Thread djaiswal
Repository: hive
Updated Tags:  refs/tags/rel/storage-release-2.7.0 [created] 5d3241345


[02/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/union_remove_5.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/union_remove_5.q.out 
b/ql/src/test/results/clientpositive/spark/union_remove_5.q.out
index 92273f8..95b4de3 100644
--- a/ql/src/test/results/clientpositive/spark/union_remove_5.q.out
+++ b/ql/src/test/results/clientpositive/spark/union_remove_5.q.out
@@ -234,6 +234,7 @@ Retention:  0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
numFiles4   
totalSize   68  
  A masked pattern was here 

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/union_remove_7.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/union_remove_7.q.out 
b/ql/src/test/results/clientpositive/spark/union_remove_7.q.out
index b0f1818..ed4559b 100644
--- a/ql/src/test/results/clientpositive/spark/union_remove_7.q.out
+++ b/ql/src/test/results/clientpositive/spark/union_remove_7.q.out
@@ -158,8 +158,9 @@ Retention:  0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
numFiles4   
-   totalSize   332 
+   totalSize   336 
  A masked pattern was here 
 
 # Storage Information   

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/union_remove_8.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/union_remove_8.q.out 
b/ql/src/test/results/clientpositive/spark/union_remove_8.q.out
index 63dd553..5d51136 100644
--- a/ql/src/test/results/clientpositive/spark/union_remove_8.q.out
+++ b/ql/src/test/results/clientpositive/spark/union_remove_8.q.out
@@ -184,8 +184,9 @@ Retention:  0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
numFiles4   
-   totalSize   348 
+   totalSize   350 
  A masked pattern was here 
 
 # Storage Information   

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/union_remove_9.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/union_remove_9.q.out 
b/ql/src/test/results/clientpositive/spark/union_remove_9.q.out
index 4dc614f..2d8c21d 100644
--- a/ql/src/test/results/clientpositive/spark/union_remove_9.q.out
+++ b/ql/src/test/results/clientpositive/spark/union_remove_9.q.out
@@ -242,8 +242,9 @@ Retention:  0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
numFiles4   
-   totalSize   348 
+   totalSize   350 
  A masked pattern was here 
 
 # Storage Information   

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/vectorization_0.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/vectorization_0.q.out 
b/ql/src/test/results/clientpositive/spark/vectorization_0.q.out
index c3201bf..8b73ef2 100644
--- a/ql/src/test/results/clientpositive/spark/vectorization_0.q.out
+++ b/ql/src/test/results/clientpositive/spark/vectorization_0.q.out
@@ -1817,6 +1817,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns 
ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2
 columns.comments 
@@ -1838,6 +1839,7 @@ STAGE PLANS:
   

[15/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/vectorization_0.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/vectorization_0.q.out 
b/ql/src/test/results/clientpositive/llap/vectorization_0.q.out
index b2db5a5..80da5a9 100644
--- a/ql/src/test/results/clientpositive/llap/vectorization_0.q.out
+++ b/ql/src/test/results/clientpositive/llap/vectorization_0.q.out
@@ -1839,6 +1839,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns 
ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2
 columns.comments 
@@ -1860,6 +1861,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns 
ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2
   columns.comments 
@@ -30660,6 +30662,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns 
ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2
 columns.comments 
@@ -30681,6 +30684,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns 
ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2
   columns.comments 
@@ -30775,6 +30779,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns 
ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2
 columns.comments 
@@ -30796,6 +30801,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns 
ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2
   columns.comments 
@@ -30890,6 +30896,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}}
 bucket_count -1
+bucketing_version

[21/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/orc_create.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/orc_create.q.out 
b/ql/src/test/results/clientpositive/llap/orc_create.q.out
index 72d94e5..df286b2 100644
--- a/ql/src/test/results/clientpositive/llap/orc_create.q.out
+++ b/ql/src/test/results/clientpositive/llap/orc_create.q.out
@@ -66,6 +66,7 @@ Retention:0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"lst\":\"true\",\"mp\":\"true\",\"str\":\"true\",\"strct\":\"true\"}}
+   bucketing_version   2   
numFiles0   
numRows 0   
rawDataSize 0   
@@ -119,6 +120,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
+   bucketing_version   2   
numFiles0   
numPartitions   0   
numRows 0   
@@ -176,6 +178,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
+   bucketing_version   2   
numFiles0   
numPartitions   0   
numRows 0   
@@ -223,6 +226,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
+   bucketing_version   2   
  A masked pattern was here 
numFiles0   
numPartitions   0   
@@ -281,6 +285,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
+   bucketing_version   2   
numFiles0   
numPartitions   0   
numRows 0   
@@ -336,6 +341,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"lst\":\"true\",\"mp\":\"true\",\"str\":\"true\",\"strct\":\"true\"}}
+   bucketing_version   2   
numFiles0   
numRows 0   
rawDataSize 0   

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/orc_llap_counters.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/orc_llap_counters.q.out 
b/ql/src/test/results/clientpositive/llap/orc_llap_counters.q.out
index af237ea..98d9e06 100644
--- a/ql/src/test/results/clientpositive/llap/orc_llap_counters.q.out
+++ b/ql/src/test/results/clientpositive/llap/orc_llap_counters.q.out
@@ -227,6 +227,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
+   bucketing_version   2   
numFiles1   
numRows 2100
orc.bloom.filter.columns*   

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/orc_llap_counters1.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/orc_llap_counters1.q.out 
b/ql/src/test/results/clientpositive/llap/orc_llap_counters1.q.out
index 01fb4af..c078326 100644
--- a/ql/src/test/results/clientpositive/llap/orc_llap_counters1.q.out
+++ b/ql/src/test/results/clientpositive/llap/orc_llap_counters1.q.out
@@ -227,6 +227,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
+   bucketing_version   2   
numFiles1   
numRows 2100
orc.bloom.filter.columns*   

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/ll

[24/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/cte_1.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/cte_1.q.out 
b/ql/src/test/results/clientpositive/llap/cte_1.q.out
index d7bc062..d98862c 100644
--- a/ql/src/test/results/clientpositive/llap/cte_1.q.out
+++ b/ql/src/test/results/clientpositive/llap/cte_1.q.out
@@ -159,178 +159,217 @@ POSTHOOK: Input: default@src
 10 val_10  val_10
 100val_100 val_100
 100val_100 val_100
+103val_103 val_103
+103val_103 val_103
 104val_104 val_104
 104val_104 val_104
-111val_111 val_111
-114val_114 val_114
-116val_116 val_116
+105val_105 val_105
+11 val_11  val_11
 126val_126 val_126
 128val_128 val_128
 128val_128 val_128
 128val_128 val_128
-131val_131 val_131
+129val_129 val_129
+129val_129 val_129
+133val_133 val_133
 134val_134 val_134
 134val_134 val_134
+136val_136 val_136
+143val_143 val_143
+149val_149 val_149
+149val_149 val_149
 152val_152 val_152
 152val_152 val_152
-155val_155 val_155
+153val_153 val_153
+157val_157 val_157
+162val_162 val_162
+166val_166 val_166
 169val_169 val_169
 169val_169 val_169
 169val_169 val_169
 169val_169 val_169
-17 val_17  val_17
 170val_170 val_170
-172val_172 val_172
-172val_172 val_172
-178val_178 val_178
-179val_179 val_179
-179val_179 val_179
-189val_189 val_189
-19 val_19  val_19
-192val_192 val_192
-194val_194 val_194
-195val_195 val_195
-195val_195 val_195
-199val_199 val_199
-199val_199 val_199
-199val_199 val_199
-20 val_20  val_20
-207val_207 val_207
-207val_207 val_207
+174val_174 val_174
+174val_174 val_174
+176val_176 val_176
+176val_176 val_176
+193val_193 val_193
+193val_193 val_193
+193val_193 val_193
+196val_196 val_196
+201val_201 val_201
+205val_205 val_205
+205val_205 val_205
+209val_209 val_209
+209val_209 val_209
 213val_213 val_213
 213val_213 val_213
 218val_218 val_218
+221val_221 val_221
+221val_221 val_221
 223val_223 val_223
 223val_223 val_223
-229val_229 val_229
-229val_229 val_229
 233val_233 val_233
 233val_233 val_233
-235val_235 val_235
+239val_239 val_239
+239val_239 val_239
 24 val_24  val_24
 24 val_24  val_24
-256val_256 val_256
-256val_256 val_256
-257val_257 val_257
+241val_241 val_241
+242val_242 val_242
+242val_242 val_242
+247val_247 val_247
+248val_248 val_248
 26 val_26  val_26
 26 val_26  val_26
-265val_265 val_265
-265val_265 val_265
-266val_266 val_266
-27 val_27  val_27
-278val_278 val_278
-278val_278 val_278
-280val_280 val_280
-280val_280 val_280
-284val_284 val_284
-291val_291 val_291
-30 val_30  val_30
+263val_263 val_263
+272val_272 val_272
+272val_272 val_272
+275val_275 val_275
+285val_285 val_285
+296val_296 val_296
+298val_298 val_298
+298val_298 val_298
+298val_298 val_298
 302val_302 val_302
-305val_305 val_305
-316val_316 val_316
-316val_316 val_316
-316val_316 val_316
-321val_321 val_321
-321val_321 val_321
+307val_307 val_307
+307val_307 val_307
+310val_310 val_310
 323val_323 val_323
-33 val_33  val_33
+325val_325 val_325
+325val_325 val_325
 333val_333 val_333
 333val_333 val_333
+335val_335 val_335
 338val_338 val_338
-344val_344 val_344
-344val_344 val_344
-353val_353 val_353
-353val_353 val_353
-356val_356 val_356
-364val_364 val_364
-368val_368 val_368
-369val_369 val_369
-369val_369 val_369
-369val_369 val_369
-37 val_37  val_37
-37 val_37  val_37
+360val_360 val_360
+362val_362 val_362
+367val_367 val_367
+367val_367 val_367
+375val_375 val_375
 377val_377 val_377
-378val_378 val_378
-386val_386 val_386
+384val_384 val_384
+384val_384 val_384
+384val_384 val_384
 392val_392 val_392
-397val_397 val_397
-397val_397 val_397
-404val_404 val_404
-404val_404 val_404
-407val_407 val_407
-411val_411 val_411
-418val_418 val_418
+399val_399 val_399
+399val_399 val_399
+401val_401 val_401
+401val_401 val_401
+401val_401 val_401
+401val_401 val_401
+401val_401 val_401
+402val_402 val_402
+403val_403 val_403
+403val_403 val_403
+403val_403 val_403
+413val_413 val_413
+413val_413 val_413
+417val_417 val_417
+417val_417 val_417
+417val_417 val_417
 419val_419 val_419
-424val_424 val_424
-424val_424 val_424
-430val_430 val_430
-430val_430 val_430
-430val_430 val_430
-432val_432 val_432
-436val_436 val_436
-

[13/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/sample1.q.out
--
diff --git a/ql/src/test/results/clientpositive/sample1.q.out 
b/ql/src/test/results/clientpositive/sample1.q.out
index 6d58658..09b9ac7 100644
--- a/ql/src/test/results/clientpositive/sample1.q.out
+++ b/ql/src/test/results/clientpositive/sample1.q.out
@@ -55,6 +55,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value,dt,hr
 columns.comments 
@@ -126,6 +127,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 'default','default'
@@ -193,6 +195,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value,dt,hr
 columns.comments 
@@ -236,6 +239,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value,dt,hr
 columns.comments 
@@ -266,6 +270,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value,dt,hr
   columns.comments 
@@ -287,6 +292,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value,dt,hr
 columns.comments 
@@ -323,6 +329,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value,dt,hr
 columns.comments 
@@ -353,6 +360,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value,dt,hr
   columns.comments 
@@ -374,6 +382,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value,dt,hr
 columns.comments 

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/sample2.q.out
--
diff --git a/ql/src/test/results/clientpositive/sample2.q.out 
b/ql/src/test/results/clientpositive/sample2.q.out
index 164ce43..53c3df2 100644
--- a/ql/src/test/results/clientpositive/sample2.q.out
+++ b/ql/src/test/results/clientpositive/sample2.q.out
@@ -54,6 +54,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -101,6 +102,7 @@ STAGE PLANS:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name

[30/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out
--
diff --git a/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out 
b/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out
index 4d8dd74..11ef8b1 100644
--- a/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out
+++ b/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out
@@ -87,6 +87,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,cnt
   columns.comments 
@@ -135,6 +136,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,val
   columns.comments 
@@ -158,6 +160,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,val
 columns.comments 
@@ -227,6 +230,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -270,6 +274,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -300,6 +305,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,cnt
   columns.comments 
@@ -321,6 +327,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -357,6 +364,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -387,6 +395,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,cnt
   columns.comments 
@@ -408,6 +417,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -522,6 +532,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,val
   columns.comments 
@@ -545,6 +556,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,val
 columns.comments 
@@ -589,6 +601,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key1,key2,cnt
 columns.comments 
@@ -647,6 +660,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","

[17/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/vector_ptf_part_simple.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/llap/vector_ptf_part_simple.q.out 
b/ql/src/test/results/clientpositive/llap/vector_ptf_part_simple.q.out
index 9aeb650..ec8611d 100644
--- a/ql/src/test/results/clientpositive/llap/vector_ptf_part_simple.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_ptf_part_simple.q.out
@@ -309,46 +309,46 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@vector_ptf_part_simple_orc
  A masked pattern was here 
 p_mfgr p_name  p_retailprice   rn  r   dr  fv  lv  c   
cs
-Manufacturer#4 almond azure aquamarine papaya violet   1290.35 1   1   
1   1290.35 1206.26 4   6
-Manufacturer#4 almond antique violet mint lemon1375.42 2   1   
1   1290.35 1206.26 4   6
-Manufacturer#4 almond aquamarine floral ivory bisque   NULL3   1   
1   1290.35 1206.26 4   6
-Manufacturer#4 almond antique gainsboro frosted violet NULL4   1   
1   1290.35 1206.26 4   6
-Manufacturer#4 almond aquamarine yellow dodger mint1844.92 5   1   
1   1290.35 1206.26 4   6
-Manufacturer#4 almond aquamarine floral ivory bisque   1206.26 6   1   
1   1290.35 1206.26 4   6
-Manufacturer#5 almond azure blanched chiffon midnight  1464.48 1   1   
1   1464.48 1788.73 6   6
-Manufacturer#5 almond aquamarine dodger light gainsboro1018.1  2   
1   1   1464.48 1788.73 6   6
-Manufacturer#5 almond antique medium spring khaki  1611.66 3   1   
1   1464.48 1788.73 6   6
-Manufacturer#5 almond antique blue firebrick mint  1789.69 4   1   
1   1464.48 1788.73 6   6
-Manufacturer#5 almond antique medium spring khaki  1611.66 5   1   
1   1464.48 1788.73 6   6
-Manufacturer#5 almond antique sky peru orange  1788.73 6   1   1   
1464.48 1788.73 6   6
-Manufacturer#2 almond aquamarine rose maroon antique   900.66  1   1   
1   900.66  1800.7  8   8
-Manufacturer#2 almond aquamarine rose maroon antique   1698.66 2   1   
1   900.66  1800.7  8   8
-Manufacturer#2 almond antique violet turquoise frosted 1800.7  3   1   
1   900.66  1800.7  8   8
-Manufacturer#2 almond antique violet chocolate turquoise   1690.68 4   
1   1   900.66  1800.7  8   8
-Manufacturer#2 almond antique violet turquoise frosted 1800.7  5   1   
1   900.66  1800.7  8   8
-Manufacturer#2 almond aquamarine sandy cyan gainsboro  1000.6  6   1   
1   900.66  1800.7  8   8
-Manufacturer#2 almond aquamarine midnight light salmon 2031.98 7   1   
1   900.66  1800.7  8   8
-Manufacturer#2 almond antique violet turquoise frosted 1800.7  8   1   
1   900.66  1800.7  8   8
-Manufacturer#1 almond antique chartreuse lavender yellow   1753.76 1   
1   1   1753.76 1632.66 11  12
-Manufacturer#1 almond aquamarine pink moccasin thistle 1632.66 2   1   
1   1753.76 1632.66 11  12
-Manufacturer#1 almond aquamarine pink moccasin thistle 1632.66 3   1   
1   1753.76 1632.66 11  12
-Manufacturer#1 almond antique chartreuse lavender yellow   1753.76 4   
1   1   1753.76 1632.66 11  12
-Manufacturer#1 almond aquamarine burnished black steel 1414.42 5   1   
1   1753.76 1632.66 11  12
-Manufacturer#1 almond antique burnished rose metallic  1173.15 6   1   
1   1753.76 1632.66 11  12
-Manufacturer#1 almond antique salmon chartreuse burlywood  1602.59 7   
1   1   1753.76 1632.66 11  12
-Manufacturer#1 almond antique chartreuse lavender yellow   1753.76 8   
1   1   1753.76 1632.66 11  12
-Manufacturer#1 almond antique burnished rose metallic  1173.15 9   1   
1   1753.76 1632.66 11  12
-Manufacturer#1 almond antique chartreuse lavender yellow   1753.76 10  
1   1   1753.76 1632.66 11  12
-Manufacturer#1 almond aquamarine pink moccasin thistle NULL11  1   
1   1753.76 1632.66 11  12
-Manufacturer#1 almond aquamarine pink moccasin thistle 1632.66 12  1   
1   1753.76 1632.66 11  12
-Manufacturer#3 almond antique forest lavender goldenrod590.27  1   
1   1   590.27  99.68   7   8
-Manufacturer#3 almond antique metallic orange dim  55.39   2   1   
1   590.27  99.68   7   8
-Manufacturer#3 almond antique forest lavender goldenrod1190.27 3   
1   1   590.27  99.68   7   8
-Manufacturer#3 almond antique olive coral navajo   1337.29 4   1   
1   590.27  99.68   7   8
-Manufacturer#3 almond antique m

[27/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out 
b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out
index 885fd18..8fa1b35 100644
--- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out
+++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out
@@ -147,6 +147,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 4
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -234,6 +235,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -282,6 +284,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -421,6 +424,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 4
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -508,6 +512,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -556,6 +561,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out 
b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out
index ff98ee5..bd7f60a 100644
--- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out
+++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out
@@ -147,6 +147,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -195,6 +196,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -283,6 +285,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 4
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -421,6 +424,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -469,6 +473,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -557,6 +562,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 4
   bucket_field_name key
+  bucketing_version 2
   colu

[11/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/smb_mapjoin_46.q.out
--
diff --git a/ql/src/test/results/clientpositive/smb_mapjoin_46.q.out 
b/ql/src/test/results/clientpositive/smb_mapjoin_46.q.out
index 383ab31..f259c06 100644
--- a/ql/src/test/results/clientpositive/smb_mapjoin_46.q.out
+++ b/ql/src/test/results/clientpositive/smb_mapjoin_46.q.out
@@ -143,11 +143,11 @@ POSTHOOK: Input: default@test2
  A masked pattern was here 
 98 NULLNoneNULLNULLNULL
 NULL   NULLNoneNULLNULLNULL
-99 0   Alice   NULLNULLNULL
 1012   Car 103 2   Ema
 1012   Car 102 2   Del
 99 2   Mat 103 2   Ema
 99 2   Mat 102 2   Del
+99 0   Alice   NULLNULLNULL
 1001   Bob NULLNULLNULL
 PREHOOK: query: EXPLAIN
 SELECT *
@@ -221,9 +221,9 @@ POSTHOOK: Input: default@test2
  A masked pattern was here 
 98 NULLNoneNULLNULLNULL
 NULL   NULLNoneNULLNULLNULL
-99 0   Alice   NULLNULLNULL
 1012   Car 102 2   Del
 99 2   Mat NULLNULLNULL
+99 0   Alice   NULLNULLNULL
 1001   Bob NULLNULLNULL
 Warning: Map Join MAPJOIN[11][bigTable=?] in task 'Stage-3:MAPRED' is a cross 
product
 PREHOOK: query: EXPLAIN
@@ -327,9 +327,9 @@ POSTHOOK: Input: default@test2
  A masked pattern was here 
 98 NULLNoneNULLNULLNULL
 NULL   NULLNoneNULLNULLNULL
-99 0   Alice   NULLNULLNULL
 1012   Car 102 2   Del
 99 2   Mat NULLNULLNULL
+99 0   Alice   NULLNULLNULL
 1001   Bob 102 2   Del
 PREHOOK: query: EXPLAIN
 SELECT *
@@ -491,12 +491,12 @@ POSTHOOK: Input: default@test2
  A masked pattern was here 
 98 NULLNoneNULLNULLNULL
 NULL   NULLNoneNULLNULLNULL
-99 0   Alice   NULLNULLNULL
 1012   Car 105 NULLNone
 1012   Car 103 2   Ema
 1012   Car 102 2   Del
 1012   Car 104 3   Fli
 99 2   Mat NULLNULLNULL
+99 0   Alice   NULLNULLNULL
 1001   Bob 105 NULLNone
 1001   Bob 103 2   Ema
 1001   Bob 102 2   Del
@@ -598,13 +598,13 @@ POSTHOOK: Input: default@test2
  A masked pattern was here 
 98 NULLNone102 2   Del
 NULL   NULLNone102 2   Del
-99 0   Alice   102 2   Del
 1012   Car 105 NULLNone
 1012   Car 103 2   Ema
 1012   Car 102 2   Del
 1012   Car 104 3   Fli
 99 2   Mat 103 2   Ema
 99 2   Mat 102 2   Del
+99 0   Alice   102 2   Del
 1001   Bob 105 NULLNone
 1001   Bob 103 2   Ema
 1001   Bob 102 2   Del
@@ -702,13 +702,13 @@ POSTHOOK: Input: default@test2
  A masked pattern was here 
 98 NULLNoneNULLNULLNULL
 NULL   NULLNoneNULLNULLNULL
-99 0   Alice   NULLNULLNULL
 1012   Car 105 NULLNone
 1012   Car 103 2   Ema
 1012   Car 102 2   Del
 1012   Car 104 3   Fli
 99 2   Mat 103 2   Ema
 99 2   Mat 102 2   Del
+99 0   Alice   NULLNULLNULL
 1001   Bob 105 NULLNone
 1001   Bob 103 2   Ema
 1001   Bob 102 2   Del
@@ -806,11 +806,11 @@ POSTHOOK: Input: default@test2
  A masked pattern was here 
 98 NULLNone102 2   Del
 NULL   NULLNone102 2   Del
-99 0   Alice   102 2   Del
 1012   Car 103 2   Ema
 1012   Car 102 2   Del
 99 2   Mat 103 2   Ema
 99 2   Mat 102 2   Del
+99 0   Alice   102 2   Del
 1001   Bob 102 2   Del
 PREHOOK: query: EXPLAIN
 SELECT *
@@ -882,10 +882,10 @@ POSTHOOK: Input: default@test2
  A masked pattern was here 
 98 NULLNoneNULLNULLNULL
 NULL   NULLNoneNULLNULLNULL
-99 0   Alice   NULLNULLNULL
 1012   Car 103 2   Ema
 1012   Car 102 2   Del
 99 2   Mat 102 2   Del
+99 0   Alice   NULLNULLNULL
 1001   Bob NULLNULLNULL
 Warning: Map Join MAPJOIN[9][bigTable=?] in task 'Stage-3:MAPRED' is a cross 
product
 PREHOOK: query: EXPLAIN
@@ -989,9 +989,9 @

[31/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/druid_basic1.q.out
--
diff --git a/ql/src/test/results/clientpositive/druid_basic1.q.out 
b/ql/src/test/results/clientpositive/druid_basic1.q.out
index 25ec0ca..a5c0687 100644
--- a/ql/src/test/results/clientpositive/druid_basic1.q.out
+++ b/ql/src/test/results/clientpositive/druid_basic1.q.out
@@ -41,6 +41,7 @@ Table Type:   EXTERNAL_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"__time\":\"true\",\"added\":\"true\",\"anonymous\":\"true\",\"count\":\"true\",\"deleted\":\"true\",\"delta\":\"true\",\"language\":\"true\",\"namespace\":\"true\",\"newpage\":\"true\",\"page\":\"true\",\"robot\":\"true\",\"unpatrolled\":\"true\",\"user\":\"true\",\"variation\":\"true\"}}
EXTERNALTRUE
+   bucketing_version   2   
druid.datasourcewikipedia   
numFiles0   
numRows 0   
@@ -102,6 +103,7 @@ Table Type: EXTERNAL_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"__time\":\"true\",\"added\":\"true\",\"anonymous\":\"true\",\"count\":\"true\",\"deleted\":\"true\",\"delta\":\"true\",\"language\":\"true\",\"namespace\":\"true\",\"newpage\":\"true\",\"page\":\"true\",\"robot\":\"true\",\"unpatrolled\":\"true\",\"user\":\"true\",\"variation\":\"true\"}}
EXTERNALTRUE
+   bucketing_version   2   
druid.datasourcewikipedia   
numFiles0   
numRows 0   

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/druid_basic2.q.out
--
diff --git a/ql/src/test/results/clientpositive/druid_basic2.q.out 
b/ql/src/test/results/clientpositive/druid_basic2.q.out
index 5ed039b..0e5e957 100644
--- a/ql/src/test/results/clientpositive/druid_basic2.q.out
+++ b/ql/src/test/results/clientpositive/druid_basic2.q.out
@@ -41,6 +41,7 @@ Table Type:   EXTERNAL_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"__time\":\"true\",\"added\":\"true\",\"anonymous\":\"true\",\"count\":\"true\",\"deleted\":\"true\",\"delta\":\"true\",\"language\":\"true\",\"namespace\":\"true\",\"newpage\":\"true\",\"page\":\"true\",\"robot\":\"true\",\"unpatrolled\":\"true\",\"user\":\"true\",\"variation\":\"true\"}}
EXTERNALTRUE
+   bucketing_version   2   
druid.datasourcewikipedia   
numFiles0   
numRows 0   
@@ -266,6 +267,7 @@ STAGE PLANS:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"__time":"true","added":"true","anonymous":"true","count":"true","deleted":"true","delta":"true","language":"true","namespace":"true","newpage":"true","page":"true","robot":"true","unpatrolled":"true","user":"true","variation":"true"}}
   EXTERNAL TRUE
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns 
__time,robot,namespace,anonymous,unpatrolled,page,language,newpage,user,count,added,delta,variation,deleted
   columns.comments 'from deserializer','from deserializer','from 
deserializer','from deserializer','from deserializer','from deserializer','from 
deserializer','from deserializer','from deserializer','from deserializer','from 
deserializer','from deserializer','from deserializer','from deserializer'
@@ -294,6 +296,7 @@ STAGE PLANS:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"__time":"true","added":"true","anonymous":"true","count":"true","deleted":"true","delta":"true","language":"true","namespace":"true","newpage":"true","page":"true","robot":"true","unpatrolled":"true","user":"true","variation":"true"}}
 EXTERNAL TRUE
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns 
__time,robot,namespace,anonymous,unpatrolled,page,language,newpage,user,count,added,delta,variation,deleted
 columns.comments 'from deserializer','from deserializer','from 
deserializer','from deserializer','from deserializer','from deserializer','from 
deserializer','from deserializer','from deserializer','from deserializer','from 
deserializer','from deserializer','from deserializer','from deserializer'
@@ -444,6 +447,7 @@ STAGE PLANS:

[36/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, 
reviewed by Jasone Dere)


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/ac6b2a3f
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/ac6b2a3f
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/ac6b2a3f

Branch: refs/heads/master
Commit: ac6b2a3fb195916e22b2e5f465add2ffbcdc7430
Parents: a095c3d
Author: Deepak Jaiswal 
Authored: Sun Apr 29 22:24:15 2018 -0700
Committer: Deepak Jaiswal 
Committed: Sun Apr 29 22:24:15 2018 -0700

--
 .../results/positive/external_table_ppd.q.out   | 1 +
 .../positive/hbase_binary_storage_queries.q.out | 2 +
 .../src/test/results/positive/hbase_ddl.q.out   | 2 +
 .../src/test/results/positive/hbasestats.q.out  | 5 +
 .../streaming/AbstractRecordWriter.java | 9 +-
 .../hive/hcatalog/streaming/TestStreaming.java  |10 +-
 .../mutate/worker/TestBucketIdResolverImpl.java | 2 +-
 .../org/apache/hive/hcatalog/api/HCatTable.java | 4 +
 .../hive/hcatalog/api/TestHCatClient.java   | 2 +
 .../insert_into_dynamic_partitions.q.out| 4 +
 .../clientpositive/insert_into_table.q.out  |11 +
 .../insert_overwrite_directory.q.out| 4 +
 .../insert_overwrite_dynamic_partitions.q.out   | 4 +
 .../clientpositive/insert_overwrite_table.q.out |11 +
 .../write_final_output_blobstore.q.out  | 8 +
 .../hive/ql/txn/compactor/TestCompactor.java| 2 +-
 .../test/resources/testconfiguration.properties | 6 +-
 .../hadoop/hive/ql/exec/FileSinkOperator.java   |16 +-
 .../hadoop/hive/ql/exec/FunctionRegistry.java   | 1 +
 .../apache/hadoop/hive/ql/exec/Operator.java| 9 +
 .../hadoop/hive/ql/exec/OperatorFactory.java| 3 +
 .../hadoop/hive/ql/exec/ReduceSinkOperator.java |44 +-
 .../apache/hadoop/hive/ql/exec/Utilities.java   |13 +
 .../bootstrap/load/table/LoadPartitions.java| 3 +-
 .../VectorKeySeriesSerializedImpl.java  | 4 +-
 .../VectorReduceSinkObjectHashOperator.java |   135 +-
 .../hadoop/hive/ql/io/HiveFileFormatUtils.java  |11 +-
 .../hadoop/hive/ql/io/orc/OrcInputFormat.java   | 8 +-
 .../apache/hadoop/hive/ql/metadata/Table.java   |11 +
 .../hive/ql/optimizer/ConvertJoinMapJoin.java   |63 +-
 .../optimizer/FixedBucketPruningOptimizer.java  |18 +-
 .../ql/optimizer/PrunerOperatorFactory.java | 3 +-
 .../optimizer/SortedDynPartitionOptimizer.java  | 6 +-
 ...tedDynPartitionTimeGranularityOptimizer.java | 9 +-
 .../annotation/OpTraitsRulesProcFactory.java|37 +-
 .../hive/ql/optimizer/physical/Vectorizer.java  |16 +-
 .../optimizer/spark/SparkMapJoinOptimizer.java  | 3 +-
 .../hadoop/hive/ql/parse/SemanticAnalyzer.java  |33 +-
 .../apache/hadoop/hive/ql/plan/OpTraits.java|24 +-
 .../apache/hadoop/hive/ql/plan/PlanUtils.java   | 2 +-
 .../hadoop/hive/ql/plan/ReduceSinkDesc.java |10 +-
 .../apache/hadoop/hive/ql/plan/TableDesc.java   | 5 +
 .../hive/ql/plan/VectorReduceSinkDesc.java  | 9 +
 .../hive/ql/udf/generic/GenericUDFHash.java | 3 +-
 .../ql/udf/generic/GenericUDFMurmurHash.java|64 +
 .../hadoop/hive/ql/TestTxnAddPartition.java |11 +-
 .../apache/hadoop/hive/ql/TestTxnCommands.java  |20 +-
 .../apache/hadoop/hive/ql/TestTxnCommands2.java |63 +-
 .../apache/hadoop/hive/ql/TestTxnNoBuckets.java |14 +-
 .../parse/authorization/TestPrivilegesV1.java   | 2 +-
 .../clientpositive/archive_excludeHadoop20.q| 4 +-
 .../test/queries/clientpositive/bucket_many.q   | 4 +-
 .../clientpositive/bucket_map_join_tez1.q   | 2 +-
 .../clientpositive/bucket_num_reducers.q| 5 +-
 .../clientpositive/bucket_num_reducers2.q   | 6 +-
 .../clientpositive/bucket_num_reducers_acid2.q  | 3 -
 .../clientpositive/insert_update_delete.q   | 1 -
 ql/src/test/queries/clientpositive/sample10.q   | 4 +-
 .../clientpositive/tez_dynpart_hashjoin_3.q | 4 +-
 ql/src/test/queries/clientpositive/tez_smb_1.q  | 2 +-
 .../clientnegative/alter_file_format.q.out  | 1 +
 .../alter_view_as_select_with_partition.q.out   | 1 +
 .../clientnegative/unset_table_property.q.out   | 1 +
 .../clientnegative/unset_view_property.q.out| 1 +
 .../results/clientpositive/acid_mapjoin.q.out   |20 +-
 .../results/clientpositive/acid_nullscan.q.out  | 2 +
 .../clientpositive/acid_table_stats.q.out   | 2 +
 .../clientpositive/alterColumnStats.q.out   | 3 +
 .../clientpositive/alterColumnStatsPart.q.out   | 1 +
 .../clientpositive/alter_file_format.q.out  | 7 +
 ...lter_numbuckets_partitioned_table2_h23.q.out | 9 +
 ...alter_numbuckets_partitioned_table_h23.q.out | 8 +
 .../alter_partition_clusterby_sortby.q.out  |  

[03/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/subquery_notin.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/subquery_notin.q.out 
b/ql/src/test/results/clientpositive/spark/subquery_notin.q.out
index ffa524e..0f2f717 100644
--- a/ql/src/test/results/clientpositive/spark/subquery_notin.q.out
+++ b/ql/src/test/results/clientpositive/spark/subquery_notin.q.out
@@ -1213,25 +1213,25 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@part
  A masked pattern was here 
 Manufacturer#1 almond aquamarine burnished black steel 28
-Manufacturer#1 almond antique chartreuse lavender yellow   34
 Manufacturer#1 almond aquamarine pink moccasin thistle 42
-Manufacturer#2 almond aquamarine rose maroon antique   25
+Manufacturer#2 almond antique violet chocolate turquoise   14
+Manufacturer#2 almond aquamarine sandy cyan gainsboro  18
 Manufacturer#2 almond antique violet turquoise frosted 40
-Manufacturer#3 almond antique chartreuse khaki white   17
 Manufacturer#3 almond antique metallic orange dim  19
+Manufacturer#4 almond aquamarine floral ivory bisque   27
+Manufacturer#4 almond antique violet mint lemon39
 Manufacturer#5 almond antique medium spring khaki  6
 Manufacturer#5 almond azure blanched chiffon midnight  23
 Manufacturer#5 almond antique blue firebrick mint  31
 Manufacturer#5 almond aquamarine dodger light gainsboro46
 Manufacturer#1 almond antique salmon chartreuse burlywood  6
-Manufacturer#2 almond antique violet chocolate turquoise   14
-Manufacturer#2 almond aquamarine sandy cyan gainsboro  18
+Manufacturer#1 almond antique chartreuse lavender yellow   34
+Manufacturer#2 almond aquamarine rose maroon antique   25
 Manufacturer#3 almond antique forest lavender goldenrod14
+Manufacturer#3 almond antique chartreuse khaki white   17
 Manufacturer#3 almond antique olive coral navajo   45
 Manufacturer#4 almond antique gainsboro frosted violet 10
 Manufacturer#4 almond azure aquamarine papaya violet   12
-Manufacturer#4 almond aquamarine floral ivory bisque   27
-Manufacturer#4 almond antique violet mint lemon39
 Warning: Shuffle Join JOIN[19][tables = [$hdt$_0, $hdt$_1]] in Work 'Reducer 
2' is a cross product
 PREHOOK: query: select li.l_partkey, count(*)
 from lineitem li
@@ -1250,21 +1250,21 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@lineitem
  A masked pattern was here 
 22630  1
-7068   1
-88362  1
-123076 1
-139636 1
-106170 1
-182052 1
-59694  1
 175839 1
-21636  1
 119767 1
+139636 1
+21636  1
 119477 1
-88035  1
-4501
-61931  1
+88362  1
+106170 1
+59694  1
+7068   1
+123076 1
 85951  1
+61931  1
+4501
+182052 1
+88035  1
 Warning: Shuffle Join JOIN[18][tables = [$hdt$_0, $hdt$_1]] in Work 'Reducer 
2' is a cross product
 PREHOOK: query: select * 
 from src 
@@ -1789,29 +1789,29 @@ POSTHOOK: query: select * from part where p_brand <> 
'Brand#14' AND p_size NOT I
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@part
  A masked pattern was here 
-17927  almond aquamarine yellow dodger mintManufacturer#4  Brand#41
ECONOMY BRUSHED COPPER  7   SM PKG  1844.92 ites. eve
 191709 almond antique violet turquoise frosted Manufacturer#2  Brand#22
ECONOMY POLISHED STEEL  40  MED BOX 1800.7   haggle
-78486  almond azure blanched chiffon midnight  Manufacturer#5  Brand#52
LARGE BRUSHED BRASS 23  MED BAG 1464.48 hely blith
+85768  almond antique chartreuse lavender yellow   Manufacturer#1  
Brand#12LARGE BRUSHED STEEL 34  SM BAG  1753.76 refull
+65667  almond aquamarine pink moccasin thistle Manufacturer#1  Brand#12
LARGE BURNISHED STEEL   42  JUMBO CASE  1632.66 e across the expr
 105685 almond antique violet chocolate turquoise   Manufacturer#2  
Brand#22MEDIUM ANODIZED COPPER  14  MED CAN 1690.68 ly pending requ
 112398 almond antique metallic orange dim  Manufacturer#3  Brand#32
MEDIUM BURNISHED BRASS  19  JUMBO JAR   1410.39 ole car
 146985 almond aquamarine midnight light salmon Manufacturer#2  Brand#23
MEDIUM BURNISHED COPPER 2   SM CASE 2031.98 s cajole caref
-90681  almond antique chartreuse khaki white   Manufacturer#3  Brand#31
MEDIUM BURNISHED TIN17  SM CASE 1671.68 are slyly after the sl
 110592 almond antique salmon chartreuse burlywood  Manufacturer#1  
Brand#15PROMO BURNISHED NICKEL  6   JUMBO PKG   1602.59  to the 
furiously
+48427  almond antique violet mint lemonManufacturer#4  Brand#42
PROMO POLISHED STEEL39  SM CASE 1375.42 hely ironic i
 49671  almond antique gainsboro frosted violet Manufacturer#4  Brand#41
SMALL BRUSHED BRASS 10  SM BOX  1620.67 ccounts run quick
-33357  almond azure aquamarine papaya violet   Manufacturer#4  Brand#41
STANDAR

[33/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/bucket1.q.out
--
diff --git a/ql/src/test/results/clientpositive/bucket1.q.out 
b/ql/src/test/results/clientpositive/bucket1.q.out
index 3060df3..07b5b9d 100644
--- a/ql/src/test/results/clientpositive/bucket1.q.out
+++ b/ql/src/test/results/clientpositive/bucket1.q.out
@@ -52,6 +52,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -73,6 +74,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 'default','default'
@@ -112,6 +114,7 @@ STAGE PLANS:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
   bucket_count 100
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -171,6 +174,7 @@ STAGE PLANS:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count 100
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/bucket2.q.out
--
diff --git a/ql/src/test/results/clientpositive/bucket2.q.out 
b/ql/src/test/results/clientpositive/bucket2.q.out
index 16218f9..e8a1dd7 100644
--- a/ql/src/test/results/clientpositive/bucket2.q.out
+++ b/ql/src/test/results/clientpositive/bucket2.q.out
@@ -51,6 +51,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -72,6 +73,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 'default','default'
@@ -111,6 +113,7 @@ STAGE PLANS:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -178,6 +181,7 @@ STAGE PLANS:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -264,66 +268,59 @@ POSTHOOK: query: select * from bucket2_1 tablesample 
(bucket 1 out of 2) s
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@bucket2_1
  A masked pattern was here 
-0  val_0
-0  val_0
-0  val_0
 10 val_10
-100val_100
-100val_100
-104val_104
-104val_104
-114val_114
+105val_105
+113val_113
+113val_113
 116val_116
-118val_118
-118val_118
-12 val_12
-12 val_12
-120val_120
-120val_120
+119val_119
+119val_119
+119val_119
 126val_126
-128val_128
-128val_128
-128val_128
+133val_133
 134val_134
 134val_134
-136val_136
-138val_138
-138val_138
-138val_138
-138val_138
+137val_137
+137val_137
 146val_146
 146val_146
-150val_150
-152val_152
-152val_152
+149val_149
+149val_149
+153val_153
 156val_156
+157val_157
 158val_158
-160val_160
 162val_162
+163val_163
 164val_164
 164val_164
-166val_166
-168val_168
-170val_170
-172val_172
-172val_172
-174val_174
-174val_174
-176val_176
-176val_176
+165val_165
+165val_165
+169val_169
+169val_169
+

[32/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out 
b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out
index b59c4bc..49598d1 100644
--- a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out
+++ b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out
@@ -128,7 +128,7 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table2
 POSTHOOK: Input: default@test_table2@ds=1
  A masked pattern was here 
-247
+243
 PREHOOK: query: select count(*) from test_table2 tablesample (bucket 2 out of 
2) s where ds = '1'
 PREHOOK: type: QUERY
 PREHOOK: Input: default@test_table2
@@ -139,7 +139,7 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table2
 POSTHOOK: Input: default@test_table2@ds=1
  A masked pattern was here 
-253
+257
 PREHOOK: query: EXPLAIN
 INSERT OVERWRITE TABLE test_table2 PARTITION (ds = '1')
 SELECT * from 
@@ -238,7 +238,7 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table2
 POSTHOOK: Input: default@test_table2@ds=1
  A masked pattern was here 
-247
+243
 PREHOOK: query: select count(*) from test_table2 tablesample (bucket 2 out of 
2) s where ds = '1'
 PREHOOK: type: QUERY
 PREHOOK: Input: default@test_table2
@@ -249,7 +249,7 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table2
 POSTHOOK: Input: default@test_table2@ds=1
  A masked pattern was here 
-253
+257
 PREHOOK: query: EXPLAIN
 INSERT OVERWRITE TABLE test_table2 PARTITION (ds = '1')
 SELECT x.key, concat(x.value, x.value) from 

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out 
b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out
index 52ef3db..d1c11f4 100644
--- a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out
+++ b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out
@@ -120,7 +120,7 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table2
 POSTHOOK: Input: default@test_table2@ds=1
  A masked pattern was here 
-247
+243
 PREHOOK: query: select count(*) from test_table2 tablesample (bucket 2 out of 
2) s where ds = '1'
 PREHOOK: type: QUERY
 PREHOOK: Input: default@test_table2
@@ -131,7 +131,7 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table2
 POSTHOOK: Input: default@test_table2@ds=1
  A masked pattern was here 
-253
+257
 PREHOOK: query: CREATE TABLE test_table3 (key INT, value STRING) PARTITIONED 
BY (ds STRING)
 CLUSTERED BY (value) SORTED BY (value) INTO 2 BUCKETS
 PREHOOK: type: CREATETABLE

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out 
b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out
index eaf85c3..72e918f 100644
--- a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out
+++ b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out
@@ -199,18 +199,7 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table3
 POSTHOOK: Input: default@test_table3@ds=1
  A masked pattern was here 
-0  0   val_0val_0  1
-0  0   val_0val_0  1
-0  0   val_0val_0  1
-0  0   val_0val_0  1
-0  0   val_0val_0  1
-0  0   val_0val_0  1
-0  0   val_0val_0  1
-0  0   val_0val_0  1
-0  0   val_0val_0  1
 2  2   val_2val_2  1
-4  4   val_4val_4  1
-8  8   val_8val_8  1
 PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s 
where ds = '1'
 PREHOOK: type: QUERY
 PREHOOK: Input: default@test_table3
@@ -221,6 +210,16 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table3
 POSTHOOK: Input: default@test_table3@ds=1
  A masked pattern was here 
+0  0   val_0val_0  1
+0  0   val_0val_0  1
+0  0   val_0val_0  1
+0  0   val_0val_0  1
+0  0   val_0val_0  1
+0  0   val_0val_0  1
+0  0   val_0val_0  1
+0  0   val_0val_0  1
+0  0   val_0val_0  1
+4  4   val_4val_4  1
 5  5   val_5val_5  1
 5  5   val_5val_5  1
 5  5   val_5val_5  1
@@ -230,6 +229,7 @@ POSTHOOK: Input: default@test_table3@ds=1
 5  5   val_5val_5  1
 5  5   val_5val_5  1
 5  5   val_5val_5  1
+8  8 

[09/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/bucketsortoptimize_insert_2.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/spark/bucketsortoptimize_insert_2.q.out 
b/ql/src/test/results/clientpositive/spark/bucketsortoptimize_insert_2.q.out
index 3e8ea6f..9b14b92 100644
--- a/ql/src/test/results/clientpositive/spark/bucketsortoptimize_insert_2.q.out
+++ b/ql/src/test/results/clientpositive/spark/bucketsortoptimize_insert_2.q.out
@@ -193,18 +193,7 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table3
 POSTHOOK: Input: default@test_table3@ds=1
  A masked pattern was here 
-0  val_0val_0  1
-0  val_0val_0  1
-0  val_0val_0  1
-0  val_0val_0  1
-0  val_0val_0  1
-0  val_0val_0  1
-0  val_0val_0  1
-0  val_0val_0  1
-0  val_0val_0  1
 2  val_2val_2  1
-4  val_4val_4  1
-8  val_8val_8  1
 PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s 
where ds = '1'
 PREHOOK: type: QUERY
 PREHOOK: Input: default@test_table3
@@ -215,6 +204,16 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table3
 POSTHOOK: Input: default@test_table3@ds=1
  A masked pattern was here 
+0  val_0val_0  1
+0  val_0val_0  1
+0  val_0val_0  1
+0  val_0val_0  1
+0  val_0val_0  1
+0  val_0val_0  1
+0  val_0val_0  1
+0  val_0val_0  1
+0  val_0val_0  1
+4  val_4val_4  1
 5  val_5val_5  1
 5  val_5val_5  1
 5  val_5val_5  1
@@ -224,6 +223,7 @@ POSTHOOK: Input: default@test_table3@ds=1
 5  val_5val_5  1
 5  val_5val_5  1
 5  val_5val_5  1
+8  val_8val_8  1
 9  val_9val_9  1
 PREHOOK: query: EXPLAIN
 INSERT OVERWRITE TABLE test_table3 PARTITION (ds = '1')
@@ -344,6 +344,18 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table3
 POSTHOOK: Input: default@test_table3@ds=1
  A masked pattern was here 
+2  val_2val_2  1
+2  val_2val_2  1
+PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s 
where ds = '1'
+PREHOOK: type: QUERY
+PREHOOK: Input: default@test_table3
+PREHOOK: Input: default@test_table3@ds=1
+ A masked pattern was here 
+POSTHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s 
where ds = '1'
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@test_table3
+POSTHOOK: Input: default@test_table3@ds=1
+ A masked pattern was here 
 0  val_0val_0  1
 0  val_0val_0  1
 0  val_0val_0  1
@@ -362,22 +374,8 @@ POSTHOOK: Input: default@test_table3@ds=1
 0  val_0val_0  1
 0  val_0val_0  1
 0  val_0val_0  1
-2  val_2val_2  1
-2  val_2val_2  1
 4  val_4val_4  1
 4  val_4val_4  1
-8  val_8val_8  1
-8  val_8val_8  1
-PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s 
where ds = '1'
-PREHOOK: type: QUERY
-PREHOOK: Input: default@test_table3
-PREHOOK: Input: default@test_table3@ds=1
- A masked pattern was here 
-POSTHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s 
where ds = '1'
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@test_table3
-POSTHOOK: Input: default@test_table3@ds=1
- A masked pattern was here 
 5  val_5val_5  1
 5  val_5val_5  1
 5  val_5val_5  1
@@ -396,6 +394,8 @@ POSTHOOK: Input: default@test_table3@ds=1
 5  val_5val_5  1
 5  val_5val_5  1
 5  val_5val_5  1
+8  val_8val_8  1
+8  val_8val_8  1
 9  val_9val_9  1
 9  val_9val_9  1
 PREHOOK: query: EXPLAIN
@@ -517,6 +517,18 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table3
 POSTHOOK: Input: default@test_table3@ds=1
  A masked pattern was here 
+2  val_2val_2  1
+2  val_2val_2  1
+PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s 
where ds = '1'
+PREHOOK: type: QUERY
+PREHOOK: Input: default@test_table3
+PREHOOK: Input: default@test_table3@ds=1
+ A masked pattern was here 
+POSTHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s 
where ds = '1'
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@test_table3
+POSTHOOK: Input: default@test_table3@ds=1
+ A masked pattern was here 
 0  val_0val_0  1
 0  val_0val_0  1
 0  val_0val_0  1
@@ -535,22 +547,8 @@ POSTHOOK: Input: default@test_table3@ds=1
 0  val_0val_0  1
 0  val_0val_0  1
 0  val_0val_0  1
-2  val_2val_2  1
-2  val_2val_2  1
 4  val_4val_4  1
 4  val_4val_4  1
-8  val_8val_8  1
-8  val_8val_8  1
-PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s 
where ds = '1'
-PREHOOK: type: QUERY
-P

[16/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/vector_windowing_range_multiorder.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/llap/vector_windowing_range_multiorder.q.out
 
b/ql/src/test/results/clientpositive/llap/vector_windowing_range_multiorder.q.out
index 256b80d..9acbf2e 100644
--- 
a/ql/src/test/results/clientpositive/llap/vector_windowing_range_multiorder.q.out
+++ 
b/ql/src/test/results/clientpositive/llap/vector_windowing_range_multiorder.q.out
@@ -900,10005 +900,10005 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@over10k
  A masked pattern was here 
 s  si  i   avg_window_0
-alice falkner  323 65669   65695.76470588235
 alice falkner  477 65722   65695.76470588235
-alice falkner  455 65718   65695.76470588235
-alice falkner  481 65709   65695.76470588235
-alice falkner  345 65773   65695.76470588235
-alice falkner  280 65597   65695.76470588235
-alice falkner  500 65775   65695.76470588235
 alice falkner  339 65785   65695.76470588235
-alice falkner  452 65596   65695.76470588235
-alice falkner  382 65690   65695.76470588235
 alice falkner  382 65622   65695.76470588235
 alice falkner  393 65611   65695.76470588235
+alice falkner  389 65699   65695.76470588235
+alice falkner  500 65775   65695.76470588235
+alice falkner  371 65710   65695.76470588235
+alice falkner  481 65709   65695.76470588235
+alice falkner  452 65596   65695.76470588235
+alice falkner  323 65669   65695.76470588235
+alice falkner  382 65690   65695.76470588235
+alice falkner  455 65718   65695.76470588235
 alice falkner  393 65685   65695.76470588235
-alice falkner  342 65752   65695.76470588235
 alice falkner  311 65715   65695.76470588235
-alice falkner  371 65710   65695.76470588235
-alice falkner  389 65699   65695.76470588235
-alice ichabod  366 65590   65654.95454545454
+alice falkner  280 65597   65695.76470588235
+alice falkner  342 65752   65695.76470588235
+alice falkner  345 65773   65695.76470588235
+alice ichabod  292 65788   65654.95454545454
 alice ichabod  458 65550   65654.95454545454
+alice ichabod  412 65718   65654.95454545454
+alice ichabod  303 65692   65654.95454545454
+alice ichabod  300 65704   65654.95454545454
+alice ichabod  366 65590   65654.95454545454
+alice ichabod  301 65693   65654.95454545454
+alice ichabod  398 65659   65654.95454545454
+alice ichabod  292 65585   65654.95454545454
+alice ichabod  305 65617   65654.95454545454
 alice ichabod  436 65738   65654.95454545454
-alice ichabod  315 65772   65654.95454545454
+alice ichabod  440 65725   65654.95454545454
 alice ichabod  453 65780   65654.95454545454
+alice ichabod  398 65785   65654.95454545454
 alice ichabod  347 65547   65654.95454545454
-alice ichabod  398 65659   65654.95454545454
-alice ichabod  338 65538   65654.95454545454
-alice ichabod  440 65725   65654.95454545454
 alice ichabod  320 65622   65654.95454545454
-alice ichabod  412 65718   65654.95454545454
-alice ichabod  305 65617   65654.95454545454
-alice ichabod  292 65585   65654.95454545454
-alice ichabod  303 65692   65654.95454545454
-alice ichabod  338 65545   65654.95454545454
 alice ichabod  398 65680   65654.95454545454
+alice ichabod  315 65772   65654.95454545454
 alice ichabod  416 65536   65654.95454545454
+alice ichabod  338 65545   65654.95454545454
 alice ichabod  344 65545   65654.95454545454
-alice ichabod  300 65704   65654.95454545454
-alice ichabod  292 65788   65654.95454545454
-alice ichabod  398 65785   65654.95454545454
-alice ichabod  301 65693   65654.95454545454
-alice polk 443 65734   65661.57142857143
-alice polk 444 65564   65661.57142857143
-alice polk 357 65550   65661.57142857143
-alice polk 321 65744   65661.57142857143
-alice polk 273 65548   65661.57142857143
-alice polk 366 65595   65661.57142857143
-alice polk 285 65761   65661.57142857143
-alice polk 466 65561   65661.57142857143
-alice polk 324 65749   65661.57142857143
-alice polk 487 65746   65661.57142857143
-alice polk 378 65598   65661.57142857143
-alice polk 395 65751   65661.57142857143
-alice polk 407 65617   65661.57142857143
-alice polk 507 65744   65661.57142857143
-alice young468 65649   65706.63636363637
-alice young489 65646   65706.63636363637
-alice young286 65705   65706.63636363637
-alice young447 65789   65706.63636363637
-alice young425 65677   65706.63636363637
-alice young282 65671   65706.63636363637
-alice young351 65776   65706.63636363637
-alice young308 65776   65706.63636363637

[04/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/smb_mapjoin_12.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/smb_mapjoin_12.q.out 
b/ql/src/test/results/clientpositive/spark/smb_mapjoin_12.q.out
index ec8b046..9b421d5 100644
--- a/ql/src/test/results/clientpositive/spark/smb_mapjoin_12.q.out
+++ b/ql/src/test/results/clientpositive/spark/smb_mapjoin_12.q.out
@@ -141,6 +141,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 16
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -181,6 +182,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 16
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -213,6 +215,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 16
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -267,7 +270,7 @@ POSTHOOK: Input: default@test_table1@ds=1
 POSTHOOK: Input: default@test_table3
 POSTHOOK: Input: default@test_table3@ds=1
  A masked pattern was here 
-879
+420
 PREHOOK: query: explain extended
 INSERT OVERWRITE TABLE test_table3 PARTITION (ds = '2') 
 SELECT /*+mapjoin(b)*/ a.key, concat(a.value, b.value) FROM test_table3 a JOIN 
test_table1 b ON a.key = b.key AND a.ds = '1' AND b.ds='1'
@@ -359,6 +362,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 16
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -399,6 +403,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 16
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -431,6 +436,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 16
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -479,4 +485,4 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table3
 POSTHOOK: Input: default@test_table3@ds=2
  A masked pattern was here 
-879
+420

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out 
b/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out
index f8d0fb2..a364374 100644
--- a/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out
+++ b/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out
@@ -124,6 +124,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 16
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -147,6 +148,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 16
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -273,6 +275,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 16
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -296,6 +299,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 16
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -373,6 +377,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 16

[07/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/list_bucket_dml_2.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/list_bucket_dml_2.q.out 
b/ql/src/test/results/clientpositive/spark/list_bucket_dml_2.q.out
index 30dc8b5..0ee050f 100644
--- a/ql/src/test/results/clientpositive/spark/list_bucket_dml_2.q.out
+++ b/ql/src/test/results/clientpositive/spark/list_bucket_dml_2.q.out
@@ -55,6 +55,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.RCFileOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -109,6 +110,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -157,6 +159,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -189,6 +192,7 @@ STAGE PLANS:
   output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -338,6 +342,7 @@ STAGE PLANS:
   output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out 
b/ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out
index 10d12a6..71996fe 100644
--- a/ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out
+++ b/ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out
@@ -173,6 +173,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -221,6 +222,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -269,6 +271,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -317,6 +320,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out 
b/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out
index 972a7a8..7fbd873 100644
--- a/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out
+++ b/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out
@@ -62,6 +62,7 @@ STAGE PLANS:
   properties:
 COLUMN_STAT

[35/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/keyseries/VectorKeySeriesSerializedImpl.java
--
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/keyseries/VectorKeySeriesSerializedImpl.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/keyseries/VectorKeySeriesSerializedImpl.java
index 86f466f..77c9ecc 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/keyseries/VectorKeySeriesSerializedImpl.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/keyseries/VectorKeySeriesSerializedImpl.java
@@ -21,9 +21,9 @@ package org.apache.hadoop.hive.ql.exec.vector.keyseries;
 import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch;
 import org.apache.hadoop.hive.serde2.ByteStream.Output;
 import org.apache.hadoop.hive.serde2.fast.SerializeWrite;
-import org.apache.hive.common.util.HashCodeUtil;
 
 import com.google.common.base.Preconditions;
+import org.apache.hive.common.util.Murmur3;
 
 /**
  * Implementation of base serialization interface.
@@ -103,7 +103,7 @@ public abstract class VectorKeySeriesSerializedImpl
 byte[] bytes = output.getData();
 for (int i = 0; i < nonNullKeyCount; i++) {
   keyLength = serializedKeyLengths[i];
-  hashCodes[i] = HashCodeUtil.murmurHash(bytes, offset, keyLength);
+  hashCodes[i] = Murmur3.hash32(bytes, offset, keyLength, 0);
   offset += keyLength;
 }
   }

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java
--
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java
index 1bc3fda..42b7784 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java
@@ -32,6 +32,7 @@ import 
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression;
 import org.apache.hadoop.hive.ql.metadata.HiveException;
 import org.apache.hadoop.hive.ql.plan.OperatorDesc;
 import org.apache.hadoop.hive.ql.plan.VectorDesc;
+import org.apache.hadoop.hive.serde2.ByteStream;
 import org.apache.hadoop.hive.serde2.ByteStream.Output;
 import 
org.apache.hadoop.hive.serde2.binarysortable.fast.BinarySortableSerializeWrite;
 import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector;
@@ -40,6 +41,7 @@ import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo;
 import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoUtils;
 
 import com.google.common.base.Preconditions;
+import org.apache.hive.common.util.Murmur3;
 
 /**
  * This class is the object hash (not Uniform Hash) operator class for native 
vectorized reduce sink.
@@ -226,61 +228,110 @@ public class VectorReduceSinkObjectHashOperator extends 
VectorReduceSinkCommonOp
   int[] selected = batch.selected;
 
   final int size = batch.size;
-  for (int logical = 0; logical < size; logical++) {
-final int batchIndex = (selectedInUse ? selected[logical] : logical);
-
-final int hashCode;
-if (isEmptyBuckets) {
-  if (isEmptyPartitions) {
-hashCode = nonPartitionRandom.nextInt();
-  } else {
+
+  // EmptyBuckets = true
+  if (isEmptyBuckets) {
+if (isEmptyPartitions) {
+  for (int logical = 0; logical< size; logical++) {
+final int batchIndex = (selectedInUse ? selected[logical] : 
logical);
+final int hashCode = nonPartitionRandom.nextInt();
+postProcess(batch, batchIndex, tag, hashCode);
+  }
+} else { // isEmptyPartition = false
+  for (int logical = 0; logical< size; logical++) {
+final int batchIndex = (selectedInUse ? selected[logical] : 
logical);
 partitionVectorExtractRow.extractRow(batch, batchIndex, 
partitionFieldValues);
-hashCode =
+final int hashCode = bucketingVersion == 2 && 
!vectorDesc.getIsAcidChange() ?
 ObjectInspectorUtils.getBucketHashCode(
+partitionFieldValues, partitionObjectInspectors) :
+ObjectInspectorUtils.getBucketHashCodeOld(
 partitionFieldValues, partitionObjectInspectors);
+postProcess(batch, batchIndex, tag, hashCode);
   }
-} else {
-  bucketVectorExtractRow.extractRow(batch, batchIndex, 
bucketFieldValues);
-  final int bucketNum =
-  ObjectInspectorUtils.getBucketNumber(
+}
+  } else { // EmptyBuckets = false
+if (isEmptyPartitions) {
+  for (int logical = 0; logical< size; logical

[28/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/join9.q.out
--
diff --git a/ql/src/test/results/clientpositive/join9.q.out 
b/ql/src/test/results/clientpositive/join9.q.out
index 50ca52c..0a86a6e 100644
--- a/ql/src/test/results/clientpositive/join9.q.out
+++ b/ql/src/test/results/clientpositive/join9.q.out
@@ -76,6 +76,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -97,6 +98,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 'default','default'
@@ -147,6 +149,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 'default','default'
@@ -192,6 +195,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -250,6 +254,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/join_filters_overlap.q.out
--
diff --git a/ql/src/test/results/clientpositive/join_filters_overlap.q.out 
b/ql/src/test/results/clientpositive/join_filters_overlap.q.out
index db0b26d..7d4855a 100644
--- a/ql/src/test/results/clientpositive/join_filters_overlap.q.out
+++ b/ql/src/test/results/clientpositive/join_filters_overlap.q.out
@@ -92,6 +92,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -113,6 +114,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -283,6 +285,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -304,6 +307,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -474,6 +478,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -495,6 +500,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -682,6 +688,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -703,6 +710,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
 bucket_count -1
+bucketing_version 2
 column.name.d

[19/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out 
b/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out
index 6a654d0..42d5528 100644
--- a/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out
+++ b/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out
@@ -651,314 +651,314 @@ POSTHOOK: query: select unionsrc.key, unionsrc.value 
FROM (select s1.key as key,
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
 POSTHOOK: Output: hdfs://### HDFS PATH ###
-0  val_0
-103val_103
-105val_105
 116val_116
+118val_118
 120val_120
-129val_129
 131val_131
-136val_136
+133val_133
+15 val_15
+150val_150
+152val_152
+153val_153
 155val_155
-157val_157
 160val_160
-180val_180
+169val_169
+175val_175
+176val_176
+178val_178
 183val_183
-190val_190
-191val_191
-193val_193
-205val_205
-207val_207
+186val_186
+194val_194
+202val_202
+203val_203
 213val_213
 223val_223
-235val_235
-238val_238
+224val_224
+226val_226
+228val_228
 239val_239
-24 val_24
-248val_248
-260val_260
+244val_244
+247val_247
+252val_252
+255val_255
+256val_256
+257val_257
+266val_266
+272val_272
+274val_274
+275val_275
 28 val_28
+284val_284
 286val_286
-298val_298
-305val_305
+289val_289
+296val_296
+302val_302
 306val_306
 307val_307
-308val_308
-315val_315
 317val_317
 321val_321
+323val_323
 33 val_33
-331val_331
-356val_356
+338val_338
+34 val_34
+35 val_35
 360val_360
-366val_366
-367val_367
-368val_368
-37 val_37
-378val_378
-379val_379
-382val_382
-389val_389
-395val_395
-400val_400
-402val_402
+365val_365
+374val_374
+377val_377
+393val_393
+399val_399
+404val_404
 406val_406
-407val_407
+409val_409
+41 val_41
+417val_417
+418val_418
 42 val_42
-430val_430
-436val_436
+431val_431
+439val_439
 44 val_44
-452val_452
-466val_466
-467val_467
-472val_472
-480val_480
-484val_484
-485val_485
-487val_487
-495val_495
-496val_496
-498val_498
-51 val_51
-57 val_57
-58 val_58
-65 val_65
+443val_443
+448val_448
+455val_455
+459val_459
+462val_462
+463val_463
+470val_470
+490val_490
+491val_491
+494val_494
+5  val_5
+53 val_53
 66 val_66
-69 val_69
-70 val_70
-78 val_78
-8  val_8
-92 val_92
-96 val_96
-104val_104
+72 val_72
+76 val_76
+77 val_77
+80 val_80
+84 val_84
+90 val_90
+10 val_10
+105val_105
 111val_111
-114val_114
-118val_118
-12 val_12
-133val_133
-138val_138
-153val_153
-156val_156
-162val_162
-168val_168
+113val_113
+119val_119
+129val_129
+136val_136
+143val_143
+149val_149
+163val_163
+167val_167
 174val_174
-176val_176
-178val_178
-181val_181
-186val_186
 187val_187
-192val_192
 197val_197
-199val_199
 2  val_2
-201val_201
-203val_203
+200val_200
 208val_208
-214val_214
+209val_209
+216val_216
+217val_217
 218val_218
-219val_219
+221val_221
 230val_230
-237val_237
-249val_249
-256val_256
-257val_257
+258val_258
+26 val_26
+263val_263
+273val_273
 277val_277
-278val_278
-284val_284
-285val_285
-310val_310
-322val_322
-327val_327
-333val_333
-336val_336
-338val_338
+281val_281
+292val_292
+30 val_30
+316val_316
+335val_335
 341val_341
+342val_342
 344val_344
-348val_348
-351val_351
+345val_345
 362val_362
-364val_364
-374val_374
+367val_367
+368val_368
+37 val_37
 392val_392
-397val_397
+394val_394
 4  val_4
-403val_403
-404val_404
-419val_419
-421val_421
-427val_427
-429val_429
+413val_413
 437val_437
-438val_438
-439val_439
-454val_454
-455val_455
+444val_444
+446val_446
+452val_452
+453val_453
+457val_457
 460val_460
-463val_463
-470val_470
-478val_478
+466val_466
+468val_468
+477val_477
 479val_479
-489val_489
-491val_491
+483val_483
+487val_487
 492val_492
-53 val_53
+493val_493
+498val_498
+51 val_51
+54 val_54
+64 val_64
+65 val_65
 67 val_67
 82 val_82
-90 val_90
-97 val_97
-10 val_10
-125val_125
-134val_134
-143val_143
+85 val_85

[18/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out 
b/ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out
index 687b4af..8bec959 100644
--- a/ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out
@@ -232,17 +232,17 @@ POSTHOOK: query: select fl_date, count(*) from 
flights_tiny_orc group by fl_date
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@flights_tiny_orc
  A masked pattern was here 
-2010-10-20 11
 2010-10-21 12
-2010-10-26 13
-2010-10-30 11
-2010-10-23 12
 2010-10-24 12
-2010-10-25 12
-2010-10-22 11
+2010-10-26 13
 2010-10-28 12
+2010-10-22 11
 2010-10-29 12
+2010-10-20 11
+2010-10-23 12
+2010-10-25 12
 2010-10-27 11
+2010-10-30 11
 2010-10-31 8
 PREHOOK: query: explain vectorization expression
 select * from flights_tiny_orc sort by fl_num, fl_date limit 25
@@ -529,17 +529,17 @@ POSTHOOK: query: select fl_date, count(*) from 
flights_tiny_orc group by fl_date
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@flights_tiny_orc
  A masked pattern was here 
-2010-10-20 11
 2010-10-21 12
-2010-10-26 13
-2010-10-30 11
-2010-10-23 12
 2010-10-24 12
-2010-10-25 12
-2010-10-22 11
+2010-10-26 13
 2010-10-28 12
+2010-10-22 11
 2010-10-29 12
+2010-10-20 11
+2010-10-23 12
+2010-10-25 12
 2010-10-27 11
+2010-10-30 11
 2010-10-31 8
 PREHOOK: query: CREATE TABLE flights_tiny_orc_partitioned_date (
   origin_city_name STRING,
@@ -908,17 +908,17 @@ POSTHOOK: Input: 
default@flights_tiny_orc_partitioned_date@fl_date=2010-10-29
 POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-30
 POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-31
  A masked pattern was here 
-2010-10-20 11
 2010-10-21 12
-2010-10-26 13
-2010-10-30 11
-2010-10-23 12
 2010-10-24 12
-2010-10-25 12
-2010-10-22 11
+2010-10-26 13
 2010-10-28 12
+2010-10-22 11
 2010-10-29 12
+2010-10-20 11
+2010-10-23 12
+2010-10-25 12
 2010-10-27 11
+2010-10-30 11
 2010-10-31 8
 PREHOOK: query: explain vectorization expression
 select * from flights_tiny_orc_partitioned_date
@@ -1484,17 +1484,17 @@ POSTHOOK: Input: 
default@flights_tiny_orc_partitioned_date@fl_date=2010-10-29
 POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-30
 POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-31
  A masked pattern was here 
-2010-10-20 11
 2010-10-21 12
-2010-10-26 13
-2010-10-30 11
-2010-10-23 12
 2010-10-24 12
-2010-10-25 12
-2010-10-22 11
+2010-10-26 13
 2010-10-28 12
+2010-10-22 11
 2010-10-29 12
+2010-10-20 11
+2010-10-23 12
+2010-10-25 12
 2010-10-27 11
+2010-10-30 11
 2010-10-31 8
 PREHOOK: query: CREATE TABLE flights_tiny_orc_partitioned_timestamp (
   origin_city_name STRING,
@@ -1863,18 +1863,18 @@ POSTHOOK: Input: 
default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-
 POSTHOOK: Input: 
default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-30 07%3A00%3A00
 POSTHOOK: Input: 
default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-31 07%3A00%3A00
  A masked pattern was here 
-2010-10-20 07:00:0011
 2010-10-23 07:00:0012
 2010-10-25 07:00:0012
 2010-10-22 07:00:0011
+2010-10-24 07:00:0012
 2010-10-26 07:00:0013
+2010-10-20 07:00:0011
+2010-10-28 07:00:0012
 2010-10-29 07:00:0012
+2010-10-30 07:00:0011
 2010-10-31 07:00:008
 2010-10-21 07:00:0012
-2010-10-24 07:00:0012
 2010-10-27 07:00:0011
-2010-10-28 07:00:0012
-2010-10-30 07:00:0011
 PREHOOK: query: explain vectorization expression
 select * from flights_tiny_orc_partitioned_timestamp
 PREHOOK: type: QUERY
@@ -2439,18 +2439,18 @@ POSTHOOK: Input: 
default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-
 POSTHOOK: Input: 
default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-30 07%3A00%3A00
 POSTHOOK: Input: 
default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-31 07%3A00%3A00
  A masked pattern was here 
-2010-10-20 07:00:0011
 2010-10-23 07:00:0012
 2010-10-25 07:00:0012
 2010-10-22 07:00:0011
+2010-10-24 07:00:0012
 2010-10-26 07:00:0013
+2010-10-20 07:00:0011
+2010-10-28 07:00:0012
 2010-10-29 07:00:0012
+2010-10-30 07:00:0011
 2010-10-31 07:00:008
 2010-10-21 07:00:0012
-2010-10-24 07:00:0012
 2010-10-27 07:00:0011
-2010-10-28 07:00:0012
-2010-10-30 07:00:

[12/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/sample7.q.out
--
diff --git a/ql/src/test/results/clientpositive/sample7.q.out 
b/ql/src/test/results/clientpositive/sample7.q.out
index f24f297..07b6f0e 100644
--- a/ql/src/test/results/clientpositive/sample7.q.out
+++ b/ql/src/test/results/clientpositive/sample7.q.out
@@ -55,6 +55,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -102,6 +103,7 @@ STAGE PLANS:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -124,6 +126,7 @@ STAGE PLANS:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -193,6 +196,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -236,6 +240,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -266,6 +271,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -287,6 +293,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -323,6 +330,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -353,6 +361,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -374,6 +383,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -424,107 +434,48 @@ order by key, value
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest1
  A masked pattern was here 
-104val_104
-104val_104
-104val_105
-104val_105
-104val_105
+110val_111
 116val_116
 116val_117
-120val_120
-120val_120
-120val_121
-128val_128
-128val_128
-128val_128
-128val_129
-128val_129
-132val_133
-132val_133
-136val_136
-136val_137
-140val_141
-144val_145
-152val_152
-152val_152
-152val_153
-152val_153
-152val_153
+126val_126
+126val_127
+126val_127
+134val_134
+134val_134
+134val_135
 156val_156
 156val_157
 156val_157
-160val_160
-160val_161
-164val_164
-164val_164
-164val_165
-164val_165
-168val_168
-168val_169
-172val_172
-172val_172
-172val_173
-176val_176
-176val_176
-180val_180
+158val_158
+178val_178
+178val_179
+178

[20/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/sysdb.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/sysdb.q.out 
b/ql/src/test/results/clientpositive/llap/sysdb.q.out
index 7d977d3..de20f2d 100644
--- a/ql/src/test/results/clientpositive/llap/sysdb.q.out
+++ b/ql/src/test/results/clientpositive/llap/sysdb.q.out
@@ -58,34 +58,10 @@ default alltypesorc hive_test_user  
USERDELETE  true-1  hive_test_user
 defaultalltypesorc hive_test_user  USERINSERT  
true-1  hive_test_user
 defaultalltypesorc hive_test_user  USERSELECT  
true-1  hive_test_user
 defaultalltypesorc hive_test_user  USERUPDATE  
true-1  hive_test_user
-defaultalltypesparquet hive_test_user  USERDELETE  
true-1  hive_test_user
-defaultalltypesparquet hive_test_user  USERINSERT  
true-1  hive_test_user
-defaultalltypesparquet hive_test_user  USERSELECT  
true-1  hive_test_user
-defaultalltypesparquet hive_test_user  USERUPDATE  
true-1  hive_test_user
-defaultcbo_t1  hive_test_user  USERDELETE  true
-1  hive_test_user
-defaultcbo_t1  hive_test_user  USERINSERT  true
-1  hive_test_user
-defaultcbo_t1  hive_test_user  USERSELECT  true
-1  hive_test_user
-defaultcbo_t1  hive_test_user  USERUPDATE  true
-1  hive_test_user
-defaultcbo_t2  hive_test_user  USERDELETE  true
-1  hive_test_user
-defaultcbo_t2  hive_test_user  USERINSERT  true
-1  hive_test_user
-defaultcbo_t2  hive_test_user  USERSELECT  true
-1  hive_test_user
-defaultcbo_t2  hive_test_user  USERUPDATE  true
-1  hive_test_user
-defaultcbo_t3  hive_test_user  USERDELETE  true
-1  hive_test_user
-defaultcbo_t3  hive_test_user  USERINSERT  true
-1  hive_test_user
-defaultcbo_t3  hive_test_user  USERSELECT  true
-1  hive_test_user
-defaultcbo_t3  hive_test_user  USERUPDATE  true
-1  hive_test_user
-defaultlineitemhive_test_user  USERDELETE  
true-1  hive_test_user
-defaultlineitemhive_test_user  USERINSERT  
true-1  hive_test_user
-defaultlineitemhive_test_user  USERSELECT  
true-1  hive_test_user
-defaultlineitemhive_test_user  USERUPDATE  
true-1  hive_test_user
 defaultmoretypes   hive_test_user  USERDELETE  
true-1  hive_test_user
 defaultmoretypes   hive_test_user  USERINSERT  
true-1  hive_test_user
 defaultmoretypes   hive_test_user  USERSELECT  
true-1  hive_test_user
 defaultmoretypes   hive_test_user  USERUPDATE  
true-1  hive_test_user
-defaultparthive_test_user  USERDELETE  true
-1  hive_test_user
-defaultparthive_test_user  USERINSERT  true
-1  hive_test_user
-defaultparthive_test_user  USERSELECT  true
-1  hive_test_user
-defaultparthive_test_user  USERUPDATE  true
-1  hive_test_user
 defaultscr_txn hive_test_user  USERDELETE  true
-1  hive_test_user
 defaultscr_txn hive_test_user  USERINSERT  true
-1  hive_test_user
 defaultscr_txn hive_test_user  USERSELECT  true
-1  hive_test_user
@@ -94,46 +70,14 @@ default src hive_test_user  USER
DELETE  true-1  hive_test_user
 defaultsrc hive_test_user  USERINSERT  true
-1  hive_test_user
 defaultsrc hive_test_user  USERSELECT  true
-1  hive_test_user
 defaultsrc hive_test_user  USERUPDATE  true
-1  hive_test_user
-defaultsrc1hive_test_user  USERDELETE  true
-1  hive_test_user
-defaultsrc1hive_test_user  USERINSERT  true
-1  hive_test_user
-defaultsrc1hive_test_user  USERSELECT  true
-1  hive_test_user
-defaultsrc1

[34/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientnegative/alter_file_format.q.out
--
diff --git a/ql/src/test/results/clientnegative/alter_file_format.q.out 
b/ql/src/test/results/clientnegative/alter_file_format.q.out
index 96259e5..3033cbe 100644
--- a/ql/src/test/results/clientnegative/alter_file_format.q.out
+++ b/ql/src/test/results/clientnegative/alter_file_format.q.out
@@ -24,6 +24,7 @@ Retention:0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
+   bucketing_version   2   
numFiles0   
numRows 0   
rawDataSize 0   

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out
--
diff --git 
a/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out 
b/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out
index 8e240ea..90c3d02 100644
--- 
a/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out
+++ 
b/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out
@@ -51,6 +51,7 @@ Retention:0
 Table Type:VIRTUAL_VIEW 
 Table Parameters:   
COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
+   bucketing_version   2   
numFiles0   
numPartitions   2   
numRows 0   

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientnegative/unset_table_property.q.out
--
diff --git a/ql/src/test/results/clientnegative/unset_table_property.q.out 
b/ql/src/test/results/clientnegative/unset_table_property.q.out
index 2dd047a..eb308eb 100644
--- a/ql/src/test/results/clientnegative/unset_table_property.q.out
+++ b/ql/src/test/results/clientnegative/unset_table_property.q.out
@@ -20,6 +20,7 @@ POSTHOOK: query: SHOW TBLPROPERTIES testTable
 POSTHOOK: type: SHOW_TBLPROPERTIES
 COLUMN_STATS_ACCURATE  
{"BASIC_STATS":"true","COLUMN_STATS":{"col1":"true","col2":"true"}}
 a  1
+bucketing_version  2
 c  3
  A masked pattern was here 
 numFiles   0

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientnegative/unset_view_property.q.out
--
diff --git a/ql/src/test/results/clientnegative/unset_view_property.q.out 
b/ql/src/test/results/clientnegative/unset_view_property.q.out
index ae7f4f6..b5038d9 100644
--- a/ql/src/test/results/clientnegative/unset_view_property.q.out
+++ b/ql/src/test/results/clientnegative/unset_view_property.q.out
@@ -21,6 +21,7 @@ PREHOOK: query: SHOW TBLPROPERTIES testView
 PREHOOK: type: SHOW_TBLPROPERTIES
 POSTHOOK: query: SHOW TBLPROPERTIES testView
 POSTHOOK: type: SHOW_TBLPROPERTIES
+bucketing_version  2
  A masked pattern was here 
 propA  100
 propB  200

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/acid_mapjoin.q.out
--
diff --git a/ql/src/test/results/clientpositive/acid_mapjoin.q.out 
b/ql/src/test/results/clientpositive/acid_mapjoin.q.out
index 5569a03..76a781e 100644
--- a/ql/src/test/results/clientpositive/acid_mapjoin.q.out
+++ b/ql/src/test/results/clientpositive/acid_mapjoin.q.out
@@ -73,21 +73,21 @@ STAGE PLANS:
   Stage: Stage-5
 Map Reduce Local Work
   Alias -> Map Local Tables:
-$hdt$_1:acid2 
+$hdt$_0:acid1 
   Fetch Operator
 limit: -1
   Alias -> Map Local Operator Tree:
-$hdt$_1:acid2 
+$hdt$_0:acid1 
   TableScan
-alias: acid2
-Statistics: Num rows: 210 Data size: 840 Basic stats: COMPLETE 
Column stats: NONE
+alias: acid1
+Statistics: Num rows: 316 Data size: 1265 Basic stats: COMPLETE 
Column stats: NONE
 Filter Operator
   predicate: key is not null (type: boolean)
-  Statistics: Num rows: 210 Data size: 840 Basic stats: COMPLETE 
Column stats: NONE
+  Statistics: Num rows: 316 Data size: 1265 Basic stats: COMPLETE 
Column stats: NONE
   Select Operator
 expressions: key (type: int)
 outputColumnNames: _col0
-Statistics: Num rows: 210 Data size: 840 Basic stats: COMPLETE 
Column stats: NONE
+  

[23/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/default_constraint.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/default_constraint.q.out 
b/ql/src/test/results/clientpositive/llap/default_constraint.q.out
index 15cbd84..d987ef2 100644
--- a/ql/src/test/results/clientpositive/llap/default_constraint.q.out
+++ b/ql/src/test/results/clientpositive/llap/default_constraint.q.out
@@ -31,6 +31,7 @@ Retention:0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
transactional   true
transactional_propertiesdefault 
  A masked pattern was here 
@@ -301,6 +302,7 @@ Retention:  0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
  A masked pattern was here 
 
 # Storage Information   
@@ -530,6 +532,7 @@ Retention:  0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
transactional   true
transactional_propertiesdefault 
  A masked pattern was here 
@@ -788,6 +791,7 @@ Retention:  0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
transactional   true
transactional_propertiesdefault 
  A masked pattern was here 
@@ -1059,6 +1063,7 @@ Retention:0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
  A masked pattern was here 
 
 # Storage Information   
@@ -1336,6 +1341,7 @@ Retention:0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
transactional   true
transactional_propertiesdefault 
  A masked pattern was here 
@@ -1490,8 +1496,9 @@ Retention:0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
numFiles1   
-   totalSize   1062
+   totalSize   1070
transactional   true
transactional_propertiesdefault 
  A masked pattern was here 
@@ -1658,9 +1665,10 @@ Retention:   0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
  A masked pattern was here 
numFiles2   
-   totalSize   2124
+   totalSize   2140
transactional   true
transactional_propertiesdefault 
  A masked pattern was here 
@@ -1737,9 +1745,10 @@ Retention:   0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
  A masked pattern was here 
numFiles2   
-   totalSize   2124
+   totalSize   2140
transactional   true
transactional_propertiesdefault 
  A masked pattern was here 
@@ -1913,9 +1922,10 @@ Retention:   0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
  A masked pattern was here 
numFiles3   
-   totalSize   3188
+   totalSize   3199
transactional   true
transactional_propertiesdefault 
  A masked pattern was here 
@@ -1991,9 +2001,10 @@ Retention:   0
  A masked pattern was here 
 Table Type:MANAGED_TAB

[22/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/intersect_distinct.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/intersect_distinct.q.out 
b/ql/src/test/results/clientpositive/llap/intersect_distinct.q.out
index 44362be..fb59f10 100644
--- a/ql/src/test/results/clientpositive/llap/intersect_distinct.q.out
+++ b/ql/src/test/results/clientpositive/llap/intersect_distinct.q.out
@@ -281,315 +281,315 @@ POSTHOOK: query: select * from src intersect distinct 
select * from src
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
 POSTHOOK: Output: hdfs://### HDFS PATH ###
-0  val_0
-103val_103
+104val_104
 105val_105
-116val_116
+113val_113
+118val_118
+12 val_12
 120val_120
-129val_129
-131val_131
+133val_133
 136val_136
 155val_155
-157val_157
+158val_158
 160val_160
+162val_162
+168val_168
+17 val_17
+170val_170
+175val_175
+18 val_18
 180val_180
-183val_183
-190val_190
-191val_191
-193val_193
-205val_205
-207val_207
-213val_213
-223val_223
+186val_186
+19 val_19
+197val_197
+200val_200
+216val_216
+218val_218
+222val_222
+224val_224
+228val_228
 235val_235
-238val_238
+237val_237
 239val_239
-24 val_24
-248val_248
-260val_260
+244val_244
+247val_247
+256val_256
+263val_263
+273val_273
 28 val_28
+283val_283
 286val_286
-298val_298
-305val_305
+292val_292
 306val_306
-307val_307
 308val_308
+327val_327
+33 val_33
+348val_348
+353val_353
+362val_362
+366val_366
+396val_396
+397val_397
+401val_401
+409val_409
+411val_411
+419val_419
+427val_427
+43 val_43
+432val_432
+436val_436
+439val_439
+443val_443
+453val_453
+460val_460
+462val_462
+47 val_47
+472val_472
+485val_485
+496val_496
+54 val_54
+64 val_64
+70 val_70
+8  val_8
+83 val_83
+84 val_84
+85 val_85
+90 val_90
+0  val_0
+103val_103
+114val_114
+125val_125
+138val_138
+146val_146
+150val_150
+152val_152
+153val_153
+156val_156
+157val_157
+165val_165
+172val_172
+177val_177
+179val_179
+187val_187
+195val_195
+196val_196
+217val_217
+242val_242
+248val_248
+252val_252
+265val_265
+27 val_27
+272val_272
+280val_280
+291val_291
+305val_305
+309val_309
+311val_311
 315val_315
 317val_317
-321val_321
-33 val_33
-331val_331
+322val_322
+333val_333
+34 val_34
+345val_345
+35 val_35
 356val_356
-360val_360
-366val_366
-367val_367
+364val_364
 368val_368
+369val_369
 37 val_37
-378val_378
-379val_379
-382val_382
-389val_389
-395val_395
-400val_400
+373val_373
+377val_377
+4  val_4
 402val_402
-406val_406
-407val_407
+404val_404
+413val_413
 42 val_42
 430val_430
-436val_436
-44 val_44
+431val_431
+444val_444
+449val_449
 452val_452
+454val_454
+457val_457
+463val_463
 466val_466
-467val_467
-472val_472
-480val_480
-484val_484
-485val_485
-487val_487
-495val_495
-496val_496
-498val_498
-51 val_51
+470val_470
+475val_475
+481val_481
+489val_489
+491val_491
 57 val_57
-58 val_58
 65 val_65
 66 val_66
-69 val_69
-70 val_70
+74 val_74
+76 val_76
 78 val_78
-8  val_8
+9  val_9
 92 val_92
-96 val_96
-104val_104
-111val_111
-114val_114
-118val_118
-12 val_12
-133val_133
-138val_138
-153val_153
-156val_156
-162val_162
-168val_168
+95 val_95
+100val_100
+119val_119
+128val_128
+129val_129
+145val_145
+167val_167
 174val_174
-176val_176
 178val_178
 181val_181
-186val_186
-187val_187
-192val_192
-197val_197
+193val_193
 199val_199
-2  val_2
+20 val_20
 201val_201
-203val_203
-208val_208
+213val_213
 214val_214
-218val_218
 219val_219
-230val_230
-237val_237
+221val_221
+223val_223
+226val_226
+233val_233
+241val_241
 249val_249
-256val_256
-257val_257
+260val_260
+262val_262
 277val_277
-278val_278
-284val_284
-285val_285
+281val_281
+287val_287
+288val_288
+298val_298
+302val_302
 310val_310
-322val_322
-327val_327
-333val_333
+323val_323
 336val_336
-338val_338
 341val_341
 344val_344
-348val_348
 351val_351
-362val_362
-364val_364
-374val_374
-392val_392
-397val_397
-4  v

[01/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
Repository: hive
Updated Branches:
  refs/heads/master a095c3d93 -> ac6b2a3fb


http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/union_remove_7.q.out
--
diff --git a/ql/src/test/results/clientpositive/union_remove_7.q.out 
b/ql/src/test/results/clientpositive/union_remove_7.q.out
index b60f7bb..437b08a 100644
--- a/ql/src/test/results/clientpositive/union_remove_7.q.out
+++ b/ql/src/test/results/clientpositive/union_remove_7.q.out
@@ -173,6 +173,7 @@ Retention:  0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
numFiles2   
totalSize   178 
  A masked pattern was here 

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/union_remove_8.q.out
--
diff --git a/ql/src/test/results/clientpositive/union_remove_8.q.out 
b/ql/src/test/results/clientpositive/union_remove_8.q.out
index 0fbaa3c..4971a74 100644
--- a/ql/src/test/results/clientpositive/union_remove_8.q.out
+++ b/ql/src/test/results/clientpositive/union_remove_8.q.out
@@ -183,6 +183,7 @@ Retention:  0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
numFiles3   
totalSize   271 
  A masked pattern was here 

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/union_remove_9.q.out
--
diff --git a/ql/src/test/results/clientpositive/union_remove_9.q.out 
b/ql/src/test/results/clientpositive/union_remove_9.q.out
index 2dadf0f..786a81f 100644
--- a/ql/src/test/results/clientpositive/union_remove_9.q.out
+++ b/ql/src/test/results/clientpositive/union_remove_9.q.out
@@ -230,6 +230,7 @@ Retention:  0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
numFiles2   
totalSize   192 
  A masked pattern was here 

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/union_stats.q.out
--
diff --git a/ql/src/test/results/clientpositive/union_stats.q.out 
b/ql/src/test/results/clientpositive/union_stats.q.out
index 5b6ada7..8bd3f44 100644
--- a/ql/src/test/results/clientpositive/union_stats.q.out
+++ b/ql/src/test/results/clientpositive/union_stats.q.out
@@ -90,6 +90,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -111,6 +112,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 'default','default'
@@ -313,6 +315,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
+   bucketing_version   2   
numFiles1   
numRows 1000
rawDataSize 10624   
@@ -361,6 +364,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
+   bucketing_version   2   
numFiles1   
numRows 1500
rawDataSize 15936   
@@ -417,6 +421,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
+   bucketing_version   2   
numFiles1   
numRows 1500
rawDataSize 15936   

http://git-wip-us.apache.o

[14/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/pcr.q.out
--
diff --git a/ql/src/test/results/clientpositive/pcr.q.out 
b/ql/src/test/results/clientpositive/pcr.q.out
index 0adee06..e30f6d1 100644
--- a/ql/src/test/results/clientpositive/pcr.q.out
+++ b/ql/src/test/results/clientpositive/pcr.q.out
@@ -116,6 +116,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -163,6 +164,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -306,6 +308,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -353,6 +356,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -400,6 +404,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -581,6 +586,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -628,6 +634,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -773,6 +780,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -820,6 +828,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -967,6 +976,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -1014,6 +1024,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -1061,6 +1072,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -1219,6 +1231,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -1266,6 +1279,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
   

[05/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/sample6.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/sample6.q.out 
b/ql/src/test/results/clientpositive/spark/sample6.q.out
index 34aee1e..cf172ea 100644
--- a/ql/src/test/results/clientpositive/spark/sample6.q.out
+++ b/ql/src/test/results/clientpositive/spark/sample6.q.out
@@ -52,6 +52,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -84,6 +85,7 @@ STAGE PLANS:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -106,6 +108,7 @@ STAGE PLANS:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -137,6 +140,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -181,172 +185,175 @@ order by key, value
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest1
  A masked pattern was here 
-0  val_0
-0  val_0
-0  val_0
-0  val_1
-0  val_1
-4  val_4
-4  val_5
-8  val_8
-8  val_9
-12 val_12
-12 val_12
-12 val_13
-16 val_17
-16 val_17
-20 val_20
-20 val_21
-20 val_21
-24 val_24
-24 val_24
-28 val_28
-32 val_33
+2  val_2
+2  val_3
+6  val_7
+6  val_7
+10 val_10
+10 val_11
+17 val_17
+21 val_22
+21 val_22
+21 val_22
+21 val_22
+27 val_27
+30 val_30
+30 val_31
+31 val_32
 40 val_41
 40 val_41
-44 val_44
-48 val_49
-48 val_49
-52 val_53
-52 val_53
-52 val_53
-52 val_53
-56 val_57
-60 val_61
+51 val_51
+51 val_51
+51 val_52
+57 val_57
+58 val_58
+58 val_58
+58 val_59
+58 val_59
 64 val_64
-68 val_69
-72 val_72
-72 val_72
-76 val_76
-76 val_76
-76 val_77
-76 val_77
-76 val_77
+65 val_65
+65 val_66
+65 val_66
+70 val_70
+70 val_70
+70 val_70
+70 val_71
 80 val_80
 80 val_81
-84 val_84
-84 val_84
-92 val_92
-96 val_96
-100val_100
-100val_100
-100val_101
-100val_101
-104val_104
-104val_104
-104val_105
-104val_105
-104val_105
+83 val_83
+83 val_83
+86 val_86
+86 val_87
+90 val_90
+90 val_90
+90 val_90
+91 val_92
+98 val_98
+98 val_98
+105val_105
+105val_106
+105val_106
+110val_111
+113val_113
+113val_113
 116val_116
 116val_117
-120val_120
-120val_120
-120val_121
-128val_128
-128val_128
-128val_128
-128val_129
-128val_129
-132val_133
-132val_133
-136val_136
-136val_137
-140val_141
-144val_145
-152val_152
-152val_152
-152val_153
-152val_153
-152val_153
+117val_118
+117val_118
+119val_119
+119val_119
+119val_119
+119val_120
+119val_120
+119val_120
+121val_122
+121val_122
+123val_124
+123val_124
+126val_126
+126val_127
+126val_127
+134val_134
+134val_134
+134val_135
+137val_137
+137val_137
+137val_138
+153val_153
+153val_154
+153val_154
 156val_156
 156val_157
 156val_157
-160val_160
-160val_161
-164val_164
-164val_164
-164val_165
-164val_165
-168val_168
-168val_169
-172val_172
-172val_172
-172val_173
-176val_176
-176val_176
-180val_180
+157val_157
+157val_158
+157val_158
+158val_158
+163val_163
+169val_169
+169val_169
+169val_169
+169val_169
+177val_177
+177val_178
+177val_178
+178val_178
+178val_179
+178val_179
 184val_185
-192val_192
-192val_193
-196   

[29/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out
--
diff --git a/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out 
b/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out
index cb897d1..5a00d10 100644
--- a/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out
+++ b/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out
@@ -87,6 +87,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,cnt
   columns.comments 
@@ -135,6 +136,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,val
   columns.comments 
@@ -158,6 +160,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,val
 columns.comments 
@@ -227,6 +230,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -270,6 +274,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -300,6 +305,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,cnt
   columns.comments 
@@ -321,6 +327,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -357,6 +364,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -387,6 +395,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,cnt
   columns.comments 
@@ -408,6 +417,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -523,6 +533,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,val
   columns.comments 
@@ -546,6 +557,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,val
 columns.comments 
@@ -659,6 +671,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key1,key2,cnt
 columns.comments 
@@ -717,6 +730,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE

[26/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/bucket_num_reducers2.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/bucket_num_reducers2.q.out 
b/ql/src/test/results/clientpositive/llap/bucket_num_reducers2.q.out
new file mode 100644
index 000..79d9755
--- /dev/null
+++ b/ql/src/test/results/clientpositive/llap/bucket_num_reducers2.q.out
@@ -0,0 +1,242 @@
+PREHOOK: query: CREATE TABLE test_table(key int, value string) CLUSTERED BY 
(key) INTO 3 BUCKETS
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@test_table
+POSTHOOK: query: CREATE TABLE test_table(key int, value string) CLUSTERED BY 
(key) INTO 3 BUCKETS
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@test_table
+PREHOOK: query: explain extended insert overwrite table test_table
+  select * from src
+PREHOOK: type: QUERY
+POSTHOOK: query: explain extended insert overwrite table test_table
+  select * from src
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-2 depends on stages: Stage-1
+  Stage-0 depends on stages: Stage-2
+  Stage-3 depends on stages: Stage-0
+
+STAGE PLANS:
+  Stage: Stage-1
+Tez
+ A masked pattern was here 
+  Edges:
+Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE)
+ A masked pattern was here 
+  Vertices:
+Map 1 
+Map Operator Tree:
+TableScan
+  alias: src
+  Statistics: Num rows: 500 Data size: 89000 Basic stats: 
COMPLETE Column stats: COMPLETE
+  GatherStats: false
+  Select Operator
+expressions: key (type: string), value (type: string)
+outputColumnNames: _col0, _col1
+Statistics: Num rows: 500 Data size: 89000 Basic stats: 
COMPLETE Column stats: COMPLETE
+Reduce Output Operator
+  null sort order: 
+  sort order: 
+  Map-reduce partition columns: UDFToInteger(_col0) (type: 
int)
+  Statistics: Num rows: 500 Data size: 89000 Basic stats: 
COMPLETE Column stats: COMPLETE
+  tag: -1
+  value expressions: _col0 (type: string), _col1 (type: 
string)
+  auto parallelism: false
+Execution mode: llap
+LLAP IO: no inputs
+Path -> Alias:
+ A masked pattern was here 
+Path -> Partition:
+ A masked pattern was here 
+Partition
+  base file name: src
+  input format: org.apache.hadoop.mapred.TextInputFormat
+  output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+  properties:
+COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
+bucket_count -1
+bucketing_version 2
+column.name.delimiter ,
+columns key,value
+columns.comments 'default','default'
+columns.types string:string
+ A masked pattern was here 
+name default.src
+numFiles 1
+numRows 500
+rawDataSize 5312
+serialization.ddl struct src { string key, string value}
+serialization.format 1
+serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+totalSize 5812
+ A masked pattern was here 
+  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+input format: org.apache.hadoop.mapred.TextInputFormat
+output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+properties:
+  COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
+  bucket_count -1
+  bucketing_version 2
+  column.name.delimiter ,
+  columns key,value
+  columns.comments 'default','default'
+  columns.types string:string
+ A masked pattern was here 
+  name default.src
+  numFiles 1
+  numRows 500
+  rawDataSize 5312
+  serialization.ddl struct src { string key, string value}
+  serialization.format 1
+  serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+  totalSize 5812
+ A masked pattern was here 
+serde:

[10/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/bucket_map_join_spark4.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/spark/bucket_map_join_spark4.q.out 
b/ql/src/test/results/clientpositive/spark/bucket_map_join_spark4.q.out
index 68889d3..6ac9dd4 100644
--- a/ql/src/test/results/clientpositive/spark/bucket_map_join_spark4.q.out
+++ b/ql/src/test/results/clientpositive/spark/bucket_map_join_spark4.q.out
@@ -110,6 +110,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -133,6 +134,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -186,6 +188,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -209,6 +212,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -310,6 +314,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -333,6 +338,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -483,6 +489,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -506,6 +513,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -559,6 +567,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -582,6 +591,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -683,6 +693,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -706,6 +717,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/bucket_map_join_tez1.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/spark/bucket_map_join_tez1.q.out 
b/ql/src/test/results/clientpositive/spark/bucket_map_join_tez1.q.out
index 158d1c3..987a98c 100644
--- a/ql/src/test/results/clientpositive/spark/bucket_map_join_tez1.q.out
+++ b/ql/src/test/results/clientpositive/spark

[25/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/llap/column_table_stats.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/column_table_stats.q.out 
b/ql/src/test/results/clientpositive/llap/column_table_stats.q.out
index 7f3cc02..1457932 100644
--- a/ql/src/test/results/clientpositive/llap/column_table_stats.q.out
+++ b/ql/src/test/results/clientpositive/llap/column_table_stats.q.out
@@ -35,6 +35,7 @@ Retention:0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
numFiles1   
numRows 0   
rawDataSize 0   
@@ -102,6 +103,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 'default','default'
@@ -122,6 +124,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -211,6 +214,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
+   bucketing_version   2   
numFiles1   
numRows 500 
rawDataSize 5312
@@ -287,6 +291,7 @@ Retention:  0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
numFiles2   
numPartitions   2   
numRows 0   
@@ -383,6 +388,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -430,6 +436,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -536,6 +543,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
+   bucketing_version   2   
numFiles2   
numPartitions   2   
numRows 1000
@@ -693,6 +701,7 @@ Retention:  0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
numFiles2   
numPartitions   2   
numRows 0   
@@ -789,6 +798,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -836,6 +846,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -942,6 +953,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   

[08/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/groupby_sort_skew_1_23.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/spark/groupby_sort_skew_1_23.q.out 
b/ql/src/test/results/clientpositive/spark/groupby_sort_skew_1_23.q.out
index 9b068ed..d8b9dca 100644
--- a/ql/src/test/results/clientpositive/spark/groupby_sort_skew_1_23.q.out
+++ b/ql/src/test/results/clientpositive/spark/groupby_sort_skew_1_23.q.out
@@ -85,6 +85,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -118,6 +119,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,val
 columns.comments 
@@ -141,6 +143,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,val
   columns.comments 
@@ -172,6 +175,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -287,6 +291,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,val
 columns.comments 
@@ -310,6 +315,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,val
   columns.comments 
@@ -375,6 +381,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key1,key2,cnt
   columns.comments 
@@ -406,6 +413,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key1,key2,cnt
 columns.comments 
@@ -506,6 +514,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -539,6 +548,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,val
 columns.comments 
@@ -562,6 +572,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,val
   columns.comments 
@@ -593,6 +604,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -691,6 +703,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
   

[06/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-04-29 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/sample1.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/sample1.q.out 
b/ql/src/test/results/clientpositive/spark/sample1.q.out
index b711292..164378b 100644
--- a/ql/src/test/results/clientpositive/spark/sample1.q.out
+++ b/ql/src/test/results/clientpositive/spark/sample1.q.out
@@ -53,6 +53,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value,dt,hr
   columns.comments 
@@ -109,6 +110,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -138,6 +140,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value,dt,hr
 columns.comments 

http://git-wip-us.apache.org/repos/asf/hive/blob/ac6b2a3f/ql/src/test/results/clientpositive/spark/sample10.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/sample10.q.out 
b/ql/src/test/results/clientpositive/spark/sample10.q.out
index 418f63a..ac28779 100644
--- a/ql/src/test/results/clientpositive/spark/sample10.q.out
+++ b/ql/src/test/results/clientpositive/spark/sample10.q.out
@@ -34,10 +34,10 @@ POSTHOOK: Lineage: srcpartbucket 
PARTITION(ds=2008-04-09,hr=11).value SIMPLE [(s
 POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-09,hr=12).key SIMPLE 
[(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
 POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-09,hr=12).value SIMPLE 
[(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
 PREHOOK: query: explain extended
-select ds, count(1) from srcpartbucket tablesample (bucket 1 out of 4 on key) 
where ds is not null group by ds ORDER BY ds ASC
+select ds, count(1) from srcpartbucket tablesample (bucket 2 out of 4 on key) 
where ds is not null group by ds ORDER BY ds ASC
 PREHOOK: type: QUERY
 POSTHOOK: query: explain extended
-select ds, count(1) from srcpartbucket tablesample (bucket 1 out of 4 on key) 
where ds is not null group by ds ORDER BY ds ASC
+select ds, count(1) from srcpartbucket tablesample (bucket 2 out of 4 on key) 
where ds is not null group by ds ORDER BY ds ASC
 POSTHOOK: type: QUERY
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
@@ -59,8 +59,8 @@ STAGE PLANS:
   GatherStats: false
   Filter Operator
 isSamplingPred: true
-predicate: (((hash(key) & 2147483647) % 4) = 0) (type: 
boolean)
-sampleDesc: BUCKET 1 OUT OF 4
+predicate: (((hash(key) & 2147483647) % 4) = 1) (type: 
boolean)
+sampleDesc: BUCKET 2 OUT OF 4
 Statistics: Num rows: 20 Data size: 120 Basic stats: 
COMPLETE Column stats: NONE
 Select Operator
   expressions: ds (type: string)
@@ -87,7 +87,7 @@ STAGE PLANS:
 Path -> Partition:
  A masked pattern was here 
 Partition
-  base file name: 00_0
+  base file name: 01_0
   input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
   output format: 
org.apache.hadoop.hive.ql.io.RCFileOutputFormat
   partition values:
@@ -120,6 +120,7 @@ STAGE PLANS:
 properties:
   bucket_count 4
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -137,7 +138,7 @@ STAGE PLANS:
   name: default.srcpartbucket
  A masked pattern was here 
 Partition
-  base file name: 00_0
+  base file name: 01_0
   input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
   output format: 
org.apache.hadoop.hive.ql.io.RCFileOutputF

[2/2] hive git commit: HIVE-19311 : Partition and bucketing support for “load data” statement (Deepak Jaiswal, reviewed by Prasanth Jayachandran)

2018-05-02 Thread djaiswal
HIVE-19311 : Partition and bucketing support for “load data” statement 
(Deepak Jaiswal, reviewed by Prasanth Jayachandran)


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/46c5580b
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/46c5580b
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/46c5580b

Branch: refs/heads/master
Commit: 46c5580b735f9b3b4f3e482d8f7a524dd88d4771
Parents: e8651cb
Author: Deepak Jaiswal 
Authored: Wed May 2 01:27:02 2018 -0700
Committer: Deepak Jaiswal 
Committed: Wed May 2 01:27:02 2018 -0700

--
 data/files/load_data_job/bucketing.txt  |  118 +
 .../load_data_job/load_data_1_partition.txt |  118 +
 .../partitions/load_data_1_partition.txt|  118 +
 .../partitions/load_data_2_partitions.txt   |  118 +
 .../partitions/subdir/load_data_1_partition.txt |  118 +
 .../subdir/load_data_2_partitions.txt   |  118 +
 .../test/resources/testconfiguration.properties |1 +
 .../java/org/apache/hadoop/hive/ql/Context.java |   12 +-
 .../org/apache/hadoop/hive/ql/ErrorMsg.java |2 +
 .../apache/hadoop/hive/ql/metadata/Table.java   |6 +-
 .../hive/ql/parse/LoadSemanticAnalyzer.java |  172 +-
 .../hadoop/hive/ql/parse/SemanticAnalyzer.java  |   43 +-
 .../ql/parse/UpdateDeleteSemanticAnalyzer.java  |   43 +-
 .../apache/hadoop/hive/ql/TestTxnLoadData.java  |5 +-
 .../queries/clientnegative/load_part_nospec.q   |2 -
 .../test/queries/clientnegative/nopart_load.q   |5 -
 .../clientpositive/load_data_using_job.q|   83 +
 .../clientnegative/load_part_nospec.q.out   |9 -
 .../results/clientnegative/nopart_load.q.out|9 -
 .../llap/load_data_using_job.q.out  | 2769 ++
 20 files changed, 3747 insertions(+), 122 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hive/blob/46c5580b/data/files/load_data_job/bucketing.txt
--
diff --git a/data/files/load_data_job/bucketing.txt 
b/data/files/load_data_job/bucketing.txt
new file mode 100644
index 000..8334415
--- /dev/null
+++ b/data/files/load_data_job/bucketing.txt
@@ -0,0 +1,118 @@
+165val_165
+484val_484
+150val_150
+224val_224
+66val_66
+213val_213
+374val_374
+495val_495
+37val_37
+327val_327
+15val_15
+338val_338
+459val_459
+466val_466
+396val_396
+309val_309
+367val_367
+0val_0
+455val_455
+316val_316
+345val_345
+129val_129
+378val_378
+4val_4
+356val_356
+169val_169
+125val_125
+437val_437
+286val_286
+187val_187
+176val_176
+459val_459
+51val_51
+103val_103
+239val_239
+213val_213
+176val_176
+275val_275
+260val_260
+404val_404
+217val_217
+84val_84
+466val_466
+8val_8
+411val_411
+172val_172
+129val_129
+158val_158
+0val_0
+26val_26
+165val_165
+327val_327
+51val_51
+404val_404
+95val_95
+282val_282
+187val_187
+316val_316
+169val_169
+77val_77
+0val_0
+118val_118
+282val_282
+419val_419
+15val_15
+118val_118
+19val_19
+224val_224
+309val_309
+389val_389
+327val_327
+242val_242
+392val_392
+242val_242
+396val_396
+95val_95
+11val_11
+143val_143
+228val_228
+33val_33
+103val_103
+367val_367
+239val_239
+480val_480
+202val_202
+316val_316
+235val_235
+80val_80
+44val_44
+466val_466
+257val_257
+190val_190
+114val_114
+396val_396
+217val_217
+125val_125
+187val_187
+480val_480
+491val_491
+305val_305
+444val_444
+169val_169
+323val_323
+480val_480
+136val_136
+172val_172
+462val_462
+26val_26
+462val_462
+341val_341
+183val_183
+84val_84
+37val_37
+448val_448
+194val_194
+477val_477
+169val_169
+400val_400

http://git-wip-us.apache.org/repos/asf/hive/blob/46c5580b/data/files/load_data_job/load_data_1_partition.txt
--
diff --git a/data/files/load_data_job/load_data_1_partition.txt 
b/data/files/load_data_job/load_data_1_partition.txt
new file mode 100644
index 000..9f34604
--- /dev/null
+++ b/data/files/load_data_job/load_data_1_partition.txt
@@ -0,0 +1,118 @@
+165val_1652008-04-08
+484val_4842008-04-08
+150val_1502008-04-08
+224val_2242008-04-08
+66val_662008-04-08
+213val_2132008-04-08
+374val_3742008-04-08
+495val_4952008-04-08
+37val_372008-04-08
+327val_3272008-04-08
+15val_152008-04-08
+338val_3382008-04-08
+459val_4592008-04-08
+466val_4662008-04-08
+396val_3962008-04-08
+309val_3092008-04-08
+367val_3672008-04-08
+0val_02008-04-08
+455val_4552008-04-08
+316val_3162008-04-08
+345val_3452008-04-08
+129val_1292008-04-08
+378val_3782008-04-08
+4val_42008-04-08
+356val_3562008-04-08
+169val_1692008-04-08
+125val_1252008-04-08
+437val_4372008-04-08
+286val_2862008-04-08
+187val_1872008-04-08
+176val_

[1/2] hive git commit: HIVE-19311 : Partition and bucketing support for “load data” statement (Deepak Jaiswal, reviewed by Prasanth Jayachandran)

2018-05-02 Thread djaiswal
Repository: hive
Updated Branches:
  refs/heads/master e8651cb97 -> 46c5580b7


http://git-wip-us.apache.org/repos/asf/hive/blob/46c5580b/ql/src/test/results/clientpositive/llap/load_data_using_job.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/load_data_using_job.q.out 
b/ql/src/test/results/clientpositive/llap/load_data_using_job.q.out
new file mode 100644
index 000..8077310
--- /dev/null
+++ b/ql/src/test/results/clientpositive/llap/load_data_using_job.q.out
@@ -0,0 +1,2769 @@
+PREHOOK: query: CREATE TABLE srcbucket_mapjoin(key int, value string) 
partitioned by (ds string) STORED AS TEXTFILE
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@srcbucket_mapjoin
+POSTHOOK: query: CREATE TABLE srcbucket_mapjoin(key int, value string) 
partitioned by (ds string) STORED AS TEXTFILE
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@srcbucket_mapjoin
+PREHOOK: query: explain load data local inpath '../../data/files/bmj/00_0' 
INTO TABLE srcbucket_mapjoin partition(ds='2008-04-08')
+PREHOOK: type: LOAD
+POSTHOOK: query: explain load data local inpath 
'../../data/files/bmj/00_0' INTO TABLE srcbucket_mapjoin 
partition(ds='2008-04-08')
+POSTHOOK: type: LOAD
+STAGE DEPENDENCIES:
+  Stage-0 is a root stage
+  Stage-1 depends on stages: Stage-0
+
+STAGE PLANS:
+  Stage: Stage-0
+Move Operator
+  tables:
+  partition:
+ds 2008-04-08
+  replace: false
+  table:
+  input format: org.apache.hadoop.mapred.TextInputFormat
+  output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+  name: default.srcbucket_mapjoin
+
+  Stage: Stage-1
+Stats Work
+  Basic Stats Work:
+
+PREHOOK: query: load data local inpath '../../data/files/bmj/00_0' INTO 
TABLE srcbucket_mapjoin partition(ds='2008-04-08')
+PREHOOK: type: LOAD
+ A masked pattern was here 
+PREHOOK: Output: default@srcbucket_mapjoin
+POSTHOOK: query: load data local inpath '../../data/files/bmj/00_0' INTO 
TABLE srcbucket_mapjoin partition(ds='2008-04-08')
+POSTHOOK: type: LOAD
+ A masked pattern was here 
+POSTHOOK: Output: default@srcbucket_mapjoin
+POSTHOOK: Output: default@srcbucket_mapjoin@ds=2008-04-08
+PREHOOK: query: select * from srcbucket_mapjoin
+PREHOOK: type: QUERY
+PREHOOK: Input: default@srcbucket_mapjoin
+PREHOOK: Input: default@srcbucket_mapjoin@ds=2008-04-08
+ A masked pattern was here 
+POSTHOOK: query: select * from srcbucket_mapjoin
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@srcbucket_mapjoin
+POSTHOOK: Input: default@srcbucket_mapjoin@ds=2008-04-08
+ A masked pattern was here 
+165val_165 2008-04-08
+484val_484 2008-04-08
+150val_150 2008-04-08
+224val_224 2008-04-08
+66 val_66  2008-04-08
+213val_213 2008-04-08
+374val_374 2008-04-08
+495val_495 2008-04-08
+37 val_37  2008-04-08
+327val_327 2008-04-08
+15 val_15  2008-04-08
+338val_338 2008-04-08
+459val_459 2008-04-08
+466val_466 2008-04-08
+396val_396 2008-04-08
+309val_309 2008-04-08
+367val_367 2008-04-08
+0  val_0   2008-04-08
+455val_455 2008-04-08
+316val_316 2008-04-08
+345val_345 2008-04-08
+129val_129 2008-04-08
+378val_378 2008-04-08
+4  val_4   2008-04-08
+356val_356 2008-04-08
+169val_169 2008-04-08
+125val_125 2008-04-08
+437val_437 2008-04-08
+286val_286 2008-04-08
+187val_187 2008-04-08
+176val_176 2008-04-08
+459val_459 2008-04-08
+51 val_51  2008-04-08
+103val_103 2008-04-08
+239val_239 2008-04-08
+213val_213 2008-04-08
+176val_176 2008-04-08
+275val_275 2008-04-08
+260val_260 2008-04-08
+404val_404 2008-04-08
+217val_217 2008-04-08
+84 val_84  2008-04-08
+466val_466 2008-04-08
+8  val_8   2008-04-08
+411val_411 2008-04-08
+172val_172 2008-04-08
+129val_129 2008-04-08
+158val_158 2008-04-08
+0  val_0   2008-04-08
+26 val_26  2008-04-08
+165val_165 2008-04-08
+327val_327 2008-04-08
+51 val_51  2008-04-08
+404val_404 2008-04-08
+95 val_95  2008-04-08
+282val_282 2008-04-08
+187val_187 2008-04-08
+316val_316 2008-04-08
+169val_169 2008-04-08
+77 val_77  2008-04-08
+0  val_0   2008-04-08
+118val_118 2008-04-08
+282val_282 2008-04-08
+419val_419 2008-04-08
+15 val_15  2008-04-08
+118val_118 2008-04-08
+19 val_19  2008-04-08
+224val_224 2008-04-08
+309val_309 2008-04-08
+389val_389 2008-04-08
+327val_327 2008-04-08
+242val_242 2008-04-08
+392val_392 2008-04-08
+242val_242 2008-04-08
+396val_396 2008-04-08
+95 val_95  2008-04-08
+11 val_11  2008-04-08
+143val_143 2008-04-08
+228val_22

[25/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/column_table_stats.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/column_table_stats.q.out 
b/ql/src/test/results/clientpositive/llap/column_table_stats.q.out
index 7f3cc02..1457932 100644
--- a/ql/src/test/results/clientpositive/llap/column_table_stats.q.out
+++ b/ql/src/test/results/clientpositive/llap/column_table_stats.q.out
@@ -35,6 +35,7 @@ Retention:0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
numFiles1   
numRows 0   
rawDataSize 0   
@@ -102,6 +103,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 'default','default'
@@ -122,6 +124,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -211,6 +214,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
+   bucketing_version   2   
numFiles1   
numRows 500 
rawDataSize 5312
@@ -287,6 +291,7 @@ Retention:  0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
numFiles2   
numPartitions   2   
numRows 0   
@@ -383,6 +388,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -430,6 +436,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -536,6 +543,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
+   bucketing_version   2   
numFiles2   
numPartitions   2   
numRows 1000
@@ -693,6 +701,7 @@ Retention:  0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
numFiles2   
numPartitions   2   
numRows 0   
@@ -789,6 +798,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -836,6 +846,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -942,6 +953,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   

[29/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out
--
diff --git a/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out 
b/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out
index cb897d1..5a00d10 100644
--- a/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out
+++ b/ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out
@@ -87,6 +87,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,cnt
   columns.comments 
@@ -135,6 +136,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,val
   columns.comments 
@@ -158,6 +160,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,val
 columns.comments 
@@ -227,6 +230,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -270,6 +274,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -300,6 +305,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,cnt
   columns.comments 
@@ -321,6 +327,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -357,6 +364,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -387,6 +395,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,cnt
   columns.comments 
@@ -408,6 +417,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -523,6 +533,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,val
   columns.comments 
@@ -546,6 +557,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,val
 columns.comments 
@@ -659,6 +671,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key1,key2,cnt
 columns.comments 
@@ -717,6 +730,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE

[22/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/intersect_distinct.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/intersect_distinct.q.out 
b/ql/src/test/results/clientpositive/llap/intersect_distinct.q.out
index 44362be..fb59f10 100644
--- a/ql/src/test/results/clientpositive/llap/intersect_distinct.q.out
+++ b/ql/src/test/results/clientpositive/llap/intersect_distinct.q.out
@@ -281,315 +281,315 @@ POSTHOOK: query: select * from src intersect distinct 
select * from src
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
 POSTHOOK: Output: hdfs://### HDFS PATH ###
-0  val_0
-103val_103
+104val_104
 105val_105
-116val_116
+113val_113
+118val_118
+12 val_12
 120val_120
-129val_129
-131val_131
+133val_133
 136val_136
 155val_155
-157val_157
+158val_158
 160val_160
+162val_162
+168val_168
+17 val_17
+170val_170
+175val_175
+18 val_18
 180val_180
-183val_183
-190val_190
-191val_191
-193val_193
-205val_205
-207val_207
-213val_213
-223val_223
+186val_186
+19 val_19
+197val_197
+200val_200
+216val_216
+218val_218
+222val_222
+224val_224
+228val_228
 235val_235
-238val_238
+237val_237
 239val_239
-24 val_24
-248val_248
-260val_260
+244val_244
+247val_247
+256val_256
+263val_263
+273val_273
 28 val_28
+283val_283
 286val_286
-298val_298
-305val_305
+292val_292
 306val_306
-307val_307
 308val_308
+327val_327
+33 val_33
+348val_348
+353val_353
+362val_362
+366val_366
+396val_396
+397val_397
+401val_401
+409val_409
+411val_411
+419val_419
+427val_427
+43 val_43
+432val_432
+436val_436
+439val_439
+443val_443
+453val_453
+460val_460
+462val_462
+47 val_47
+472val_472
+485val_485
+496val_496
+54 val_54
+64 val_64
+70 val_70
+8  val_8
+83 val_83
+84 val_84
+85 val_85
+90 val_90
+0  val_0
+103val_103
+114val_114
+125val_125
+138val_138
+146val_146
+150val_150
+152val_152
+153val_153
+156val_156
+157val_157
+165val_165
+172val_172
+177val_177
+179val_179
+187val_187
+195val_195
+196val_196
+217val_217
+242val_242
+248val_248
+252val_252
+265val_265
+27 val_27
+272val_272
+280val_280
+291val_291
+305val_305
+309val_309
+311val_311
 315val_315
 317val_317
-321val_321
-33 val_33
-331val_331
+322val_322
+333val_333
+34 val_34
+345val_345
+35 val_35
 356val_356
-360val_360
-366val_366
-367val_367
+364val_364
 368val_368
+369val_369
 37 val_37
-378val_378
-379val_379
-382val_382
-389val_389
-395val_395
-400val_400
+373val_373
+377val_377
+4  val_4
 402val_402
-406val_406
-407val_407
+404val_404
+413val_413
 42 val_42
 430val_430
-436val_436
-44 val_44
+431val_431
+444val_444
+449val_449
 452val_452
+454val_454
+457val_457
+463val_463
 466val_466
-467val_467
-472val_472
-480val_480
-484val_484
-485val_485
-487val_487
-495val_495
-496val_496
-498val_498
-51 val_51
+470val_470
+475val_475
+481val_481
+489val_489
+491val_491
 57 val_57
-58 val_58
 65 val_65
 66 val_66
-69 val_69
-70 val_70
+74 val_74
+76 val_76
 78 val_78
-8  val_8
+9  val_9
 92 val_92
-96 val_96
-104val_104
-111val_111
-114val_114
-118val_118
-12 val_12
-133val_133
-138val_138
-153val_153
-156val_156
-162val_162
-168val_168
+95 val_95
+100val_100
+119val_119
+128val_128
+129val_129
+145val_145
+167val_167
 174val_174
-176val_176
 178val_178
 181val_181
-186val_186
-187val_187
-192val_192
-197val_197
+193val_193
 199val_199
-2  val_2
+20 val_20
 201val_201
-203val_203
-208val_208
+213val_213
 214val_214
-218val_218
 219val_219
-230val_230
-237val_237
+221val_221
+223val_223
+226val_226
+233val_233
+241val_241
 249val_249
-256val_256
-257val_257
+260val_260
+262val_262
 277val_277
-278val_278
-284val_284
-285val_285
+281val_281
+287val_287
+288val_288
+298val_298
+302val_302
 310val_310
-322val_322
-327val_327
-333val_333
+323val_323
 336val_336
-338val_338
 341val_341
 344val_344
-348val_348
 351val_351
-362val_362
-364val_364
-374val_374
-392val_392
-397val_397
-4  v

[31/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/druid_basic1.q.out
--
diff --git a/ql/src/test/results/clientpositive/druid_basic1.q.out 
b/ql/src/test/results/clientpositive/druid_basic1.q.out
index 25ec0ca..a5c0687 100644
--- a/ql/src/test/results/clientpositive/druid_basic1.q.out
+++ b/ql/src/test/results/clientpositive/druid_basic1.q.out
@@ -41,6 +41,7 @@ Table Type:   EXTERNAL_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"__time\":\"true\",\"added\":\"true\",\"anonymous\":\"true\",\"count\":\"true\",\"deleted\":\"true\",\"delta\":\"true\",\"language\":\"true\",\"namespace\":\"true\",\"newpage\":\"true\",\"page\":\"true\",\"robot\":\"true\",\"unpatrolled\":\"true\",\"user\":\"true\",\"variation\":\"true\"}}
EXTERNALTRUE
+   bucketing_version   2   
druid.datasourcewikipedia   
numFiles0   
numRows 0   
@@ -102,6 +103,7 @@ Table Type: EXTERNAL_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"__time\":\"true\",\"added\":\"true\",\"anonymous\":\"true\",\"count\":\"true\",\"deleted\":\"true\",\"delta\":\"true\",\"language\":\"true\",\"namespace\":\"true\",\"newpage\":\"true\",\"page\":\"true\",\"robot\":\"true\",\"unpatrolled\":\"true\",\"user\":\"true\",\"variation\":\"true\"}}
EXTERNALTRUE
+   bucketing_version   2   
druid.datasourcewikipedia   
numFiles0   
numRows 0   

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/druid_basic2.q.out
--
diff --git a/ql/src/test/results/clientpositive/druid_basic2.q.out 
b/ql/src/test/results/clientpositive/druid_basic2.q.out
index 5ed039b..0e5e957 100644
--- a/ql/src/test/results/clientpositive/druid_basic2.q.out
+++ b/ql/src/test/results/clientpositive/druid_basic2.q.out
@@ -41,6 +41,7 @@ Table Type:   EXTERNAL_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"__time\":\"true\",\"added\":\"true\",\"anonymous\":\"true\",\"count\":\"true\",\"deleted\":\"true\",\"delta\":\"true\",\"language\":\"true\",\"namespace\":\"true\",\"newpage\":\"true\",\"page\":\"true\",\"robot\":\"true\",\"unpatrolled\":\"true\",\"user\":\"true\",\"variation\":\"true\"}}
EXTERNALTRUE
+   bucketing_version   2   
druid.datasourcewikipedia   
numFiles0   
numRows 0   
@@ -266,6 +267,7 @@ STAGE PLANS:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"__time":"true","added":"true","anonymous":"true","count":"true","deleted":"true","delta":"true","language":"true","namespace":"true","newpage":"true","page":"true","robot":"true","unpatrolled":"true","user":"true","variation":"true"}}
   EXTERNAL TRUE
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns 
__time,robot,namespace,anonymous,unpatrolled,page,language,newpage,user,count,added,delta,variation,deleted
   columns.comments 'from deserializer','from deserializer','from 
deserializer','from deserializer','from deserializer','from deserializer','from 
deserializer','from deserializer','from deserializer','from deserializer','from 
deserializer','from deserializer','from deserializer','from deserializer'
@@ -294,6 +296,7 @@ STAGE PLANS:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"__time":"true","added":"true","anonymous":"true","count":"true","deleted":"true","delta":"true","language":"true","namespace":"true","newpage":"true","page":"true","robot":"true","unpatrolled":"true","user":"true","variation":"true"}}
 EXTERNAL TRUE
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns 
__time,robot,namespace,anonymous,unpatrolled,page,language,newpage,user,count,added,delta,variation,deleted
 columns.comments 'from deserializer','from deserializer','from 
deserializer','from deserializer','from deserializer','from deserializer','from 
deserializer','from deserializer','from deserializer','from deserializer','from 
deserializer','from deserializer','from deserializer','from deserializer'
@@ -444,6 +447,7 @@ STAGE PLANS:

[26/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/bucket_num_reducers2.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/bucket_num_reducers2.q.out 
b/ql/src/test/results/clientpositive/llap/bucket_num_reducers2.q.out
new file mode 100644
index 000..d0ca23c
--- /dev/null
+++ b/ql/src/test/results/clientpositive/llap/bucket_num_reducers2.q.out
@@ -0,0 +1,242 @@
+PREHOOK: query: CREATE TABLE test_table(key int, value string) CLUSTERED BY 
(key) INTO 3 BUCKETS
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@test_table
+POSTHOOK: query: CREATE TABLE test_table(key int, value string) CLUSTERED BY 
(key) INTO 3 BUCKETS
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@test_table
+PREHOOK: query: explain extended insert overwrite table test_table
+  select * from src
+PREHOOK: type: QUERY
+POSTHOOK: query: explain extended insert overwrite table test_table
+  select * from src
+POSTHOOK: type: QUERY
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-2 depends on stages: Stage-1
+  Stage-0 depends on stages: Stage-2
+  Stage-3 depends on stages: Stage-0
+
+STAGE PLANS:
+  Stage: Stage-1
+Tez
+ A masked pattern was here 
+  Edges:
+Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE)
+ A masked pattern was here 
+  Vertices:
+Map 1 
+Map Operator Tree:
+TableScan
+  alias: src
+  Statistics: Num rows: 500 Data size: 89000 Basic stats: 
COMPLETE Column stats: COMPLETE
+  GatherStats: false
+  Select Operator
+expressions: key (type: string), value (type: string)
+outputColumnNames: _col0, _col1
+Statistics: Num rows: 500 Data size: 89000 Basic stats: 
COMPLETE Column stats: COMPLETE
+Reduce Output Operator
+  null sort order: 
+  sort order: 
+  Map-reduce partition columns: UDFToInteger(_col0) (type: 
int)
+  Statistics: Num rows: 500 Data size: 89000 Basic stats: 
COMPLETE Column stats: COMPLETE
+  tag: -1
+  value expressions: _col0 (type: string), _col1 (type: 
string)
+  auto parallelism: false
+Execution mode: vectorized, llap
+LLAP IO: no inputs
+Path -> Alias:
+ A masked pattern was here 
+Path -> Partition:
+ A masked pattern was here 
+Partition
+  base file name: src
+  input format: org.apache.hadoop.mapred.TextInputFormat
+  output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+  properties:
+COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
+bucket_count -1
+bucketing_version 2
+column.name.delimiter ,
+columns key,value
+columns.comments 'default','default'
+columns.types string:string
+ A masked pattern was here 
+name default.src
+numFiles 1
+numRows 500
+rawDataSize 5312
+serialization.ddl struct src { string key, string value}
+serialization.format 1
+serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+totalSize 5812
+ A masked pattern was here 
+  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+
+input format: org.apache.hadoop.mapred.TextInputFormat
+output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+properties:
+  COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
+  bucket_count -1
+  bucketing_version 2
+  column.name.delimiter ,
+  columns key,value
+  columns.comments 'default','default'
+  columns.types string:string
+ A masked pattern was here 
+  name default.src
+  numFiles 1
+  numRows 500
+  rawDataSize 5312
+  serialization.ddl struct src { string key, string value}
+  serialization.format 1
+  serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+  totalSize 5812
+ A masked pattern was here 
+  

[30/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out
--
diff --git a/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out 
b/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out
index 4d8dd74..11ef8b1 100644
--- a/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out
+++ b/ql/src/test/results/clientpositive/groupby_sort_1_23.q.out
@@ -87,6 +87,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,cnt
   columns.comments 
@@ -135,6 +136,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,val
   columns.comments 
@@ -158,6 +160,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,val
 columns.comments 
@@ -227,6 +230,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -270,6 +274,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -300,6 +305,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,cnt
   columns.comments 
@@ -321,6 +327,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -357,6 +364,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -387,6 +395,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,cnt
   columns.comments 
@@ -408,6 +417,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -522,6 +532,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,val
   columns.comments 
@@ -545,6 +556,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,val
 columns.comments 
@@ -589,6 +601,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key1,key2,cnt
 columns.comments 
@@ -647,6 +660,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","

[35/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/keyseries/VectorKeySeriesSerializedImpl.java
--
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/keyseries/VectorKeySeriesSerializedImpl.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/keyseries/VectorKeySeriesSerializedImpl.java
index 86f466f..77c9ecc 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/keyseries/VectorKeySeriesSerializedImpl.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/keyseries/VectorKeySeriesSerializedImpl.java
@@ -21,9 +21,9 @@ package org.apache.hadoop.hive.ql.exec.vector.keyseries;
 import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch;
 import org.apache.hadoop.hive.serde2.ByteStream.Output;
 import org.apache.hadoop.hive.serde2.fast.SerializeWrite;
-import org.apache.hive.common.util.HashCodeUtil;
 
 import com.google.common.base.Preconditions;
+import org.apache.hive.common.util.Murmur3;
 
 /**
  * Implementation of base serialization interface.
@@ -103,7 +103,7 @@ public abstract class VectorKeySeriesSerializedImpl
 byte[] bytes = output.getData();
 for (int i = 0; i < nonNullKeyCount; i++) {
   keyLength = serializedKeyLengths[i];
-  hashCodes[i] = HashCodeUtil.murmurHash(bytes, offset, keyLength);
+  hashCodes[i] = Murmur3.hash32(bytes, offset, keyLength, 0);
   offset += keyLength;
 }
   }

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java
--
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java
index 1bc3fda..42b7784 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java
@@ -32,6 +32,7 @@ import 
org.apache.hadoop.hive.ql.exec.vector.expressions.VectorExpression;
 import org.apache.hadoop.hive.ql.metadata.HiveException;
 import org.apache.hadoop.hive.ql.plan.OperatorDesc;
 import org.apache.hadoop.hive.ql.plan.VectorDesc;
+import org.apache.hadoop.hive.serde2.ByteStream;
 import org.apache.hadoop.hive.serde2.ByteStream.Output;
 import 
org.apache.hadoop.hive.serde2.binarysortable.fast.BinarySortableSerializeWrite;
 import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector;
@@ -40,6 +41,7 @@ import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo;
 import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoUtils;
 
 import com.google.common.base.Preconditions;
+import org.apache.hive.common.util.Murmur3;
 
 /**
  * This class is the object hash (not Uniform Hash) operator class for native 
vectorized reduce sink.
@@ -226,61 +228,110 @@ public class VectorReduceSinkObjectHashOperator extends 
VectorReduceSinkCommonOp
   int[] selected = batch.selected;
 
   final int size = batch.size;
-  for (int logical = 0; logical < size; logical++) {
-final int batchIndex = (selectedInUse ? selected[logical] : logical);
-
-final int hashCode;
-if (isEmptyBuckets) {
-  if (isEmptyPartitions) {
-hashCode = nonPartitionRandom.nextInt();
-  } else {
+
+  // EmptyBuckets = true
+  if (isEmptyBuckets) {
+if (isEmptyPartitions) {
+  for (int logical = 0; logical< size; logical++) {
+final int batchIndex = (selectedInUse ? selected[logical] : 
logical);
+final int hashCode = nonPartitionRandom.nextInt();
+postProcess(batch, batchIndex, tag, hashCode);
+  }
+} else { // isEmptyPartition = false
+  for (int logical = 0; logical< size; logical++) {
+final int batchIndex = (selectedInUse ? selected[logical] : 
logical);
 partitionVectorExtractRow.extractRow(batch, batchIndex, 
partitionFieldValues);
-hashCode =
+final int hashCode = bucketingVersion == 2 && 
!vectorDesc.getIsAcidChange() ?
 ObjectInspectorUtils.getBucketHashCode(
+partitionFieldValues, partitionObjectInspectors) :
+ObjectInspectorUtils.getBucketHashCodeOld(
 partitionFieldValues, partitionObjectInspectors);
+postProcess(batch, batchIndex, tag, hashCode);
   }
-} else {
-  bucketVectorExtractRow.extractRow(batch, batchIndex, 
bucketFieldValues);
-  final int bucketNum =
-  ObjectInspectorUtils.getBucketNumber(
+}
+  } else { // EmptyBuckets = false
+if (isEmptyPartitions) {
+  for (int logical = 0; logical< size; logical

[08/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/groupby_sort_skew_1_23.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/spark/groupby_sort_skew_1_23.q.out 
b/ql/src/test/results/clientpositive/spark/groupby_sort_skew_1_23.q.out
index 9b068ed..d8b9dca 100644
--- a/ql/src/test/results/clientpositive/spark/groupby_sort_skew_1_23.q.out
+++ b/ql/src/test/results/clientpositive/spark/groupby_sort_skew_1_23.q.out
@@ -85,6 +85,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -118,6 +119,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,val
 columns.comments 
@@ -141,6 +143,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,val
   columns.comments 
@@ -172,6 +175,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -287,6 +291,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,val
 columns.comments 
@@ -310,6 +315,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,val
   columns.comments 
@@ -375,6 +381,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key1,key2,cnt
   columns.comments 
@@ -406,6 +413,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key1,key2,cnt
 columns.comments 
@@ -506,6 +514,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -539,6 +548,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,val
 columns.comments 
@@ -562,6 +572,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,val
   columns.comments 
@@ -593,6 +604,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,cnt
 columns.comments 
@@ -691,6 +703,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
   

[05/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/sample6.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/sample6.q.out 
b/ql/src/test/results/clientpositive/spark/sample6.q.out
index 34aee1e..cf172ea 100644
--- a/ql/src/test/results/clientpositive/spark/sample6.q.out
+++ b/ql/src/test/results/clientpositive/spark/sample6.q.out
@@ -52,6 +52,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -84,6 +85,7 @@ STAGE PLANS:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -106,6 +108,7 @@ STAGE PLANS:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -137,6 +140,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -181,172 +185,175 @@ order by key, value
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@dest1
  A masked pattern was here 
-0  val_0
-0  val_0
-0  val_0
-0  val_1
-0  val_1
-4  val_4
-4  val_5
-8  val_8
-8  val_9
-12 val_12
-12 val_12
-12 val_13
-16 val_17
-16 val_17
-20 val_20
-20 val_21
-20 val_21
-24 val_24
-24 val_24
-28 val_28
-32 val_33
+2  val_2
+2  val_3
+6  val_7
+6  val_7
+10 val_10
+10 val_11
+17 val_17
+21 val_22
+21 val_22
+21 val_22
+21 val_22
+27 val_27
+30 val_30
+30 val_31
+31 val_32
 40 val_41
 40 val_41
-44 val_44
-48 val_49
-48 val_49
-52 val_53
-52 val_53
-52 val_53
-52 val_53
-56 val_57
-60 val_61
+51 val_51
+51 val_51
+51 val_52
+57 val_57
+58 val_58
+58 val_58
+58 val_59
+58 val_59
 64 val_64
-68 val_69
-72 val_72
-72 val_72
-76 val_76
-76 val_76
-76 val_77
-76 val_77
-76 val_77
+65 val_65
+65 val_66
+65 val_66
+70 val_70
+70 val_70
+70 val_70
+70 val_71
 80 val_80
 80 val_81
-84 val_84
-84 val_84
-92 val_92
-96 val_96
-100val_100
-100val_100
-100val_101
-100val_101
-104val_104
-104val_104
-104val_105
-104val_105
-104val_105
+83 val_83
+83 val_83
+86 val_86
+86 val_87
+90 val_90
+90 val_90
+90 val_90
+91 val_92
+98 val_98
+98 val_98
+105val_105
+105val_106
+105val_106
+110val_111
+113val_113
+113val_113
 116val_116
 116val_117
-120val_120
-120val_120
-120val_121
-128val_128
-128val_128
-128val_128
-128val_129
-128val_129
-132val_133
-132val_133
-136val_136
-136val_137
-140val_141
-144val_145
-152val_152
-152val_152
-152val_153
-152val_153
-152val_153
+117val_118
+117val_118
+119val_119
+119val_119
+119val_119
+119val_120
+119val_120
+119val_120
+121val_122
+121val_122
+123val_124
+123val_124
+126val_126
+126val_127
+126val_127
+134val_134
+134val_134
+134val_135
+137val_137
+137val_137
+137val_138
+153val_153
+153val_154
+153val_154
 156val_156
 156val_157
 156val_157
-160val_160
-160val_161
-164val_164
-164val_164
-164val_165
-164val_165
-168val_168
-168val_169
-172val_172
-172val_172
-172val_173
-176val_176
-176val_176
-180val_180
+157val_157
+157val_158
+157val_158
+158val_158
+163val_163
+169val_169
+169val_169
+169val_169
+169val_169
+177val_177
+177val_178
+177val_178
+178val_178
+178val_179
+178val_179
 184val_185
-192val_192
-192val_193
-196   

[28/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/join9.q.out
--
diff --git a/ql/src/test/results/clientpositive/join9.q.out 
b/ql/src/test/results/clientpositive/join9.q.out
index 50ca52c..0a86a6e 100644
--- a/ql/src/test/results/clientpositive/join9.q.out
+++ b/ql/src/test/results/clientpositive/join9.q.out
@@ -76,6 +76,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -97,6 +98,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 'default','default'
@@ -147,6 +149,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 'default','default'
@@ -192,6 +195,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -250,6 +254,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/join_filters_overlap.q.out
--
diff --git a/ql/src/test/results/clientpositive/join_filters_overlap.q.out 
b/ql/src/test/results/clientpositive/join_filters_overlap.q.out
index db0b26d..7d4855a 100644
--- a/ql/src/test/results/clientpositive/join_filters_overlap.q.out
+++ b/ql/src/test/results/clientpositive/join_filters_overlap.q.out
@@ -92,6 +92,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -113,6 +114,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -283,6 +285,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -304,6 +307,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -474,6 +478,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -495,6 +500,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -682,6 +688,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -703,6 +710,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"}
 bucket_count -1
+bucketing_version 2
 column.name.d

[18/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out 
b/ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out
index 687b4af..8bec959 100644
--- a/ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out
@@ -232,17 +232,17 @@ POSTHOOK: query: select fl_date, count(*) from 
flights_tiny_orc group by fl_date
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@flights_tiny_orc
  A masked pattern was here 
-2010-10-20 11
 2010-10-21 12
-2010-10-26 13
-2010-10-30 11
-2010-10-23 12
 2010-10-24 12
-2010-10-25 12
-2010-10-22 11
+2010-10-26 13
 2010-10-28 12
+2010-10-22 11
 2010-10-29 12
+2010-10-20 11
+2010-10-23 12
+2010-10-25 12
 2010-10-27 11
+2010-10-30 11
 2010-10-31 8
 PREHOOK: query: explain vectorization expression
 select * from flights_tiny_orc sort by fl_num, fl_date limit 25
@@ -529,17 +529,17 @@ POSTHOOK: query: select fl_date, count(*) from 
flights_tiny_orc group by fl_date
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@flights_tiny_orc
  A masked pattern was here 
-2010-10-20 11
 2010-10-21 12
-2010-10-26 13
-2010-10-30 11
-2010-10-23 12
 2010-10-24 12
-2010-10-25 12
-2010-10-22 11
+2010-10-26 13
 2010-10-28 12
+2010-10-22 11
 2010-10-29 12
+2010-10-20 11
+2010-10-23 12
+2010-10-25 12
 2010-10-27 11
+2010-10-30 11
 2010-10-31 8
 PREHOOK: query: CREATE TABLE flights_tiny_orc_partitioned_date (
   origin_city_name STRING,
@@ -908,17 +908,17 @@ POSTHOOK: Input: 
default@flights_tiny_orc_partitioned_date@fl_date=2010-10-29
 POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-30
 POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-31
  A masked pattern was here 
-2010-10-20 11
 2010-10-21 12
-2010-10-26 13
-2010-10-30 11
-2010-10-23 12
 2010-10-24 12
-2010-10-25 12
-2010-10-22 11
+2010-10-26 13
 2010-10-28 12
+2010-10-22 11
 2010-10-29 12
+2010-10-20 11
+2010-10-23 12
+2010-10-25 12
 2010-10-27 11
+2010-10-30 11
 2010-10-31 8
 PREHOOK: query: explain vectorization expression
 select * from flights_tiny_orc_partitioned_date
@@ -1484,17 +1484,17 @@ POSTHOOK: Input: 
default@flights_tiny_orc_partitioned_date@fl_date=2010-10-29
 POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-30
 POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-31
  A masked pattern was here 
-2010-10-20 11
 2010-10-21 12
-2010-10-26 13
-2010-10-30 11
-2010-10-23 12
 2010-10-24 12
-2010-10-25 12
-2010-10-22 11
+2010-10-26 13
 2010-10-28 12
+2010-10-22 11
 2010-10-29 12
+2010-10-20 11
+2010-10-23 12
+2010-10-25 12
 2010-10-27 11
+2010-10-30 11
 2010-10-31 8
 PREHOOK: query: CREATE TABLE flights_tiny_orc_partitioned_timestamp (
   origin_city_name STRING,
@@ -1863,18 +1863,18 @@ POSTHOOK: Input: 
default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-
 POSTHOOK: Input: 
default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-30 07%3A00%3A00
 POSTHOOK: Input: 
default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-31 07%3A00%3A00
  A masked pattern was here 
-2010-10-20 07:00:0011
 2010-10-23 07:00:0012
 2010-10-25 07:00:0012
 2010-10-22 07:00:0011
+2010-10-24 07:00:0012
 2010-10-26 07:00:0013
+2010-10-20 07:00:0011
+2010-10-28 07:00:0012
 2010-10-29 07:00:0012
+2010-10-30 07:00:0011
 2010-10-31 07:00:008
 2010-10-21 07:00:0012
-2010-10-24 07:00:0012
 2010-10-27 07:00:0011
-2010-10-28 07:00:0012
-2010-10-30 07:00:0011
 PREHOOK: query: explain vectorization expression
 select * from flights_tiny_orc_partitioned_timestamp
 PREHOOK: type: QUERY
@@ -2439,18 +2439,18 @@ POSTHOOK: Input: 
default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-
 POSTHOOK: Input: 
default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-30 07%3A00%3A00
 POSTHOOK: Input: 
default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-31 07%3A00%3A00
  A masked pattern was here 
-2010-10-20 07:00:0011
 2010-10-23 07:00:0012
 2010-10-25 07:00:0012
 2010-10-22 07:00:0011
+2010-10-24 07:00:0012
 2010-10-26 07:00:0013
+2010-10-20 07:00:0011
+2010-10-28 07:00:0012
 2010-10-29 07:00:0012
+2010-10-30 07:00:0011
 2010-10-31 07:00:008
 2010-10-21 07:00:0012
-2010-10-24 07:00:0012
 2010-10-27 07:00:0011
-2010-10-28 07:00:0012
-2010-10-30 07:00:

[13/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/sample1.q.out
--
diff --git a/ql/src/test/results/clientpositive/sample1.q.out 
b/ql/src/test/results/clientpositive/sample1.q.out
index 6d58658..09b9ac7 100644
--- a/ql/src/test/results/clientpositive/sample1.q.out
+++ b/ql/src/test/results/clientpositive/sample1.q.out
@@ -55,6 +55,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value,dt,hr
 columns.comments 
@@ -126,6 +127,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 'default','default'
@@ -193,6 +195,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value,dt,hr
 columns.comments 
@@ -236,6 +239,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value,dt,hr
 columns.comments 
@@ -266,6 +270,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value,dt,hr
   columns.comments 
@@ -287,6 +292,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value,dt,hr
 columns.comments 
@@ -323,6 +329,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value,dt,hr
 columns.comments 
@@ -353,6 +360,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value,dt,hr
   columns.comments 
@@ -374,6 +382,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value,dt,hr
 columns.comments 

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/sample2.q.out
--
diff --git a/ql/src/test/results/clientpositive/sample2.q.out 
b/ql/src/test/results/clientpositive/sample2.q.out
index 164ce43..17e5491 100644
--- a/ql/src/test/results/clientpositive/sample2.q.out
+++ b/ql/src/test/results/clientpositive/sample2.q.out
@@ -54,6 +54,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -101,6 +102,7 @@ STAGE PLANS:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name

[33/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/bucket1.q.out
--
diff --git a/ql/src/test/results/clientpositive/bucket1.q.out 
b/ql/src/test/results/clientpositive/bucket1.q.out
index 3060df3..07b5b9d 100644
--- a/ql/src/test/results/clientpositive/bucket1.q.out
+++ b/ql/src/test/results/clientpositive/bucket1.q.out
@@ -52,6 +52,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -73,6 +74,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 'default','default'
@@ -112,6 +114,7 @@ STAGE PLANS:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
   bucket_count 100
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -171,6 +174,7 @@ STAGE PLANS:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count 100
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/bucket2.q.out
--
diff --git a/ql/src/test/results/clientpositive/bucket2.q.out 
b/ql/src/test/results/clientpositive/bucket2.q.out
index 16218f9..e8a1dd7 100644
--- a/ql/src/test/results/clientpositive/bucket2.q.out
+++ b/ql/src/test/results/clientpositive/bucket2.q.out
@@ -51,6 +51,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -72,6 +73,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 'default','default'
@@ -111,6 +113,7 @@ STAGE PLANS:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -178,6 +181,7 @@ STAGE PLANS:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}}
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -264,66 +268,59 @@ POSTHOOK: query: select * from bucket2_1 tablesample 
(bucket 1 out of 2) s
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@bucket2_1
  A masked pattern was here 
-0  val_0
-0  val_0
-0  val_0
 10 val_10
-100val_100
-100val_100
-104val_104
-104val_104
-114val_114
+105val_105
+113val_113
+113val_113
 116val_116
-118val_118
-118val_118
-12 val_12
-12 val_12
-120val_120
-120val_120
+119val_119
+119val_119
+119val_119
 126val_126
-128val_128
-128val_128
-128val_128
+133val_133
 134val_134
 134val_134
-136val_136
-138val_138
-138val_138
-138val_138
-138val_138
+137val_137
+137val_137
 146val_146
 146val_146
-150val_150
-152val_152
-152val_152
+149val_149
+149val_149
+153val_153
 156val_156
+157val_157
 158val_158
-160val_160
 162val_162
+163val_163
 164val_164
 164val_164
-166val_166
-168val_168
-170val_170
-172val_172
-172val_172
-174val_174
-174val_174
-176val_176
-176val_176
+165val_165
+165val_165
+169val_169
+169val_169
+

[10/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/bucket_map_join_tez1.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/spark/bucket_map_join_tez1.q.out 
b/ql/src/test/results/clientpositive/spark/bucket_map_join_tez1.q.out
index 158d1c3..987a98c 100644
--- a/ql/src/test/results/clientpositive/spark/bucket_map_join_tez1.q.out
+++ b/ql/src/test/results/clientpositive/spark/bucket_map_join_tez1.q.out
@@ -4323,110 +4323,104 @@ PREHOOK: type: QUERY
 POSTHOOK: query: explain select a.key, b.key from tab_part a join tab_part c 
on a.key = c.key join tab_part b on a.value = b.value
 POSTHOOK: type: QUERY
 STAGE DEPENDENCIES:
-  Stage-1 is a root stage
+  Stage-2 is a root stage
+  Stage-1 depends on stages: Stage-2
   Stage-0 depends on stages: Stage-1
 
 STAGE PLANS:
-  Stage: Stage-1
+  Stage: Stage-2
 Spark
-  Edges:
-Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 2), Map 4 (PARTITION-LEVEL 
SORT, 2)
-Reducer 3 <- Map 5 (PARTITION-LEVEL SORT, 2), Reducer 2 
(PARTITION-LEVEL SORT, 2)
  A masked pattern was here 
   Vertices:
-Map 1 
+Map 2 
 Map Operator Tree:
 TableScan
-  alias: a
+  alias: c
   Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
   Filter Operator
-predicate: (key is not null and value is not null) (type: 
boolean)
+predicate: key is not null (type: boolean)
 Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
 Select Operator
-  expressions: key (type: int), value (type: string)
-  outputColumnNames: _col0, _col1
+  expressions: key (type: int)
+  outputColumnNames: _col0
   Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
-  Reduce Output Operator
-key expressions: _col0 (type: int)
-sort order: +
-Map-reduce partition columns: _col0 (type: int)
-Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
-value expressions: _col1 (type: string)
+  Spark HashTable Sink Operator
+keys:
+  0 _col0 (type: int)
+  1 _col0 (type: int)
 Execution mode: vectorized
-Map 4 
+Local Work:
+  Map Reduce Local Work
+Map 3 
 Map Operator Tree:
 TableScan
-  alias: c
+  alias: b
   Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
   Filter Operator
-predicate: key is not null (type: boolean)
+predicate: value is not null (type: boolean)
 Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
 Select Operator
-  expressions: key (type: int)
-  outputColumnNames: _col0
+  expressions: key (type: int), value (type: string)
+  outputColumnNames: _col0, _col1
   Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
-  Reduce Output Operator
-key expressions: _col0 (type: int)
-sort order: +
-Map-reduce partition columns: _col0 (type: int)
-Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
+  Spark HashTable Sink Operator
+keys:
+  0 _col1 (type: string)
+  1 _col1 (type: string)
 Execution mode: vectorized
-Map 5 
+Local Work:
+  Map Reduce Local Work
+
+  Stage: Stage-1
+Spark
+ A masked pattern was here 
+  Vertices:
+Map 1 
 Map Operator Tree:
 TableScan
-  alias: b
+  alias: a
   Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
   Filter Operator
-predicate: value is not null (type: boolean)
+predicate: (key is not null and value is not null) (type: 
boolean)
 Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
 Select Operator
   expressions: key (type

[27/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out 
b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out
index 885fd18..8fa1b35 100644
--- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out
+++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out
@@ -147,6 +147,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 4
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -234,6 +235,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -282,6 +284,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -421,6 +424,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 4
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -508,6 +512,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -556,6 +561,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out 
b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out
index ff98ee5..bd7f60a 100644
--- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out
+++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out
@@ -147,6 +147,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -195,6 +196,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -283,6 +285,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 4
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -421,6 +424,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -469,6 +473,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -557,6 +562,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 4
   bucket_field_name key
+  bucketing_version 2
   colu

[04/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/smb_mapjoin_12.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/smb_mapjoin_12.q.out 
b/ql/src/test/results/clientpositive/spark/smb_mapjoin_12.q.out
index ec8b046..9b421d5 100644
--- a/ql/src/test/results/clientpositive/spark/smb_mapjoin_12.q.out
+++ b/ql/src/test/results/clientpositive/spark/smb_mapjoin_12.q.out
@@ -141,6 +141,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 16
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -181,6 +182,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 16
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -213,6 +215,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 16
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -267,7 +270,7 @@ POSTHOOK: Input: default@test_table1@ds=1
 POSTHOOK: Input: default@test_table3
 POSTHOOK: Input: default@test_table3@ds=1
  A masked pattern was here 
-879
+420
 PREHOOK: query: explain extended
 INSERT OVERWRITE TABLE test_table3 PARTITION (ds = '2') 
 SELECT /*+mapjoin(b)*/ a.key, concat(a.value, b.value) FROM test_table3 a JOIN 
test_table1 b ON a.key = b.key AND a.ds = '1' AND b.ds='1'
@@ -359,6 +362,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 16
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -399,6 +403,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 16
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -431,6 +436,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 16
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -479,4 +485,4 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table3
 POSTHOOK: Input: default@test_table3@ds=2
  A masked pattern was here 
-879
+420

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out 
b/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out
index f8d0fb2..a364374 100644
--- a/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out
+++ b/ql/src/test/results/clientpositive/spark/smb_mapjoin_13.q.out
@@ -124,6 +124,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 16
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -147,6 +148,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 16
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -273,6 +275,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 16
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -296,6 +299,7 @@ STAGE PLANS:
   SORTBUCKETCOLSPREFIX TRUE
   bucket_count 16
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -373,6 +377,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 16

[06/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/sample1.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/sample1.q.out 
b/ql/src/test/results/clientpositive/spark/sample1.q.out
index b711292..164378b 100644
--- a/ql/src/test/results/clientpositive/spark/sample1.q.out
+++ b/ql/src/test/results/clientpositive/spark/sample1.q.out
@@ -53,6 +53,7 @@ STAGE PLANS:
 properties:
   COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}}
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value,dt,hr
   columns.comments 
@@ -109,6 +110,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -138,6 +140,7 @@ STAGE PLANS:
   properties:
 COLUMN_STATS_ACCURATE 
{"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}}
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value,dt,hr
 columns.comments 

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/sample10.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/sample10.q.out 
b/ql/src/test/results/clientpositive/spark/sample10.q.out
index 418f63a..ac28779 100644
--- a/ql/src/test/results/clientpositive/spark/sample10.q.out
+++ b/ql/src/test/results/clientpositive/spark/sample10.q.out
@@ -34,10 +34,10 @@ POSTHOOK: Lineage: srcpartbucket 
PARTITION(ds=2008-04-09,hr=11).value SIMPLE [(s
 POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-09,hr=12).key SIMPLE 
[(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
 POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-09,hr=12).value SIMPLE 
[(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
 PREHOOK: query: explain extended
-select ds, count(1) from srcpartbucket tablesample (bucket 1 out of 4 on key) 
where ds is not null group by ds ORDER BY ds ASC
+select ds, count(1) from srcpartbucket tablesample (bucket 2 out of 4 on key) 
where ds is not null group by ds ORDER BY ds ASC
 PREHOOK: type: QUERY
 POSTHOOK: query: explain extended
-select ds, count(1) from srcpartbucket tablesample (bucket 1 out of 4 on key) 
where ds is not null group by ds ORDER BY ds ASC
+select ds, count(1) from srcpartbucket tablesample (bucket 2 out of 4 on key) 
where ds is not null group by ds ORDER BY ds ASC
 POSTHOOK: type: QUERY
 STAGE DEPENDENCIES:
   Stage-1 is a root stage
@@ -59,8 +59,8 @@ STAGE PLANS:
   GatherStats: false
   Filter Operator
 isSamplingPred: true
-predicate: (((hash(key) & 2147483647) % 4) = 0) (type: 
boolean)
-sampleDesc: BUCKET 1 OUT OF 4
+predicate: (((hash(key) & 2147483647) % 4) = 1) (type: 
boolean)
+sampleDesc: BUCKET 2 OUT OF 4
 Statistics: Num rows: 20 Data size: 120 Basic stats: 
COMPLETE Column stats: NONE
 Select Operator
   expressions: ds (type: string)
@@ -87,7 +87,7 @@ STAGE PLANS:
 Path -> Partition:
  A masked pattern was here 
 Partition
-  base file name: 00_0
+  base file name: 01_0
   input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
   output format: 
org.apache.hadoop.hive.ql.io.RCFileOutputFormat
   partition values:
@@ -120,6 +120,7 @@ STAGE PLANS:
 properties:
   bucket_count 4
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -137,7 +138,7 @@ STAGE PLANS:
   name: default.srcpartbucket
  A masked pattern was here 
 Partition
-  base file name: 00_0
+  base file name: 01_0
   input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
   output format: 
org.apache.hadoop.hive.ql.io.RCFileOutputF

[02/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/union_remove_2.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/union_remove_2.q.out 
b/ql/src/test/results/clientpositive/spark/union_remove_2.q.out
index f681966..18563c7 100644
--- a/ql/src/test/results/clientpositive/spark/union_remove_2.q.out
+++ b/ql/src/test/results/clientpositive/spark/union_remove_2.q.out
@@ -184,6 +184,7 @@ Retention:  0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
numFiles4   
totalSize   68  
  A masked pattern was here 

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/union_remove_20.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/union_remove_20.q.out 
b/ql/src/test/results/clientpositive/spark/union_remove_20.q.out
index 402d2b6..9f2bf0e 100644
--- a/ql/src/test/results/clientpositive/spark/union_remove_20.q.out
+++ b/ql/src/test/results/clientpositive/spark/union_remove_20.q.out
@@ -166,6 +166,7 @@ Retention:  0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
numFiles4   
totalSize   40  
  A masked pattern was here 

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/union_remove_21.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/union_remove_21.q.out 
b/ql/src/test/results/clientpositive/spark/union_remove_21.q.out
index 0df7318..6ef6d4e 100644
--- a/ql/src/test/results/clientpositive/spark/union_remove_21.q.out
+++ b/ql/src/test/results/clientpositive/spark/union_remove_21.q.out
@@ -152,6 +152,7 @@ Retention:  0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
numFiles4   
totalSize   20  
  A masked pattern was here 

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/union_remove_22.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/union_remove_22.q.out 
b/ql/src/test/results/clientpositive/spark/union_remove_22.q.out
index 967d6fe..16d52bb 100644
--- a/ql/src/test/results/clientpositive/spark/union_remove_22.q.out
+++ b/ql/src/test/results/clientpositive/spark/union_remove_22.q.out
@@ -168,6 +168,7 @@ Retention:  0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
numFiles4   
totalSize   60  
  A masked pattern was here 

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/union_remove_23.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/union_remove_23.q.out 
b/ql/src/test/results/clientpositive/spark/union_remove_23.q.out
index 19a3262..0e25378 100644
--- a/ql/src/test/results/clientpositive/spark/union_remove_23.q.out
+++ b/ql/src/test/results/clientpositive/spark/union_remove_23.q.out
@@ -221,6 +221,7 @@ Retention:  0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
numFiles4   
totalSize   40  
  A masked pattern was here 

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/union_remove_24.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/union_remove_24.q.out 
b/ql/src/test/results/clientpositive/spark/union_remove_24.q.out
index 998ffa9..16b9272 100644
--- a/ql/src/test/results/clientpositive/spark/union_remove_24.q.out
+++ b/ql/src/test/results/clientpositive/spark/union_remove_24.q.out
@@ -166,6 +166,7 @@ Retention:  0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Paramete

[19/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out 
b/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out
index 6a654d0..42d5528 100644
--- a/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out
+++ b/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out
@@ -651,314 +651,314 @@ POSTHOOK: query: select unionsrc.key, unionsrc.value 
FROM (select s1.key as key,
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@src
 POSTHOOK: Output: hdfs://### HDFS PATH ###
-0  val_0
-103val_103
-105val_105
 116val_116
+118val_118
 120val_120
-129val_129
 131val_131
-136val_136
+133val_133
+15 val_15
+150val_150
+152val_152
+153val_153
 155val_155
-157val_157
 160val_160
-180val_180
+169val_169
+175val_175
+176val_176
+178val_178
 183val_183
-190val_190
-191val_191
-193val_193
-205val_205
-207val_207
+186val_186
+194val_194
+202val_202
+203val_203
 213val_213
 223val_223
-235val_235
-238val_238
+224val_224
+226val_226
+228val_228
 239val_239
-24 val_24
-248val_248
-260val_260
+244val_244
+247val_247
+252val_252
+255val_255
+256val_256
+257val_257
+266val_266
+272val_272
+274val_274
+275val_275
 28 val_28
+284val_284
 286val_286
-298val_298
-305val_305
+289val_289
+296val_296
+302val_302
 306val_306
 307val_307
-308val_308
-315val_315
 317val_317
 321val_321
+323val_323
 33 val_33
-331val_331
-356val_356
+338val_338
+34 val_34
+35 val_35
 360val_360
-366val_366
-367val_367
-368val_368
-37 val_37
-378val_378
-379val_379
-382val_382
-389val_389
-395val_395
-400val_400
-402val_402
+365val_365
+374val_374
+377val_377
+393val_393
+399val_399
+404val_404
 406val_406
-407val_407
+409val_409
+41 val_41
+417val_417
+418val_418
 42 val_42
-430val_430
-436val_436
+431val_431
+439val_439
 44 val_44
-452val_452
-466val_466
-467val_467
-472val_472
-480val_480
-484val_484
-485val_485
-487val_487
-495val_495
-496val_496
-498val_498
-51 val_51
-57 val_57
-58 val_58
-65 val_65
+443val_443
+448val_448
+455val_455
+459val_459
+462val_462
+463val_463
+470val_470
+490val_490
+491val_491
+494val_494
+5  val_5
+53 val_53
 66 val_66
-69 val_69
-70 val_70
-78 val_78
-8  val_8
-92 val_92
-96 val_96
-104val_104
+72 val_72
+76 val_76
+77 val_77
+80 val_80
+84 val_84
+90 val_90
+10 val_10
+105val_105
 111val_111
-114val_114
-118val_118
-12 val_12
-133val_133
-138val_138
-153val_153
-156val_156
-162val_162
-168val_168
+113val_113
+119val_119
+129val_129
+136val_136
+143val_143
+149val_149
+163val_163
+167val_167
 174val_174
-176val_176
-178val_178
-181val_181
-186val_186
 187val_187
-192val_192
 197val_197
-199val_199
 2  val_2
-201val_201
-203val_203
+200val_200
 208val_208
-214val_214
+209val_209
+216val_216
+217val_217
 218val_218
-219val_219
+221val_221
 230val_230
-237val_237
-249val_249
-256val_256
-257val_257
+258val_258
+26 val_26
+263val_263
+273val_273
 277val_277
-278val_278
-284val_284
-285val_285
-310val_310
-322val_322
-327val_327
-333val_333
-336val_336
-338val_338
+281val_281
+292val_292
+30 val_30
+316val_316
+335val_335
 341val_341
+342val_342
 344val_344
-348val_348
-351val_351
+345val_345
 362val_362
-364val_364
-374val_374
+367val_367
+368val_368
+37 val_37
 392val_392
-397val_397
+394val_394
 4  val_4
-403val_403
-404val_404
-419val_419
-421val_421
-427val_427
-429val_429
+413val_413
 437val_437
-438val_438
-439val_439
-454val_454
-455val_455
+444val_444
+446val_446
+452val_452
+453val_453
+457val_457
 460val_460
-463val_463
-470val_470
-478val_478
+466val_466
+468val_468
+477val_477
 479val_479
-489val_489
-491val_491
+483val_483
+487val_487
 492val_492
-53 val_53
+493val_493
+498val_498
+51 val_51
+54 val_54
+64 val_64
+65 val_65
 67 val_67
 82 val_82
-90 val_90
-97 val_97
-10 val_10
-125val_125
-134val_134
-143val_143
+85 val_85

[16/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/vector_windowing_range_multiorder.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/llap/vector_windowing_range_multiorder.q.out
 
b/ql/src/test/results/clientpositive/llap/vector_windowing_range_multiorder.q.out
index 256b80d..9acbf2e 100644
--- 
a/ql/src/test/results/clientpositive/llap/vector_windowing_range_multiorder.q.out
+++ 
b/ql/src/test/results/clientpositive/llap/vector_windowing_range_multiorder.q.out
@@ -900,10005 +900,10005 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@over10k
  A masked pattern was here 
 s  si  i   avg_window_0
-alice falkner  323 65669   65695.76470588235
 alice falkner  477 65722   65695.76470588235
-alice falkner  455 65718   65695.76470588235
-alice falkner  481 65709   65695.76470588235
-alice falkner  345 65773   65695.76470588235
-alice falkner  280 65597   65695.76470588235
-alice falkner  500 65775   65695.76470588235
 alice falkner  339 65785   65695.76470588235
-alice falkner  452 65596   65695.76470588235
-alice falkner  382 65690   65695.76470588235
 alice falkner  382 65622   65695.76470588235
 alice falkner  393 65611   65695.76470588235
+alice falkner  389 65699   65695.76470588235
+alice falkner  500 65775   65695.76470588235
+alice falkner  371 65710   65695.76470588235
+alice falkner  481 65709   65695.76470588235
+alice falkner  452 65596   65695.76470588235
+alice falkner  323 65669   65695.76470588235
+alice falkner  382 65690   65695.76470588235
+alice falkner  455 65718   65695.76470588235
 alice falkner  393 65685   65695.76470588235
-alice falkner  342 65752   65695.76470588235
 alice falkner  311 65715   65695.76470588235
-alice falkner  371 65710   65695.76470588235
-alice falkner  389 65699   65695.76470588235
-alice ichabod  366 65590   65654.95454545454
+alice falkner  280 65597   65695.76470588235
+alice falkner  342 65752   65695.76470588235
+alice falkner  345 65773   65695.76470588235
+alice ichabod  292 65788   65654.95454545454
 alice ichabod  458 65550   65654.95454545454
+alice ichabod  412 65718   65654.95454545454
+alice ichabod  303 65692   65654.95454545454
+alice ichabod  300 65704   65654.95454545454
+alice ichabod  366 65590   65654.95454545454
+alice ichabod  301 65693   65654.95454545454
+alice ichabod  398 65659   65654.95454545454
+alice ichabod  292 65585   65654.95454545454
+alice ichabod  305 65617   65654.95454545454
 alice ichabod  436 65738   65654.95454545454
-alice ichabod  315 65772   65654.95454545454
+alice ichabod  440 65725   65654.95454545454
 alice ichabod  453 65780   65654.95454545454
+alice ichabod  398 65785   65654.95454545454
 alice ichabod  347 65547   65654.95454545454
-alice ichabod  398 65659   65654.95454545454
-alice ichabod  338 65538   65654.95454545454
-alice ichabod  440 65725   65654.95454545454
 alice ichabod  320 65622   65654.95454545454
-alice ichabod  412 65718   65654.95454545454
-alice ichabod  305 65617   65654.95454545454
-alice ichabod  292 65585   65654.95454545454
-alice ichabod  303 65692   65654.95454545454
-alice ichabod  338 65545   65654.95454545454
 alice ichabod  398 65680   65654.95454545454
+alice ichabod  315 65772   65654.95454545454
 alice ichabod  416 65536   65654.95454545454
+alice ichabod  338 65545   65654.95454545454
 alice ichabod  344 65545   65654.95454545454
-alice ichabod  300 65704   65654.95454545454
-alice ichabod  292 65788   65654.95454545454
-alice ichabod  398 65785   65654.95454545454
-alice ichabod  301 65693   65654.95454545454
-alice polk 443 65734   65661.57142857143
-alice polk 444 65564   65661.57142857143
-alice polk 357 65550   65661.57142857143
-alice polk 321 65744   65661.57142857143
-alice polk 273 65548   65661.57142857143
-alice polk 366 65595   65661.57142857143
-alice polk 285 65761   65661.57142857143
-alice polk 466 65561   65661.57142857143
-alice polk 324 65749   65661.57142857143
-alice polk 487 65746   65661.57142857143
-alice polk 378 65598   65661.57142857143
-alice polk 395 65751   65661.57142857143
-alice polk 407 65617   65661.57142857143
-alice polk 507 65744   65661.57142857143
-alice young468 65649   65706.63636363637
-alice young489 65646   65706.63636363637
-alice young286 65705   65706.63636363637
-alice young447 65789   65706.63636363637
-alice young425 65677   65706.63636363637
-alice young282 65671   65706.63636363637
-alice young351 65776   65706.63636363637
-alice young308 65776   65706.63636363637

[20/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/sysdb.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/sysdb.q.out 
b/ql/src/test/results/clientpositive/llap/sysdb.q.out
index 7d977d3..de20f2d 100644
--- a/ql/src/test/results/clientpositive/llap/sysdb.q.out
+++ b/ql/src/test/results/clientpositive/llap/sysdb.q.out
@@ -58,34 +58,10 @@ default alltypesorc hive_test_user  
USERDELETE  true-1  hive_test_user
 defaultalltypesorc hive_test_user  USERINSERT  
true-1  hive_test_user
 defaultalltypesorc hive_test_user  USERSELECT  
true-1  hive_test_user
 defaultalltypesorc hive_test_user  USERUPDATE  
true-1  hive_test_user
-defaultalltypesparquet hive_test_user  USERDELETE  
true-1  hive_test_user
-defaultalltypesparquet hive_test_user  USERINSERT  
true-1  hive_test_user
-defaultalltypesparquet hive_test_user  USERSELECT  
true-1  hive_test_user
-defaultalltypesparquet hive_test_user  USERUPDATE  
true-1  hive_test_user
-defaultcbo_t1  hive_test_user  USERDELETE  true
-1  hive_test_user
-defaultcbo_t1  hive_test_user  USERINSERT  true
-1  hive_test_user
-defaultcbo_t1  hive_test_user  USERSELECT  true
-1  hive_test_user
-defaultcbo_t1  hive_test_user  USERUPDATE  true
-1  hive_test_user
-defaultcbo_t2  hive_test_user  USERDELETE  true
-1  hive_test_user
-defaultcbo_t2  hive_test_user  USERINSERT  true
-1  hive_test_user
-defaultcbo_t2  hive_test_user  USERSELECT  true
-1  hive_test_user
-defaultcbo_t2  hive_test_user  USERUPDATE  true
-1  hive_test_user
-defaultcbo_t3  hive_test_user  USERDELETE  true
-1  hive_test_user
-defaultcbo_t3  hive_test_user  USERINSERT  true
-1  hive_test_user
-defaultcbo_t3  hive_test_user  USERSELECT  true
-1  hive_test_user
-defaultcbo_t3  hive_test_user  USERUPDATE  true
-1  hive_test_user
-defaultlineitemhive_test_user  USERDELETE  
true-1  hive_test_user
-defaultlineitemhive_test_user  USERINSERT  
true-1  hive_test_user
-defaultlineitemhive_test_user  USERSELECT  
true-1  hive_test_user
-defaultlineitemhive_test_user  USERUPDATE  
true-1  hive_test_user
 defaultmoretypes   hive_test_user  USERDELETE  
true-1  hive_test_user
 defaultmoretypes   hive_test_user  USERINSERT  
true-1  hive_test_user
 defaultmoretypes   hive_test_user  USERSELECT  
true-1  hive_test_user
 defaultmoretypes   hive_test_user  USERUPDATE  
true-1  hive_test_user
-defaultparthive_test_user  USERDELETE  true
-1  hive_test_user
-defaultparthive_test_user  USERINSERT  true
-1  hive_test_user
-defaultparthive_test_user  USERSELECT  true
-1  hive_test_user
-defaultparthive_test_user  USERUPDATE  true
-1  hive_test_user
 defaultscr_txn hive_test_user  USERDELETE  true
-1  hive_test_user
 defaultscr_txn hive_test_user  USERINSERT  true
-1  hive_test_user
 defaultscr_txn hive_test_user  USERSELECT  true
-1  hive_test_user
@@ -94,46 +70,14 @@ default src hive_test_user  USER
DELETE  true-1  hive_test_user
 defaultsrc hive_test_user  USERINSERT  true
-1  hive_test_user
 defaultsrc hive_test_user  USERSELECT  true
-1  hive_test_user
 defaultsrc hive_test_user  USERUPDATE  true
-1  hive_test_user
-defaultsrc1hive_test_user  USERDELETE  true
-1  hive_test_user
-defaultsrc1hive_test_user  USERINSERT  true
-1  hive_test_user
-defaultsrc1hive_test_user  USERSELECT  true
-1  hive_test_user
-defaultsrc1

[03/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/statsfs.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/statsfs.q.out 
b/ql/src/test/results/clientpositive/spark/statsfs.q.out
index fef2040..1af1495 100644
--- a/ql/src/test/results/clientpositive/spark/statsfs.q.out
+++ b/ql/src/test/results/clientpositive/spark/statsfs.q.out
@@ -274,6 +274,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
+   bucketing_version   2   
numFiles1   
numRows 500 
rawDataSize 5312
@@ -334,6 +335,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
+   bucketing_version   2   
numFiles1   
numRows 500 
rawDataSize 5312

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/subquery_nested_subquery.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/spark/subquery_nested_subquery.q.out 
b/ql/src/test/results/clientpositive/spark/subquery_nested_subquery.q.out
index 15f33f0..0d894e3 100644
--- a/ql/src/test/results/clientpositive/spark/subquery_nested_subquery.q.out
+++ b/ql/src/test/results/clientpositive/spark/subquery_nested_subquery.q.out
@@ -11,28 +11,28 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@part
  A masked pattern was here 
 192697 almond antique blue firebrick mint  Manufacturer#5  Brand#52
MEDIUM BURNISHED TIN31  LG DRUM 1789.69 ickly ir
+121152 almond antique burnished rose metallic  Manufacturer#1  Brand#14
PROMO PLATED TIN2   JUMBO BOX   1173.15 e pinto beans h
+121152 almond antique burnished rose metallic  Manufacturer#1  Brand#14
PROMO PLATED TIN2   JUMBO BOX   1173.15 e pinto beans h
 90681  almond antique chartreuse khaki white   Manufacturer#3  Brand#31
MEDIUM BURNISHED TIN17  SM CASE 1671.68 are slyly after the sl
-85768  almond antique chartreuse lavender yellow   Manufacturer#1  
Brand#12LARGE BRUSHED STEEL 34  SM BAG  1753.76 refull
+49671  almond antique gainsboro frosted violet Manufacturer#4  Brand#41
SMALL BRUSHED BRASS 10  SM BOX  1620.67 ccounts run quick
 42669  almond antique medium spring khaki  Manufacturer#5  Brand#51
STANDARD BURNISHED TIN  6   MED CAN 1611.66 sits haggl
-105685 almond antique violet chocolate turquoise   Manufacturer#2  
Brand#22MEDIUM ANODIZED COPPER  14  MED CAN 1690.68 ly pending requ
+112398 almond antique metallic orange dim  Manufacturer#3  Brand#32
MEDIUM BURNISHED BRASS  19  JUMBO JAR   1410.39 ole car
+110592 almond antique salmon chartreuse burlywood  Manufacturer#1  
Brand#15PROMO BURNISHED NICKEL  6   JUMBO PKG   1602.59  to the 
furiously
+155733 almond antique sky peru orange  Manufacturer#5  Brand#53SMALL 
PLATED BRASS  2   WRAP DRUM   1788.73 furiously. bra
 48427  almond antique violet mint lemonManufacturer#4  Brand#42
PROMO POLISHED STEEL39  SM CASE 1375.42 hely ironic i
 86428  almond aquamarine burnished black steel Manufacturer#1  Brand#12
STANDARD ANODIZED STEEL 28  WRAP BAG1414.42 arefully 
 15103  almond aquamarine dodger light gainsboroManufacturer#5  
Brand#53ECONOMY BURNISHED STEEL 46  LG PACK 1018.1  packages hinder 
carefu
-45261  almond aquamarine floral ivory bisque   Manufacturer#4  Brand#42
SMALL PLATED STEEL  27  WRAP CASE   1206.26 careful
-65667  almond aquamarine pink moccasin thistle Manufacturer#1  Brand#12
LARGE BURNISHED STEEL   42  JUMBO CASE  1632.66 e across the expr
-132666 almond aquamarine rose maroon antique   Manufacturer#2  Brand#24
SMALL POLISHED NICKEL   25  MED BOX 1698.66 even 
 195606 almond aquamarine sandy cyan gainsboro  Manufacturer#2  Brand#25
STANDARD PLATED TIN 18  SM PKG  1701.6  ic de
-17927  almond aquamarine yellow dodger mintManufacturer#4  Brand#41
ECONOMY BRUSHED COPPER  7   SM PKG  1844.92 ites. eve
 33357  almond azure aquamarine papaya violet   Manufacturer#4  Brand#41
STANDARD ANODIZED TIN   12  WRAP CASE   1290.35 reful
 78486  almond azure blanched chiffon midnight  Manufacturer#5  Brand#52
LARGE BRUSHED BRASS 23  MED BAG 1464.48 hel

[23/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/default_constraint.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/default_constraint.q.out 
b/ql/src/test/results/clientpositive/llap/default_constraint.q.out
index 15cbd84..d987ef2 100644
--- a/ql/src/test/results/clientpositive/llap/default_constraint.q.out
+++ b/ql/src/test/results/clientpositive/llap/default_constraint.q.out
@@ -31,6 +31,7 @@ Retention:0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
transactional   true
transactional_propertiesdefault 
  A masked pattern was here 
@@ -301,6 +302,7 @@ Retention:  0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
  A masked pattern was here 
 
 # Storage Information   
@@ -530,6 +532,7 @@ Retention:  0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
transactional   true
transactional_propertiesdefault 
  A masked pattern was here 
@@ -788,6 +791,7 @@ Retention:  0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
transactional   true
transactional_propertiesdefault 
  A masked pattern was here 
@@ -1059,6 +1063,7 @@ Retention:0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
  A masked pattern was here 
 
 # Storage Information   
@@ -1336,6 +1341,7 @@ Retention:0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
transactional   true
transactional_propertiesdefault 
  A masked pattern was here 
@@ -1490,8 +1496,9 @@ Retention:0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
numFiles1   
-   totalSize   1062
+   totalSize   1070
transactional   true
transactional_propertiesdefault 
  A masked pattern was here 
@@ -1658,9 +1665,10 @@ Retention:   0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
  A masked pattern was here 
numFiles2   
-   totalSize   2124
+   totalSize   2140
transactional   true
transactional_propertiesdefault 
  A masked pattern was here 
@@ -1737,9 +1745,10 @@ Retention:   0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
  A masked pattern was here 
numFiles2   
-   totalSize   2124
+   totalSize   2140
transactional   true
transactional_propertiesdefault 
  A masked pattern was here 
@@ -1913,9 +1922,10 @@ Retention:   0
  A masked pattern was here 
 Table Type:MANAGED_TABLE
 Table Parameters:   
+   bucketing_version   2   
  A masked pattern was here 
numFiles3   
-   totalSize   3188
+   totalSize   3199
transactional   true
transactional_propertiesdefault 
  A masked pattern was here 
@@ -1991,9 +2001,10 @@ Retention:   0
  A masked pattern was here 
 Table Type:MANAGED_TAB

[07/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/list_bucket_dml_2.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/list_bucket_dml_2.q.out 
b/ql/src/test/results/clientpositive/spark/list_bucket_dml_2.q.out
index 30dc8b5..0ee050f 100644
--- a/ql/src/test/results/clientpositive/spark/list_bucket_dml_2.q.out
+++ b/ql/src/test/results/clientpositive/spark/list_bucket_dml_2.q.out
@@ -55,6 +55,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.RCFileOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -109,6 +110,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -157,6 +159,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -189,6 +192,7 @@ STAGE PLANS:
   output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -338,6 +342,7 @@ STAGE PLANS:
   output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out 
b/ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out
index 10d12a6..71996fe 100644
--- a/ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out
+++ b/ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out
@@ -173,6 +173,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -221,6 +222,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -269,6 +271,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'
@@ -317,6 +320,7 @@ STAGE PLANS:
 output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 properties:
   bucket_count -1
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 'default','default'

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out
--
diff --git a/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out 
b/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out
index 972a7a8..7fbd873 100644
--- a/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out
+++ b/ql/src/test/results/clientpositive/spark/louter_join_ppr.q.out
@@ -62,6 +62,7 @@ STAGE PLANS:
   properties:
 COLUMN_STAT

[12/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/sample8.q.out
--
diff --git a/ql/src/test/results/clientpositive/sample8.q.out 
b/ql/src/test/results/clientpositive/sample8.q.out
index 365b7cf..3280cb5 100644
--- a/ql/src/test/results/clientpositive/sample8.q.out
+++ b/ql/src/test/results/clientpositive/sample8.q.out
@@ -86,6 +86,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 'default','default'
@@ -134,6 +135,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 'default','default'
@@ -182,6 +184,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 'default','default'
@@ -230,6 +233,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 'default','default'
@@ -319,358 +323,334 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
 POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11
 POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12
  A masked pattern was here 
-105val_105
-105val_105
-105val_105
-105val_105
-114val_114
-114val_114
-114val_114
-114val_114
-150val_150
-150val_150
-150val_150
-150val_150
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-169val_169
-178val_178
-178val_178
-178val_178
-178val_178
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
-187val_187
+149val_149
+149val_149
+149val_149
+149val_149
+149val_149
+149val_149
+149val_149
+149val_149
+149val_149
+149val_149
+149val_149
+149val_149
+149val_149
+149val_149
+149val_149
+149val_149
 196val_196
 196val_196
 196val_196
 196val_196
-2  val_2
-2  val_2
-2  val_2
-2  val_2
-213val_213
-213val_213
-213val_213
-213val_213
-213val_213
-213val_213
-213val_213
-213val_213
-213val_213
-213val_213
-213val_213
-213val_213
-213val_213
-213val_213
-213val_213
-213val_213
-222val_222
-222val_222
-222val_222
-222val_222
-277val_277
-277val_277
-277val_277
-277val_277
-277val_277
-277val_277
-277val_277
-277val_277
-277val_277
-277val_277
-277val_277
-277val_277
-277val_277
-277val_277
-277val_277
-277val_277
-277val_277
-277val_277
-277val_277
-277val_277
-277val_277
-277val_277
-277val_277
-277val_277
-277val

[24/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/cte_1.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/cte_1.q.out 
b/ql/src/test/results/clientpositive/llap/cte_1.q.out
index d7bc062..d98862c 100644
--- a/ql/src/test/results/clientpositive/llap/cte_1.q.out
+++ b/ql/src/test/results/clientpositive/llap/cte_1.q.out
@@ -159,178 +159,217 @@ POSTHOOK: Input: default@src
 10 val_10  val_10
 100val_100 val_100
 100val_100 val_100
+103val_103 val_103
+103val_103 val_103
 104val_104 val_104
 104val_104 val_104
-111val_111 val_111
-114val_114 val_114
-116val_116 val_116
+105val_105 val_105
+11 val_11  val_11
 126val_126 val_126
 128val_128 val_128
 128val_128 val_128
 128val_128 val_128
-131val_131 val_131
+129val_129 val_129
+129val_129 val_129
+133val_133 val_133
 134val_134 val_134
 134val_134 val_134
+136val_136 val_136
+143val_143 val_143
+149val_149 val_149
+149val_149 val_149
 152val_152 val_152
 152val_152 val_152
-155val_155 val_155
+153val_153 val_153
+157val_157 val_157
+162val_162 val_162
+166val_166 val_166
 169val_169 val_169
 169val_169 val_169
 169val_169 val_169
 169val_169 val_169
-17 val_17  val_17
 170val_170 val_170
-172val_172 val_172
-172val_172 val_172
-178val_178 val_178
-179val_179 val_179
-179val_179 val_179
-189val_189 val_189
-19 val_19  val_19
-192val_192 val_192
-194val_194 val_194
-195val_195 val_195
-195val_195 val_195
-199val_199 val_199
-199val_199 val_199
-199val_199 val_199
-20 val_20  val_20
-207val_207 val_207
-207val_207 val_207
+174val_174 val_174
+174val_174 val_174
+176val_176 val_176
+176val_176 val_176
+193val_193 val_193
+193val_193 val_193
+193val_193 val_193
+196val_196 val_196
+201val_201 val_201
+205val_205 val_205
+205val_205 val_205
+209val_209 val_209
+209val_209 val_209
 213val_213 val_213
 213val_213 val_213
 218val_218 val_218
+221val_221 val_221
+221val_221 val_221
 223val_223 val_223
 223val_223 val_223
-229val_229 val_229
-229val_229 val_229
 233val_233 val_233
 233val_233 val_233
-235val_235 val_235
+239val_239 val_239
+239val_239 val_239
 24 val_24  val_24
 24 val_24  val_24
-256val_256 val_256
-256val_256 val_256
-257val_257 val_257
+241val_241 val_241
+242val_242 val_242
+242val_242 val_242
+247val_247 val_247
+248val_248 val_248
 26 val_26  val_26
 26 val_26  val_26
-265val_265 val_265
-265val_265 val_265
-266val_266 val_266
-27 val_27  val_27
-278val_278 val_278
-278val_278 val_278
-280val_280 val_280
-280val_280 val_280
-284val_284 val_284
-291val_291 val_291
-30 val_30  val_30
+263val_263 val_263
+272val_272 val_272
+272val_272 val_272
+275val_275 val_275
+285val_285 val_285
+296val_296 val_296
+298val_298 val_298
+298val_298 val_298
+298val_298 val_298
 302val_302 val_302
-305val_305 val_305
-316val_316 val_316
-316val_316 val_316
-316val_316 val_316
-321val_321 val_321
-321val_321 val_321
+307val_307 val_307
+307val_307 val_307
+310val_310 val_310
 323val_323 val_323
-33 val_33  val_33
+325val_325 val_325
+325val_325 val_325
 333val_333 val_333
 333val_333 val_333
+335val_335 val_335
 338val_338 val_338
-344val_344 val_344
-344val_344 val_344
-353val_353 val_353
-353val_353 val_353
-356val_356 val_356
-364val_364 val_364
-368val_368 val_368
-369val_369 val_369
-369val_369 val_369
-369val_369 val_369
-37 val_37  val_37
-37 val_37  val_37
+360val_360 val_360
+362val_362 val_362
+367val_367 val_367
+367val_367 val_367
+375val_375 val_375
 377val_377 val_377
-378val_378 val_378
-386val_386 val_386
+384val_384 val_384
+384val_384 val_384
+384val_384 val_384
 392val_392 val_392
-397val_397 val_397
-397val_397 val_397
-404val_404 val_404
-404val_404 val_404
-407val_407 val_407
-411val_411 val_411
-418val_418 val_418
+399val_399 val_399
+399val_399 val_399
+401val_401 val_401
+401val_401 val_401
+401val_401 val_401
+401val_401 val_401
+401val_401 val_401
+402val_402 val_402
+403val_403 val_403
+403val_403 val_403
+403val_403 val_403
+413val_413 val_413
+413val_413 val_413
+417val_417 val_417
+417val_417 val_417
+417val_417 val_417
 419val_419 val_419
-424val_424 val_424
-424val_424 val_424
-430val_430 val_430
-430val_430 val_430
-430val_430 val_430
-432val_432 val_432
-436val_436 val_436
-

[11/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/sort_merge_join_desc_5.q.out
--
diff --git a/ql/src/test/results/clientpositive/sort_merge_join_desc_5.q.out 
b/ql/src/test/results/clientpositive/sort_merge_join_desc_5.q.out
index 5754a74..38d2b96 100644
--- a/ql/src/test/results/clientpositive/sort_merge_join_desc_5.q.out
+++ b/ql/src/test/results/clientpositive/sort_merge_join_desc_5.q.out
@@ -128,6 +128,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 1
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/sort_merge_join_desc_6.q.out
--
diff --git a/ql/src/test/results/clientpositive/sort_merge_join_desc_6.q.out 
b/ql/src/test/results/clientpositive/sort_merge_join_desc_6.q.out
index 23cb97a..64bf678 100644
--- a/ql/src/test/results/clientpositive/sort_merge_join_desc_6.q.out
+++ b/ql/src/test/results/clientpositive/sort_merge_join_desc_6.q.out
@@ -101,6 +101,7 @@ STAGE PLANS:
 properties:
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -212,6 +213,7 @@ STAGE PLANS:
 SORTBUCKETCOLSPREFIX TRUE
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/sort_merge_join_desc_7.q.out
--
diff --git a/ql/src/test/results/clientpositive/sort_merge_join_desc_7.q.out 
b/ql/src/test/results/clientpositive/sort_merge_join_desc_7.q.out
index 426f63f..77f7792 100644
--- a/ql/src/test/results/clientpositive/sort_merge_join_desc_7.q.out
+++ b/ql/src/test/results/clientpositive/sort_merge_join_desc_7.q.out
@@ -137,6 +137,7 @@ STAGE PLANS:
 properties:
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -184,6 +185,7 @@ STAGE PLANS:
 properties:
   bucket_count 2
   bucket_field_name key
+  bucketing_version 2
   column.name.delimiter ,
   columns key,value
   columns.comments 
@@ -286,6 +288,7 @@ STAGE PLANS:
   properties:
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -335,6 +338,7 @@ STAGE PLANS:
   properties:
 bucket_count 2
 bucket_field_name key
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/alter_merge_stats_orc.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/spark/alter_merge_stats_orc.q.out 
b/ql/src/test/results/clientpositive/spark/alter_merge_stats_orc.q.out
index 2b069ec..9337387 100644
--- a/ql/src/test/results/clientpositive/spark/alter_merge_stats_orc.q.out
+++ b/ql/src/test/results/clientpositive/spark/alter_merge_stats_orc.q.out
@@ -89,6 +89,7 @@ Retention:0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
+   bucketing_version   2   
numFiles3   
numRows 1500
rawDataSize 141000  
@@ -139,6 +140,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
+   bucketing_version   2   
numFiles1   
numRows 1500
rawDataSize 141000  

http://git-wip-us.apache.org

[32/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out 
b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out
index b59c4bc..49598d1 100644
--- a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out
+++ b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_1.q.out
@@ -128,7 +128,7 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table2
 POSTHOOK: Input: default@test_table2@ds=1
  A masked pattern was here 
-247
+243
 PREHOOK: query: select count(*) from test_table2 tablesample (bucket 2 out of 
2) s where ds = '1'
 PREHOOK: type: QUERY
 PREHOOK: Input: default@test_table2
@@ -139,7 +139,7 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table2
 POSTHOOK: Input: default@test_table2@ds=1
  A masked pattern was here 
-253
+257
 PREHOOK: query: EXPLAIN
 INSERT OVERWRITE TABLE test_table2 PARTITION (ds = '1')
 SELECT * from 
@@ -238,7 +238,7 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table2
 POSTHOOK: Input: default@test_table2@ds=1
  A masked pattern was here 
-247
+243
 PREHOOK: query: select count(*) from test_table2 tablesample (bucket 2 out of 
2) s where ds = '1'
 PREHOOK: type: QUERY
 PREHOOK: Input: default@test_table2
@@ -249,7 +249,7 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table2
 POSTHOOK: Input: default@test_table2@ds=1
  A masked pattern was here 
-253
+257
 PREHOOK: query: EXPLAIN
 INSERT OVERWRITE TABLE test_table2 PARTITION (ds = '1')
 SELECT x.key, concat(x.value, x.value) from 

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out 
b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out
index 52ef3db..d1c11f4 100644
--- a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out
+++ b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_3.q.out
@@ -120,7 +120,7 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table2
 POSTHOOK: Input: default@test_table2@ds=1
  A masked pattern was here 
-247
+243
 PREHOOK: query: select count(*) from test_table2 tablesample (bucket 2 out of 
2) s where ds = '1'
 PREHOOK: type: QUERY
 PREHOOK: Input: default@test_table2
@@ -131,7 +131,7 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table2
 POSTHOOK: Input: default@test_table2@ds=1
  A masked pattern was here 
-253
+257
 PREHOOK: query: CREATE TABLE test_table3 (key INT, value STRING) PARTITIONED 
BY (ds STRING)
 CLUSTERED BY (value) SORTED BY (value) INTO 2 BUCKETS
 PREHOOK: type: CREATETABLE

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out 
b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out
index eaf85c3..72e918f 100644
--- a/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out
+++ b/ql/src/test/results/clientpositive/bucketsortoptimize_insert_4.q.out
@@ -199,18 +199,7 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table3
 POSTHOOK: Input: default@test_table3@ds=1
  A masked pattern was here 
-0  0   val_0val_0  1
-0  0   val_0val_0  1
-0  0   val_0val_0  1
-0  0   val_0val_0  1
-0  0   val_0val_0  1
-0  0   val_0val_0  1
-0  0   val_0val_0  1
-0  0   val_0val_0  1
-0  0   val_0val_0  1
 2  2   val_2val_2  1
-4  4   val_4val_4  1
-8  8   val_8val_8  1
 PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s 
where ds = '1'
 PREHOOK: type: QUERY
 PREHOOK: Input: default@test_table3
@@ -221,6 +210,16 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table3
 POSTHOOK: Input: default@test_table3@ds=1
  A masked pattern was here 
+0  0   val_0val_0  1
+0  0   val_0val_0  1
+0  0   val_0val_0  1
+0  0   val_0val_0  1
+0  0   val_0val_0  1
+0  0   val_0val_0  1
+0  0   val_0val_0  1
+0  0   val_0val_0  1
+0  0   val_0val_0  1
+4  4   val_4val_4  1
 5  5   val_5val_5  1
 5  5   val_5val_5  1
 5  5   val_5val_5  1
@@ -230,6 +229,7 @@ POSTHOOK: Input: default@test_table3@ds=1
 5  5   val_5val_5  1
 5  5   val_5val_5  1
 5  5   val_5val_5  1
+8  8 

[14/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/pcr.q.out
--
diff --git a/ql/src/test/results/clientpositive/pcr.q.out 
b/ql/src/test/results/clientpositive/pcr.q.out
index 0adee06..e30f6d1 100644
--- a/ql/src/test/results/clientpositive/pcr.q.out
+++ b/ql/src/test/results/clientpositive/pcr.q.out
@@ -116,6 +116,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -163,6 +164,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -306,6 +308,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -353,6 +356,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -400,6 +404,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -581,6 +586,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -628,6 +634,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -773,6 +780,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -820,6 +828,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -967,6 +976,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -1014,6 +1024,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -1061,6 +1072,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -1219,6 +1231,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
 columns key,value
 columns.comments 
@@ -1266,6 +1279,7 @@ STAGE PLANS:
   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
   properties:
 bucket_count -1
+bucketing_version 2
 column.name.delimiter ,
   

[21/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/orc_create.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/orc_create.q.out 
b/ql/src/test/results/clientpositive/llap/orc_create.q.out
index 72d94e5..df286b2 100644
--- a/ql/src/test/results/clientpositive/llap/orc_create.q.out
+++ b/ql/src/test/results/clientpositive/llap/orc_create.q.out
@@ -66,6 +66,7 @@ Retention:0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"lst\":\"true\",\"mp\":\"true\",\"str\":\"true\",\"strct\":\"true\"}}
+   bucketing_version   2   
numFiles0   
numRows 0   
rawDataSize 0   
@@ -119,6 +120,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
+   bucketing_version   2   
numFiles0   
numPartitions   0   
numRows 0   
@@ -176,6 +178,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
+   bucketing_version   2   
numFiles0   
numPartitions   0   
numRows 0   
@@ -223,6 +226,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
+   bucketing_version   2   
  A masked pattern was here 
numFiles0   
numPartitions   0   
@@ -281,6 +285,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
+   bucketing_version   2   
numFiles0   
numPartitions   0   
numRows 0   
@@ -336,6 +341,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"lst\":\"true\",\"mp\":\"true\",\"str\":\"true\",\"strct\":\"true\"}}
+   bucketing_version   2   
numFiles0   
numRows 0   
rawDataSize 0   

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/orc_llap_counters.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/orc_llap_counters.q.out 
b/ql/src/test/results/clientpositive/llap/orc_llap_counters.q.out
index af237ea..98d9e06 100644
--- a/ql/src/test/results/clientpositive/llap/orc_llap_counters.q.out
+++ b/ql/src/test/results/clientpositive/llap/orc_llap_counters.q.out
@@ -227,6 +227,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
+   bucketing_version   2   
numFiles1   
numRows 2100
orc.bloom.filter.columns*   

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/orc_llap_counters1.q.out
--
diff --git a/ql/src/test/results/clientpositive/llap/orc_llap_counters1.q.out 
b/ql/src/test/results/clientpositive/llap/orc_llap_counters1.q.out
index 01fb4af..c078326 100644
--- a/ql/src/test/results/clientpositive/llap/orc_llap_counters1.q.out
+++ b/ql/src/test/results/clientpositive/llap/orc_llap_counters1.q.out
@@ -227,6 +227,7 @@ Retention:  0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
+   bucketing_version   2   
numFiles1   
numRows 2100
orc.bloom.filter.columns*   

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/ll

[09/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/spark/bucketsortoptimize_insert_2.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/spark/bucketsortoptimize_insert_2.q.out 
b/ql/src/test/results/clientpositive/spark/bucketsortoptimize_insert_2.q.out
index 3e8ea6f..9b14b92 100644
--- a/ql/src/test/results/clientpositive/spark/bucketsortoptimize_insert_2.q.out
+++ b/ql/src/test/results/clientpositive/spark/bucketsortoptimize_insert_2.q.out
@@ -193,18 +193,7 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table3
 POSTHOOK: Input: default@test_table3@ds=1
  A masked pattern was here 
-0  val_0val_0  1
-0  val_0val_0  1
-0  val_0val_0  1
-0  val_0val_0  1
-0  val_0val_0  1
-0  val_0val_0  1
-0  val_0val_0  1
-0  val_0val_0  1
-0  val_0val_0  1
 2  val_2val_2  1
-4  val_4val_4  1
-8  val_8val_8  1
 PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s 
where ds = '1'
 PREHOOK: type: QUERY
 PREHOOK: Input: default@test_table3
@@ -215,6 +204,16 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table3
 POSTHOOK: Input: default@test_table3@ds=1
  A masked pattern was here 
+0  val_0val_0  1
+0  val_0val_0  1
+0  val_0val_0  1
+0  val_0val_0  1
+0  val_0val_0  1
+0  val_0val_0  1
+0  val_0val_0  1
+0  val_0val_0  1
+0  val_0val_0  1
+4  val_4val_4  1
 5  val_5val_5  1
 5  val_5val_5  1
 5  val_5val_5  1
@@ -224,6 +223,7 @@ POSTHOOK: Input: default@test_table3@ds=1
 5  val_5val_5  1
 5  val_5val_5  1
 5  val_5val_5  1
+8  val_8val_8  1
 9  val_9val_9  1
 PREHOOK: query: EXPLAIN
 INSERT OVERWRITE TABLE test_table3 PARTITION (ds = '1')
@@ -344,6 +344,18 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table3
 POSTHOOK: Input: default@test_table3@ds=1
  A masked pattern was here 
+2  val_2val_2  1
+2  val_2val_2  1
+PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s 
where ds = '1'
+PREHOOK: type: QUERY
+PREHOOK: Input: default@test_table3
+PREHOOK: Input: default@test_table3@ds=1
+ A masked pattern was here 
+POSTHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s 
where ds = '1'
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@test_table3
+POSTHOOK: Input: default@test_table3@ds=1
+ A masked pattern was here 
 0  val_0val_0  1
 0  val_0val_0  1
 0  val_0val_0  1
@@ -362,22 +374,8 @@ POSTHOOK: Input: default@test_table3@ds=1
 0  val_0val_0  1
 0  val_0val_0  1
 0  val_0val_0  1
-2  val_2val_2  1
-2  val_2val_2  1
 4  val_4val_4  1
 4  val_4val_4  1
-8  val_8val_8  1
-8  val_8val_8  1
-PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s 
where ds = '1'
-PREHOOK: type: QUERY
-PREHOOK: Input: default@test_table3
-PREHOOK: Input: default@test_table3@ds=1
- A masked pattern was here 
-POSTHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s 
where ds = '1'
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@test_table3
-POSTHOOK: Input: default@test_table3@ds=1
- A masked pattern was here 
 5  val_5val_5  1
 5  val_5val_5  1
 5  val_5val_5  1
@@ -396,6 +394,8 @@ POSTHOOK: Input: default@test_table3@ds=1
 5  val_5val_5  1
 5  val_5val_5  1
 5  val_5val_5  1
+8  val_8val_8  1
+8  val_8val_8  1
 9  val_9val_9  1
 9  val_9val_9  1
 PREHOOK: query: EXPLAIN
@@ -517,6 +517,18 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@test_table3
 POSTHOOK: Input: default@test_table3@ds=1
  A masked pattern was here 
+2  val_2val_2  1
+2  val_2val_2  1
+PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s 
where ds = '1'
+PREHOOK: type: QUERY
+PREHOOK: Input: default@test_table3
+PREHOOK: Input: default@test_table3@ds=1
+ A masked pattern was here 
+POSTHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s 
where ds = '1'
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@test_table3
+POSTHOOK: Input: default@test_table3@ds=1
+ A masked pattern was here 
 0  val_0val_0  1
 0  val_0val_0  1
 0  val_0val_0  1
@@ -535,22 +547,8 @@ POSTHOOK: Input: default@test_table3@ds=1
 0  val_0val_0  1
 0  val_0val_0  1
 0  val_0val_0  1
-2  val_2val_2  1
-2  val_2val_2  1
 4  val_4val_4  1
 4  val_4val_4  1
-8  val_8val_8  1
-8  val_8val_8  1
-PREHOOK: query: select * from test_table3 tablesample (bucket 2 out of 2) s 
where ds = '1'
-PREHOOK: type: QUERY
-P

[17/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/llap/vector_ptf_part_simple.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/llap/vector_ptf_part_simple.q.out 
b/ql/src/test/results/clientpositive/llap/vector_ptf_part_simple.q.out
index 9aeb650..ec8611d 100644
--- a/ql/src/test/results/clientpositive/llap/vector_ptf_part_simple.q.out
+++ b/ql/src/test/results/clientpositive/llap/vector_ptf_part_simple.q.out
@@ -309,46 +309,46 @@ POSTHOOK: type: QUERY
 POSTHOOK: Input: default@vector_ptf_part_simple_orc
  A masked pattern was here 
 p_mfgr p_name  p_retailprice   rn  r   dr  fv  lv  c   
cs
-Manufacturer#4 almond azure aquamarine papaya violet   1290.35 1   1   
1   1290.35 1206.26 4   6
-Manufacturer#4 almond antique violet mint lemon1375.42 2   1   
1   1290.35 1206.26 4   6
-Manufacturer#4 almond aquamarine floral ivory bisque   NULL3   1   
1   1290.35 1206.26 4   6
-Manufacturer#4 almond antique gainsboro frosted violet NULL4   1   
1   1290.35 1206.26 4   6
-Manufacturer#4 almond aquamarine yellow dodger mint1844.92 5   1   
1   1290.35 1206.26 4   6
-Manufacturer#4 almond aquamarine floral ivory bisque   1206.26 6   1   
1   1290.35 1206.26 4   6
-Manufacturer#5 almond azure blanched chiffon midnight  1464.48 1   1   
1   1464.48 1788.73 6   6
-Manufacturer#5 almond aquamarine dodger light gainsboro1018.1  2   
1   1   1464.48 1788.73 6   6
-Manufacturer#5 almond antique medium spring khaki  1611.66 3   1   
1   1464.48 1788.73 6   6
-Manufacturer#5 almond antique blue firebrick mint  1789.69 4   1   
1   1464.48 1788.73 6   6
-Manufacturer#5 almond antique medium spring khaki  1611.66 5   1   
1   1464.48 1788.73 6   6
-Manufacturer#5 almond antique sky peru orange  1788.73 6   1   1   
1464.48 1788.73 6   6
-Manufacturer#2 almond aquamarine rose maroon antique   900.66  1   1   
1   900.66  1800.7  8   8
-Manufacturer#2 almond aquamarine rose maroon antique   1698.66 2   1   
1   900.66  1800.7  8   8
-Manufacturer#2 almond antique violet turquoise frosted 1800.7  3   1   
1   900.66  1800.7  8   8
-Manufacturer#2 almond antique violet chocolate turquoise   1690.68 4   
1   1   900.66  1800.7  8   8
-Manufacturer#2 almond antique violet turquoise frosted 1800.7  5   1   
1   900.66  1800.7  8   8
-Manufacturer#2 almond aquamarine sandy cyan gainsboro  1000.6  6   1   
1   900.66  1800.7  8   8
-Manufacturer#2 almond aquamarine midnight light salmon 2031.98 7   1   
1   900.66  1800.7  8   8
-Manufacturer#2 almond antique violet turquoise frosted 1800.7  8   1   
1   900.66  1800.7  8   8
-Manufacturer#1 almond antique chartreuse lavender yellow   1753.76 1   
1   1   1753.76 1632.66 11  12
-Manufacturer#1 almond aquamarine pink moccasin thistle 1632.66 2   1   
1   1753.76 1632.66 11  12
-Manufacturer#1 almond aquamarine pink moccasin thistle 1632.66 3   1   
1   1753.76 1632.66 11  12
-Manufacturer#1 almond antique chartreuse lavender yellow   1753.76 4   
1   1   1753.76 1632.66 11  12
-Manufacturer#1 almond aquamarine burnished black steel 1414.42 5   1   
1   1753.76 1632.66 11  12
-Manufacturer#1 almond antique burnished rose metallic  1173.15 6   1   
1   1753.76 1632.66 11  12
-Manufacturer#1 almond antique salmon chartreuse burlywood  1602.59 7   
1   1   1753.76 1632.66 11  12
-Manufacturer#1 almond antique chartreuse lavender yellow   1753.76 8   
1   1   1753.76 1632.66 11  12
-Manufacturer#1 almond antique burnished rose metallic  1173.15 9   1   
1   1753.76 1632.66 11  12
-Manufacturer#1 almond antique chartreuse lavender yellow   1753.76 10  
1   1   1753.76 1632.66 11  12
-Manufacturer#1 almond aquamarine pink moccasin thistle NULL11  1   
1   1753.76 1632.66 11  12
-Manufacturer#1 almond aquamarine pink moccasin thistle 1632.66 12  1   
1   1753.76 1632.66 11  12
-Manufacturer#3 almond antique forest lavender goldenrod590.27  1   
1   1   590.27  99.68   7   8
-Manufacturer#3 almond antique metallic orange dim  55.39   2   1   
1   590.27  99.68   7   8
-Manufacturer#3 almond antique forest lavender goldenrod1190.27 3   
1   1   590.27  99.68   7   8
-Manufacturer#3 almond antique olive coral navajo   1337.29 4   1   
1   590.27  99.68   7   8
-Manufacturer#3 almond antique m

[34/36] hive git commit: HIVE-18910 : Migrate to Murmur hash for shuffle and bucketing (Deepak Jaiswal, reviewed by Jasone Dere)

2018-05-02 Thread djaiswal
http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientnegative/alter_file_format.q.out
--
diff --git a/ql/src/test/results/clientnegative/alter_file_format.q.out 
b/ql/src/test/results/clientnegative/alter_file_format.q.out
index 96259e5..3033cbe 100644
--- a/ql/src/test/results/clientnegative/alter_file_format.q.out
+++ b/ql/src/test/results/clientnegative/alter_file_format.q.out
@@ -24,6 +24,7 @@ Retention:0
 Table Type:MANAGED_TABLE
 Table Parameters:   
COLUMN_STATS_ACCURATE   
{\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}}
+   bucketing_version   2   
numFiles0   
numRows 0   
rawDataSize 0   

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out
--
diff --git 
a/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out 
b/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out
index 8e240ea..90c3d02 100644
--- 
a/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out
+++ 
b/ql/src/test/results/clientnegative/alter_view_as_select_with_partition.q.out
@@ -51,6 +51,7 @@ Retention:0
 Table Type:VIRTUAL_VIEW 
 Table Parameters:   
COLUMN_STATS_ACCURATE   {\"BASIC_STATS\":\"true\"}
+   bucketing_version   2   
numFiles0   
numPartitions   2   
numRows 0   

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientnegative/unset_table_property.q.out
--
diff --git a/ql/src/test/results/clientnegative/unset_table_property.q.out 
b/ql/src/test/results/clientnegative/unset_table_property.q.out
index 2dd047a..eb308eb 100644
--- a/ql/src/test/results/clientnegative/unset_table_property.q.out
+++ b/ql/src/test/results/clientnegative/unset_table_property.q.out
@@ -20,6 +20,7 @@ POSTHOOK: query: SHOW TBLPROPERTIES testTable
 POSTHOOK: type: SHOW_TBLPROPERTIES
 COLUMN_STATS_ACCURATE  
{"BASIC_STATS":"true","COLUMN_STATS":{"col1":"true","col2":"true"}}
 a  1
+bucketing_version  2
 c  3
  A masked pattern was here 
 numFiles   0

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientnegative/unset_view_property.q.out
--
diff --git a/ql/src/test/results/clientnegative/unset_view_property.q.out 
b/ql/src/test/results/clientnegative/unset_view_property.q.out
index ae7f4f6..b5038d9 100644
--- a/ql/src/test/results/clientnegative/unset_view_property.q.out
+++ b/ql/src/test/results/clientnegative/unset_view_property.q.out
@@ -21,6 +21,7 @@ PREHOOK: query: SHOW TBLPROPERTIES testView
 PREHOOK: type: SHOW_TBLPROPERTIES
 POSTHOOK: query: SHOW TBLPROPERTIES testView
 POSTHOOK: type: SHOW_TBLPROPERTIES
+bucketing_version  2
  A masked pattern was here 
 propA  100
 propB  200

http://git-wip-us.apache.org/repos/asf/hive/blob/54651c78/ql/src/test/results/clientpositive/acid_mapjoin.q.out
--
diff --git a/ql/src/test/results/clientpositive/acid_mapjoin.q.out 
b/ql/src/test/results/clientpositive/acid_mapjoin.q.out
index 5569a03..76a781e 100644
--- a/ql/src/test/results/clientpositive/acid_mapjoin.q.out
+++ b/ql/src/test/results/clientpositive/acid_mapjoin.q.out
@@ -73,21 +73,21 @@ STAGE PLANS:
   Stage: Stage-5
 Map Reduce Local Work
   Alias -> Map Local Tables:
-$hdt$_1:acid2 
+$hdt$_0:acid1 
   Fetch Operator
 limit: -1
   Alias -> Map Local Operator Tree:
-$hdt$_1:acid2 
+$hdt$_0:acid1 
   TableScan
-alias: acid2
-Statistics: Num rows: 210 Data size: 840 Basic stats: COMPLETE 
Column stats: NONE
+alias: acid1
+Statistics: Num rows: 316 Data size: 1265 Basic stats: COMPLETE 
Column stats: NONE
 Filter Operator
   predicate: key is not null (type: boolean)
-  Statistics: Num rows: 210 Data size: 840 Basic stats: COMPLETE 
Column stats: NONE
+  Statistics: Num rows: 316 Data size: 1265 Basic stats: COMPLETE 
Column stats: NONE
   Select Operator
 expressions: key (type: int)
 outputColumnNames: _col0
-Statistics: Num rows: 210 Data size: 840 Basic stats: COMPLETE 
Column stats: NONE
+  

  1   2   3   >