Repository: hive Updated Branches: refs/heads/master 7c4599baa -> cc8893e48
http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/BucketingSortingOpProcFactory.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/BucketingSortingOpProcFactory.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/BucketingSortingOpProcFactory.java index 9159120..391cfda 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/BucketingSortingOpProcFactory.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/BucketingSortingOpProcFactory.java @@ -166,7 +166,6 @@ public class BucketingSortingOpProcFactory { } String sortOrder = rsDesc.getOrder(); - String nullSortOrder = rsDesc.getNullOrder(); List<ExprNodeDesc> keyCols = rsDesc.getKeyCols(); List<ExprNodeDesc> valCols = ExprNodeDescUtils.backtrack(joinValues, jop, parent); @@ -187,8 +186,7 @@ public class BucketingSortingOpProcFactory { newSortCols[keyIndex].addAlias(vname, vindex); } else { newBucketCols[keyIndex] = new BucketCol(vname, vindex); - newSortCols[keyIndex] = new SortCol(vname, vindex, sortOrder.charAt(keyIndex), - nullSortOrder.charAt(keyIndex)); + newSortCols[keyIndex] = new SortCol(vname, vindex, sortOrder.charAt(keyIndex)); } } } @@ -313,8 +311,7 @@ public class BucketingSortingOpProcFactory { int sortIndex = indexOfColName(sortCols, columnExpr.getColumn()); if (sortIndex != -1) { if (newSortCols[sortIndex] == null) { - newSortCols[sortIndex] = new SortCol(sortCols.get(sortIndex).getSortOrder(), - sortCols.get(sortIndex).getNullSortOrder()); + newSortCols[sortIndex] = new SortCol(sortCols.get(sortIndex).getSortOrder()); } newSortCols[sortIndex].addAlias( colInfos.get(colInfosIndex).getInternalName(), colInfosIndex); @@ -439,7 +436,7 @@ public class BucketingSortingOpProcFactory { private static List<SortCol> getNewSortCols(List<SortCol> sortCols, List<ColumnInfo> colInfos) { List<SortCol> newSortCols = new ArrayList<SortCol>(sortCols.size()); for (int i = 0; i < sortCols.size(); i++) { - SortCol sortCol = new SortCol(sortCols.get(i).getSortOrder(), sortCols.get(i).getNullSortOrder()); + SortCol sortCol = new SortCol(sortCols.get(i).getSortOrder()); for (Integer index : sortCols.get(i).getIndexes()) { // The only time this condition should be false is in the case of dynamic partitioning if (index < colInfos.size()) { @@ -540,7 +537,6 @@ public class BucketingSortingOpProcFactory { static List<SortCol> extractSortCols(ReduceSinkOperator rop, List<ExprNodeDesc> outputValues) { String sortOrder = rop.getConf().getOrder(); - String nullSortOrder = rop.getConf().getNullOrder(); List<SortCol> sortCols = new ArrayList<SortCol>(); ArrayList<ExprNodeDesc> keyCols = rop.getConf().getKeyCols(); for (int i = 0; i < keyCols.size(); i++) { @@ -553,7 +549,7 @@ public class BucketingSortingOpProcFactory { break; } sortCols.add(new SortCol(((ExprNodeColumnDesc) keyCol).getColumn(), index, - sortOrder.charAt(i), nullSortOrder.charAt(i))); + sortOrder.charAt(i))); } // If the sorted columns can't all be found in the values then the data is only sorted on // the columns seen up until now @@ -654,7 +650,6 @@ public class BucketingSortingOpProcFactory { GroupByDesc groupByDesc = gop.getConf(); String sortOrder = rop.getConf().getOrder(); - String nullSortOrder = rop.getConf().getNullOrder(); List<BucketCol> bucketCols = new ArrayList<BucketCol>(); List<SortCol> sortCols = new ArrayList<SortCol>(); assert rop.getConf().getKeyCols().size() <= rop.getSchema().getSignature().size(); @@ -665,7 +660,7 @@ public class BucketingSortingOpProcFactory { } String colName = rop.getSchema().getSignature().get(i).getInternalName(); bucketCols.add(new BucketCol(colName, i)); - sortCols.add(new SortCol(colName, i, sortOrder.charAt(i), nullSortOrder.charAt(i))); + sortCols.add(new SortCol(colName, i, sortOrder.charAt(i))); } bctx.setBucketedCols(rop, bucketCols); bctx.setSortedCols(rop, sortCols); http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java index f10a40a..b36a9a0 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java @@ -653,7 +653,7 @@ public abstract class BaseSemanticAnalyzer { return colList; } - protected List<Order> getColumnNamesOrder(ASTNode ast) { + protected List<Order> getColumnNamesOrder(ASTNode ast) throws SemanticException { List<Order> colList = new ArrayList<Order>(); int numCh = ast.getChildCount(); for (int i = 0; i < numCh; i++) { @@ -662,19 +662,19 @@ public abstract class BaseSemanticAnalyzer { child = (ASTNode) child.getChild(0); if (child.getToken().getType() == HiveParser.TOK_NULLS_FIRST) { colList.add(new Order(unescapeIdentifier(child.getChild(0).getText()).toLowerCase(), - HIVE_COLUMN_ORDER_ASC, HIVE_COLUMN_NULLS_FIRST)); + HIVE_COLUMN_ORDER_ASC)); } else { - colList.add(new Order(unescapeIdentifier(child.getChild(0).getText()).toLowerCase(), - HIVE_COLUMN_ORDER_ASC, HIVE_COLUMN_NULLS_LAST)); + throw new SemanticException("create/alter table: " + + "not supported NULLS LAST for ORDER BY in ASC order"); } } else { child = (ASTNode) child.getChild(0); if (child.getToken().getType() == HiveParser.TOK_NULLS_LAST) { colList.add(new Order(unescapeIdentifier(child.getChild(0).getText()).toLowerCase(), - HIVE_COLUMN_ORDER_DESC, HIVE_COLUMN_NULLS_LAST)); + HIVE_COLUMN_ORDER_DESC)); } else { - colList.add(new Order(unescapeIdentifier(child.getChild(0).getText()).toLowerCase(), - HIVE_COLUMN_ORDER_DESC, HIVE_COLUMN_NULLS_FIRST)); + throw new SemanticException("create/alter table: " + + "not supported NULLS FIRST for ORDER BY in DESC order"); } } } http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java index 3e91e10..9ab091d 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java @@ -117,7 +117,6 @@ import org.apache.hadoop.hive.ql.metadata.Partition; import org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient; import org.apache.hadoop.hive.ql.metadata.Table; import org.apache.hadoop.hive.ql.metadata.VirtualColumn; -import org.apache.hadoop.hive.ql.optimizer.ColumnPruner; import org.apache.hadoop.hive.ql.optimizer.Optimizer; import org.apache.hadoop.hive.ql.optimizer.Transform; import org.apache.hadoop.hive.ql.optimizer.calcite.CalciteSemanticException; @@ -6358,7 +6357,6 @@ public class SemanticAnalyzer extends BaseSemanticAnalyzer { (dest_tab.getSortCols().size() > 0)) { sortCols = getSortCols(dest, qb, dest_tab, table_desc, input, true); sortOrders = getSortOrders(dest, qb, dest_tab, input); - nullSortOrders = getNullSortOrders(dest, qb, dest_tab, input); if (!enforceBucketing && !dest_tab.isIndexTable()) { throw new SemanticException(ErrorMsg.TBL_SORTED_NOT_BUCKETED.getErrorCodedMsg(dest_tab.getCompleteName())); } else { @@ -6393,12 +6391,10 @@ public class SemanticAnalyzer extends BaseSemanticAnalyzer { } StringBuilder order = new StringBuilder(); + StringBuilder nullOrder = new StringBuilder(); for (int sortOrder : sortOrders) { order.append(sortOrder == BaseSemanticAnalyzer.HIVE_COLUMN_ORDER_ASC ? '+' : '-'); - } - StringBuilder nullOrder = new StringBuilder(); - for (int pos : nullSortOrders) { - nullOrder.append(pos == BaseSemanticAnalyzer.HIVE_COLUMN_NULLS_FIRST ? 'a' : 'z'); + nullOrder.append(sortOrder == BaseSemanticAnalyzer.HIVE_COLUMN_ORDER_ASC ? 'a' : 'z'); } input = genReduceSinkPlan(input, partnCols, sortCols, order.toString(), nullOrder.toString(), maxReducers, (AcidUtils.isAcidTable(dest_tab) ? getAcidType() : AcidUtils.Operation.NOT_ACID)); @@ -7362,23 +7358,6 @@ public class SemanticAnalyzer extends BaseSemanticAnalyzer { return orders; } - private ArrayList<Integer> getNullSortOrders(String dest, QB qb, Table tab, Operator input) - throws SemanticException { - List<Order> tabSortCols = tab.getSortCols(); - List<FieldSchema> tabCols = tab.getCols(); - - ArrayList<Integer> orders = new ArrayList<Integer>(); - for (Order sortCol : tabSortCols) { - for (FieldSchema tabCol : tabCols) { - if (sortCol.getCol().equals(tabCol.getName())) { - orders.add(sortCol.getNullOrder()); - break; - } - } - } - return orders; - } - private Operator genReduceSinkPlan(String dest, QB qb, Operator<?> input, int numReducers, boolean hasOrderBy) throws SemanticException { http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/org/apache/hadoop/hive/ql/txn/compactor/TestWorker.java ---------------------------------------------------------------------- diff --git a/ql/src/test/org/apache/hadoop/hive/ql/txn/compactor/TestWorker.java b/ql/src/test/org/apache/hadoop/hive/ql/txn/compactor/TestWorker.java index 7f6430f..cf7eb70 100644 --- a/ql/src/test/org/apache/hadoop/hive/ql/txn/compactor/TestWorker.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/txn/compactor/TestWorker.java @@ -224,7 +224,7 @@ public class TestWorker extends CompactorTest { @Test public void sortedTable() throws Exception { List<Order> sortCols = new ArrayList<Order>(1); - sortCols.add(new Order("b", 1, 0)); + sortCols.add(new Order("b", 1)); Table t = newTable("default", "st", false, new HashMap<String, String>(), sortCols, false); @@ -249,7 +249,7 @@ public class TestWorker extends CompactorTest { @Test public void sortedPartition() throws Exception { List<Order> sortCols = new ArrayList<Order>(1); - sortCols.add(new Order("b", 1, 0)); + sortCols.add(new Order("b", 1)); Table t = newTable("default", "sp", true, new HashMap<String, String>(), sortCols, false); Partition p = newPartition(t, "today", sortCols); http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table2_h23.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table2_h23.q.out b/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table2_h23.q.out index 9eea7f7..3b71598 100644 --- a/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table2_h23.q.out +++ b/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table2_h23.q.out @@ -232,7 +232,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 8 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:0, nullOrder:1)] +Sort Columns: [Order(col:key, order:0)] Storage Desc Params: serialization.format 1 PREHOOK: query: INSERT OVERWRITE TABLE tst1 PARTITION (ds = '1') SELECT key, value FROM src @@ -281,7 +281,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 8 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:0, nullOrder:1)] +Sort Columns: [Order(col:key, order:0)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test changing the bucket columns @@ -326,7 +326,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 8 Bucket Columns: [value] -Sort Columns: [Order(col:key, order:0, nullOrder:1)] +Sort Columns: [Order(col:key, order:0)] Storage Desc Params: serialization.format 1 PREHOOK: query: INSERT OVERWRITE TABLE tst1 PARTITION (ds = '1') SELECT key, value FROM src @@ -375,7 +375,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 8 Bucket Columns: [value] -Sort Columns: [Order(col:key, order:0, nullOrder:1)] +Sort Columns: [Order(col:key, order:0)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test changing the number of buckets @@ -420,7 +420,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [value] -Sort Columns: [Order(col:key, order:0, nullOrder:1)] +Sort Columns: [Order(col:key, order:0)] Storage Desc Params: serialization.format 1 PREHOOK: query: INSERT OVERWRITE TABLE tst1 PARTITION (ds = '1') SELECT key, value FROM src @@ -469,7 +469,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [value] -Sort Columns: [Order(col:key, order:0, nullOrder:1)] +Sort Columns: [Order(col:key, order:0)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test changing the sort columns @@ -514,7 +514,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [value] -Sort Columns: [Order(col:value, order:0, nullOrder:1)] +Sort Columns: [Order(col:value, order:0)] Storage Desc Params: serialization.format 1 PREHOOK: query: INSERT OVERWRITE TABLE tst1 PARTITION (ds = '1') SELECT key, value FROM src @@ -563,7 +563,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [value] -Sort Columns: [Order(col:value, order:0, nullOrder:1)] +Sort Columns: [Order(col:value, order:0)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test changing the sort order @@ -608,7 +608,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [value] -Sort Columns: [Order(col:value, order:1, nullOrder:0)] +Sort Columns: [Order(col:value, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: INSERT OVERWRITE TABLE tst1 PARTITION (ds = '1') SELECT key, value FROM src @@ -657,7 +657,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [value] -Sort Columns: [Order(col:value, order:1, nullOrder:0)] +Sort Columns: [Order(col:value, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test a sorted partition gets converted to unsorted http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table_h23.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table_h23.q.out b/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table_h23.q.out index af2f47a..cab3de4 100644 --- a/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table_h23.q.out +++ b/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table_h23.q.out @@ -408,7 +408,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 12 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test changing sort order @@ -455,7 +455,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 12 Bucket Columns: [key] -Sort Columns: [Order(col:value, order:0, nullOrder:1)] +Sort Columns: [Order(col:value, order:0)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test removing test order http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/alter_partition_clusterby_sortby.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/alter_partition_clusterby_sortby.q.out b/ql/src/test/results/clientpositive/alter_partition_clusterby_sortby.q.out index 46d5b34..184d2e4 100644 --- a/ql/src/test/results/clientpositive/alter_partition_clusterby_sortby.q.out +++ b/ql/src/test/results/clientpositive/alter_partition_clusterby_sortby.q.out @@ -110,7 +110,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [b] -Sort Columns: [Order(col:b, order:0, nullOrder:1)] +Sort Columns: [Order(col:b, order:0)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Turn off clustering for a partition @@ -200,7 +200,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [a, b] -Sort Columns: [Order(col:a, order:0, nullOrder:1), Order(col:b, order:1, nullOrder:0)] +Sort Columns: [Order(col:a, order:0), Order(col:b, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: drop table alter_table_partition_clusterby_sortby http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/alter_table_not_sorted.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/alter_table_not_sorted.q.out b/ql/src/test/results/clientpositive/alter_table_not_sorted.q.out index 8a88bc4..6e1ec59 100644 --- a/ql/src/test/results/clientpositive/alter_table_not_sorted.q.out +++ b/ql/src/test/results/clientpositive/alter_table_not_sorted.q.out @@ -34,7 +34,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [a] -Sort Columns: [Order(col:a, order:1, nullOrder:0)] +Sort Columns: [Order(col:a, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: alter table alter_table_not_sorted not sorted http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/authorization_index.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/authorization_index.q.out b/ql/src/test/results/clientpositive/authorization_index.q.out index a6ec998..adc02ad 100644 --- a/ql/src/test/results/clientpositive/authorization_index.q.out +++ b/ql/src/test/results/clientpositive/authorization_index.q.out @@ -41,7 +41,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat Compressed: No Num Buckets: -1 Bucket Columns: [] -Sort Columns: [Order(col:a, order:1, nullOrder:0)] +Sort Columns: [Order(col:a, order:1)] PREHOOK: query: alter index t1_index on t1 rebuild PREHOOK: type: ALTERINDEX_REBUILD PREHOOK: Input: default@t1 http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/bucket5.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/bucket5.q.out b/ql/src/test/results/clientpositive/bucket5.q.out index aa47ba8..fa926aa 100644 --- a/ql/src/test/results/clientpositive/bucket5.q.out +++ b/ql/src/test/results/clientpositive/bucket5.q.out @@ -535,7 +535,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 2 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: SELECT * FROM bucketed_table TABLESAMPLE (BUCKET 1 OUT OF 2) s LIMIT 10 http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/dynpart_sort_opt_vectorization.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/dynpart_sort_opt_vectorization.q.out b/ql/src/test/results/clientpositive/dynpart_sort_opt_vectorization.q.out index c5d695e..d03bfe4 100644 --- a/ql/src/test/results/clientpositive/dynpart_sort_opt_vectorization.q.out +++ b/ql/src/test/results/clientpositive/dynpart_sort_opt_vectorization.q.out @@ -1155,7 +1155,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: desc formatted over1k_part_buck_sort_orc partition(t="__HIVE_DEFAULT_PARTITION__") @@ -1196,7 +1196,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: select count(*) from over1k_part_orc @@ -2129,7 +2129,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: desc formatted over1k_part_buck_sort2_orc partition(t="__HIVE_DEFAULT_PARTITION__") @@ -2170,7 +2170,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: explain select * from over1k_part_buck_sort2_orc @@ -2339,7 +2339,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: desc formatted over1k_part_buck_sort2_orc partition(t="__HIVE_DEFAULT_PARTITION__") @@ -2380,7 +2380,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: explain select * from over1k_part_buck_sort2_orc http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/dynpart_sort_optimization.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/dynpart_sort_optimization.q.out b/ql/src/test/results/clientpositive/dynpart_sort_optimization.q.out index 9a222b1..857d609 100644 --- a/ql/src/test/results/clientpositive/dynpart_sort_optimization.q.out +++ b/ql/src/test/results/clientpositive/dynpart_sort_optimization.q.out @@ -1060,7 +1060,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: desc formatted over1k_part_buck_sort partition(t="__HIVE_DEFAULT_PARTITION__") @@ -1101,7 +1101,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: select count(*) from over1k_part @@ -2027,7 +2027,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: desc formatted over1k_part_buck_sort2 partition(t="__HIVE_DEFAULT_PARTITION__") @@ -2068,7 +2068,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: select * from over1k_part_buck_sort2 @@ -2170,7 +2170,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: desc formatted over1k_part_buck_sort2 partition(t="__HIVE_DEFAULT_PARTITION__") @@ -2211,7 +2211,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: select * from over1k_part_buck_sort2 http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/index_skewtable.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/index_skewtable.q.out b/ql/src/test/results/clientpositive/index_skewtable.q.out index 45e6834..972789d 100644 --- a/ql/src/test/results/clientpositive/index_skewtable.q.out +++ b/ql/src/test/results/clientpositive/index_skewtable.q.out @@ -57,7 +57,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat Compressed: No Num Buckets: -1 Bucket Columns: [] -Sort Columns: [Order(col:value, order:1, nullOrder:0)] +Sort Columns: [Order(col:value, order:1)] PREHOOK: query: ALTER INDEX kv_index ON kv REBUILD PREHOOK: type: ALTERINDEX_REBUILD PREHOOK: Input: default@kv http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/infer_bucket_sort.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/infer_bucket_sort.q.out b/ql/src/test/results/clientpositive/infer_bucket_sort.q.out index 709ca10..182dd3a 100644 --- a/ql/src/test/results/clientpositive/infer_bucket_sort.q.out +++ b/ql/src/test/results/clientpositive/infer_bucket_sort.q.out @@ -62,7 +62,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test group by where a key isn't selected, should not be bucketed or sorted @@ -168,7 +168,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test join with two keys, should be bucketed and sorted by join keys @@ -221,7 +221,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key, value] -Sort Columns: [Order(col:key, order:1, nullOrder:0), Order(col:value, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1), Order(col:value, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test join with two keys and only one selected, should not be bucketed or sorted @@ -327,7 +327,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test join on three tables on different keys, should be bucketed and sorted by latter key @@ -380,7 +380,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test distribute by, should only be bucketed by key @@ -486,7 +486,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: -1 Bucket Columns: [] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test sort by desc, should be sorted by key @@ -539,7 +539,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: -1 Bucket Columns: [] -Sort Columns: [Order(col:key, order:0, nullOrder:1)] +Sort Columns: [Order(col:key, order:0)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test cluster by, should be bucketed and sorted by key @@ -592,7 +592,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test distribute by and sort by different keys, should be bucketed by one key sorted by the other @@ -645,7 +645,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key] -Sort Columns: [Order(col:value, order:1, nullOrder:0)] +Sort Columns: [Order(col:value, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test join in simple subquery, should be bucketed and sorted on key @@ -698,7 +698,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test join in simple subquery renaming key column, should be bucketed and sorted on key @@ -751,7 +751,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test group by in simple subquery, should be bucketed and sorted on key @@ -804,7 +804,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test group by in simple subquery renaming key column, should be bucketed and sorted on key @@ -857,7 +857,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test group by in subquery with where outside, should still be bucketed and sorted on key @@ -910,7 +910,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test group by in subquery with expression on value, should still be bucketed and sorted on key @@ -963,7 +963,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test group by in subquery with lateral view outside, should still be bucketed and sorted @@ -1016,7 +1016,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test group by in subquery with another group by outside, should be bucketed and sorted by the @@ -1071,7 +1071,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [value] -Sort Columns: [Order(col:value, order:1, nullOrder:0)] +Sort Columns: [Order(col:value, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test group by in subquery with select on outside reordering the columns, should be bucketed and @@ -1126,7 +1126,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [value] -Sort Columns: [Order(col:value, order:1, nullOrder:0)] +Sort Columns: [Order(col:value, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test group by in subquery followed by distribute by, should only be bucketed by the distribute key @@ -1179,7 +1179,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test group by in subquery followed by sort by, should only be sorted by the sort key @@ -1232,7 +1232,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test group by in subquery followed by transform script, should not be bucketed or sorted @@ -1338,6 +1338,6 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key, value] -Sort Columns: [Order(col:key, order:1, nullOrder:0), Order(col:value, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1), Order(col:value, order:1)] Storage Desc Params: serialization.format 1 http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/infer_bucket_sort_bucketed_table.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/infer_bucket_sort_bucketed_table.q.out b/ql/src/test/results/clientpositive/infer_bucket_sort_bucketed_table.q.out index 6db9428..33d795b 100644 --- a/ql/src/test/results/clientpositive/infer_bucket_sort_bucketed_table.q.out +++ b/ql/src/test/results/clientpositive/infer_bucket_sort_bucketed_table.q.out @@ -64,7 +64,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 3 Bucket Columns: [value] -Sort Columns: [Order(col:value, order:1, nullOrder:0)] +Sort Columns: [Order(col:value, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- If the count(*) from sampling the buckets matches the count(*) from each file, the table is http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/infer_bucket_sort_convert_join.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/infer_bucket_sort_convert_join.q.out b/ql/src/test/results/clientpositive/infer_bucket_sort_convert_join.q.out index e5626a4..2f7e538 100644 --- a/ql/src/test/results/clientpositive/infer_bucket_sort_convert_join.q.out +++ b/ql/src/test/results/clientpositive/infer_bucket_sort_convert_join.q.out @@ -123,6 +123,6 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/infer_bucket_sort_dyn_part.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/infer_bucket_sort_dyn_part.q.out b/ql/src/test/results/clientpositive/infer_bucket_sort_dyn_part.q.out index 5f983c7..7e3b48f 100644 --- a/ql/src/test/results/clientpositive/infer_bucket_sort_dyn_part.q.out +++ b/ql/src/test/results/clientpositive/infer_bucket_sort_dyn_part.q.out @@ -302,7 +302,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.RCFileOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: DESCRIBE FORMATTED test_table PARTITION (ds='2008-04-08', hr='12') @@ -342,7 +342,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.RCFileOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: CREATE TABLE srcpart_merge_dp LIKE srcpart http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/infer_bucket_sort_grouping_operators.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/infer_bucket_sort_grouping_operators.q.out b/ql/src/test/results/clientpositive/infer_bucket_sort_grouping_operators.q.out index 90a71ae..ebfce60 100644 --- a/ql/src/test/results/clientpositive/infer_bucket_sort_grouping_operators.q.out +++ b/ql/src/test/results/clientpositive/infer_bucket_sort_grouping_operators.q.out @@ -144,7 +144,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key, value] -Sort Columns: [Order(col:key, order:1, nullOrder:0), Order(col:value, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1), Order(col:value, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test rollup, should be bucketed and sorted on key, value, grouping_key @@ -203,7 +203,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key, value, grouping_key] -Sort Columns: [Order(col:key, order:1, nullOrder:0), Order(col:value, order:1, nullOrder:0), Order(col:grouping_key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1), Order(col:value, order:1), Order(col:grouping_key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test cube, should not be bucketed or sorted because its missing the grouping ID @@ -328,7 +328,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key, value] -Sort Columns: [Order(col:key, order:1, nullOrder:0), Order(col:value, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1), Order(col:value, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test cube, should be bucketed and sorted on key, value, grouping_key @@ -387,7 +387,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key, value, grouping_key] -Sort Columns: [Order(col:key, order:1, nullOrder:0), Order(col:value, order:1, nullOrder:0), Order(col:grouping_key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1), Order(col:value, order:1), Order(col:grouping_key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test grouping sets, should not be bucketed or sorted because its missing the grouping ID @@ -512,7 +512,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key, value] -Sort Columns: [Order(col:key, order:1, nullOrder:0), Order(col:value, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1), Order(col:value, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test grouping sets, should be bucketed and sorted on key, value, grouping_key @@ -571,6 +571,6 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key, value, grouping_key] -Sort Columns: [Order(col:key, order:1, nullOrder:0), Order(col:value, order:1, nullOrder:0), Order(col:grouping_key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1), Order(col:value, order:1), Order(col:grouping_key, order:1)] Storage Desc Params: serialization.format 1 http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/infer_bucket_sort_map_operators.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/infer_bucket_sort_map_operators.q.out b/ql/src/test/results/clientpositive/infer_bucket_sort_map_operators.q.out index 5ebcf41..4ca6517 100644 --- a/ql/src/test/results/clientpositive/infer_bucket_sort_map_operators.q.out +++ b/ql/src/test/results/clientpositive/infer_bucket_sort_map_operators.q.out @@ -361,7 +361,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [value] -Sort Columns: [Order(col:value, order:1, nullOrder:0)] +Sort Columns: [Order(col:value, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test SMB join doesn't affect inference, should not be bucketed or sorted @@ -640,6 +640,6 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/infer_bucket_sort_merge.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/infer_bucket_sort_merge.q.out b/ql/src/test/results/clientpositive/infer_bucket_sort_merge.q.out index 97cb92d..53407c5 100644 --- a/ql/src/test/results/clientpositive/infer_bucket_sort_merge.q.out +++ b/ql/src/test/results/clientpositive/infer_bucket_sort_merge.q.out @@ -117,6 +117,6 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 2 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/infer_bucket_sort_multi_insert.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/infer_bucket_sort_multi_insert.q.out b/ql/src/test/results/clientpositive/infer_bucket_sort_multi_insert.q.out index e59026e..d4c22f4 100644 --- a/ql/src/test/results/clientpositive/infer_bucket_sort_multi_insert.q.out +++ b/ql/src/test/results/clientpositive/infer_bucket_sort_multi_insert.q.out @@ -172,7 +172,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: DESCRIBE FORMATTED test_table PARTITION (part = '2') @@ -211,7 +211,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [value] -Sort Columns: [Order(col:value, order:1, nullOrder:0)] +Sort Columns: [Order(col:value, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- The first partition should be bucketed and sorted, the second should not @@ -272,7 +272,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: DESCRIBE FORMATTED test_table PARTITION (part = '2') @@ -372,7 +372,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: DESCRIBE FORMATTED test_table PARTITION (part = '2') @@ -411,6 +411,6 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/infer_bucket_sort_reducers_power_two.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/infer_bucket_sort_reducers_power_two.q.out b/ql/src/test/results/clientpositive/infer_bucket_sort_reducers_power_two.q.out index 91466f1..1e4db29 100644 --- a/ql/src/test/results/clientpositive/infer_bucket_sort_reducers_power_two.q.out +++ b/ql/src/test/results/clientpositive/infer_bucket_sort_reducers_power_two.q.out @@ -64,7 +64,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test join, should be bucketed and sorted by join key @@ -117,7 +117,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test join with two keys, should be bucketed and sorted by join keys @@ -170,7 +170,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [key, value] -Sort Columns: [Order(col:key, order:1, nullOrder:0), Order(col:value, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1), Order(col:value, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test join on three tables on same key, should be bucketed and sorted by join key @@ -223,7 +223,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test join on three tables on different keys, should be bucketed and sorted by latter key @@ -276,7 +276,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 16 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- Test group by in subquery with another group by outside, should be bucketed and sorted by the @@ -331,6 +331,6 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [value] -Sort Columns: [Order(col:value, order:1, nullOrder:0)] +Sort Columns: [Order(col:value, order:1)] Storage Desc Params: serialization.format 1 http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/orc_analyze.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/orc_analyze.q.out b/ql/src/test/results/clientpositive/orc_analyze.q.out index 9b7e7b7..87855fa 100644 --- a/ql/src/test/results/clientpositive/orc_analyze.q.out +++ b/ql/src/test/results/clientpositive/orc_analyze.q.out @@ -917,7 +917,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [first_name] -Sort Columns: [Order(col:last_name, order:1, nullOrder:0)] +Sort Columns: [Order(col:last_name, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: desc formatted orc_create_people partition(state="Or") @@ -960,7 +960,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [first_name] -Sort Columns: [Order(col:last_name, order:1, nullOrder:0)] +Sort Columns: [Order(col:last_name, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: analyze table orc_create_people partition(state) compute statistics partialscan @@ -1015,7 +1015,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [first_name] -Sort Columns: [Order(col:last_name, order:1, nullOrder:0)] +Sort Columns: [Order(col:last_name, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: desc formatted orc_create_people partition(state="Or") @@ -1058,7 +1058,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [first_name] -Sort Columns: [Order(col:last_name, order:1, nullOrder:0)] +Sort Columns: [Order(col:last_name, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: analyze table orc_create_people partition(state) compute statistics noscan @@ -1113,7 +1113,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [first_name] -Sort Columns: [Order(col:last_name, order:1, nullOrder:0)] +Sort Columns: [Order(col:last_name, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: desc formatted orc_create_people partition(state="Or") @@ -1156,7 +1156,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [first_name] -Sort Columns: [Order(col:last_name, order:1, nullOrder:0)] +Sort Columns: [Order(col:last_name, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: drop table orc_create_people @@ -1262,7 +1262,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [first_name] -Sort Columns: [Order(col:last_name, order:1, nullOrder:0)] +Sort Columns: [Order(col:last_name, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: desc formatted orc_create_people partition(state="Or") @@ -1305,7 +1305,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [first_name] -Sort Columns: [Order(col:last_name, order:1, nullOrder:0)] +Sort Columns: [Order(col:last_name, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: drop table orc_create_people http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/partition_coltype_literals.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/partition_coltype_literals.q.out b/ql/src/test/results/clientpositive/partition_coltype_literals.q.out index 1a93b7a..bc159eb 100644 --- a/ql/src/test/results/clientpositive/partition_coltype_literals.q.out +++ b/ql/src/test/results/clientpositive/partition_coltype_literals.q.out @@ -154,7 +154,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.RCFileOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:0, nullOrder:1)] +Sort Columns: [Order(col:key, order:0)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- rename partition @@ -205,7 +205,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.RCFileOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:0, nullOrder:1)] +Sort Columns: [Order(col:key, order:0)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- insert partition http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/show_create_table_alter.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/show_create_table_alter.q.out b/ql/src/test/results/clientpositive/show_create_table_alter.q.out index 78a34de..32819ea 100644 --- a/ql/src/test/results/clientpositive/show_create_table_alter.q.out +++ b/ql/src/test/results/clientpositive/show_create_table_alter.q.out @@ -24,7 +24,7 @@ CREATE EXTERNAL TABLE `tmp_showcrt1`( CLUSTERED BY ( key) SORTED BY ( - value DESC NULLS LAST) + value DESC) INTO 5 BUCKETS ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' @@ -59,7 +59,7 @@ COMMENT 'temporary table' CLUSTERED BY ( key) SORTED BY ( - value DESC NULLS LAST) + value DESC) INTO 5 BUCKETS ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' @@ -98,7 +98,7 @@ COMMENT 'changed comment' CLUSTERED BY ( key) SORTED BY ( - value DESC NULLS LAST) + value DESC) INTO 5 BUCKETS ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' @@ -136,7 +136,7 @@ COMMENT 'changed comment' CLUSTERED BY ( key) SORTED BY ( - value DESC NULLS LAST) + value DESC) INTO 5 BUCKETS ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' @@ -174,7 +174,7 @@ COMMENT 'changed comment' CLUSTERED BY ( key) SORTED BY ( - value DESC NULLS LAST) + value DESC) INTO 5 BUCKETS ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/spark/bucket5.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/bucket5.q.out b/ql/src/test/results/clientpositive/spark/bucket5.q.out index af6d399..5baf054 100644 --- a/ql/src/test/results/clientpositive/spark/bucket5.q.out +++ b/ql/src/test/results/clientpositive/spark/bucket5.q.out @@ -391,7 +391,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 2 Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1, nullOrder:0)] +Sort Columns: [Order(col:key, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: SELECT * FROM bucketed_table TABLESAMPLE (BUCKET 1 OUT OF 2) s LIMIT 10 http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/spark/infer_bucket_sort_bucketed_table.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/infer_bucket_sort_bucketed_table.q.out b/ql/src/test/results/clientpositive/spark/infer_bucket_sort_bucketed_table.q.out index 6db9428..33d795b 100644 --- a/ql/src/test/results/clientpositive/spark/infer_bucket_sort_bucketed_table.q.out +++ b/ql/src/test/results/clientpositive/spark/infer_bucket_sort_bucketed_table.q.out @@ -64,7 +64,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 3 Bucket Columns: [value] -Sort Columns: [Order(col:value, order:1, nullOrder:0)] +Sort Columns: [Order(col:value, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: -- If the count(*) from sampling the buckets matches the count(*) from each file, the table is http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out b/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out index f7a94ac..a90e3f6 100644 --- a/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out +++ b/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out @@ -1215,7 +1215,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: desc formatted over1k_part_buck_sort_orc partition(t="__HIVE_DEFAULT_PARTITION__") @@ -1256,7 +1256,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: select count(*) from over1k_part_orc @@ -2255,7 +2255,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: desc formatted over1k_part_buck_sort2_orc partition(t="__HIVE_DEFAULT_PARTITION__") @@ -2296,7 +2296,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: explain select * from over1k_part_buck_sort2_orc @@ -2471,7 +2471,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: desc formatted over1k_part_buck_sort2_orc partition(t="__HIVE_DEFAULT_PARTITION__") @@ -2512,7 +2512,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: explain select * from over1k_part_buck_sort2_orc http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/tez/dynpart_sort_optimization.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/tez/dynpart_sort_optimization.q.out b/ql/src/test/results/clientpositive/tez/dynpart_sort_optimization.q.out index 01dc4ae..5292106 100644 --- a/ql/src/test/results/clientpositive/tez/dynpart_sort_optimization.q.out +++ b/ql/src/test/results/clientpositive/tez/dynpart_sort_optimization.q.out @@ -1128,7 +1128,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: desc formatted over1k_part_buck_sort partition(t="__HIVE_DEFAULT_PARTITION__") @@ -1169,7 +1169,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: select count(*) from over1k_part @@ -2152,7 +2152,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: desc formatted over1k_part_buck_sort2 partition(t="__HIVE_DEFAULT_PARTITION__") @@ -2193,7 +2193,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: select * from over1k_part_buck_sort2 @@ -2295,7 +2295,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: desc formatted over1k_part_buck_sort2 partition(t="__HIVE_DEFAULT_PARTITION__") @@ -2336,7 +2336,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No Num Buckets: 1 Bucket Columns: [si] -Sort Columns: [Order(col:f, order:1, nullOrder:0)] +Sort Columns: [Order(col:f, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: select * from over1k_part_buck_sort2 http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/tez/orc_analyze.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/tez/orc_analyze.q.out b/ql/src/test/results/clientpositive/tez/orc_analyze.q.out index 9b7e7b7..87855fa 100644 --- a/ql/src/test/results/clientpositive/tez/orc_analyze.q.out +++ b/ql/src/test/results/clientpositive/tez/orc_analyze.q.out @@ -917,7 +917,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [first_name] -Sort Columns: [Order(col:last_name, order:1, nullOrder:0)] +Sort Columns: [Order(col:last_name, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: desc formatted orc_create_people partition(state="Or") @@ -960,7 +960,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [first_name] -Sort Columns: [Order(col:last_name, order:1, nullOrder:0)] +Sort Columns: [Order(col:last_name, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: analyze table orc_create_people partition(state) compute statistics partialscan @@ -1015,7 +1015,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [first_name] -Sort Columns: [Order(col:last_name, order:1, nullOrder:0)] +Sort Columns: [Order(col:last_name, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: desc formatted orc_create_people partition(state="Or") @@ -1058,7 +1058,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [first_name] -Sort Columns: [Order(col:last_name, order:1, nullOrder:0)] +Sort Columns: [Order(col:last_name, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: analyze table orc_create_people partition(state) compute statistics noscan @@ -1113,7 +1113,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [first_name] -Sort Columns: [Order(col:last_name, order:1, nullOrder:0)] +Sort Columns: [Order(col:last_name, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: desc formatted orc_create_people partition(state="Or") @@ -1156,7 +1156,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [first_name] -Sort Columns: [Order(col:last_name, order:1, nullOrder:0)] +Sort Columns: [Order(col:last_name, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: drop table orc_create_people @@ -1262,7 +1262,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [first_name] -Sort Columns: [Order(col:last_name, order:1, nullOrder:0)] +Sort Columns: [Order(col:last_name, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: desc formatted orc_create_people partition(state="Or") @@ -1305,7 +1305,7 @@ OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat Compressed: No Num Buckets: 4 Bucket Columns: [first_name] -Sort Columns: [Order(col:last_name, order:1, nullOrder:0)] +Sort Columns: [Order(col:last_name, order:1)] Storage Desc Params: serialization.format 1 PREHOOK: query: drop table orc_create_people http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/serde/if/serde.thrift ---------------------------------------------------------------------- diff --git a/serde/if/serde.thrift b/serde/if/serde.thrift index 0b3804d..6caad36 100644 --- a/serde/if/serde.thrift +++ b/serde/if/serde.thrift @@ -30,7 +30,7 @@ const string SERIALIZATION_NULL_FORMAT = "serialization.null.format" const string SERIALIZATION_ESCAPE_CRLF = "serialization.escape.crlf" const string SERIALIZATION_LAST_COLUMN_TAKES_REST = "serialization.last.column.takes.rest" const string SERIALIZATION_SORT_ORDER = "serialization.sort.order" -const string SERIALIZATION_NULL_POSITION = "serialization.null.position"; +const string SERIALIZATION_NULL_SORT_ORDER = "serialization.sort.order.null"; const string SERIALIZATION_USE_JSON_OBJECTS = "serialization.use.json.object" const string SERIALIZATION_ENCODING = "serialization.encoding" http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/serde/src/gen/thrift/gen-cpp/serde_constants.cpp ---------------------------------------------------------------------- diff --git a/serde/src/gen/thrift/gen-cpp/serde_constants.cpp b/serde/src/gen/thrift/gen-cpp/serde_constants.cpp index 75701e2..3a675bf 100644 --- a/serde/src/gen/thrift/gen-cpp/serde_constants.cpp +++ b/serde/src/gen/thrift/gen-cpp/serde_constants.cpp @@ -27,7 +27,7 @@ serdeConstants::serdeConstants() { SERIALIZATION_SORT_ORDER = "serialization.sort.order"; - SERIALIZATION_NULL_POSITION = "serialization.null.position"; + SERIALIZATION_NULL_SORT_ORDER = "serialization.sort.order.null"; SERIALIZATION_USE_JSON_OBJECTS = "serialization.use.json.object"; http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/serde/src/gen/thrift/gen-cpp/serde_constants.h ---------------------------------------------------------------------- diff --git a/serde/src/gen/thrift/gen-cpp/serde_constants.h b/serde/src/gen/thrift/gen-cpp/serde_constants.h index 6d85928..a5f33fb 100644 --- a/serde/src/gen/thrift/gen-cpp/serde_constants.h +++ b/serde/src/gen/thrift/gen-cpp/serde_constants.h @@ -23,7 +23,7 @@ class serdeConstants { std::string SERIALIZATION_ESCAPE_CRLF; std::string SERIALIZATION_LAST_COLUMN_TAKES_REST; std::string SERIALIZATION_SORT_ORDER; - std::string SERIALIZATION_NULL_POSITION; + std::string SERIALIZATION_NULL_SORT_ORDER; std::string SERIALIZATION_USE_JSON_OBJECTS; std::string SERIALIZATION_ENCODING; std::string FIELD_DELIM; http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/serde/src/gen/thrift/gen-php/org/apache/hadoop/hive/serde/Types.php ---------------------------------------------------------------------- diff --git a/serde/src/gen/thrift/gen-php/org/apache/hadoop/hive/serde/Types.php b/serde/src/gen/thrift/gen-php/org/apache/hadoop/hive/serde/Types.php index 0bc6dd7..18c3991 100644 --- a/serde/src/gen/thrift/gen-php/org/apache/hadoop/hive/serde/Types.php +++ b/serde/src/gen/thrift/gen-php/org/apache/hadoop/hive/serde/Types.php @@ -26,7 +26,7 @@ final class Constant extends \Thrift\Type\TConstant { static protected $SERIALIZATION_ESCAPE_CRLF; static protected $SERIALIZATION_LAST_COLUMN_TAKES_REST; static protected $SERIALIZATION_SORT_ORDER; - static protected $SERIALIZATION_NULL_POSITION; + static protected $SERIALIZATION_NULL_SORT_ORDER; static protected $SERIALIZATION_USE_JSON_OBJECTS; static protected $SERIALIZATION_ENCODING; static protected $FIELD_DELIM; @@ -98,8 +98,8 @@ final class Constant extends \Thrift\Type\TConstant { return "serialization.sort.order"; } - static protected function init_SERIALIZATION_NULL_POSITION() { - return "serialization.null.position"; + static protected function init_SERIALIZATION_NULL_SORT_ORDER() { + return "serialization.sort.order.null"; } static protected function init_SERIALIZATION_USE_JSON_OBJECTS() { http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/serde/src/gen/thrift/gen-py/org_apache_hadoop_hive_serde/constants.py ---------------------------------------------------------------------- diff --git a/serde/src/gen/thrift/gen-py/org_apache_hadoop_hive_serde/constants.py b/serde/src/gen/thrift/gen-py/org_apache_hadoop_hive_serde/constants.py index 7939791..fafdc24 100644 --- a/serde/src/gen/thrift/gen-py/org_apache_hadoop_hive_serde/constants.py +++ b/serde/src/gen/thrift/gen-py/org_apache_hadoop_hive_serde/constants.py @@ -17,7 +17,7 @@ SERIALIZATION_NULL_FORMAT = "serialization.null.format" SERIALIZATION_ESCAPE_CRLF = "serialization.escape.crlf" SERIALIZATION_LAST_COLUMN_TAKES_REST = "serialization.last.column.takes.rest" SERIALIZATION_SORT_ORDER = "serialization.sort.order" -SERIALIZATION_NULL_POSITION = "serialization.null.position" +SERIALIZATION_NULL_SORT_ORDER = "serialization.sort.order.null" SERIALIZATION_USE_JSON_OBJECTS = "serialization.use.json.object" SERIALIZATION_ENCODING = "serialization.encoding" FIELD_DELIM = "field.delim" http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/serde/src/gen/thrift/gen-rb/serde_constants.rb ---------------------------------------------------------------------- diff --git a/serde/src/gen/thrift/gen-rb/serde_constants.rb b/serde/src/gen/thrift/gen-rb/serde_constants.rb index d09e3c2..0ce9f27 100644 --- a/serde/src/gen/thrift/gen-rb/serde_constants.rb +++ b/serde/src/gen/thrift/gen-rb/serde_constants.rb @@ -23,7 +23,7 @@ SERIALIZATION_LAST_COLUMN_TAKES_REST = %q"serialization.last.column.takes.rest" SERIALIZATION_SORT_ORDER = %q"serialization.sort.order" -SERIALIZATION_NULL_POSITION = %q"serialization.null.position" +SERIALIZATION_NULL_SORT_ORDER = %q"serialization.sort.order.null" SERIALIZATION_USE_JSON_OBJECTS = %q"serialization.use.json.object"
