Repository: hive
Updated Branches:
  refs/heads/master 7c4599baa -> cc8893e48


http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/BucketingSortingOpProcFactory.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/BucketingSortingOpProcFactory.java
 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/BucketingSortingOpProcFactory.java
index 9159120..391cfda 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/BucketingSortingOpProcFactory.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/BucketingSortingOpProcFactory.java
@@ -166,7 +166,6 @@ public class BucketingSortingOpProcFactory {
         }
 
         String sortOrder = rsDesc.getOrder();
-        String nullSortOrder = rsDesc.getNullOrder();
         List<ExprNodeDesc> keyCols = rsDesc.getKeyCols();
         List<ExprNodeDesc> valCols = ExprNodeDescUtils.backtrack(joinValues, 
jop, parent);
 
@@ -187,8 +186,7 @@ public class BucketingSortingOpProcFactory {
               newSortCols[keyIndex].addAlias(vname, vindex);
             } else {
               newBucketCols[keyIndex] = new BucketCol(vname, vindex);
-              newSortCols[keyIndex] = new SortCol(vname, vindex, 
sortOrder.charAt(keyIndex),
-                      nullSortOrder.charAt(keyIndex));
+              newSortCols[keyIndex] = new SortCol(vname, vindex, 
sortOrder.charAt(keyIndex));
             }
           }
         }
@@ -313,8 +311,7 @@ public class BucketingSortingOpProcFactory {
         int sortIndex = indexOfColName(sortCols, columnExpr.getColumn());
         if (sortIndex != -1) {
           if (newSortCols[sortIndex] == null) {
-            newSortCols[sortIndex] = new 
SortCol(sortCols.get(sortIndex).getSortOrder(),
-                    sortCols.get(sortIndex).getNullSortOrder());
+            newSortCols[sortIndex] = new 
SortCol(sortCols.get(sortIndex).getSortOrder());
           }
           newSortCols[sortIndex].addAlias(
               colInfos.get(colInfosIndex).getInternalName(), colInfosIndex);
@@ -439,7 +436,7 @@ public class BucketingSortingOpProcFactory {
   private static List<SortCol> getNewSortCols(List<SortCol> sortCols, 
List<ColumnInfo> colInfos) {
     List<SortCol> newSortCols = new ArrayList<SortCol>(sortCols.size());
     for (int i = 0; i < sortCols.size(); i++) {
-      SortCol sortCol = new SortCol(sortCols.get(i).getSortOrder(), 
sortCols.get(i).getNullSortOrder());
+      SortCol sortCol = new SortCol(sortCols.get(i).getSortOrder());
       for (Integer index : sortCols.get(i).getIndexes()) {
         // The only time this condition should be false is in the case of 
dynamic partitioning
         if (index < colInfos.size()) {
@@ -540,7 +537,6 @@ public class BucketingSortingOpProcFactory {
 
   static List<SortCol> extractSortCols(ReduceSinkOperator rop, 
List<ExprNodeDesc> outputValues) {
     String sortOrder = rop.getConf().getOrder();
-    String nullSortOrder = rop.getConf().getNullOrder();
     List<SortCol> sortCols = new ArrayList<SortCol>();
     ArrayList<ExprNodeDesc> keyCols = rop.getConf().getKeyCols();
     for (int i = 0; i < keyCols.size(); i++) {
@@ -553,7 +549,7 @@ public class BucketingSortingOpProcFactory {
         break;
       }
       sortCols.add(new SortCol(((ExprNodeColumnDesc) keyCol).getColumn(), 
index,
-              sortOrder.charAt(i), nullSortOrder.charAt(i)));
+              sortOrder.charAt(i)));
     }
     // If the sorted columns can't all be found in the values then the data is 
only sorted on
     // the columns seen up until now
@@ -654,7 +650,6 @@ public class BucketingSortingOpProcFactory {
 
       GroupByDesc groupByDesc = gop.getConf();
       String sortOrder = rop.getConf().getOrder();
-      String nullSortOrder = rop.getConf().getNullOrder();
       List<BucketCol> bucketCols = new ArrayList<BucketCol>();
       List<SortCol> sortCols = new ArrayList<SortCol>();
       assert rop.getConf().getKeyCols().size() <= 
rop.getSchema().getSignature().size();
@@ -665,7 +660,7 @@ public class BucketingSortingOpProcFactory {
         }
         String colName = 
rop.getSchema().getSignature().get(i).getInternalName();
         bucketCols.add(new BucketCol(colName, i));
-        sortCols.add(new SortCol(colName, i, sortOrder.charAt(i), 
nullSortOrder.charAt(i)));
+        sortCols.add(new SortCol(colName, i, sortOrder.charAt(i)));
       }
       bctx.setBucketedCols(rop, bucketCols);
       bctx.setSortedCols(rop, sortCols);

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java 
b/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java
index f10a40a..b36a9a0 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java
@@ -653,7 +653,7 @@ public abstract class BaseSemanticAnalyzer {
     return colList;
   }
 
-  protected List<Order> getColumnNamesOrder(ASTNode ast) {
+  protected List<Order> getColumnNamesOrder(ASTNode ast) throws 
SemanticException {
     List<Order> colList = new ArrayList<Order>();
     int numCh = ast.getChildCount();
     for (int i = 0; i < numCh; i++) {
@@ -662,19 +662,19 @@ public abstract class BaseSemanticAnalyzer {
         child = (ASTNode) child.getChild(0);
         if (child.getToken().getType() == HiveParser.TOK_NULLS_FIRST) {
           colList.add(new 
Order(unescapeIdentifier(child.getChild(0).getText()).toLowerCase(),
-              HIVE_COLUMN_ORDER_ASC, HIVE_COLUMN_NULLS_FIRST));
+              HIVE_COLUMN_ORDER_ASC));
         } else {
-          colList.add(new 
Order(unescapeIdentifier(child.getChild(0).getText()).toLowerCase(),
-              HIVE_COLUMN_ORDER_ASC, HIVE_COLUMN_NULLS_LAST));
+          throw new SemanticException("create/alter table: "
+                  + "not supported NULLS LAST for ORDER BY in ASC order");
         }
       } else {
         child = (ASTNode) child.getChild(0);
         if (child.getToken().getType() == HiveParser.TOK_NULLS_LAST) {
           colList.add(new 
Order(unescapeIdentifier(child.getChild(0).getText()).toLowerCase(),
-              HIVE_COLUMN_ORDER_DESC, HIVE_COLUMN_NULLS_LAST));
+              HIVE_COLUMN_ORDER_DESC));
         } else {
-          colList.add(new 
Order(unescapeIdentifier(child.getChild(0).getText()).toLowerCase(),
-              HIVE_COLUMN_ORDER_DESC, HIVE_COLUMN_NULLS_FIRST));
+          throw new SemanticException("create/alter table: "
+                  + "not supported NULLS FIRST for ORDER BY in DESC order");
         }
       }
     }

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java 
b/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java
index 3e91e10..9ab091d 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java
@@ -117,7 +117,6 @@ import org.apache.hadoop.hive.ql.metadata.Partition;
 import org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient;
 import org.apache.hadoop.hive.ql.metadata.Table;
 import org.apache.hadoop.hive.ql.metadata.VirtualColumn;
-import org.apache.hadoop.hive.ql.optimizer.ColumnPruner;
 import org.apache.hadoop.hive.ql.optimizer.Optimizer;
 import org.apache.hadoop.hive.ql.optimizer.Transform;
 import org.apache.hadoop.hive.ql.optimizer.calcite.CalciteSemanticException;
@@ -6358,7 +6357,6 @@ public class SemanticAnalyzer extends 
BaseSemanticAnalyzer {
         (dest_tab.getSortCols().size() > 0)) {
       sortCols = getSortCols(dest, qb, dest_tab, table_desc, input, true);
       sortOrders = getSortOrders(dest, qb, dest_tab, input);
-      nullSortOrders = getNullSortOrders(dest, qb, dest_tab, input);
       if (!enforceBucketing && !dest_tab.isIndexTable()) {
         throw new 
SemanticException(ErrorMsg.TBL_SORTED_NOT_BUCKETED.getErrorCodedMsg(dest_tab.getCompleteName()));
       } else {
@@ -6393,12 +6391,10 @@ public class SemanticAnalyzer extends 
BaseSemanticAnalyzer {
       }
 
       StringBuilder order = new StringBuilder();
+      StringBuilder nullOrder = new StringBuilder();
       for (int sortOrder : sortOrders) {
         order.append(sortOrder == BaseSemanticAnalyzer.HIVE_COLUMN_ORDER_ASC ? 
'+' : '-');
-      }
-      StringBuilder nullOrder = new StringBuilder();
-      for (int pos : nullSortOrders) {
-        nullOrder.append(pos == BaseSemanticAnalyzer.HIVE_COLUMN_NULLS_FIRST ? 
'a' : 'z');
+        nullOrder.append(sortOrder == 
BaseSemanticAnalyzer.HIVE_COLUMN_ORDER_ASC ? 'a' : 'z');
       }
       input = genReduceSinkPlan(input, partnCols, sortCols, order.toString(), 
nullOrder.toString(),
               maxReducers, (AcidUtils.isAcidTable(dest_tab) ? getAcidType() : 
AcidUtils.Operation.NOT_ACID));
@@ -7362,23 +7358,6 @@ public class SemanticAnalyzer extends 
BaseSemanticAnalyzer {
     return orders;
   }
 
-  private ArrayList<Integer> getNullSortOrders(String dest, QB qb, Table tab, 
Operator input)
-      throws SemanticException {
-    List<Order> tabSortCols = tab.getSortCols();
-    List<FieldSchema> tabCols = tab.getCols();
-
-    ArrayList<Integer> orders = new ArrayList<Integer>();
-    for (Order sortCol : tabSortCols) {
-      for (FieldSchema tabCol : tabCols) {
-        if (sortCol.getCol().equals(tabCol.getName())) {
-          orders.add(sortCol.getNullOrder());
-          break;
-        }
-      }
-    }
-    return orders;
-  }
-
   private Operator genReduceSinkPlan(String dest, QB qb, Operator<?> input,
       int numReducers, boolean hasOrderBy) throws SemanticException {
 

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/org/apache/hadoop/hive/ql/txn/compactor/TestWorker.java
----------------------------------------------------------------------
diff --git 
a/ql/src/test/org/apache/hadoop/hive/ql/txn/compactor/TestWorker.java 
b/ql/src/test/org/apache/hadoop/hive/ql/txn/compactor/TestWorker.java
index 7f6430f..cf7eb70 100644
--- a/ql/src/test/org/apache/hadoop/hive/ql/txn/compactor/TestWorker.java
+++ b/ql/src/test/org/apache/hadoop/hive/ql/txn/compactor/TestWorker.java
@@ -224,7 +224,7 @@ public class TestWorker extends CompactorTest {
   @Test
   public void sortedTable() throws Exception {
     List<Order> sortCols = new ArrayList<Order>(1);
-    sortCols.add(new Order("b", 1, 0));
+    sortCols.add(new Order("b", 1));
 
     Table t = newTable("default", "st", false, new HashMap<String, String>(), 
sortCols, false);
 
@@ -249,7 +249,7 @@ public class TestWorker extends CompactorTest {
   @Test
   public void sortedPartition() throws Exception {
     List<Order> sortCols = new ArrayList<Order>(1);
-    sortCols.add(new Order("b", 1, 0));
+    sortCols.add(new Order("b", 1));
 
     Table t = newTable("default", "sp", true, new HashMap<String, String>(), 
sortCols, false);
     Partition p = newPartition(t, "today", sortCols);

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table2_h23.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table2_h23.q.out
 
b/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table2_h23.q.out
index 9eea7f7..3b71598 100644
--- 
a/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table2_h23.q.out
+++ 
b/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table2_h23.q.out
@@ -232,7 +232,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           8                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:0, nullOrder:1)]   
+Sort Columns:          [Order(col:key, order:0)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: INSERT OVERWRITE TABLE tst1 PARTITION (ds = '1') SELECT key, 
value FROM src
@@ -281,7 +281,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           8                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:0, nullOrder:1)]   
+Sort Columns:          [Order(col:key, order:0)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test changing the bucket columns
@@ -326,7 +326,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           8                        
 Bucket Columns:        [value]                  
-Sort Columns:          [Order(col:key, order:0, nullOrder:1)]   
+Sort Columns:          [Order(col:key, order:0)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: INSERT OVERWRITE TABLE tst1 PARTITION (ds = '1') SELECT key, 
value FROM src
@@ -375,7 +375,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           8                        
 Bucket Columns:        [value]                  
-Sort Columns:          [Order(col:key, order:0, nullOrder:1)]   
+Sort Columns:          [Order(col:key, order:0)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test changing the number of buckets
@@ -420,7 +420,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [value]                  
-Sort Columns:          [Order(col:key, order:0, nullOrder:1)]   
+Sort Columns:          [Order(col:key, order:0)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: INSERT OVERWRITE TABLE tst1 PARTITION (ds = '1') SELECT key, 
value FROM src
@@ -469,7 +469,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [value]                  
-Sort Columns:          [Order(col:key, order:0, nullOrder:1)]   
+Sort Columns:          [Order(col:key, order:0)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test changing the sort columns
@@ -514,7 +514,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [value]                  
-Sort Columns:          [Order(col:value, order:0, nullOrder:1)]         
+Sort Columns:          [Order(col:value, order:0)]      
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: INSERT OVERWRITE TABLE tst1 PARTITION (ds = '1') SELECT key, 
value FROM src
@@ -563,7 +563,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [value]                  
-Sort Columns:          [Order(col:value, order:0, nullOrder:1)]         
+Sort Columns:          [Order(col:value, order:0)]      
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test changing the sort order
@@ -608,7 +608,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [value]                  
-Sort Columns:          [Order(col:value, order:1, nullOrder:0)]         
+Sort Columns:          [Order(col:value, order:1)]      
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: INSERT OVERWRITE TABLE tst1 PARTITION (ds = '1') SELECT key, 
value FROM src
@@ -657,7 +657,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [value]                  
-Sort Columns:          [Order(col:value, order:1, nullOrder:0)]         
+Sort Columns:          [Order(col:value, order:1)]      
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test a sorted partition gets converted to unsorted

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table_h23.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table_h23.q.out
 
b/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table_h23.q.out
index af2f47a..cab3de4 100644
--- 
a/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table_h23.q.out
+++ 
b/ql/src/test/results/clientpositive/alter_numbuckets_partitioned_table_h23.q.out
@@ -408,7 +408,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           12                       
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test changing sort order
@@ -455,7 +455,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           12                       
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:value, order:0, nullOrder:1)]         
+Sort Columns:          [Order(col:value, order:0)]      
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test removing test order

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/alter_partition_clusterby_sortby.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/alter_partition_clusterby_sortby.q.out 
b/ql/src/test/results/clientpositive/alter_partition_clusterby_sortby.q.out
index 46d5b34..184d2e4 100644
--- a/ql/src/test/results/clientpositive/alter_partition_clusterby_sortby.q.out
+++ b/ql/src/test/results/clientpositive/alter_partition_clusterby_sortby.q.out
@@ -110,7 +110,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [b]                      
-Sort Columns:          [Order(col:b, order:0, nullOrder:1)]     
+Sort Columns:          [Order(col:b, order:0)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Turn off clustering for a partition
@@ -200,7 +200,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [a, b]                   
-Sort Columns:          [Order(col:a, order:0, nullOrder:1), Order(col:b, 
order:1, nullOrder:0)]         
+Sort Columns:          [Order(col:a, order:0), Order(col:b, order:1)]   
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: drop table alter_table_partition_clusterby_sortby

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/alter_table_not_sorted.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/alter_table_not_sorted.q.out 
b/ql/src/test/results/clientpositive/alter_table_not_sorted.q.out
index 8a88bc4..6e1ec59 100644
--- a/ql/src/test/results/clientpositive/alter_table_not_sorted.q.out
+++ b/ql/src/test/results/clientpositive/alter_table_not_sorted.q.out
@@ -34,7 +34,7 @@ OutputFormat:         
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [a]                      
-Sort Columns:          [Order(col:a, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:a, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: alter table alter_table_not_sorted not sorted

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/authorization_index.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/authorization_index.q.out 
b/ql/src/test/results/clientpositive/authorization_index.q.out
index a6ec998..adc02ad 100644
--- a/ql/src/test/results/clientpositive/authorization_index.q.out
+++ b/ql/src/test/results/clientpositive/authorization_index.q.out
@@ -41,7 +41,7 @@ OutputFormat:         
org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           -1                       
 Bucket Columns:        []                       
-Sort Columns:          [Order(col:a, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:a, order:1)]  
 PREHOOK: query: alter index t1_index on t1 rebuild
 PREHOOK: type: ALTERINDEX_REBUILD
 PREHOOK: Input: default@t1

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/bucket5.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/bucket5.q.out 
b/ql/src/test/results/clientpositive/bucket5.q.out
index aa47ba8..fa926aa 100644
--- a/ql/src/test/results/clientpositive/bucket5.q.out
+++ b/ql/src/test/results/clientpositive/bucket5.q.out
@@ -535,7 +535,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           2                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: SELECT * FROM bucketed_table TABLESAMPLE (BUCKET 1 OUT OF 2) s 
LIMIT 10

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/dynpart_sort_opt_vectorization.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/dynpart_sort_opt_vectorization.q.out 
b/ql/src/test/results/clientpositive/dynpart_sort_opt_vectorization.q.out
index c5d695e..d03bfe4 100644
--- a/ql/src/test/results/clientpositive/dynpart_sort_opt_vectorization.q.out
+++ b/ql/src/test/results/clientpositive/dynpart_sort_opt_vectorization.q.out
@@ -1155,7 +1155,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: desc formatted over1k_part_buck_sort_orc 
partition(t="__HIVE_DEFAULT_PARTITION__")
@@ -1196,7 +1196,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: select count(*) from over1k_part_orc
@@ -2129,7 +2129,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: desc formatted over1k_part_buck_sort2_orc 
partition(t="__HIVE_DEFAULT_PARTITION__")
@@ -2170,7 +2170,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: explain select * from over1k_part_buck_sort2_orc
@@ -2339,7 +2339,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: desc formatted over1k_part_buck_sort2_orc 
partition(t="__HIVE_DEFAULT_PARTITION__")
@@ -2380,7 +2380,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: explain select * from over1k_part_buck_sort2_orc

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/dynpart_sort_optimization.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/dynpart_sort_optimization.q.out 
b/ql/src/test/results/clientpositive/dynpart_sort_optimization.q.out
index 9a222b1..857d609 100644
--- a/ql/src/test/results/clientpositive/dynpart_sort_optimization.q.out
+++ b/ql/src/test/results/clientpositive/dynpart_sort_optimization.q.out
@@ -1060,7 +1060,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: desc formatted over1k_part_buck_sort 
partition(t="__HIVE_DEFAULT_PARTITION__")
@@ -1101,7 +1101,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: select count(*) from over1k_part
@@ -2027,7 +2027,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: desc formatted over1k_part_buck_sort2 
partition(t="__HIVE_DEFAULT_PARTITION__")
@@ -2068,7 +2068,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: select * from over1k_part_buck_sort2
@@ -2170,7 +2170,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: desc formatted over1k_part_buck_sort2 
partition(t="__HIVE_DEFAULT_PARTITION__")
@@ -2211,7 +2211,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: select * from over1k_part_buck_sort2

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/index_skewtable.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/index_skewtable.q.out 
b/ql/src/test/results/clientpositive/index_skewtable.q.out
index 45e6834..972789d 100644
--- a/ql/src/test/results/clientpositive/index_skewtable.q.out
+++ b/ql/src/test/results/clientpositive/index_skewtable.q.out
@@ -57,7 +57,7 @@ OutputFormat:         
org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           -1                       
 Bucket Columns:        []                       
-Sort Columns:          [Order(col:value, order:1, nullOrder:0)]         
+Sort Columns:          [Order(col:value, order:1)]      
 PREHOOK: query: ALTER INDEX kv_index ON kv REBUILD
 PREHOOK: type: ALTERINDEX_REBUILD
 PREHOOK: Input: default@kv

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/infer_bucket_sort.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/infer_bucket_sort.q.out 
b/ql/src/test/results/clientpositive/infer_bucket_sort.q.out
index 709ca10..182dd3a 100644
--- a/ql/src/test/results/clientpositive/infer_bucket_sort.q.out
+++ b/ql/src/test/results/clientpositive/infer_bucket_sort.q.out
@@ -62,7 +62,7 @@ OutputFormat:         
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test group by where a key isn't selected, should not be 
bucketed or sorted
@@ -168,7 +168,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test join with two keys, should be bucketed and sorted by 
join keys
@@ -221,7 +221,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key, value]             
-Sort Columns:          [Order(col:key, order:1, nullOrder:0), Order(col:value, 
order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1), Order(col:value, order:1)]    
 
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test join with two keys and only one selected, should not 
be bucketed or sorted
@@ -327,7 +327,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test join on three tables on different keys, should be 
bucketed and sorted by latter key
@@ -380,7 +380,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test distribute by, should only be bucketed by key
@@ -486,7 +486,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           -1                       
 Bucket Columns:        []                       
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test sort by desc, should be sorted by key
@@ -539,7 +539,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           -1                       
 Bucket Columns:        []                       
-Sort Columns:          [Order(col:key, order:0, nullOrder:1)]   
+Sort Columns:          [Order(col:key, order:0)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test cluster by, should be bucketed and sorted by key
@@ -592,7 +592,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test distribute by and sort by different keys, should be 
bucketed by one key sorted by the other
@@ -645,7 +645,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:value, order:1, nullOrder:0)]         
+Sort Columns:          [Order(col:value, order:1)]      
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test join in simple subquery, should be bucketed and sorted 
on key
@@ -698,7 +698,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test join in simple subquery renaming key column, should be 
bucketed and sorted on key
@@ -751,7 +751,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test group by in simple subquery, should be bucketed and 
sorted on key
@@ -804,7 +804,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test group by in simple subquery renaming key column, 
should be bucketed and sorted on key
@@ -857,7 +857,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test group by in subquery with where outside, should still 
be bucketed and sorted on key
@@ -910,7 +910,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test group by in subquery with expression on value, should 
still be bucketed and sorted on key
@@ -963,7 +963,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test group by in subquery with lateral view outside, should 
still be bucketed and sorted
@@ -1016,7 +1016,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test group by in subquery with another group by outside, 
should be bucketed and sorted by the
@@ -1071,7 +1071,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [value]                  
-Sort Columns:          [Order(col:value, order:1, nullOrder:0)]         
+Sort Columns:          [Order(col:value, order:1)]      
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test group by in subquery with select on outside reordering 
the columns, should be bucketed and
@@ -1126,7 +1126,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [value]                  
-Sort Columns:          [Order(col:value, order:1, nullOrder:0)]         
+Sort Columns:          [Order(col:value, order:1)]      
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test group by in subquery followed by distribute by, should 
only be bucketed by the distribute key
@@ -1179,7 +1179,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test group by in subquery followed by sort by, should only 
be sorted by the sort key
@@ -1232,7 +1232,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test group by in subquery followed by transform script, 
should not be bucketed or sorted
@@ -1338,6 +1338,6 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key, value]             
-Sort Columns:          [Order(col:key, order:1, nullOrder:0), Order(col:value, 
order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1), Order(col:value, order:1)]    
 
 Storage Desc Params:            
        serialization.format    1                   

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/infer_bucket_sort_bucketed_table.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/infer_bucket_sort_bucketed_table.q.out 
b/ql/src/test/results/clientpositive/infer_bucket_sort_bucketed_table.q.out
index 6db9428..33d795b 100644
--- a/ql/src/test/results/clientpositive/infer_bucket_sort_bucketed_table.q.out
+++ b/ql/src/test/results/clientpositive/infer_bucket_sort_bucketed_table.q.out
@@ -64,7 +64,7 @@ OutputFormat:         
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           3                        
 Bucket Columns:        [value]                  
-Sort Columns:          [Order(col:value, order:1, nullOrder:0)]         
+Sort Columns:          [Order(col:value, order:1)]      
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- If the count(*) from sampling the buckets matches the 
count(*) from each file, the table is

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/infer_bucket_sort_convert_join.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/infer_bucket_sort_convert_join.q.out 
b/ql/src/test/results/clientpositive/infer_bucket_sort_convert_join.q.out
index e5626a4..2f7e538 100644
--- a/ql/src/test/results/clientpositive/infer_bucket_sort_convert_join.q.out
+++ b/ql/src/test/results/clientpositive/infer_bucket_sort_convert_join.q.out
@@ -123,6 +123,6 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/infer_bucket_sort_dyn_part.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/infer_bucket_sort_dyn_part.q.out 
b/ql/src/test/results/clientpositive/infer_bucket_sort_dyn_part.q.out
index 5f983c7..7e3b48f 100644
--- a/ql/src/test/results/clientpositive/infer_bucket_sort_dyn_part.q.out
+++ b/ql/src/test/results/clientpositive/infer_bucket_sort_dyn_part.q.out
@@ -302,7 +302,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.RCFileOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: DESCRIBE FORMATTED test_table PARTITION (ds='2008-04-08', 
hr='12')
@@ -342,7 +342,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.RCFileOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: CREATE TABLE srcpart_merge_dp LIKE srcpart

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/infer_bucket_sort_grouping_operators.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/infer_bucket_sort_grouping_operators.q.out 
b/ql/src/test/results/clientpositive/infer_bucket_sort_grouping_operators.q.out
index 90a71ae..ebfce60 100644
--- 
a/ql/src/test/results/clientpositive/infer_bucket_sort_grouping_operators.q.out
+++ 
b/ql/src/test/results/clientpositive/infer_bucket_sort_grouping_operators.q.out
@@ -144,7 +144,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key, value]             
-Sort Columns:          [Order(col:key, order:1, nullOrder:0), Order(col:value, 
order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1), Order(col:value, order:1)]    
 
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test rollup, should be bucketed and sorted on key, value, 
grouping_key
@@ -203,7 +203,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key, value, grouping_key]       
-Sort Columns:          [Order(col:key, order:1, nullOrder:0), Order(col:value, 
order:1, nullOrder:0), Order(col:grouping_key, order:1, nullOrder:0)]    
+Sort Columns:          [Order(col:key, order:1), Order(col:value, order:1), 
Order(col:grouping_key, order:1)]   
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test cube, should not be bucketed or sorted because its 
missing the grouping ID
@@ -328,7 +328,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key, value]             
-Sort Columns:          [Order(col:key, order:1, nullOrder:0), Order(col:value, 
order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1), Order(col:value, order:1)]    
 
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test cube, should be bucketed and sorted on key, value, 
grouping_key
@@ -387,7 +387,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key, value, grouping_key]       
-Sort Columns:          [Order(col:key, order:1, nullOrder:0), Order(col:value, 
order:1, nullOrder:0), Order(col:grouping_key, order:1, nullOrder:0)]    
+Sort Columns:          [Order(col:key, order:1), Order(col:value, order:1), 
Order(col:grouping_key, order:1)]   
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test grouping sets, should not be bucketed or sorted 
because its missing the grouping ID
@@ -512,7 +512,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key, value]             
-Sort Columns:          [Order(col:key, order:1, nullOrder:0), Order(col:value, 
order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1), Order(col:value, order:1)]    
 
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test grouping sets, should be bucketed and sorted on key, 
value, grouping_key
@@ -571,6 +571,6 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key, value, grouping_key]       
-Sort Columns:          [Order(col:key, order:1, nullOrder:0), Order(col:value, 
order:1, nullOrder:0), Order(col:grouping_key, order:1, nullOrder:0)]    
+Sort Columns:          [Order(col:key, order:1), Order(col:value, order:1), 
Order(col:grouping_key, order:1)]   
 Storage Desc Params:            
        serialization.format    1                   

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/infer_bucket_sort_map_operators.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/infer_bucket_sort_map_operators.q.out 
b/ql/src/test/results/clientpositive/infer_bucket_sort_map_operators.q.out
index 5ebcf41..4ca6517 100644
--- a/ql/src/test/results/clientpositive/infer_bucket_sort_map_operators.q.out
+++ b/ql/src/test/results/clientpositive/infer_bucket_sort_map_operators.q.out
@@ -361,7 +361,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [value]                  
-Sort Columns:          [Order(col:value, order:1, nullOrder:0)]         
+Sort Columns:          [Order(col:value, order:1)]      
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test SMB join doesn't affect inference, should not be 
bucketed or sorted
@@ -640,6 +640,6 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/infer_bucket_sort_merge.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/infer_bucket_sort_merge.q.out 
b/ql/src/test/results/clientpositive/infer_bucket_sort_merge.q.out
index 97cb92d..53407c5 100644
--- a/ql/src/test/results/clientpositive/infer_bucket_sort_merge.q.out
+++ b/ql/src/test/results/clientpositive/infer_bucket_sort_merge.q.out
@@ -117,6 +117,6 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           2                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/infer_bucket_sort_multi_insert.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/infer_bucket_sort_multi_insert.q.out 
b/ql/src/test/results/clientpositive/infer_bucket_sort_multi_insert.q.out
index e59026e..d4c22f4 100644
--- a/ql/src/test/results/clientpositive/infer_bucket_sort_multi_insert.q.out
+++ b/ql/src/test/results/clientpositive/infer_bucket_sort_multi_insert.q.out
@@ -172,7 +172,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: DESCRIBE FORMATTED test_table PARTITION (part = '2')
@@ -211,7 +211,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [value]                  
-Sort Columns:          [Order(col:value, order:1, nullOrder:0)]         
+Sort Columns:          [Order(col:value, order:1)]      
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- The first partition should be bucketed and sorted, the 
second should not
@@ -272,7 +272,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: DESCRIBE FORMATTED test_table PARTITION (part = '2')
@@ -372,7 +372,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: DESCRIBE FORMATTED test_table PARTITION (part = '2')
@@ -411,6 +411,6 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/infer_bucket_sort_reducers_power_two.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/infer_bucket_sort_reducers_power_two.q.out 
b/ql/src/test/results/clientpositive/infer_bucket_sort_reducers_power_two.q.out
index 91466f1..1e4db29 100644
--- 
a/ql/src/test/results/clientpositive/infer_bucket_sort_reducers_power_two.q.out
+++ 
b/ql/src/test/results/clientpositive/infer_bucket_sort_reducers_power_two.q.out
@@ -64,7 +64,7 @@ OutputFormat:         
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test join, should be bucketed and sorted by join key
@@ -117,7 +117,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test join with two keys, should be bucketed and sorted by 
join keys
@@ -170,7 +170,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [key, value]             
-Sort Columns:          [Order(col:key, order:1, nullOrder:0), Order(col:value, 
order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1), Order(col:value, order:1)]    
 
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test join on three tables on same key, should be bucketed 
and sorted by join key
@@ -223,7 +223,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test join on three tables on different keys, should be 
bucketed and sorted by latter key
@@ -276,7 +276,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           16                       
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- Test group by in subquery with another group by outside, 
should be bucketed and sorted by the
@@ -331,6 +331,6 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [value]                  
-Sort Columns:          [Order(col:value, order:1, nullOrder:0)]         
+Sort Columns:          [Order(col:value, order:1)]      
 Storage Desc Params:            
        serialization.format    1                   

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/orc_analyze.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/orc_analyze.q.out 
b/ql/src/test/results/clientpositive/orc_analyze.q.out
index 9b7e7b7..87855fa 100644
--- a/ql/src/test/results/clientpositive/orc_analyze.q.out
+++ b/ql/src/test/results/clientpositive/orc_analyze.q.out
@@ -917,7 +917,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [first_name]             
-Sort Columns:          [Order(col:last_name, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:last_name, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: desc formatted orc_create_people partition(state="Or")
@@ -960,7 +960,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [first_name]             
-Sort Columns:          [Order(col:last_name, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:last_name, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: analyze table orc_create_people partition(state) compute 
statistics partialscan
@@ -1015,7 +1015,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [first_name]             
-Sort Columns:          [Order(col:last_name, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:last_name, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: desc formatted orc_create_people partition(state="Or")
@@ -1058,7 +1058,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [first_name]             
-Sort Columns:          [Order(col:last_name, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:last_name, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: analyze table orc_create_people partition(state) compute 
statistics noscan
@@ -1113,7 +1113,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [first_name]             
-Sort Columns:          [Order(col:last_name, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:last_name, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: desc formatted orc_create_people partition(state="Or")
@@ -1156,7 +1156,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [first_name]             
-Sort Columns:          [Order(col:last_name, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:last_name, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: drop table orc_create_people
@@ -1262,7 +1262,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [first_name]             
-Sort Columns:          [Order(col:last_name, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:last_name, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: desc formatted orc_create_people partition(state="Or")
@@ -1305,7 +1305,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [first_name]             
-Sort Columns:          [Order(col:last_name, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:last_name, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: drop table orc_create_people

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/partition_coltype_literals.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/partition_coltype_literals.q.out 
b/ql/src/test/results/clientpositive/partition_coltype_literals.q.out
index 1a93b7a..bc159eb 100644
--- a/ql/src/test/results/clientpositive/partition_coltype_literals.q.out
+++ b/ql/src/test/results/clientpositive/partition_coltype_literals.q.out
@@ -154,7 +154,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.RCFileOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:0, nullOrder:1)]   
+Sort Columns:          [Order(col:key, order:0)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- rename partition
@@ -205,7 +205,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.RCFileOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:0, nullOrder:1)]   
+Sort Columns:          [Order(col:key, order:0)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- insert partition

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/show_create_table_alter.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/show_create_table_alter.q.out 
b/ql/src/test/results/clientpositive/show_create_table_alter.q.out
index 78a34de..32819ea 100644
--- a/ql/src/test/results/clientpositive/show_create_table_alter.q.out
+++ b/ql/src/test/results/clientpositive/show_create_table_alter.q.out
@@ -24,7 +24,7 @@ CREATE EXTERNAL TABLE `tmp_showcrt1`(
 CLUSTERED BY ( 
   key) 
 SORTED BY ( 
-  value DESC NULLS LAST) 
+  value DESC) 
 INTO 5 BUCKETS
 ROW FORMAT SERDE 
   'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' 
@@ -59,7 +59,7 @@ COMMENT 'temporary table'
 CLUSTERED BY ( 
   key) 
 SORTED BY ( 
-  value DESC NULLS LAST) 
+  value DESC) 
 INTO 5 BUCKETS
 ROW FORMAT SERDE 
   'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' 
@@ -98,7 +98,7 @@ COMMENT 'changed comment'
 CLUSTERED BY ( 
   key) 
 SORTED BY ( 
-  value DESC NULLS LAST) 
+  value DESC) 
 INTO 5 BUCKETS
 ROW FORMAT SERDE 
   'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' 
@@ -136,7 +136,7 @@ COMMENT 'changed comment'
 CLUSTERED BY ( 
   key) 
 SORTED BY ( 
-  value DESC NULLS LAST) 
+  value DESC) 
 INTO 5 BUCKETS
 ROW FORMAT SERDE 
   'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' 
@@ -174,7 +174,7 @@ COMMENT 'changed comment'
 CLUSTERED BY ( 
   key) 
 SORTED BY ( 
-  value DESC NULLS LAST) 
+  value DESC) 
 INTO 5 BUCKETS
 ROW FORMAT SERDE 
   'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' 

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/spark/bucket5.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/bucket5.q.out 
b/ql/src/test/results/clientpositive/spark/bucket5.q.out
index af6d399..5baf054 100644
--- a/ql/src/test/results/clientpositive/spark/bucket5.q.out
+++ b/ql/src/test/results/clientpositive/spark/bucket5.q.out
@@ -391,7 +391,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           2                        
 Bucket Columns:        [key]                    
-Sort Columns:          [Order(col:key, order:1, nullOrder:0)]   
+Sort Columns:          [Order(col:key, order:1)]        
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: SELECT * FROM bucketed_table TABLESAMPLE (BUCKET 1 OUT OF 2) s 
LIMIT 10

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/spark/infer_bucket_sort_bucketed_table.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/spark/infer_bucket_sort_bucketed_table.q.out
 
b/ql/src/test/results/clientpositive/spark/infer_bucket_sort_bucketed_table.q.out
index 6db9428..33d795b 100644
--- 
a/ql/src/test/results/clientpositive/spark/infer_bucket_sort_bucketed_table.q.out
+++ 
b/ql/src/test/results/clientpositive/spark/infer_bucket_sort_bucketed_table.q.out
@@ -64,7 +64,7 @@ OutputFormat:         
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           3                        
 Bucket Columns:        [value]                  
-Sort Columns:          [Order(col:value, order:1, nullOrder:0)]         
+Sort Columns:          [Order(col:value, order:1)]      
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: -- If the count(*) from sampling the buckets matches the 
count(*) from each file, the table is

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out 
b/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out
index f7a94ac..a90e3f6 100644
--- 
a/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out
+++ 
b/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out
@@ -1215,7 +1215,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: desc formatted over1k_part_buck_sort_orc 
partition(t="__HIVE_DEFAULT_PARTITION__")
@@ -1256,7 +1256,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: select count(*) from over1k_part_orc
@@ -2255,7 +2255,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: desc formatted over1k_part_buck_sort2_orc 
partition(t="__HIVE_DEFAULT_PARTITION__")
@@ -2296,7 +2296,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: explain select * from over1k_part_buck_sort2_orc
@@ -2471,7 +2471,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: desc formatted over1k_part_buck_sort2_orc 
partition(t="__HIVE_DEFAULT_PARTITION__")
@@ -2512,7 +2512,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: explain select * from over1k_part_buck_sort2_orc

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/tez/dynpart_sort_optimization.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/tez/dynpart_sort_optimization.q.out 
b/ql/src/test/results/clientpositive/tez/dynpart_sort_optimization.q.out
index 01dc4ae..5292106 100644
--- a/ql/src/test/results/clientpositive/tez/dynpart_sort_optimization.q.out
+++ b/ql/src/test/results/clientpositive/tez/dynpart_sort_optimization.q.out
@@ -1128,7 +1128,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: desc formatted over1k_part_buck_sort 
partition(t="__HIVE_DEFAULT_PARTITION__")
@@ -1169,7 +1169,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: select count(*) from over1k_part
@@ -2152,7 +2152,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: desc formatted over1k_part_buck_sort2 
partition(t="__HIVE_DEFAULT_PARTITION__")
@@ -2193,7 +2193,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: select * from over1k_part_buck_sort2
@@ -2295,7 +2295,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: desc formatted over1k_part_buck_sort2 
partition(t="__HIVE_DEFAULT_PARTITION__")
@@ -2336,7 +2336,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
 Compressed:            No                       
 Num Buckets:           1                        
 Bucket Columns:        [si]                     
-Sort Columns:          [Order(col:f, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:f, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: select * from over1k_part_buck_sort2

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/ql/src/test/results/clientpositive/tez/orc_analyze.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/tez/orc_analyze.q.out 
b/ql/src/test/results/clientpositive/tez/orc_analyze.q.out
index 9b7e7b7..87855fa 100644
--- a/ql/src/test/results/clientpositive/tez/orc_analyze.q.out
+++ b/ql/src/test/results/clientpositive/tez/orc_analyze.q.out
@@ -917,7 +917,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [first_name]             
-Sort Columns:          [Order(col:last_name, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:last_name, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: desc formatted orc_create_people partition(state="Or")
@@ -960,7 +960,7 @@ OutputFormat:               
org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [first_name]             
-Sort Columns:          [Order(col:last_name, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:last_name, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: analyze table orc_create_people partition(state) compute 
statistics partialscan
@@ -1015,7 +1015,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [first_name]             
-Sort Columns:          [Order(col:last_name, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:last_name, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: desc formatted orc_create_people partition(state="Or")
@@ -1058,7 +1058,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [first_name]             
-Sort Columns:          [Order(col:last_name, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:last_name, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: analyze table orc_create_people partition(state) compute 
statistics noscan
@@ -1113,7 +1113,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [first_name]             
-Sort Columns:          [Order(col:last_name, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:last_name, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: desc formatted orc_create_people partition(state="Or")
@@ -1156,7 +1156,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [first_name]             
-Sort Columns:          [Order(col:last_name, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:last_name, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: drop table orc_create_people
@@ -1262,7 +1262,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [first_name]             
-Sort Columns:          [Order(col:last_name, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:last_name, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: desc formatted orc_create_people partition(state="Or")
@@ -1305,7 +1305,7 @@ OutputFormat:             
org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat
 Compressed:            No                       
 Num Buckets:           4                        
 Bucket Columns:        [first_name]             
-Sort Columns:          [Order(col:last_name, order:1, nullOrder:0)]     
+Sort Columns:          [Order(col:last_name, order:1)]  
 Storage Desc Params:            
        serialization.format    1                   
 PREHOOK: query: drop table orc_create_people

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/serde/if/serde.thrift
----------------------------------------------------------------------
diff --git a/serde/if/serde.thrift b/serde/if/serde.thrift
index 0b3804d..6caad36 100644
--- a/serde/if/serde.thrift
+++ b/serde/if/serde.thrift
@@ -30,7 +30,7 @@ const string SERIALIZATION_NULL_FORMAT = 
"serialization.null.format"
 const string SERIALIZATION_ESCAPE_CRLF = "serialization.escape.crlf"
 const string SERIALIZATION_LAST_COLUMN_TAKES_REST = 
"serialization.last.column.takes.rest"
 const string SERIALIZATION_SORT_ORDER = "serialization.sort.order"
-const string SERIALIZATION_NULL_POSITION = "serialization.null.position";
+const string SERIALIZATION_NULL_SORT_ORDER = "serialization.sort.order.null";
 const string SERIALIZATION_USE_JSON_OBJECTS = "serialization.use.json.object"
 const string SERIALIZATION_ENCODING = "serialization.encoding"
 

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/serde/src/gen/thrift/gen-cpp/serde_constants.cpp
----------------------------------------------------------------------
diff --git a/serde/src/gen/thrift/gen-cpp/serde_constants.cpp 
b/serde/src/gen/thrift/gen-cpp/serde_constants.cpp
index 75701e2..3a675bf 100644
--- a/serde/src/gen/thrift/gen-cpp/serde_constants.cpp
+++ b/serde/src/gen/thrift/gen-cpp/serde_constants.cpp
@@ -27,7 +27,7 @@ serdeConstants::serdeConstants() {
 
   SERIALIZATION_SORT_ORDER = "serialization.sort.order";
 
-  SERIALIZATION_NULL_POSITION = "serialization.null.position";
+  SERIALIZATION_NULL_SORT_ORDER = "serialization.sort.order.null";
 
   SERIALIZATION_USE_JSON_OBJECTS = "serialization.use.json.object";
 

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/serde/src/gen/thrift/gen-cpp/serde_constants.h
----------------------------------------------------------------------
diff --git a/serde/src/gen/thrift/gen-cpp/serde_constants.h 
b/serde/src/gen/thrift/gen-cpp/serde_constants.h
index 6d85928..a5f33fb 100644
--- a/serde/src/gen/thrift/gen-cpp/serde_constants.h
+++ b/serde/src/gen/thrift/gen-cpp/serde_constants.h
@@ -23,7 +23,7 @@ class serdeConstants {
   std::string SERIALIZATION_ESCAPE_CRLF;
   std::string SERIALIZATION_LAST_COLUMN_TAKES_REST;
   std::string SERIALIZATION_SORT_ORDER;
-  std::string SERIALIZATION_NULL_POSITION;
+  std::string SERIALIZATION_NULL_SORT_ORDER;
   std::string SERIALIZATION_USE_JSON_OBJECTS;
   std::string SERIALIZATION_ENCODING;
   std::string FIELD_DELIM;

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/serde/src/gen/thrift/gen-php/org/apache/hadoop/hive/serde/Types.php
----------------------------------------------------------------------
diff --git 
a/serde/src/gen/thrift/gen-php/org/apache/hadoop/hive/serde/Types.php 
b/serde/src/gen/thrift/gen-php/org/apache/hadoop/hive/serde/Types.php
index 0bc6dd7..18c3991 100644
--- a/serde/src/gen/thrift/gen-php/org/apache/hadoop/hive/serde/Types.php
+++ b/serde/src/gen/thrift/gen-php/org/apache/hadoop/hive/serde/Types.php
@@ -26,7 +26,7 @@ final class Constant extends \Thrift\Type\TConstant {
   static protected $SERIALIZATION_ESCAPE_CRLF;
   static protected $SERIALIZATION_LAST_COLUMN_TAKES_REST;
   static protected $SERIALIZATION_SORT_ORDER;
-  static protected $SERIALIZATION_NULL_POSITION;
+  static protected $SERIALIZATION_NULL_SORT_ORDER;
   static protected $SERIALIZATION_USE_JSON_OBJECTS;
   static protected $SERIALIZATION_ENCODING;
   static protected $FIELD_DELIM;
@@ -98,8 +98,8 @@ final class Constant extends \Thrift\Type\TConstant {
     return "serialization.sort.order";
   }
 
-  static protected function init_SERIALIZATION_NULL_POSITION() {
-    return "serialization.null.position";
+  static protected function init_SERIALIZATION_NULL_SORT_ORDER() {
+    return "serialization.sort.order.null";
   }
 
   static protected function init_SERIALIZATION_USE_JSON_OBJECTS() {

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/serde/src/gen/thrift/gen-py/org_apache_hadoop_hive_serde/constants.py
----------------------------------------------------------------------
diff --git 
a/serde/src/gen/thrift/gen-py/org_apache_hadoop_hive_serde/constants.py 
b/serde/src/gen/thrift/gen-py/org_apache_hadoop_hive_serde/constants.py
index 7939791..fafdc24 100644
--- a/serde/src/gen/thrift/gen-py/org_apache_hadoop_hive_serde/constants.py
+++ b/serde/src/gen/thrift/gen-py/org_apache_hadoop_hive_serde/constants.py
@@ -17,7 +17,7 @@ SERIALIZATION_NULL_FORMAT = "serialization.null.format"
 SERIALIZATION_ESCAPE_CRLF = "serialization.escape.crlf"
 SERIALIZATION_LAST_COLUMN_TAKES_REST = "serialization.last.column.takes.rest"
 SERIALIZATION_SORT_ORDER = "serialization.sort.order"
-SERIALIZATION_NULL_POSITION = "serialization.null.position"
+SERIALIZATION_NULL_SORT_ORDER = "serialization.sort.order.null"
 SERIALIZATION_USE_JSON_OBJECTS = "serialization.use.json.object"
 SERIALIZATION_ENCODING = "serialization.encoding"
 FIELD_DELIM = "field.delim"

http://git-wip-us.apache.org/repos/asf/hive/blob/cc8893e4/serde/src/gen/thrift/gen-rb/serde_constants.rb
----------------------------------------------------------------------
diff --git a/serde/src/gen/thrift/gen-rb/serde_constants.rb 
b/serde/src/gen/thrift/gen-rb/serde_constants.rb
index d09e3c2..0ce9f27 100644
--- a/serde/src/gen/thrift/gen-rb/serde_constants.rb
+++ b/serde/src/gen/thrift/gen-rb/serde_constants.rb
@@ -23,7 +23,7 @@ SERIALIZATION_LAST_COLUMN_TAKES_REST = 
%q"serialization.last.column.takes.rest"
 
 SERIALIZATION_SORT_ORDER = %q"serialization.sort.order"
 
-SERIALIZATION_NULL_POSITION = %q"serialization.null.position"
+SERIALIZATION_NULL_SORT_ORDER = %q"serialization.sort.order.null"
 
 SERIALIZATION_USE_JSON_OBJECTS = %q"serialization.use.json.object"
 

Reply via email to