Added: 
hive/branches/spark/ql/src/test/results/clientpositive/spark/infer_bucket_sort_reducers_power_two.q.out
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/infer_bucket_sort_reducers_power_two.q.out?rev=1656573&view=auto
==============================================================================
--- 
hive/branches/spark/ql/src/test/results/clientpositive/spark/infer_bucket_sort_reducers_power_two.q.out
 (added)
+++ 
hive/branches/spark/ql/src/test/results/clientpositive/spark/infer_bucket_sort_reducers_power_two.q.out
 Mon Feb  2 21:10:08 2015
@@ -0,0 +1,348 @@
+PREHOOK: query: -- This tests inferring how data is bucketed/sorted from the 
operators in the reducer
+-- and populating that information in partitions' metadata, it also verifies 
that the
+-- number of reducers chosen will be a power of two
+
+CREATE TABLE test_table (key STRING, value STRING) PARTITIONED BY (part STRING)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@test_table
+POSTHOOK: query: -- This tests inferring how data is bucketed/sorted from the 
operators in the reducer
+-- and populating that information in partitions' metadata, it also verifies 
that the
+-- number of reducers chosen will be a power of two
+
+CREATE TABLE test_table (key STRING, value STRING) PARTITIONED BY (part STRING)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@test_table
+PREHOOK: query: -- Test group by, should be bucketed and sorted by group by key
+INSERT OVERWRITE TABLE test_table PARTITION (part = '1') 
+SELECT key, count(*) FROM src GROUP BY key
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@test_table@part=1
+POSTHOOK: query: -- Test group by, should be bucketed and sorted by group by 
key
+INSERT OVERWRITE TABLE test_table PARTITION (part = '1') 
+SELECT key, count(*) FROM src GROUP BY key
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@test_table@part=1
+POSTHOOK: Lineage: test_table PARTITION(part=1).key SIMPLE 
[(src)src.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: test_table PARTITION(part=1).value EXPRESSION 
[(src)src.null, ]
+PREHOOK: query: DESCRIBE FORMATTED test_table PARTITION (part = '1')
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@test_table
+POSTHOOK: query: DESCRIBE FORMATTED test_table PARTITION (part = '1')
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@test_table
+# col_name             data_type               comment             
+                
+key                    string                                      
+value                  string                                      
+                
+# Partition Information                 
+# col_name             data_type               comment             
+                
+part                   string                                      
+                
+# Detailed Partition Information                
+Partition Value:       [1]                      
+Database:              default                  
+Table:                 test_table               
+#### A masked pattern was here ####
+Protect Mode:          None                     
+#### A masked pattern was here ####
+Partition Parameters:           
+       COLUMN_STATS_ACCURATE   true                
+       numFiles                5                   
+       numRows                 309                 
+       rawDataSize             1482                
+       totalSize               1791                
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
+PREHOOK: query: -- Test join, should be bucketed and sorted by join key
+INSERT OVERWRITE TABLE test_table PARTITION (part = '1') 
+SELECT a.key, a.value FROM src a JOIN src b ON a.key = b.key
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@test_table@part=1
+POSTHOOK: query: -- Test join, should be bucketed and sorted by join key
+INSERT OVERWRITE TABLE test_table PARTITION (part = '1') 
+SELECT a.key, a.value FROM src a JOIN src b ON a.key = b.key
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@test_table@part=1
+POSTHOOK: Lineage: test_table PARTITION(part=1).key SIMPLE 
[(src)a.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: test_table PARTITION(part=1).value SIMPLE 
[(src)a.FieldSchema(name:value, type:string, comment:default), ]
+PREHOOK: query: DESCRIBE FORMATTED test_table PARTITION (part = '1')
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@test_table
+POSTHOOK: query: DESCRIBE FORMATTED test_table PARTITION (part = '1')
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@test_table
+# col_name             data_type               comment             
+                
+key                    string                                      
+value                  string                                      
+                
+# Partition Information                 
+# col_name             data_type               comment             
+                
+part                   string                                      
+                
+# Detailed Partition Information                
+Partition Value:       [1]                      
+Database:              default                  
+Table:                 test_table               
+#### A masked pattern was here ####
+Protect Mode:          None                     
+#### A masked pattern was here ####
+Partition Parameters:           
+       COLUMN_STATS_ACCURATE   true                
+       numFiles                5                   
+       numRows                 1028                
+       rawDataSize             10968               
+       totalSize               11996               
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
+PREHOOK: query: -- Test join with two keys, should be bucketed and sorted by 
join keys
+INSERT OVERWRITE TABLE test_table PARTITION (part = '1') 
+SELECT a.key, a.value FROM src a JOIN src b ON a.key = b.key AND a.value = 
b.value
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@test_table@part=1
+POSTHOOK: query: -- Test join with two keys, should be bucketed and sorted by 
join keys
+INSERT OVERWRITE TABLE test_table PARTITION (part = '1') 
+SELECT a.key, a.value FROM src a JOIN src b ON a.key = b.key AND a.value = 
b.value
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@test_table@part=1
+POSTHOOK: Lineage: test_table PARTITION(part=1).key SIMPLE 
[(src)a.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: test_table PARTITION(part=1).value SIMPLE 
[(src)a.FieldSchema(name:value, type:string, comment:default), ]
+PREHOOK: query: DESCRIBE FORMATTED test_table PARTITION (part = '1')
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@test_table
+POSTHOOK: query: DESCRIBE FORMATTED test_table PARTITION (part = '1')
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@test_table
+# col_name             data_type               comment             
+                
+key                    string                                      
+value                  string                                      
+                
+# Partition Information                 
+# col_name             data_type               comment             
+                
+part                   string                                      
+                
+# Detailed Partition Information                
+Partition Value:       [1]                      
+Database:              default                  
+Table:                 test_table               
+#### A masked pattern was here ####
+Protect Mode:          None                     
+#### A masked pattern was here ####
+Partition Parameters:           
+       COLUMN_STATS_ACCURATE   true                
+       numFiles                4                   
+       numRows                 1028                
+       rawDataSize             10968               
+       totalSize               11996               
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
+PREHOOK: query: -- Test join on three tables on same key, should be bucketed 
and sorted by join key
+INSERT OVERWRITE TABLE test_table PARTITION (part = '1') 
+SELECT a.key, c.value FROM src a JOIN src b ON (a.key = b.key) JOIN src c ON 
(b.key = c.key)
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@test_table@part=1
+POSTHOOK: query: -- Test join on three tables on same key, should be bucketed 
and sorted by join key
+INSERT OVERWRITE TABLE test_table PARTITION (part = '1') 
+SELECT a.key, c.value FROM src a JOIN src b ON (a.key = b.key) JOIN src c ON 
(b.key = c.key)
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@test_table@part=1
+POSTHOOK: Lineage: test_table PARTITION(part=1).key SIMPLE 
[(src)a.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: test_table PARTITION(part=1).value SIMPLE 
[(src)a.FieldSchema(name:value, type:string, comment:default), ]
+PREHOOK: query: DESCRIBE FORMATTED test_table PARTITION (part = '1')
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@test_table
+POSTHOOK: query: DESCRIBE FORMATTED test_table PARTITION (part = '1')
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@test_table
+# col_name             data_type               comment             
+                
+key                    string                                      
+value                  string                                      
+                
+# Partition Information                 
+# col_name             data_type               comment             
+                
+part                   string                                      
+                
+# Detailed Partition Information                
+Partition Value:       [1]                      
+Database:              default                  
+Table:                 test_table               
+#### A masked pattern was here ####
+Protect Mode:          None                     
+#### A masked pattern was here ####
+Partition Parameters:           
+       COLUMN_STATS_ACCURATE   true                
+       numFiles                7                   
+       numRows                 2654                
+       rawDataSize             28466               
+       totalSize               31120               
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
+PREHOOK: query: -- Test join on three tables on different keys, should be 
bucketed and sorted by latter key
+INSERT OVERWRITE TABLE test_table PARTITION (part = '1') 
+SELECT a.key, c.value FROM src a JOIN src b ON (a.key = b.key) JOIN src c ON 
(b.value = c.value)
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@test_table@part=1
+POSTHOOK: query: -- Test join on three tables on different keys, should be 
bucketed and sorted by latter key
+INSERT OVERWRITE TABLE test_table PARTITION (part = '1') 
+SELECT a.key, c.value FROM src a JOIN src b ON (a.key = b.key) JOIN src c ON 
(b.value = c.value)
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@test_table@part=1
+POSTHOOK: Lineage: test_table PARTITION(part=1).key SIMPLE 
[(src)a.FieldSchema(name:key, type:string, comment:default), ]
+POSTHOOK: Lineage: test_table PARTITION(part=1).value SIMPLE 
[(src)a.FieldSchema(name:value, type:string, comment:default), ]
+PREHOOK: query: DESCRIBE FORMATTED test_table PARTITION (part = '1')
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@test_table
+POSTHOOK: query: DESCRIBE FORMATTED test_table PARTITION (part = '1')
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@test_table
+# col_name             data_type               comment             
+                
+key                    string                                      
+value                  string                                      
+                
+# Partition Information                 
+# col_name             data_type               comment             
+                
+part                   string                                      
+                
+# Detailed Partition Information                
+Partition Value:       [1]                      
+Database:              default                  
+Table:                 test_table               
+#### A masked pattern was here ####
+Protect Mode:          None                     
+#### A masked pattern was here ####
+Partition Parameters:           
+       COLUMN_STATS_ACCURATE   true                
+       numFiles                5                   
+       numRows                 2654                
+       rawDataSize             28466               
+       totalSize               31120               
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
+PREHOOK: query: -- Test group by in subquery with another group by outside, 
should be bucketed and sorted by the
+-- key of the outer group by
+INSERT OVERWRITE TABLE test_table PARTITION (part = '1')
+SELECT count(1), value FROM (SELECT key, count(1) as value FROM src group by 
key) a group by value
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@test_table@part=1
+POSTHOOK: query: -- Test group by in subquery with another group by outside, 
should be bucketed and sorted by the
+-- key of the outer group by
+INSERT OVERWRITE TABLE test_table PARTITION (part = '1')
+SELECT count(1), value FROM (SELECT key, count(1) as value FROM src group by 
key) a group by value
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@test_table@part=1
+POSTHOOK: Lineage: test_table PARTITION(part=1).key EXPRESSION [(src)src.null, 
]
+POSTHOOK: Lineage: test_table PARTITION(part=1).value EXPRESSION 
[(src)src.null, ]
+PREHOOK: query: DESCRIBE FORMATTED test_table PARTITION (part = '1')
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@test_table
+POSTHOOK: query: DESCRIBE FORMATTED test_table PARTITION (part = '1')
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@test_table
+# col_name             data_type               comment             
+                
+key                    string                                      
+value                  string                                      
+                
+# Partition Information                 
+# col_name             data_type               comment             
+                
+part                   string                                      
+                
+# Detailed Partition Information                
+Partition Value:       [1]                      
+Database:              default                  
+Table:                 test_table               
+#### A masked pattern was here ####
+Protect Mode:          None                     
+#### A masked pattern was here ####
+Partition Parameters:           
+       COLUMN_STATS_ACCURATE   true                
+       numFiles                4                   
+       numRows                 5                   
+       rawDataSize             19                  
+       totalSize               24                  
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   

Added: 
hive/branches/spark/ql/src/test/results/clientpositive/spark/input16_cc.q.out
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/input16_cc.q.out?rev=1656573&view=auto
==============================================================================
--- 
hive/branches/spark/ql/src/test/results/clientpositive/spark/input16_cc.q.out 
(added)
+++ 
hive/branches/spark/ql/src/test/results/clientpositive/spark/input16_cc.q.out 
Mon Feb  2 21:10:08 2015
@@ -0,0 +1,534 @@
+PREHOOK: query: -- TestSerDe is a user defined serde where the default 
delimiter is Ctrl-B
+-- the user is overwriting it with ctrlC
+
+DROP TABLE INPUT16_CC
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: -- TestSerDe is a user defined serde where the default 
delimiter is Ctrl-B
+-- the user is overwriting it with ctrlC
+
+DROP TABLE INPUT16_CC
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: CREATE TABLE INPUT16_CC(KEY STRING, VALUE STRING) ROW FORMAT 
SERDE 'org.apache.hadoop.hive.serde2.TestSerDe'  with serdeproperties 
('testserde.default.serialization.format'='\003', 
'dummy.prop.not.used'='dummyy.val') STORED AS TEXTFILE
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@INPUT16_CC
+POSTHOOK: query: CREATE TABLE INPUT16_CC(KEY STRING, VALUE STRING) ROW FORMAT 
SERDE 'org.apache.hadoop.hive.serde2.TestSerDe'  with serdeproperties 
('testserde.default.serialization.format'='\003', 
'dummy.prop.not.used'='dummyy.val') STORED AS TEXTFILE
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@INPUT16_CC
+PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/kv1_cc.txt' INTO 
TABLE INPUT16_CC
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@input16_cc
+POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/kv1_cc.txt' INTO 
TABLE INPUT16_CC
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@input16_cc
+PREHOOK: query: SELECT INPUT16_CC.VALUE, INPUT16_CC.KEY FROM INPUT16_CC
+PREHOOK: type: QUERY
+PREHOOK: Input: default@input16_cc
+#### A masked pattern was here ####
+POSTHOOK: query: SELECT INPUT16_CC.VALUE, INPUT16_CC.KEY FROM INPUT16_CC
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@input16_cc
+#### A masked pattern was here ####
+val_238        238
+val_86 86
+val_311        311
+val_27 27
+val_165        165
+val_409        409
+val_255        255
+val_278        278
+val_98 98
+val_484        484
+val_265        265
+val_193        193
+val_401        401
+val_150        150
+val_273        273
+val_224        224
+val_369        369
+val_66 66
+val_128        128
+val_213        213
+val_146        146
+val_406        406
+val_429        429
+val_374        374
+val_152        152
+val_469        469
+val_145        145
+val_495        495
+val_37 37
+val_327        327
+val_281        281
+val_277        277
+val_209        209
+val_15 15
+val_82 82
+val_403        403
+val_166        166
+val_417        417
+val_430        430
+val_252        252
+val_292        292
+val_219        219
+val_287        287
+val_153        153
+val_193        193
+val_338        338
+val_446        446
+val_459        459
+val_394        394
+val_237        237
+val_482        482
+val_174        174
+val_413        413
+val_494        494
+val_207        207
+val_199        199
+val_466        466
+val_208        208
+val_174        174
+val_399        399
+val_396        396
+val_247        247
+val_417        417
+val_489        489
+val_162        162
+val_377        377
+val_397        397
+val_309        309
+val_365        365
+val_266        266
+val_439        439
+val_342        342
+val_367        367
+val_325        325
+val_167        167
+val_195        195
+val_475        475
+val_17 17
+val_113        113
+val_155        155
+val_203        203
+val_339        339
+val_0  0
+val_455        455
+val_128        128
+val_311        311
+val_316        316
+val_57 57
+val_302        302
+val_205        205
+val_149        149
+val_438        438
+val_345        345
+val_129        129
+val_170        170
+val_20 20
+val_489        489
+val_157        157
+val_378        378
+val_221        221
+val_92 92
+val_111        111
+val_47 47
+val_72 72
+val_4  4
+val_280        280
+val_35 35
+val_427        427
+val_277        277
+val_208        208
+val_356        356
+val_399        399
+val_169        169
+val_382        382
+val_498        498
+val_125        125
+val_386        386
+val_437        437
+val_469        469
+val_192        192
+val_286        286
+val_187        187
+val_176        176
+val_54 54
+val_459        459
+val_51 51
+val_138        138
+val_103        103
+val_239        239
+val_213        213
+val_216        216
+val_430        430
+val_278        278
+val_176        176
+val_289        289
+val_221        221
+val_65 65
+val_318        318
+val_332        332
+val_311        311
+val_275        275
+val_137        137
+val_241        241
+val_83 83
+val_333        333
+val_180        180
+val_284        284
+val_12 12
+val_230        230
+val_181        181
+val_67 67
+val_260        260
+val_404        404
+val_384        384
+val_489        489
+val_353        353
+val_373        373
+val_272        272
+val_138        138
+val_217        217
+val_84 84
+val_348        348
+val_466        466
+val_58 58
+val_8  8
+val_411        411
+val_230        230
+val_208        208
+val_348        348
+val_24 24
+val_463        463
+val_431        431
+val_179        179
+val_172        172
+val_42 42
+val_129        129
+val_158        158
+val_119        119
+val_496        496
+val_0  0
+val_322        322
+val_197        197
+val_468        468
+val_393        393
+val_454        454
+val_100        100
+val_298        298
+val_199        199
+val_191        191
+val_418        418
+val_96 96
+val_26 26
+val_165        165
+val_327        327
+val_230        230
+val_205        205
+val_120        120
+val_131        131
+val_51 51
+val_404        404
+val_43 43
+val_436        436
+val_156        156
+val_469        469
+val_468        468
+val_308        308
+val_95 95
+val_196        196
+val_288        288
+val_481        481
+val_457        457
+val_98 98
+val_282        282
+val_197        197
+val_187        187
+val_318        318
+val_318        318
+val_409        409
+val_470        470
+val_137        137
+val_369        369
+val_316        316
+val_169        169
+val_413        413
+val_85 85
+val_77 77
+val_0  0
+val_490        490
+val_87 87
+val_364        364
+val_179        179
+val_118        118
+val_134        134
+val_395        395
+val_282        282
+val_138        138
+val_238        238
+val_419        419
+val_15 15
+val_118        118
+val_72 72
+val_90 90
+val_307        307
+val_19 19
+val_435        435
+val_10 10
+val_277        277
+val_273        273
+val_306        306
+val_224        224
+val_309        309
+val_389        389
+val_327        327
+val_242        242
+val_369        369
+val_392        392
+val_272        272
+val_331        331
+val_401        401
+val_242        242
+val_452        452
+val_177        177
+val_226        226
+val_5  5
+val_497        497
+val_402        402
+val_396        396
+val_317        317
+val_395        395
+val_58 58
+val_35 35
+val_336        336
+val_95 95
+val_11 11
+val_168        168
+val_34 34
+val_229        229
+val_233        233
+val_143        143
+val_472        472
+val_322        322
+val_498        498
+val_160        160
+val_195        195
+val_42 42
+val_321        321
+val_430        430
+val_119        119
+val_489        489
+val_458        458
+val_78 78
+val_76 76
+val_41 41
+val_223        223
+val_492        492
+val_149        149
+val_449        449
+val_218        218
+val_228        228
+val_138        138
+val_453        453
+val_30 30
+val_209        209
+val_64 64
+val_468        468
+val_76 76
+val_74 74
+val_342        342
+val_69 69
+val_230        230
+val_33 33
+val_368        368
+val_103        103
+val_296        296
+val_113        113
+val_216        216
+val_367        367
+val_344        344
+val_167        167
+val_274        274
+val_219        219
+val_239        239
+val_485        485
+val_116        116
+val_223        223
+val_256        256
+val_263        263
+val_70 70
+val_487        487
+val_480        480
+val_401        401
+val_288        288
+val_191        191
+val_5  5
+val_244        244
+val_438        438
+val_128        128
+val_467        467
+val_432        432
+val_202        202
+val_316        316
+val_229        229
+val_469        469
+val_463        463
+val_280        280
+val_2  2
+val_35 35
+val_283        283
+val_331        331
+val_235        235
+val_80 80
+val_44 44
+val_193        193
+val_321        321
+val_335        335
+val_104        104
+val_466        466
+val_366        366
+val_175        175
+val_403        403
+val_483        483
+val_53 53
+val_105        105
+val_257        257
+val_406        406
+val_409        409
+val_190        190
+val_406        406
+val_401        401
+val_114        114
+val_258        258
+val_90 90
+val_203        203
+val_262        262
+val_348        348
+val_424        424
+val_12 12
+val_396        396
+val_201        201
+val_217        217
+val_164        164
+val_431        431
+val_454        454
+val_478        478
+val_298        298
+val_125        125
+val_431        431
+val_164        164
+val_424        424
+val_187        187
+val_382        382
+val_5  5
+val_70 70
+val_397        397
+val_480        480
+val_291        291
+val_24 24
+val_351        351
+val_255        255
+val_104        104
+val_70 70
+val_163        163
+val_438        438
+val_119        119
+val_414        414
+val_200        200
+val_491        491
+val_237        237
+val_439        439
+val_360        360
+val_248        248
+val_479        479
+val_305        305
+val_417        417
+val_199        199
+val_444        444
+val_120        120
+val_429        429
+val_169        169
+val_443        443
+val_323        323
+val_325        325
+val_277        277
+val_230        230
+val_478        478
+val_178        178
+val_468        468
+val_310        310
+val_317        317
+val_333        333
+val_493        493
+val_460        460
+val_207        207
+val_249        249
+val_265        265
+val_480        480
+val_83 83
+val_136        136
+val_353        353
+val_172        172
+val_214        214
+val_462        462
+val_233        233
+val_406        406
+val_133        133
+val_175        175
+val_189        189
+val_454        454
+val_375        375
+val_401        401
+val_421        421
+val_407        407
+val_384        384
+val_256        256
+val_26 26
+val_134        134
+val_67 67
+val_384        384
+val_379        379
+val_18 18
+val_462        462
+val_492        492
+val_100        100
+val_298        298
+val_9  9
+val_341        341
+val_498        498
+val_146        146
+val_458        458
+val_362        362
+val_186        186
+val_285        285
+val_348        348
+val_167        167
+val_18 18
+val_273        273
+val_183        183
+val_281        281
+val_344        344
+val_97 97
+val_469        469
+val_315        315
+val_84 84
+val_28 28
+val_37 37
+val_448        448
+val_152        152
+val_348        348
+val_307        307
+val_194        194
+val_414        414
+val_477        477
+val_222        222
+val_126        126
+val_90 90
+val_169        169
+val_403        403
+val_400        400
+val_200        200
+val_97 97

Added: 
hive/branches/spark/ql/src/test/results/clientpositive/spark/list_bucket_dml_10.q.java1.7.out
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/list_bucket_dml_10.q.java1.7.out?rev=1656573&view=auto
==============================================================================
--- 
hive/branches/spark/ql/src/test/results/clientpositive/spark/list_bucket_dml_10.q.java1.7.out
 (added)
+++ 
hive/branches/spark/ql/src/test/results/clientpositive/spark/list_bucket_dml_10.q.java1.7.out
 Mon Feb  2 21:10:08 2015
@@ -0,0 +1,282 @@
+PREHOOK: query: -- run this test case in minimr to ensure it works in cluster
+-- JAVA_VERSION_SPECIFIC_OUTPUT
+
+-- list bucketing DML: static partition. multiple skewed columns.
+-- ds=2008-04-08/hr=11/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME:
+--  5263 000000_0
+--  5263 000001_0
+-- ds=2008-04-08/hr=11/key=103/value=val_103:
+-- 99 000000_0
+-- 99 000001_0
+-- ds=2008-04-08/hr=11/key=484/value=val_484:
+-- 87 000000_0
+-- 87 000001_0
+
+-- create a skewed table
+create table list_bucketing_static_part (key String, value String) 
+    partitioned by (ds String, hr String) 
+    skewed by (key) on ('484','51','103')
+    stored as DIRECTORIES
+    STORED AS RCFILE
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@list_bucketing_static_part
+POSTHOOK: query: -- run this test case in minimr to ensure it works in cluster
+-- JAVA_VERSION_SPECIFIC_OUTPUT
+
+-- list bucketing DML: static partition. multiple skewed columns.
+-- ds=2008-04-08/hr=11/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME:
+--  5263 000000_0
+--  5263 000001_0
+-- ds=2008-04-08/hr=11/key=103/value=val_103:
+-- 99 000000_0
+-- 99 000001_0
+-- ds=2008-04-08/hr=11/key=484/value=val_484:
+-- 87 000000_0
+-- 87 000001_0
+
+-- create a skewed table
+create table list_bucketing_static_part (key String, value String) 
+    partitioned by (ds String, hr String) 
+    skewed by (key) on ('484','51','103')
+    stored as DIRECTORIES
+    STORED AS RCFILE
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@list_bucketing_static_part
+PREHOOK: query: -- list bucketing DML without merge. use bucketize to generate 
a few small files.
+explain extended
+insert overwrite table list_bucketing_static_part partition (ds = 
'2008-04-08',  hr = '11')
+select key, value from src
+PREHOOK: type: QUERY
+POSTHOOK: query: -- list bucketing DML without merge. use bucketize to 
generate a few small files.
+explain extended
+insert overwrite table list_bucketing_static_part partition (ds = 
'2008-04-08',  hr = '11')
+select key, value from src
+POSTHOOK: type: QUERY
+ABSTRACT SYNTAX TREE:
+  
+TOK_QUERY
+   TOK_FROM
+      TOK_TABREF
+         TOK_TABNAME
+            src
+   TOK_INSERT
+      TOK_DESTINATION
+         TOK_TAB
+            TOK_TABNAME
+               list_bucketing_static_part
+            TOK_PARTSPEC
+               TOK_PARTVAL
+                  ds
+                  '2008-04-08'
+               TOK_PARTVAL
+                  hr
+                  '11'
+      TOK_SELECT
+         TOK_SELEXPR
+            TOK_TABLE_OR_COL
+               key
+         TOK_SELEXPR
+            TOK_TABLE_OR_COL
+               value
+
+
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 depends on stages: Stage-1
+  Stage-2 depends on stages: Stage-0
+
+STAGE PLANS:
+  Stage: Stage-1
+    Spark
+#### A masked pattern was here ####
+      Vertices:
+        Map 1 
+            Map Operator Tree:
+                TableScan
+                  alias: src
+                  Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
+                  GatherStats: false
+                  Select Operator
+                    expressions: key (type: string), value (type: string)
+                    outputColumnNames: _col0, _col1
+                    Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
+                    File Output Operator
+                      compressed: false
+                      GlobalTableId: 1
+#### A masked pattern was here ####
+                      NumFilesPerFileSink: 1
+                      Static Partition Specification: ds=2008-04-08/hr=11/
+                      Statistics: Num rows: 500 Data size: 5312 Basic stats: 
COMPLETE Column stats: NONE
+#### A masked pattern was here ####
+                      table:
+                          input format: 
org.apache.hadoop.hive.ql.io.RCFileInputFormat
+                          output format: 
org.apache.hadoop.hive.ql.io.RCFileOutputFormat
+                          properties:
+                            bucket_count -1
+                            columns key,value
+                            columns.comments 
+                            columns.types string:string
+#### A masked pattern was here ####
+                            name default.list_bucketing_static_part
+                            partition_columns ds/hr
+                            partition_columns.types string:string
+                            serialization.ddl struct 
list_bucketing_static_part { string key, string value}
+                            serialization.format 1
+                            serialization.lib 
org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
+#### A masked pattern was here ####
+                          serde: 
org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
+                          name: default.list_bucketing_static_part
+                      TotalFiles: 1
+                      GatherStats: true
+                      MultiFileSpray: false
+            Path -> Alias:
+#### A masked pattern was here ####
+            Path -> Partition:
+#### A masked pattern was here ####
+                Partition
+                  base file name: src
+                  input format: org.apache.hadoop.mapred.TextInputFormat
+                  output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                  properties:
+                    COLUMN_STATS_ACCURATE true
+                    bucket_count -1
+                    columns key,value
+                    columns.comments 'default','default'
+                    columns.types string:string
+#### A masked pattern was here ####
+                    name default.src
+                    numFiles 1
+                    numRows 500
+                    rawDataSize 5312
+                    serialization.ddl struct src { string key, string value}
+                    serialization.format 1
+                    serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                    totalSize 5812
+#### A masked pattern was here ####
+                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                
+                    input format: org.apache.hadoop.mapred.TextInputFormat
+                    output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                    properties:
+                      COLUMN_STATS_ACCURATE true
+                      bucket_count -1
+                      columns key,value
+                      columns.comments 'default','default'
+                      columns.types string:string
+#### A masked pattern was here ####
+                      name default.src
+                      numFiles 1
+                      numRows 500
+                      rawDataSize 5312
+                      serialization.ddl struct src { string key, string value}
+                      serialization.format 1
+                      serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                      totalSize 5812
+#### A masked pattern was here ####
+                    serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                    name: default.src
+                  name: default.src
+            Truncated Path -> Alias:
+              /src [src]
+
+  Stage: Stage-0
+    Move Operator
+      tables:
+          partition:
+            ds 2008-04-08
+            hr 11
+          replace: true
+#### A masked pattern was here ####
+          table:
+              input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
+              output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat
+              properties:
+                bucket_count -1
+                columns key,value
+                columns.comments 
+                columns.types string:string
+#### A masked pattern was here ####
+                name default.list_bucketing_static_part
+                partition_columns ds/hr
+                partition_columns.types string:string
+                serialization.ddl struct list_bucketing_static_part { string 
key, string value}
+                serialization.format 1
+                serialization.lib 
org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
+#### A masked pattern was here ####
+              serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
+              name: default.list_bucketing_static_part
+
+  Stage: Stage-2
+    Stats-Aggr Operator
+#### A masked pattern was here ####
+
+PREHOOK: query: insert overwrite table list_bucketing_static_part partition 
(ds = '2008-04-08', hr = '11')
+select key, value from src
+PREHOOK: type: QUERY
+PREHOOK: Input: default@src
+PREHOOK: Output: default@list_bucketing_static_part@ds=2008-04-08/hr=11
+POSTHOOK: query: insert overwrite table list_bucketing_static_part partition 
(ds = '2008-04-08', hr = '11')
+select key, value from src
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@src
+POSTHOOK: Output: default@list_bucketing_static_part@ds=2008-04-08/hr=11
+POSTHOOK: Lineage: list_bucketing_static_part 
PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, 
type:string, comment:default), ]
+POSTHOOK: Lineage: list_bucketing_static_part 
PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, 
type:string, comment:default), ]
+PREHOOK: query: -- check DML result
+show partitions list_bucketing_static_part
+PREHOOK: type: SHOWPARTITIONS
+PREHOOK: Input: default@list_bucketing_static_part
+POSTHOOK: query: -- check DML result
+show partitions list_bucketing_static_part
+POSTHOOK: type: SHOWPARTITIONS
+POSTHOOK: Input: default@list_bucketing_static_part
+ds=2008-04-08/hr=11
+PREHOOK: query: desc formatted list_bucketing_static_part partition 
(ds='2008-04-08', hr='11')
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@list_bucketing_static_part
+POSTHOOK: query: desc formatted list_bucketing_static_part partition 
(ds='2008-04-08', hr='11')
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@list_bucketing_static_part
+# col_name             data_type               comment             
+                
+key                    string                                      
+value                  string                                      
+                
+# Partition Information                 
+# col_name             data_type               comment             
+                
+ds                     string                                      
+hr                     string                                      
+                
+# Detailed Partition Information                
+Partition Value:       [2008-04-08, 11]         
+Database:              default                  
+Table:                 list_bucketing_static_part       
+#### A masked pattern was here ####
+Protect Mode:          None                     
+#### A masked pattern was here ####
+Partition Parameters:           
+       COLUMN_STATS_ACCURATE   true                
+       numFiles                4                   
+       numRows                 500                 
+       rawDataSize             4812                
+       totalSize               5520                
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe    
 
+InputFormat:           org.apache.hadoop.hive.ql.io.RCFileInputFormat   
+OutputFormat:          org.apache.hadoop.hive.ql.io.RCFileOutputFormat  
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Stored As SubDirectories:      Yes                      
+Skewed Columns:        [key]                    
+Skewed Values:         [[484], [51], [103]]     
+#### A masked pattern was here ####
+Skewed Value to Truncated Path:        
{[484]=/list_bucketing_static_part/ds=2008-04-08/hr=11/key=484, 
[103]=/list_bucketing_static_part/ds=2008-04-08/hr=11/key=103, 
[51]=/list_bucketing_static_part/ds=2008-04-08/hr=11/key=51}      
+Storage Desc Params:            
+       serialization.format    1                   

Added: 
hive/branches/spark/ql/src/test/results/clientpositive/spark/load_fs2.q.out
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/load_fs2.q.out?rev=1656573&view=auto
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientpositive/spark/load_fs2.q.out 
(added)
+++ hive/branches/spark/ql/src/test/results/clientpositive/spark/load_fs2.q.out 
Mon Feb  2 21:10:08 2015
@@ -0,0 +1,121 @@
+PREHOOK: query: -- HIVE-3300 [jira] LOAD DATA INPATH fails if a hdfs file with 
same name is added to table
+-- 'loader' table is used only for uploading kv1.txt to HDFS (!hdfs -put is 
not working on minMRDriver)
+
+create table result (key string, value string)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@result
+POSTHOOK: query: -- HIVE-3300 [jira] LOAD DATA INPATH fails if a hdfs file 
with same name is added to table
+-- 'loader' table is used only for uploading kv1.txt to HDFS (!hdfs -put is 
not working on minMRDriver)
+
+create table result (key string, value string)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@result
+PREHOOK: query: create table loader (key string, value string)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@loader
+POSTHOOK: query: create table loader (key string, value string)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@loader
+PREHOOK: query: load data local inpath '../../data/files/kv1.txt' into table 
loader
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@loader
+POSTHOOK: query: load data local inpath '../../data/files/kv1.txt' into table 
loader
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@loader
+PREHOOK: query: load data inpath 
'/build/ql/test/data/warehouse/loader/kv1.txt' into table result
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@result
+POSTHOOK: query: load data inpath 
'/build/ql/test/data/warehouse/loader/kv1.txt' into table result
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@result
+PREHOOK: query: show table extended like result
+PREHOOK: type: SHOW_TABLESTATUS
+POSTHOOK: query: show table extended like result
+POSTHOOK: type: SHOW_TABLESTATUS
+tableName:result
+#### A masked pattern was here ####
+inputformat:org.apache.hadoop.mapred.TextInputFormat
+outputformat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+columns:struct columns { string key, string value}
+partitioned:false
+partitionColumns:
+totalNumberFiles:1
+totalFileSize:5812
+maxFileSize:5812
+minFileSize:5812
+#### A masked pattern was here ####
+
+PREHOOK: query: load data local inpath '../../data/files/kv1.txt' into table 
loader
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@loader
+POSTHOOK: query: load data local inpath '../../data/files/kv1.txt' into table 
loader
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@loader
+PREHOOK: query: load data inpath 
'/build/ql/test/data/warehouse/loader/kv1.txt' into table result
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@result
+POSTHOOK: query: load data inpath 
'/build/ql/test/data/warehouse/loader/kv1.txt' into table result
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@result
+PREHOOK: query: show table extended like result
+PREHOOK: type: SHOW_TABLESTATUS
+POSTHOOK: query: show table extended like result
+POSTHOOK: type: SHOW_TABLESTATUS
+tableName:result
+#### A masked pattern was here ####
+inputformat:org.apache.hadoop.mapred.TextInputFormat
+outputformat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+columns:struct columns { string key, string value}
+partitioned:false
+partitionColumns:
+totalNumberFiles:2
+totalFileSize:11624
+maxFileSize:5812
+minFileSize:5812
+#### A masked pattern was here ####
+
+PREHOOK: query: load data local inpath '../../data/files/kv1.txt' into table 
loader
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@loader
+POSTHOOK: query: load data local inpath '../../data/files/kv1.txt' into table 
loader
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@loader
+PREHOOK: query: load data inpath 
'/build/ql/test/data/warehouse/loader/kv1.txt' into table result
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@result
+POSTHOOK: query: load data inpath 
'/build/ql/test/data/warehouse/loader/kv1.txt' into table result
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@result
+PREHOOK: query: show table extended like result
+PREHOOK: type: SHOW_TABLESTATUS
+POSTHOOK: query: show table extended like result
+POSTHOOK: type: SHOW_TABLESTATUS
+tableName:result
+#### A masked pattern was here ####
+inputformat:org.apache.hadoop.mapred.TextInputFormat
+outputformat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+columns:struct columns { string key, string value}
+partitioned:false
+partitionColumns:
+totalNumberFiles:3
+totalFileSize:17436
+maxFileSize:5812
+minFileSize:5812
+#### A masked pattern was here ####
+

Added: 
hive/branches/spark/ql/src/test/results/clientpositive/spark/load_hdfs_file_with_space_in_the_name.q.out
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/load_hdfs_file_with_space_in_the_name.q.out?rev=1656573&view=auto
==============================================================================
--- 
hive/branches/spark/ql/src/test/results/clientpositive/spark/load_hdfs_file_with_space_in_the_name.q.out
 (added)
+++ 
hive/branches/spark/ql/src/test/results/clientpositive/spark/load_hdfs_file_with_space_in_the_name.q.out
 Mon Feb  2 21:10:08 2015
@@ -0,0 +1,25 @@
+PREHOOK: query: CREATE TABLE load_file_with_space_in_the_name(name STRING, age 
INT)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@load_file_with_space_in_the_name
+POSTHOOK: query: CREATE TABLE load_file_with_space_in_the_name(name STRING, 
age INT)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@load_file_with_space_in_the_name
+#### A masked pattern was here ####
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@load_file_with_space_in_the_name
+#### A masked pattern was here ####
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@load_file_with_space_in_the_name
+#### A masked pattern was here ####
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@load_file_with_space_in_the_name
+#### A masked pattern was here ####
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@load_file_with_space_in_the_name
+#### A masked pattern was here ####

Added: 
hive/branches/spark/ql/src/test/results/clientpositive/spark/parallel_orderby.q.out
URL: 
http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/parallel_orderby.q.out?rev=1656573&view=auto
==============================================================================
--- 
hive/branches/spark/ql/src/test/results/clientpositive/spark/parallel_orderby.q.out
 (added)
+++ 
hive/branches/spark/ql/src/test/results/clientpositive/spark/parallel_orderby.q.out
 Mon Feb  2 21:10:08 2015
@@ -0,0 +1,305 @@
+PREHOOK: query: create table src5 (key string, value string)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@src5
+POSTHOOK: query: create table src5 (key string, value string)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@src5
+PREHOOK: query: load data local inpath '../../data/files/kv5.txt' into table 
src5
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@src5
+POSTHOOK: query: load data local inpath '../../data/files/kv5.txt' into table 
src5
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@src5
+PREHOOK: query: load data local inpath '../../data/files/kv5.txt' into table 
src5
+PREHOOK: type: LOAD
+#### A masked pattern was here ####
+PREHOOK: Output: default@src5
+POSTHOOK: query: load data local inpath '../../data/files/kv5.txt' into table 
src5
+POSTHOOK: type: LOAD
+#### A masked pattern was here ####
+POSTHOOK: Output: default@src5
+PREHOOK: query: explain
+create table total_ordered as select * from src5 order by key, value
+PREHOOK: type: CREATETABLE_AS_SELECT
+POSTHOOK: query: explain
+create table total_ordered as select * from src5 order by key, value
+POSTHOOK: type: CREATETABLE_AS_SELECT
+STAGE DEPENDENCIES:
+  Stage-1 is a root stage
+  Stage-0 depends on stages: Stage-1
+  Stage-3 depends on stages: Stage-0
+  Stage-2 depends on stages: Stage-3
+
+STAGE PLANS:
+  Stage: Stage-1
+    Spark
+      Edges:
+        Reducer 2 <- Map 1 (SORT, 1)
+#### A masked pattern was here ####
+      Vertices:
+        Map 1 
+            Map Operator Tree:
+                TableScan
+                  alias: src5
+                  Statistics: Num rows: 2 Data size: 560 Basic stats: COMPLETE 
Column stats: NONE
+                  Select Operator
+                    expressions: key (type: string), value (type: string)
+                    outputColumnNames: _col0, _col1
+                    Statistics: Num rows: 2 Data size: 560 Basic stats: 
COMPLETE Column stats: NONE
+                    Reduce Output Operator
+                      key expressions: _col0 (type: string), _col1 (type: 
string)
+                      sort order: ++
+                      Statistics: Num rows: 2 Data size: 560 Basic stats: 
COMPLETE Column stats: NONE
+        Reducer 2 
+            Reduce Operator Tree:
+              Select Operator
+                expressions: KEY.reducesinkkey0 (type: string), 
KEY.reducesinkkey1 (type: string)
+                outputColumnNames: _col0, _col1
+                Statistics: Num rows: 2 Data size: 560 Basic stats: COMPLETE 
Column stats: NONE
+                File Output Operator
+                  compressed: false
+                  Statistics: Num rows: 2 Data size: 560 Basic stats: COMPLETE 
Column stats: NONE
+                  table:
+                      input format: org.apache.hadoop.mapred.TextInputFormat
+                      output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
+                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+                      name: default.total_ordered
+
+  Stage: Stage-0
+    Move Operator
+      files:
+          hdfs directory: true
+#### A masked pattern was here ####
+
+  Stage: Stage-3
+      Create Table Operator:
+        Create Table
+          columns: key string, value string
+          input format: org.apache.hadoop.mapred.TextInputFormat
+          output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat
+          serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
+          name: default.total_ordered
+
+  Stage: Stage-2
+    Stats-Aggr Operator
+
+PREHOOK: query: create table total_ordered as select * from src5 order by key, 
value
+PREHOOK: type: CREATETABLE_AS_SELECT
+PREHOOK: Input: default@src5
+PREHOOK: Output: database:default
+PREHOOK: Output: default@total_ordered
+POSTHOOK: query: create table total_ordered as select * from src5 order by 
key, value
+POSTHOOK: type: CREATETABLE_AS_SELECT
+POSTHOOK: Input: default@src5
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@total_ordered
+PREHOOK: query: desc formatted total_ordered
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@total_ordered
+POSTHOOK: query: desc formatted total_ordered
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@total_ordered
+# col_name             data_type               comment             
+                
+key                    string                                      
+value                  string                                      
+                
+# Detailed Table Information            
+Database:              default                  
+#### A masked pattern was here ####
+Protect Mode:          None                     
+Retention:             0                        
+#### A masked pattern was here ####
+Table Type:            MANAGED_TABLE            
+Table Parameters:               
+       COLUMN_STATS_ACCURATE   true                
+       numFiles                1                   
+       numRows                 48                  
+       rawDataSize             512                 
+       totalSize               560                 
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
+PREHOOK: query: select * from total_ordered
+PREHOOK: type: QUERY
+PREHOOK: Input: default@total_ordered
+#### A masked pattern was here ####
+POSTHOOK: query: select * from total_ordered
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@total_ordered
+#### A masked pattern was here ####
+128    val_128
+128    val_128
+150    val_150
+150    val_150
+165    val_165
+165    val_165
+193    val_193
+193    val_193
+213    val_213
+213    val_213
+213    val_213
+213    val_213
+213    val_214
+213    val_214
+224    val_224
+224    val_224
+238    val_238
+238    val_238
+238    val_239
+238    val_239
+238    val_240
+238    val_240
+255    val_255
+255    val_255
+265    val_265
+265    val_265
+27     val_27
+27     val_27
+273    val_273
+273    val_273
+278    val_278
+278    val_278
+311    val_311
+311    val_311
+369    val_369
+369    val_369
+401    val_401
+401    val_401
+409    val_409
+409    val_409
+484    val_484
+484    val_484
+66     val_66
+66     val_66
+86     val_86
+86     val_86
+98     val_98
+98     val_98
+PREHOOK: query: -- rolling back to single task in case that the number of 
sample is not enough
+
+drop table total_ordered
+PREHOOK: type: DROPTABLE
+PREHOOK: Input: default@total_ordered
+PREHOOK: Output: default@total_ordered
+POSTHOOK: query: -- rolling back to single task in case that the number of 
sample is not enough
+
+drop table total_ordered
+POSTHOOK: type: DROPTABLE
+POSTHOOK: Input: default@total_ordered
+POSTHOOK: Output: default@total_ordered
+PREHOOK: query: create table total_ordered as select * from src5 order by key, 
value
+PREHOOK: type: CREATETABLE_AS_SELECT
+PREHOOK: Input: default@src5
+PREHOOK: Output: database:default
+PREHOOK: Output: default@total_ordered
+POSTHOOK: query: create table total_ordered as select * from src5 order by 
key, value
+POSTHOOK: type: CREATETABLE_AS_SELECT
+POSTHOOK: Input: default@src5
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@total_ordered
+PREHOOK: query: desc formatted total_ordered
+PREHOOK: type: DESCTABLE
+PREHOOK: Input: default@total_ordered
+POSTHOOK: query: desc formatted total_ordered
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: default@total_ordered
+# col_name             data_type               comment             
+                
+key                    string                                      
+value                  string                                      
+                
+# Detailed Table Information            
+Database:              default                  
+#### A masked pattern was here ####
+Protect Mode:          None                     
+Retention:             0                        
+#### A masked pattern was here ####
+Table Type:            MANAGED_TABLE            
+Table Parameters:               
+       COLUMN_STATS_ACCURATE   true                
+       numFiles                1                   
+       numRows                 48                  
+       rawDataSize             512                 
+       totalSize               560                 
+#### A masked pattern was here ####
+                
+# Storage Information           
+SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
+InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
+OutputFormat:          
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat       
+Compressed:            No                       
+Num Buckets:           -1                       
+Bucket Columns:        []                       
+Sort Columns:          []                       
+Storage Desc Params:            
+       serialization.format    1                   
+PREHOOK: query: select * from total_ordered
+PREHOOK: type: QUERY
+PREHOOK: Input: default@total_ordered
+#### A masked pattern was here ####
+POSTHOOK: query: select * from total_ordered
+POSTHOOK: type: QUERY
+POSTHOOK: Input: default@total_ordered
+#### A masked pattern was here ####
+128    val_128
+128    val_128
+150    val_150
+150    val_150
+165    val_165
+165    val_165
+193    val_193
+193    val_193
+213    val_213
+213    val_213
+213    val_213
+213    val_213
+213    val_214
+213    val_214
+224    val_224
+224    val_224
+238    val_238
+238    val_238
+238    val_239
+238    val_239
+238    val_240
+238    val_240
+255    val_255
+255    val_255
+265    val_265
+265    val_265
+27     val_27
+27     val_27
+273    val_273
+273    val_273
+278    val_278
+278    val_278
+311    val_311
+311    val_311
+369    val_369
+369    val_369
+401    val_401
+401    val_401
+409    val_409
+409    val_409
+484    val_484
+484    val_484
+66     val_66
+66     val_66
+86     val_86
+86     val_86
+98     val_98
+98     val_98


Reply via email to