http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/input_part1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/input_part1.q.out 
b/ql/src/test/results/beelinepositive/input_part1.q.out
deleted file mode 100644
index ed01e63..0000000
--- a/ql/src/test/results/beelinepositive/input_part1.q.out
+++ /dev/null
@@ -1,421 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/input_part1.q.raw". Enter "record" 
with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/input_part1.q
->>>  CREATE TABLE dest1(key INT, value STRING, hr STRING, ds STRING) STORED AS 
TEXTFILE;
-No rows affected 
->>>  
->>>  EXPLAIN EXTENDED 
-FROM srcpart 
-INSERT OVERWRITE TABLE dest1 SELECT srcpart.key, srcpart.value, srcpart.hr, 
srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr 
= '12';
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME srcpart))) (TOK_INSERT 
(TOK_DESTINATION (TOK_TAB (TOK_TABNAME dest1))) (TOK_SELECT (TOK_SELEXPR (. 
(TOK_TABLE_OR_COL srcpart) key)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL srcpart) 
value)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL srcpart) hr)) (TOK_SELEXPR (. 
(TOK_TABLE_OR_COL srcpart) ds))) (TOK_WHERE (and (and (< (. (TOK_TABLE_OR_COL 
srcpart) key) 100) (= (. (TOK_TABLE_OR_COL srcpart) ds) '2008-04-08')) (= (. 
(TOK_TABLE_OR_COL srcpart) hr) '12')))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-1 is a root stage'
-'  Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5'
-'  Stage-4'
-'  Stage-0 depends on stages: Stage-4, Stage-3, Stage-6'
-'  Stage-2 depends on stages: Stage-0'
-'  Stage-3'
-'  Stage-5'
-'  Stage-6 depends on stages: Stage-5'
-''
-'STAGE PLANS:'
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        srcpart '
-'          TableScan'
-'            alias: srcpart'
-'            GatherStats: false'
-'            Filter Operator'
-'              isSamplingPred: false'
-'              predicate:'
-'                  expr: (key < 100.0)'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                      expr: hr'
-'                      type: string'
-'                      expr: ds'
-'                      type: string'
-'                outputColumnNames: _col0, _col1, _col2, _col3'
-'                Select Operator'
-'                  expressions:'
-'                        expr: UDFToInteger(_col0)'
-'                        type: int'
-'                        expr: _col1'
-'                        type: string'
-'                        expr: _col2'
-'                        type: string'
-'                        expr: _col3'
-'                        type: string'
-'                  outputColumnNames: _col0, _col1, _col2, _col3'
-'                  File Output Operator'
-'                    compressed: false'
-'                    GlobalTableId: 1'
-'                    directory: pfile:!!{hive.exec.scratchdir}!!'
-'                    NumFilesPerFileSink: 1'
-'                    Stats Publishing Key Prefix: 
pfile:!!{hive.exec.scratchdir}!!'
-'                    table:'
-'                        input format: 
org.apache.hadoop.mapred.TextInputFormat'
-'                        output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                        properties:'
-'                          bucket_count -1'
-'                          columns key,value,hr,ds'
-'                          columns.types int:string:string:string'
-'                          file.inputformat 
org.apache.hadoop.mapred.TextInputFormat'
-'                          file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                          location 
!!{hive.metastore.warehouse.dir}!!/input_part1.db/dest1'
-'                          name input_part1.dest1'
-'                          serialization.ddl struct dest1 { i32 key, string 
value, string hr, string ds}'
-'                          serialization.format 1'
-'                          serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                          transient_lastDdlTime !!UNIXTIME!!'
-'                        serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                        name: input_part1.dest1'
-'                    TotalFiles: 1'
-'                    GatherStats: true'
-'                    MultiFileSpray: false'
-'      Needs Tagging: false'
-'      Path -> Alias:'
-'        
!!{hive.metastore.warehouse.dir}!!/input_part1.db/srcpart/ds=2008-04-08/hr=12 
[srcpart]'
-'      Path -> Partition:'
-'        
!!{hive.metastore.warehouse.dir}!!/input_part1.db/srcpart/ds=2008-04-08/hr=12 '
-'          Partition'
-'            base file name: hr=12'
-'            input format: org.apache.hadoop.mapred.TextInputFormat'
-'            output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'            partition values:'
-'              ds 2008-04-08'
-'              hr 12'
-'            properties:'
-'              bucket_count -1'
-'              columns key,value'
-'              columns.types string:string'
-'              file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'              file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              location 
!!{hive.metastore.warehouse.dir}!!/input_part1.db/srcpart/ds=2008-04-08/hr=12'
-'              name input_part1.srcpart'
-'              numFiles 1'
-'              numPartitions 4'
-'              numRows 0'
-'              partition_columns ds/hr'
-'              rawDataSize 0'
-'              serialization.ddl struct srcpart { string key, string value}'
-'              serialization.format 1'
-'              serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              totalSize 5812'
-'              transient_lastDdlTime !!UNIXTIME!!'
-'            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'          '
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              properties:'
-'                bucket_count -1'
-'                columns key,value'
-'                columns.types string:string'
-'                file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'                file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                location 
!!{hive.metastore.warehouse.dir}!!/input_part1.db/srcpart'
-'                name input_part1.srcpart'
-'                numFiles 4'
-'                numPartitions 4'
-'                numRows 0'
-'                partition_columns ds/hr'
-'                rawDataSize 0'
-'                serialization.ddl struct srcpart { string key, string value}'
-'                serialization.format 1'
-'                serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                totalSize 23248'
-'                transient_lastDdlTime !!UNIXTIME!!'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: input_part1.srcpart'
-'            name: input_part1.srcpart'
-''
-'  Stage: Stage-7'
-'    Conditional Operator'
-''
-'  Stage: Stage-4'
-'    Move Operator'
-'      files:'
-'          hdfs directory: true'
-'          source: pfile:!!{hive.exec.scratchdir}!!'
-'          destination: pfile:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-0'
-'    Move Operator'
-'      tables:'
-'          replace: true'
-'          source: pfile:!!{hive.exec.scratchdir}!!'
-'          table:'
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              properties:'
-'                bucket_count -1'
-'                columns key,value,hr,ds'
-'                columns.types int:string:string:string'
-'                file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'                file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                location 
!!{hive.metastore.warehouse.dir}!!/input_part1.db/dest1'
-'                name input_part1.dest1'
-'                serialization.ddl struct dest1 { i32 key, string value, 
string hr, string ds}'
-'                serialization.format 1'
-'                serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                transient_lastDdlTime !!UNIXTIME!!'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: input_part1.dest1'
-'          tmp directory: pfile:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-2'
-'    Stats-Aggr Operator'
-'      Stats Aggregation Key Prefix: pfile:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-3'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        pfile:!!{hive.exec.scratchdir}!! '
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              directory: pfile:!!{hive.exec.scratchdir}!!'
-'              NumFilesPerFileSink: 1'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                  properties:'
-'                    bucket_count -1'
-'                    columns key,value,hr,ds'
-'                    columns.types int:string:string:string'
-'                    file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'                    file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                    location 
!!{hive.metastore.warehouse.dir}!!/input_part1.db/dest1'
-'                    name input_part1.dest1'
-'                    serialization.ddl struct dest1 { i32 key, string value, 
string hr, string ds}'
-'                    serialization.format 1'
-'                    serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                    transient_lastDdlTime !!UNIXTIME!!'
-'                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                  name: input_part1.dest1'
-'              TotalFiles: 1'
-'              GatherStats: false'
-'              MultiFileSpray: false'
-'      Needs Tagging: false'
-'      Path -> Alias:'
-'        pfile:!!{hive.exec.scratchdir}!! [pfile:!!{hive.exec.scratchdir}!!]'
-'      Path -> Partition:'
-'        pfile:!!{hive.exec.scratchdir}!! '
-'          Partition'
-'            base file name: -ext-10002'
-'            input format: org.apache.hadoop.mapred.TextInputFormat'
-'            output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'            properties:'
-'              bucket_count -1'
-'              columns key,value,hr,ds'
-'              columns.types int:string:string:string'
-'              file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'              file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              location 
!!{hive.metastore.warehouse.dir}!!/input_part1.db/dest1'
-'              name input_part1.dest1'
-'              serialization.ddl struct dest1 { i32 key, string value, string 
hr, string ds}'
-'              serialization.format 1'
-'              serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              transient_lastDdlTime !!UNIXTIME!!'
-'            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'          '
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              properties:'
-'                bucket_count -1'
-'                columns key,value,hr,ds'
-'                columns.types int:string:string:string'
-'                file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'                file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                location 
!!{hive.metastore.warehouse.dir}!!/input_part1.db/dest1'
-'                name input_part1.dest1'
-'                serialization.ddl struct dest1 { i32 key, string value, 
string hr, string ds}'
-'                serialization.format 1'
-'                serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                transient_lastDdlTime !!UNIXTIME!!'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: input_part1.dest1'
-'            name: input_part1.dest1'
-''
-'  Stage: Stage-5'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        pfile:!!{hive.exec.scratchdir}!! '
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              directory: pfile:!!{hive.exec.scratchdir}!!'
-'              NumFilesPerFileSink: 1'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                  properties:'
-'                    bucket_count -1'
-'                    columns key,value,hr,ds'
-'                    columns.types int:string:string:string'
-'                    file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'                    file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                    location 
!!{hive.metastore.warehouse.dir}!!/input_part1.db/dest1'
-'                    name input_part1.dest1'
-'                    serialization.ddl struct dest1 { i32 key, string value, 
string hr, string ds}'
-'                    serialization.format 1'
-'                    serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                    transient_lastDdlTime !!UNIXTIME!!'
-'                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                  name: input_part1.dest1'
-'              TotalFiles: 1'
-'              GatherStats: false'
-'              MultiFileSpray: false'
-'      Needs Tagging: false'
-'      Path -> Alias:'
-'        pfile:!!{hive.exec.scratchdir}!! [pfile:!!{hive.exec.scratchdir}!!]'
-'      Path -> Partition:'
-'        pfile:!!{hive.exec.scratchdir}!! '
-'          Partition'
-'            base file name: -ext-10002'
-'            input format: org.apache.hadoop.mapred.TextInputFormat'
-'            output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'            properties:'
-'              bucket_count -1'
-'              columns key,value,hr,ds'
-'              columns.types int:string:string:string'
-'              file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'              file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              location 
!!{hive.metastore.warehouse.dir}!!/input_part1.db/dest1'
-'              name input_part1.dest1'
-'              serialization.ddl struct dest1 { i32 key, string value, string 
hr, string ds}'
-'              serialization.format 1'
-'              serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              transient_lastDdlTime !!UNIXTIME!!'
-'            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'          '
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              properties:'
-'                bucket_count -1'
-'                columns key,value,hr,ds'
-'                columns.types int:string:string:string'
-'                file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'                file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                location 
!!{hive.metastore.warehouse.dir}!!/input_part1.db/dest1'
-'                name input_part1.dest1'
-'                serialization.ddl struct dest1 { i32 key, string value, 
string hr, string ds}'
-'                serialization.format 1'
-'                serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                transient_lastDdlTime !!UNIXTIME!!'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: input_part1.dest1'
-'            name: input_part1.dest1'
-''
-'  Stage: Stage-6'
-'    Move Operator'
-'      files:'
-'          hdfs directory: true'
-'          source: pfile:!!{hive.exec.scratchdir}!!'
-'          destination: pfile:!!{hive.exec.scratchdir}!!'
-''
-''
-316 rows selected 
->>>  
->>>  FROM srcpart 
-INSERT OVERWRITE TABLE dest1 SELECT srcpart.key, srcpart.value, srcpart.hr, 
srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr 
= '12';
-'_col0','_col1','_col2','_col3'
-No rows selected 
->>>  
->>>  SELECT dest1.* FROM dest1;
-'key','value','hr','ds'
-'86','val_86','12','2008-04-08'
-'27','val_27','12','2008-04-08'
-'98','val_98','12','2008-04-08'
-'66','val_66','12','2008-04-08'
-'37','val_37','12','2008-04-08'
-'15','val_15','12','2008-04-08'
-'82','val_82','12','2008-04-08'
-'17','val_17','12','2008-04-08'
-'0','val_0','12','2008-04-08'
-'57','val_57','12','2008-04-08'
-'20','val_20','12','2008-04-08'
-'92','val_92','12','2008-04-08'
-'47','val_47','12','2008-04-08'
-'72','val_72','12','2008-04-08'
-'4','val_4','12','2008-04-08'
-'35','val_35','12','2008-04-08'
-'54','val_54','12','2008-04-08'
-'51','val_51','12','2008-04-08'
-'65','val_65','12','2008-04-08'
-'83','val_83','12','2008-04-08'
-'12','val_12','12','2008-04-08'
-'67','val_67','12','2008-04-08'
-'84','val_84','12','2008-04-08'
-'58','val_58','12','2008-04-08'
-'8','val_8','12','2008-04-08'
-'24','val_24','12','2008-04-08'
-'42','val_42','12','2008-04-08'
-'0','val_0','12','2008-04-08'
-'96','val_96','12','2008-04-08'
-'26','val_26','12','2008-04-08'
-'51','val_51','12','2008-04-08'
-'43','val_43','12','2008-04-08'
-'95','val_95','12','2008-04-08'
-'98','val_98','12','2008-04-08'
-'85','val_85','12','2008-04-08'
-'77','val_77','12','2008-04-08'
-'0','val_0','12','2008-04-08'
-'87','val_87','12','2008-04-08'
-'15','val_15','12','2008-04-08'
-'72','val_72','12','2008-04-08'
-'90','val_90','12','2008-04-08'
-'19','val_19','12','2008-04-08'
-'10','val_10','12','2008-04-08'
-'5','val_5','12','2008-04-08'
-'58','val_58','12','2008-04-08'
-'35','val_35','12','2008-04-08'
-'95','val_95','12','2008-04-08'
-'11','val_11','12','2008-04-08'
-'34','val_34','12','2008-04-08'
-'42','val_42','12','2008-04-08'
-'78','val_78','12','2008-04-08'
-'76','val_76','12','2008-04-08'
-'41','val_41','12','2008-04-08'
-'30','val_30','12','2008-04-08'
-'64','val_64','12','2008-04-08'
-'76','val_76','12','2008-04-08'
-'74','val_74','12','2008-04-08'
-'69','val_69','12','2008-04-08'
-'33','val_33','12','2008-04-08'
-'70','val_70','12','2008-04-08'
-'5','val_5','12','2008-04-08'
-'2','val_2','12','2008-04-08'
-'35','val_35','12','2008-04-08'
-'80','val_80','12','2008-04-08'
-'44','val_44','12','2008-04-08'
-'53','val_53','12','2008-04-08'
-'90','val_90','12','2008-04-08'
-'12','val_12','12','2008-04-08'
-'5','val_5','12','2008-04-08'
-'70','val_70','12','2008-04-08'
-'24','val_24','12','2008-04-08'
-'70','val_70','12','2008-04-08'
-'83','val_83','12','2008-04-08'
-'26','val_26','12','2008-04-08'
-'67','val_67','12','2008-04-08'
-'18','val_18','12','2008-04-08'
-'9','val_9','12','2008-04-08'
-'18','val_18','12','2008-04-08'
-'97','val_97','12','2008-04-08'
-'84','val_84','12','2008-04-08'
-'28','val_28','12','2008-04-08'
-'37','val_37','12','2008-04-08'
-'90','val_90','12','2008-04-08'
-'97','val_97','12','2008-04-08'
-84 rows selected 
->>>  
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/input_part10.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/input_part10.q.out 
b/ql/src/test/results/beelinepositive/input_part10.q.out
deleted file mode 100644
index e5fa5d2..0000000
--- a/ql/src/test/results/beelinepositive/input_part10.q.out
+++ /dev/null
@@ -1,99 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/input_part10.q.raw". Enter 
"record" with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/input_part10.q
->>>  CREATE TABLE part_special ( 
-a STRING, 
-b STRING 
-) PARTITIONED BY ( 
-ds STRING, 
-ts STRING 
-);
-No rows affected 
->>>  
->>>  EXPLAIN 
-INSERT OVERWRITE TABLE part_special PARTITION(ds='2008 04 08', ts = 
'10:11:12=455') 
-SELECT 1, 2 FROM src LIMIT 1;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT 
(TOK_DESTINATION (TOK_TAB (TOK_TABNAME part_special) (TOK_PARTSPEC (TOK_PARTVAL 
ds '2008 04 08') (TOK_PARTVAL ts '10:11:12=455')))) (TOK_SELECT (TOK_SELEXPR 1) 
(TOK_SELEXPR 2)) (TOK_LIMIT 1)))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-1 is a root stage'
-'  Stage-0 depends on stages: Stage-1'
-'  Stage-2 depends on stages: Stage-0'
-''
-'STAGE PLANS:'
-'  Stage: Stage-1'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        src '
-'          TableScan'
-'            alias: src'
-'            Select Operator'
-'              expressions:'
-'                    expr: 1'
-'                    type: int'
-'                    expr: 2'
-'                    type: int'
-'              outputColumnNames: _col0, _col1'
-'              Limit'
-'                Reduce Output Operator'
-'                  sort order: '
-'                  tag: -1'
-'                  value expressions:'
-'                        expr: _col0'
-'                        type: int'
-'                        expr: _col1'
-'                        type: int'
-'      Reduce Operator Tree:'
-'        Extract'
-'          Limit'
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 1'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                  name: input_part10.part_special'
-''
-'  Stage: Stage-0'
-'    Move Operator'
-'      tables:'
-'          partition:'
-'            ds 2008 04 08'
-'            ts 10:11:12=455'
-'          replace: true'
-'          table:'
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: input_part10.part_special'
-''
-'  Stage: Stage-2'
-'    Stats-Aggr Operator'
-''
-''
-60 rows selected 
->>>  
->>>  INSERT OVERWRITE TABLE part_special PARTITION(ds='2008 04 08', ts = 
'10:11:12=455') 
-SELECT 1, 2 FROM src LIMIT 1;
-'_c0','_c1'
-No rows selected 
->>>  
->>>  DESCRIBE EXTENDED part_special PARTITION(ds='2008 04 08', ts = 
'10:11:12=455');
-'col_name','data_type','comment'
-'a','string',''
-'b','string',''
-'ds','string',''
-'ts','string',''
-'','',''
-'Detailed Partition Information','Partition(values:[2008 04 08, 10:11:12=455], 
dbName:input_part10, tableName:part_special, createTime:!!UNIXTIME!!, 
lastAccessTime:0, sd:StorageDescriptor(cols:[FieldSchema(name:a, type:string, 
comment:null), FieldSchema(name:b, type:string, comment:null), 
FieldSchema(name:ds, type:string, comment:null), FieldSchema(name:ts, 
type:string, comment:null)], 
location:!!{hive.metastore.warehouse.dir}!!/input_part10.db/part_special/ds=2008
 04 08/ts=10%3A11%3A12%3D455, 
inputFormat:org.apache.hadoop.mapred.TextInputFormat, 
outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, 
compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, 
serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, 
parameters:{serialization.format=1}), bucketCols:[], sortCols:[], 
parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], 
skewedColValueLocationMaps:{})), parameters:{numFiles=1, 
transient_lastDdlTime=!!UNIXTIME!!, tot
 alSize=4, numRows=1, rawDataSize=3})',''
-6 rows selected 
->>>  
->>>  SELECT * FROM part_special WHERE ds='2008 04 08' AND ts = '10:11:12=455';
-'a','b','ds','ts'
-'1','2','2008 04 08','10:11:12=455'
-1 row selected 
->>>  
->>>  
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/input_part2.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/input_part2.q.out 
b/ql/src/test/results/beelinepositive/input_part2.q.out
deleted file mode 100644
index 167d719..0000000
--- a/ql/src/test/results/beelinepositive/input_part2.q.out
+++ /dev/null
@@ -1,810 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/input_part2.q.raw". Enter "record" 
with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/input_part2.q
->>>  CREATE TABLE dest1(key INT, value STRING, hr STRING, ds STRING) STORED AS 
TEXTFILE;
-No rows affected 
->>>  CREATE TABLE dest2(key INT, value STRING, hr STRING, ds STRING) STORED AS 
TEXTFILE;
-No rows affected 
->>>  
->>>  EXPLAIN EXTENDED 
-FROM srcpart 
-INSERT OVERWRITE TABLE dest1 SELECT srcpart.key, srcpart.value, srcpart.hr, 
srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr 
= '12' 
-INSERT OVERWRITE TABLE dest2 SELECT srcpart.key, srcpart.value, srcpart.hr, 
srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-09' and srcpart.hr 
= '12';
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME srcpart))) (TOK_INSERT 
(TOK_DESTINATION (TOK_TAB (TOK_TABNAME dest1))) (TOK_SELECT (TOK_SELEXPR (. 
(TOK_TABLE_OR_COL srcpart) key)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL srcpart) 
value)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL srcpart) hr)) (TOK_SELEXPR (. 
(TOK_TABLE_OR_COL srcpart) ds))) (TOK_WHERE (and (and (< (. (TOK_TABLE_OR_COL 
srcpart) key) 100) (= (. (TOK_TABLE_OR_COL srcpart) ds) '2008-04-08')) (= (. 
(TOK_TABLE_OR_COL srcpart) hr) '12')))) (TOK_INSERT (TOK_DESTINATION (TOK_TAB 
(TOK_TABNAME dest2))) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL srcpart) 
key)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL srcpart) value)) (TOK_SELEXPR (. 
(TOK_TABLE_OR_COL srcpart) hr)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL srcpart) 
ds))) (TOK_WHERE (and (and (< (. (TOK_TABLE_OR_COL srcpart) key) 100) (= (. 
(TOK_TABLE_OR_COL srcpart) ds) '2008-04-09')) (= (. (TOK_TABLE_OR_COL srcpart) 
hr) '12')))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-2 is a root stage'
-'  Stage-8 depends on stages: Stage-2 , consists of Stage-5, Stage-4, Stage-6'
-'  Stage-5'
-'  Stage-0 depends on stages: Stage-5, Stage-4, Stage-7'
-'  Stage-3 depends on stages: Stage-0'
-'  Stage-4'
-'  Stage-6'
-'  Stage-7 depends on stages: Stage-6'
-'  Stage-14 depends on stages: Stage-2 , consists of Stage-11, Stage-10, 
Stage-12'
-'  Stage-11'
-'  Stage-1 depends on stages: Stage-11, Stage-10, Stage-13'
-'  Stage-9 depends on stages: Stage-1'
-'  Stage-10'
-'  Stage-12'
-'  Stage-13 depends on stages: Stage-12'
-''
-'STAGE PLANS:'
-'  Stage: Stage-2'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        srcpart '
-'          TableScan'
-'            alias: srcpart'
-'            GatherStats: false'
-'            Filter Operator'
-'              isSamplingPred: false'
-'              predicate:'
-'                  expr: ((key < 100.0) and (ds = '2008-04-08'))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                      expr: hr'
-'                      type: string'
-'                      expr: ds'
-'                      type: string'
-'                outputColumnNames: _col0, _col1, _col2, _col3'
-'                Select Operator'
-'                  expressions:'
-'                        expr: UDFToInteger(_col0)'
-'                        type: int'
-'                        expr: _col1'
-'                        type: string'
-'                        expr: _col2'
-'                        type: string'
-'                        expr: _col3'
-'                        type: string'
-'                  outputColumnNames: _col0, _col1, _col2, _col3'
-'                  File Output Operator'
-'                    compressed: false'
-'                    GlobalTableId: 1'
-'                    directory: pfile:!!{hive.exec.scratchdir}!!'
-'                    NumFilesPerFileSink: 1'
-'                    Stats Publishing Key Prefix: 
pfile:!!{hive.exec.scratchdir}!!'
-'                    table:'
-'                        input format: 
org.apache.hadoop.mapred.TextInputFormat'
-'                        output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                        properties:'
-'                          bucket_count -1'
-'                          columns key,value,hr,ds'
-'                          columns.types int:string:string:string'
-'                          file.inputformat 
org.apache.hadoop.mapred.TextInputFormat'
-'                          file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                          location 
!!{hive.metastore.warehouse.dir}!!/input_part2.db/dest1'
-'                          name input_part2.dest1'
-'                          serialization.ddl struct dest1 { i32 key, string 
value, string hr, string ds}'
-'                          serialization.format 1'
-'                          serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                          transient_lastDdlTime !!UNIXTIME!!'
-'                        serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                        name: input_part2.dest1'
-'                    TotalFiles: 1'
-'                    GatherStats: true'
-'                    MultiFileSpray: false'
-'            Filter Operator'
-'              isSamplingPred: false'
-'              predicate:'
-'                  expr: ((key < 100.0) and (ds = '2008-04-09'))'
-'                  type: boolean'
-'              Select Operator'
-'                expressions:'
-'                      expr: key'
-'                      type: string'
-'                      expr: value'
-'                      type: string'
-'                      expr: hr'
-'                      type: string'
-'                      expr: ds'
-'                      type: string'
-'                outputColumnNames: _col0, _col1, _col2, _col3'
-'                Select Operator'
-'                  expressions:'
-'                        expr: UDFToInteger(_col0)'
-'                        type: int'
-'                        expr: _col1'
-'                        type: string'
-'                        expr: _col2'
-'                        type: string'
-'                        expr: _col3'
-'                        type: string'
-'                  outputColumnNames: _col0, _col1, _col2, _col3'
-'                  File Output Operator'
-'                    compressed: false'
-'                    GlobalTableId: 2'
-'                    directory: pfile:!!{hive.exec.scratchdir}!!'
-'                    NumFilesPerFileSink: 1'
-'                    Stats Publishing Key Prefix: 
pfile:!!{hive.exec.scratchdir}!!'
-'                    table:'
-'                        input format: 
org.apache.hadoop.mapred.TextInputFormat'
-'                        output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                        properties:'
-'                          bucket_count -1'
-'                          columns key,value,hr,ds'
-'                          columns.types int:string:string:string'
-'                          file.inputformat 
org.apache.hadoop.mapred.TextInputFormat'
-'                          file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                          location 
!!{hive.metastore.warehouse.dir}!!/input_part2.db/dest2'
-'                          name input_part2.dest2'
-'                          serialization.ddl struct dest2 { i32 key, string 
value, string hr, string ds}'
-'                          serialization.format 1'
-'                          serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                          transient_lastDdlTime !!UNIXTIME!!'
-'                        serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                        name: input_part2.dest2'
-'                    TotalFiles: 1'
-'                    GatherStats: true'
-'                    MultiFileSpray: false'
-'      Needs Tagging: false'
-'      Path -> Alias:'
-'        
!!{hive.metastore.warehouse.dir}!!/input_part2.db/srcpart/ds=2008-04-08/hr=12 
[srcpart]'
-'        
!!{hive.metastore.warehouse.dir}!!/input_part2.db/srcpart/ds=2008-04-09/hr=12 
[srcpart]'
-'      Path -> Partition:'
-'        
!!{hive.metastore.warehouse.dir}!!/input_part2.db/srcpart/ds=2008-04-08/hr=12 '
-'          Partition'
-'            base file name: hr=12'
-'            input format: org.apache.hadoop.mapred.TextInputFormat'
-'            output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'            partition values:'
-'              ds 2008-04-08'
-'              hr 12'
-'            properties:'
-'              bucket_count -1'
-'              columns key,value'
-'              columns.types string:string'
-'              file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'              file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              location 
!!{hive.metastore.warehouse.dir}!!/input_part2.db/srcpart/ds=2008-04-08/hr=12'
-'              name input_part2.srcpart'
-'              numFiles 1'
-'              numPartitions 4'
-'              numRows 0'
-'              partition_columns ds/hr'
-'              rawDataSize 0'
-'              serialization.ddl struct srcpart { string key, string value}'
-'              serialization.format 1'
-'              serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              totalSize 5812'
-'              transient_lastDdlTime !!UNIXTIME!!'
-'            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'          '
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              properties:'
-'                bucket_count -1'
-'                columns key,value'
-'                columns.types string:string'
-'                file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'                file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                location 
!!{hive.metastore.warehouse.dir}!!/input_part2.db/srcpart'
-'                name input_part2.srcpart'
-'                numFiles 4'
-'                numPartitions 4'
-'                numRows 0'
-'                partition_columns ds/hr'
-'                rawDataSize 0'
-'                serialization.ddl struct srcpart { string key, string value}'
-'                serialization.format 1'
-'                serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                totalSize 23248'
-'                transient_lastDdlTime !!UNIXTIME!!'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: input_part2.srcpart'
-'            name: input_part2.srcpart'
-'        
!!{hive.metastore.warehouse.dir}!!/input_part2.db/srcpart/ds=2008-04-09/hr=12 '
-'          Partition'
-'            base file name: hr=12'
-'            input format: org.apache.hadoop.mapred.TextInputFormat'
-'            output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'            partition values:'
-'              ds 2008-04-09'
-'              hr 12'
-'            properties:'
-'              bucket_count -1'
-'              columns key,value'
-'              columns.types string:string'
-'              file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'              file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              location 
!!{hive.metastore.warehouse.dir}!!/input_part2.db/srcpart/ds=2008-04-09/hr=12'
-'              name input_part2.srcpart'
-'              numFiles 1'
-'              numPartitions 4'
-'              numRows 0'
-'              partition_columns ds/hr'
-'              rawDataSize 0'
-'              serialization.ddl struct srcpart { string key, string value}'
-'              serialization.format 1'
-'              serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              totalSize 5812'
-'              transient_lastDdlTime !!UNIXTIME!!'
-'            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'          '
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              properties:'
-'                bucket_count -1'
-'                columns key,value'
-'                columns.types string:string'
-'                file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'                file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                location 
!!{hive.metastore.warehouse.dir}!!/input_part2.db/srcpart'
-'                name input_part2.srcpart'
-'                numFiles 4'
-'                numPartitions 4'
-'                numRows 0'
-'                partition_columns ds/hr'
-'                rawDataSize 0'
-'                serialization.ddl struct srcpart { string key, string value}'
-'                serialization.format 1'
-'                serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                totalSize 23248'
-'                transient_lastDdlTime !!UNIXTIME!!'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: input_part2.srcpart'
-'            name: input_part2.srcpart'
-''
-'  Stage: Stage-8'
-'    Conditional Operator'
-''
-'  Stage: Stage-5'
-'    Move Operator'
-'      files:'
-'          hdfs directory: true'
-'          source: pfile:!!{hive.exec.scratchdir}!!'
-'          destination: pfile:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-0'
-'    Move Operator'
-'      tables:'
-'          replace: true'
-'          source: pfile:!!{hive.exec.scratchdir}!!'
-'          table:'
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              properties:'
-'                bucket_count -1'
-'                columns key,value,hr,ds'
-'                columns.types int:string:string:string'
-'                file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'                file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                location 
!!{hive.metastore.warehouse.dir}!!/input_part2.db/dest1'
-'                name input_part2.dest1'
-'                serialization.ddl struct dest1 { i32 key, string value, 
string hr, string ds}'
-'                serialization.format 1'
-'                serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                transient_lastDdlTime !!UNIXTIME!!'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: input_part2.dest1'
-'          tmp directory: pfile:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-3'
-'    Stats-Aggr Operator'
-'      Stats Aggregation Key Prefix: pfile:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-4'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        pfile:!!{hive.exec.scratchdir}!! '
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              directory: pfile:!!{hive.exec.scratchdir}!!'
-'              NumFilesPerFileSink: 1'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                  properties:'
-'                    bucket_count -1'
-'                    columns key,value,hr,ds'
-'                    columns.types int:string:string:string'
-'                    file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'                    file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                    location 
!!{hive.metastore.warehouse.dir}!!/input_part2.db/dest1'
-'                    name input_part2.dest1'
-'                    serialization.ddl struct dest1 { i32 key, string value, 
string hr, string ds}'
-'                    serialization.format 1'
-'                    serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                    transient_lastDdlTime !!UNIXTIME!!'
-'                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                  name: input_part2.dest1'
-'              TotalFiles: 1'
-'              GatherStats: false'
-'              MultiFileSpray: false'
-'      Needs Tagging: false'
-'      Path -> Alias:'
-'        pfile:!!{hive.exec.scratchdir}!! [pfile:!!{hive.exec.scratchdir}!!]'
-'      Path -> Partition:'
-'        pfile:!!{hive.exec.scratchdir}!! '
-'          Partition'
-'            base file name: -ext-10004'
-'            input format: org.apache.hadoop.mapred.TextInputFormat'
-'            output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'            properties:'
-'              bucket_count -1'
-'              columns key,value,hr,ds'
-'              columns.types int:string:string:string'
-'              file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'              file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              location 
!!{hive.metastore.warehouse.dir}!!/input_part2.db/dest1'
-'              name input_part2.dest1'
-'              serialization.ddl struct dest1 { i32 key, string value, string 
hr, string ds}'
-'              serialization.format 1'
-'              serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              transient_lastDdlTime !!UNIXTIME!!'
-'            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'          '
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              properties:'
-'                bucket_count -1'
-'                columns key,value,hr,ds'
-'                columns.types int:string:string:string'
-'                file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'                file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                location 
!!{hive.metastore.warehouse.dir}!!/input_part2.db/dest1'
-'                name input_part2.dest1'
-'                serialization.ddl struct dest1 { i32 key, string value, 
string hr, string ds}'
-'                serialization.format 1'
-'                serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                transient_lastDdlTime !!UNIXTIME!!'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: input_part2.dest1'
-'            name: input_part2.dest1'
-''
-'  Stage: Stage-6'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        pfile:!!{hive.exec.scratchdir}!! '
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              directory: pfile:!!{hive.exec.scratchdir}!!'
-'              NumFilesPerFileSink: 1'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                  properties:'
-'                    bucket_count -1'
-'                    columns key,value,hr,ds'
-'                    columns.types int:string:string:string'
-'                    file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'                    file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                    location 
!!{hive.metastore.warehouse.dir}!!/input_part2.db/dest1'
-'                    name input_part2.dest1'
-'                    serialization.ddl struct dest1 { i32 key, string value, 
string hr, string ds}'
-'                    serialization.format 1'
-'                    serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                    transient_lastDdlTime !!UNIXTIME!!'
-'                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                  name: input_part2.dest1'
-'              TotalFiles: 1'
-'              GatherStats: false'
-'              MultiFileSpray: false'
-'      Needs Tagging: false'
-'      Path -> Alias:'
-'        pfile:!!{hive.exec.scratchdir}!! [pfile:!!{hive.exec.scratchdir}!!]'
-'      Path -> Partition:'
-'        pfile:!!{hive.exec.scratchdir}!! '
-'          Partition'
-'            base file name: -ext-10004'
-'            input format: org.apache.hadoop.mapred.TextInputFormat'
-'            output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'            properties:'
-'              bucket_count -1'
-'              columns key,value,hr,ds'
-'              columns.types int:string:string:string'
-'              file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'              file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              location 
!!{hive.metastore.warehouse.dir}!!/input_part2.db/dest1'
-'              name input_part2.dest1'
-'              serialization.ddl struct dest1 { i32 key, string value, string 
hr, string ds}'
-'              serialization.format 1'
-'              serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              transient_lastDdlTime !!UNIXTIME!!'
-'            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'          '
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              properties:'
-'                bucket_count -1'
-'                columns key,value,hr,ds'
-'                columns.types int:string:string:string'
-'                file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'                file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                location 
!!{hive.metastore.warehouse.dir}!!/input_part2.db/dest1'
-'                name input_part2.dest1'
-'                serialization.ddl struct dest1 { i32 key, string value, 
string hr, string ds}'
-'                serialization.format 1'
-'                serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                transient_lastDdlTime !!UNIXTIME!!'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: input_part2.dest1'
-'            name: input_part2.dest1'
-''
-'  Stage: Stage-7'
-'    Move Operator'
-'      files:'
-'          hdfs directory: true'
-'          source: pfile:!!{hive.exec.scratchdir}!!'
-'          destination: pfile:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-14'
-'    Conditional Operator'
-''
-'  Stage: Stage-11'
-'    Move Operator'
-'      files:'
-'          hdfs directory: true'
-'          source: pfile:!!{hive.exec.scratchdir}!!'
-'          destination: pfile:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-1'
-'    Move Operator'
-'      tables:'
-'          replace: true'
-'          source: pfile:!!{hive.exec.scratchdir}!!'
-'          table:'
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              properties:'
-'                bucket_count -1'
-'                columns key,value,hr,ds'
-'                columns.types int:string:string:string'
-'                file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'                file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                location 
!!{hive.metastore.warehouse.dir}!!/input_part2.db/dest2'
-'                name input_part2.dest2'
-'                serialization.ddl struct dest2 { i32 key, string value, 
string hr, string ds}'
-'                serialization.format 1'
-'                serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                transient_lastDdlTime !!UNIXTIME!!'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: input_part2.dest2'
-'          tmp directory: pfile:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-9'
-'    Stats-Aggr Operator'
-'      Stats Aggregation Key Prefix: pfile:!!{hive.exec.scratchdir}!!'
-''
-'  Stage: Stage-10'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        pfile:!!{hive.exec.scratchdir}!! '
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              directory: pfile:!!{hive.exec.scratchdir}!!'
-'              NumFilesPerFileSink: 1'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                  properties:'
-'                    bucket_count -1'
-'                    columns key,value,hr,ds'
-'                    columns.types int:string:string:string'
-'                    file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'                    file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                    location 
!!{hive.metastore.warehouse.dir}!!/input_part2.db/dest2'
-'                    name input_part2.dest2'
-'                    serialization.ddl struct dest2 { i32 key, string value, 
string hr, string ds}'
-'                    serialization.format 1'
-'                    serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                    transient_lastDdlTime !!UNIXTIME!!'
-'                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                  name: input_part2.dest2'
-'              TotalFiles: 1'
-'              GatherStats: false'
-'              MultiFileSpray: false'
-'      Needs Tagging: false'
-'      Path -> Alias:'
-'        pfile:!!{hive.exec.scratchdir}!! [pfile:!!{hive.exec.scratchdir}!!]'
-'      Path -> Partition:'
-'        pfile:!!{hive.exec.scratchdir}!! '
-'          Partition'
-'            base file name: -ext-10005'
-'            input format: org.apache.hadoop.mapred.TextInputFormat'
-'            output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'            properties:'
-'              bucket_count -1'
-'              columns key,value,hr,ds'
-'              columns.types int:string:string:string'
-'              file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'              file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              location 
!!{hive.metastore.warehouse.dir}!!/input_part2.db/dest2'
-'              name input_part2.dest2'
-'              serialization.ddl struct dest2 { i32 key, string value, string 
hr, string ds}'
-'              serialization.format 1'
-'              serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              transient_lastDdlTime !!UNIXTIME!!'
-'            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'          '
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              properties:'
-'                bucket_count -1'
-'                columns key,value,hr,ds'
-'                columns.types int:string:string:string'
-'                file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'                file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                location 
!!{hive.metastore.warehouse.dir}!!/input_part2.db/dest2'
-'                name input_part2.dest2'
-'                serialization.ddl struct dest2 { i32 key, string value, 
string hr, string ds}'
-'                serialization.format 1'
-'                serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                transient_lastDdlTime !!UNIXTIME!!'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: input_part2.dest2'
-'            name: input_part2.dest2'
-''
-'  Stage: Stage-12'
-'    Map Reduce'
-'      Alias -> Map Operator Tree:'
-'        pfile:!!{hive.exec.scratchdir}!! '
-'            File Output Operator'
-'              compressed: false'
-'              GlobalTableId: 0'
-'              directory: pfile:!!{hive.exec.scratchdir}!!'
-'              NumFilesPerFileSink: 1'
-'              table:'
-'                  input format: org.apache.hadoop.mapred.TextInputFormat'
-'                  output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                  properties:'
-'                    bucket_count -1'
-'                    columns key,value,hr,ds'
-'                    columns.types int:string:string:string'
-'                    file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'                    file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                    location 
!!{hive.metastore.warehouse.dir}!!/input_part2.db/dest2'
-'                    name input_part2.dest2'
-'                    serialization.ddl struct dest2 { i32 key, string value, 
string hr, string ds}'
-'                    serialization.format 1'
-'                    serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                    transient_lastDdlTime !!UNIXTIME!!'
-'                  serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                  name: input_part2.dest2'
-'              TotalFiles: 1'
-'              GatherStats: false'
-'              MultiFileSpray: false'
-'      Needs Tagging: false'
-'      Path -> Alias:'
-'        pfile:!!{hive.exec.scratchdir}!! [pfile:!!{hive.exec.scratchdir}!!]'
-'      Path -> Partition:'
-'        pfile:!!{hive.exec.scratchdir}!! '
-'          Partition'
-'            base file name: -ext-10005'
-'            input format: org.apache.hadoop.mapred.TextInputFormat'
-'            output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'            properties:'
-'              bucket_count -1'
-'              columns key,value,hr,ds'
-'              columns.types int:string:string:string'
-'              file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'              file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              location 
!!{hive.metastore.warehouse.dir}!!/input_part2.db/dest2'
-'              name input_part2.dest2'
-'              serialization.ddl struct dest2 { i32 key, string value, string 
hr, string ds}'
-'              serialization.format 1'
-'              serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              transient_lastDdlTime !!UNIXTIME!!'
-'            serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'          '
-'              input format: org.apache.hadoop.mapred.TextInputFormat'
-'              output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'              properties:'
-'                bucket_count -1'
-'                columns key,value,hr,ds'
-'                columns.types int:string:string:string'
-'                file.inputformat org.apache.hadoop.mapred.TextInputFormat'
-'                file.outputformat 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
-'                location 
!!{hive.metastore.warehouse.dir}!!/input_part2.db/dest2'
-'                name input_part2.dest2'
-'                serialization.ddl struct dest2 { i32 key, string value, 
string hr, string ds}'
-'                serialization.format 1'
-'                serialization.lib 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'                transient_lastDdlTime !!UNIXTIME!!'
-'              serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
-'              name: input_part2.dest2'
-'            name: input_part2.dest2'
-''
-'  Stage: Stage-13'
-'    Move Operator'
-'      files:'
-'          hdfs directory: true'
-'          source: pfile:!!{hive.exec.scratchdir}!!'
-'          destination: pfile:!!{hive.exec.scratchdir}!!'
-''
-''
-613 rows selected 
->>>  
->>>  FROM srcpart 
-INSERT OVERWRITE TABLE dest1 SELECT srcpart.key, srcpart.value, srcpart.hr, 
srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr 
= '12' 
-INSERT OVERWRITE TABLE dest2 SELECT srcpart.key, srcpart.value, srcpart.hr, 
srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-09' and srcpart.hr 
= '12';
-'_col0','_col1','_col2','_col3'
-No rows selected 
->>>  
->>>  SELECT dest1.* FROM dest1 sort by key,value,ds,hr;
-'key','value','hr','ds'
-'0','val_0','12','2008-04-08'
-'0','val_0','12','2008-04-08'
-'0','val_0','12','2008-04-08'
-'2','val_2','12','2008-04-08'
-'4','val_4','12','2008-04-08'
-'5','val_5','12','2008-04-08'
-'5','val_5','12','2008-04-08'
-'5','val_5','12','2008-04-08'
-'8','val_8','12','2008-04-08'
-'9','val_9','12','2008-04-08'
-'10','val_10','12','2008-04-08'
-'11','val_11','12','2008-04-08'
-'12','val_12','12','2008-04-08'
-'12','val_12','12','2008-04-08'
-'15','val_15','12','2008-04-08'
-'15','val_15','12','2008-04-08'
-'17','val_17','12','2008-04-08'
-'18','val_18','12','2008-04-08'
-'18','val_18','12','2008-04-08'
-'19','val_19','12','2008-04-08'
-'20','val_20','12','2008-04-08'
-'24','val_24','12','2008-04-08'
-'24','val_24','12','2008-04-08'
-'26','val_26','12','2008-04-08'
-'26','val_26','12','2008-04-08'
-'27','val_27','12','2008-04-08'
-'28','val_28','12','2008-04-08'
-'30','val_30','12','2008-04-08'
-'33','val_33','12','2008-04-08'
-'34','val_34','12','2008-04-08'
-'35','val_35','12','2008-04-08'
-'35','val_35','12','2008-04-08'
-'35','val_35','12','2008-04-08'
-'37','val_37','12','2008-04-08'
-'37','val_37','12','2008-04-08'
-'41','val_41','12','2008-04-08'
-'42','val_42','12','2008-04-08'
-'42','val_42','12','2008-04-08'
-'43','val_43','12','2008-04-08'
-'44','val_44','12','2008-04-08'
-'47','val_47','12','2008-04-08'
-'51','val_51','12','2008-04-08'
-'51','val_51','12','2008-04-08'
-'53','val_53','12','2008-04-08'
-'54','val_54','12','2008-04-08'
-'57','val_57','12','2008-04-08'
-'58','val_58','12','2008-04-08'
-'58','val_58','12','2008-04-08'
-'64','val_64','12','2008-04-08'
-'65','val_65','12','2008-04-08'
-'66','val_66','12','2008-04-08'
-'67','val_67','12','2008-04-08'
-'67','val_67','12','2008-04-08'
-'69','val_69','12','2008-04-08'
-'70','val_70','12','2008-04-08'
-'70','val_70','12','2008-04-08'
-'70','val_70','12','2008-04-08'
-'72','val_72','12','2008-04-08'
-'72','val_72','12','2008-04-08'
-'74','val_74','12','2008-04-08'
-'76','val_76','12','2008-04-08'
-'76','val_76','12','2008-04-08'
-'77','val_77','12','2008-04-08'
-'78','val_78','12','2008-04-08'
-'80','val_80','12','2008-04-08'
-'82','val_82','12','2008-04-08'
-'83','val_83','12','2008-04-08'
-'83','val_83','12','2008-04-08'
-'84','val_84','12','2008-04-08'
-'84','val_84','12','2008-04-08'
-'85','val_85','12','2008-04-08'
-'86','val_86','12','2008-04-08'
-'87','val_87','12','2008-04-08'
-'90','val_90','12','2008-04-08'
-'90','val_90','12','2008-04-08'
-'90','val_90','12','2008-04-08'
-'92','val_92','12','2008-04-08'
-'95','val_95','12','2008-04-08'
-'95','val_95','12','2008-04-08'
-'96','val_96','12','2008-04-08'
-'97','val_97','12','2008-04-08'
-'97','val_97','12','2008-04-08'
-'98','val_98','12','2008-04-08'
-'98','val_98','12','2008-04-08'
-84 rows selected 
->>>  SELECT dest2.* FROM dest2 sort by key,value,ds,hr;
-'key','value','hr','ds'
-'0','val_0','12','2008-04-09'
-'0','val_0','12','2008-04-09'
-'0','val_0','12','2008-04-09'
-'2','val_2','12','2008-04-09'
-'4','val_4','12','2008-04-09'
-'5','val_5','12','2008-04-09'
-'5','val_5','12','2008-04-09'
-'5','val_5','12','2008-04-09'
-'8','val_8','12','2008-04-09'
-'9','val_9','12','2008-04-09'
-'10','val_10','12','2008-04-09'
-'11','val_11','12','2008-04-09'
-'12','val_12','12','2008-04-09'
-'12','val_12','12','2008-04-09'
-'15','val_15','12','2008-04-09'
-'15','val_15','12','2008-04-09'
-'17','val_17','12','2008-04-09'
-'18','val_18','12','2008-04-09'
-'18','val_18','12','2008-04-09'
-'19','val_19','12','2008-04-09'
-'20','val_20','12','2008-04-09'
-'24','val_24','12','2008-04-09'
-'24','val_24','12','2008-04-09'
-'26','val_26','12','2008-04-09'
-'26','val_26','12','2008-04-09'
-'27','val_27','12','2008-04-09'
-'28','val_28','12','2008-04-09'
-'30','val_30','12','2008-04-09'
-'33','val_33','12','2008-04-09'
-'34','val_34','12','2008-04-09'
-'35','val_35','12','2008-04-09'
-'35','val_35','12','2008-04-09'
-'35','val_35','12','2008-04-09'
-'37','val_37','12','2008-04-09'
-'37','val_37','12','2008-04-09'
-'41','val_41','12','2008-04-09'
-'42','val_42','12','2008-04-09'
-'42','val_42','12','2008-04-09'
-'43','val_43','12','2008-04-09'
-'44','val_44','12','2008-04-09'
-'47','val_47','12','2008-04-09'
-'51','val_51','12','2008-04-09'
-'51','val_51','12','2008-04-09'
-'53','val_53','12','2008-04-09'
-'54','val_54','12','2008-04-09'
-'57','val_57','12','2008-04-09'
-'58','val_58','12','2008-04-09'
-'58','val_58','12','2008-04-09'
-'64','val_64','12','2008-04-09'
-'65','val_65','12','2008-04-09'
-'66','val_66','12','2008-04-09'
-'67','val_67','12','2008-04-09'
-'67','val_67','12','2008-04-09'
-'69','val_69','12','2008-04-09'
-'70','val_70','12','2008-04-09'
-'70','val_70','12','2008-04-09'
-'70','val_70','12','2008-04-09'
-'72','val_72','12','2008-04-09'
-'72','val_72','12','2008-04-09'
-'74','val_74','12','2008-04-09'
-'76','val_76','12','2008-04-09'
-'76','val_76','12','2008-04-09'
-'77','val_77','12','2008-04-09'
-'78','val_78','12','2008-04-09'
-'80','val_80','12','2008-04-09'
-'82','val_82','12','2008-04-09'
-'83','val_83','12','2008-04-09'
-'83','val_83','12','2008-04-09'
-'84','val_84','12','2008-04-09'
-'84','val_84','12','2008-04-09'
-'85','val_85','12','2008-04-09'
-'86','val_86','12','2008-04-09'
-'87','val_87','12','2008-04-09'
-'90','val_90','12','2008-04-09'
-'90','val_90','12','2008-04-09'
-'90','val_90','12','2008-04-09'
-'92','val_92','12','2008-04-09'
-'95','val_95','12','2008-04-09'
-'95','val_95','12','2008-04-09'
-'96','val_96','12','2008-04-09'
-'97','val_97','12','2008-04-09'
-'97','val_97','12','2008-04-09'
-'98','val_98','12','2008-04-09'
-'98','val_98','12','2008-04-09'
-84 rows selected 
->>>  
->>>  
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/input_part3.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/input_part3.q.out 
b/ql/src/test/results/beelinepositive/input_part3.q.out
deleted file mode 100644
index 8851485..0000000
--- a/ql/src/test/results/beelinepositive/input_part3.q.out
+++ /dev/null
@@ -1,538 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/input_part3.q.raw". Enter "record" 
with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/input_part3.q
->>>  EXPLAIN 
-SELECT x.* FROM SRCPART x WHERE x.ds = '2008-04-08' and x.hr = 11;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME SRCPART) x)) (TOK_INSERT 
(TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR 
(TOK_ALLCOLREF (TOK_TABNAME x)))) (TOK_WHERE (and (= (. (TOK_TABLE_OR_COL x) 
ds) '2008-04-08') (= (. (TOK_TABLE_OR_COL x) hr) 11)))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: -1'
-'      Processor Tree:'
-'        TableScan'
-'          alias: x'
-'          Select Operator'
-'            expressions:'
-'                  expr: key'
-'                  type: string'
-'                  expr: value'
-'                  type: string'
-'                  expr: ds'
-'                  type: string'
-'                  expr: hr'
-'                  type: string'
-'            outputColumnNames: _col0, _col1, _col2, _col3'
-'            ListSink'
-''
-''
-27 rows selected 
->>>  
->>>  SELECT x.* FROM SRCPART x WHERE x.ds = '2008-04-08' and x.hr = 11;
-'key','value','ds','hr'
-'238','val_238','2008-04-08','11'
-'86','val_86','2008-04-08','11'
-'311','val_311','2008-04-08','11'
-'27','val_27','2008-04-08','11'
-'165','val_165','2008-04-08','11'
-'409','val_409','2008-04-08','11'
-'255','val_255','2008-04-08','11'
-'278','val_278','2008-04-08','11'
-'98','val_98','2008-04-08','11'
-'484','val_484','2008-04-08','11'
-'265','val_265','2008-04-08','11'
-'193','val_193','2008-04-08','11'
-'401','val_401','2008-04-08','11'
-'150','val_150','2008-04-08','11'
-'273','val_273','2008-04-08','11'
-'224','val_224','2008-04-08','11'
-'369','val_369','2008-04-08','11'
-'66','val_66','2008-04-08','11'
-'128','val_128','2008-04-08','11'
-'213','val_213','2008-04-08','11'
-'146','val_146','2008-04-08','11'
-'406','val_406','2008-04-08','11'
-'429','val_429','2008-04-08','11'
-'374','val_374','2008-04-08','11'
-'152','val_152','2008-04-08','11'
-'469','val_469','2008-04-08','11'
-'145','val_145','2008-04-08','11'
-'495','val_495','2008-04-08','11'
-'37','val_37','2008-04-08','11'
-'327','val_327','2008-04-08','11'
-'281','val_281','2008-04-08','11'
-'277','val_277','2008-04-08','11'
-'209','val_209','2008-04-08','11'
-'15','val_15','2008-04-08','11'
-'82','val_82','2008-04-08','11'
-'403','val_403','2008-04-08','11'
-'166','val_166','2008-04-08','11'
-'417','val_417','2008-04-08','11'
-'430','val_430','2008-04-08','11'
-'252','val_252','2008-04-08','11'
-'292','val_292','2008-04-08','11'
-'219','val_219','2008-04-08','11'
-'287','val_287','2008-04-08','11'
-'153','val_153','2008-04-08','11'
-'193','val_193','2008-04-08','11'
-'338','val_338','2008-04-08','11'
-'446','val_446','2008-04-08','11'
-'459','val_459','2008-04-08','11'
-'394','val_394','2008-04-08','11'
-'237','val_237','2008-04-08','11'
-'482','val_482','2008-04-08','11'
-'174','val_174','2008-04-08','11'
-'413','val_413','2008-04-08','11'
-'494','val_494','2008-04-08','11'
-'207','val_207','2008-04-08','11'
-'199','val_199','2008-04-08','11'
-'466','val_466','2008-04-08','11'
-'208','val_208','2008-04-08','11'
-'174','val_174','2008-04-08','11'
-'399','val_399','2008-04-08','11'
-'396','val_396','2008-04-08','11'
-'247','val_247','2008-04-08','11'
-'417','val_417','2008-04-08','11'
-'489','val_489','2008-04-08','11'
-'162','val_162','2008-04-08','11'
-'377','val_377','2008-04-08','11'
-'397','val_397','2008-04-08','11'
-'309','val_309','2008-04-08','11'
-'365','val_365','2008-04-08','11'
-'266','val_266','2008-04-08','11'
-'439','val_439','2008-04-08','11'
-'342','val_342','2008-04-08','11'
-'367','val_367','2008-04-08','11'
-'325','val_325','2008-04-08','11'
-'167','val_167','2008-04-08','11'
-'195','val_195','2008-04-08','11'
-'475','val_475','2008-04-08','11'
-'17','val_17','2008-04-08','11'
-'113','val_113','2008-04-08','11'
-'155','val_155','2008-04-08','11'
-'203','val_203','2008-04-08','11'
-'339','val_339','2008-04-08','11'
-'0','val_0','2008-04-08','11'
-'455','val_455','2008-04-08','11'
-'128','val_128','2008-04-08','11'
-'311','val_311','2008-04-08','11'
-'316','val_316','2008-04-08','11'
-'57','val_57','2008-04-08','11'
-'302','val_302','2008-04-08','11'
-'205','val_205','2008-04-08','11'
-'149','val_149','2008-04-08','11'
-'438','val_438','2008-04-08','11'
-'345','val_345','2008-04-08','11'
-'129','val_129','2008-04-08','11'
-'170','val_170','2008-04-08','11'
-'20','val_20','2008-04-08','11'
-'489','val_489','2008-04-08','11'
-'157','val_157','2008-04-08','11'
-'378','val_378','2008-04-08','11'
-'221','val_221','2008-04-08','11'
-'92','val_92','2008-04-08','11'
-'111','val_111','2008-04-08','11'
-'47','val_47','2008-04-08','11'
-'72','val_72','2008-04-08','11'
-'4','val_4','2008-04-08','11'
-'280','val_280','2008-04-08','11'
-'35','val_35','2008-04-08','11'
-'427','val_427','2008-04-08','11'
-'277','val_277','2008-04-08','11'
-'208','val_208','2008-04-08','11'
-'356','val_356','2008-04-08','11'
-'399','val_399','2008-04-08','11'
-'169','val_169','2008-04-08','11'
-'382','val_382','2008-04-08','11'
-'498','val_498','2008-04-08','11'
-'125','val_125','2008-04-08','11'
-'386','val_386','2008-04-08','11'
-'437','val_437','2008-04-08','11'
-'469','val_469','2008-04-08','11'
-'192','val_192','2008-04-08','11'
-'286','val_286','2008-04-08','11'
-'187','val_187','2008-04-08','11'
-'176','val_176','2008-04-08','11'
-'54','val_54','2008-04-08','11'
-'459','val_459','2008-04-08','11'
-'51','val_51','2008-04-08','11'
-'138','val_138','2008-04-08','11'
-'103','val_103','2008-04-08','11'
-'239','val_239','2008-04-08','11'
-'213','val_213','2008-04-08','11'
-'216','val_216','2008-04-08','11'
-'430','val_430','2008-04-08','11'
-'278','val_278','2008-04-08','11'
-'176','val_176','2008-04-08','11'
-'289','val_289','2008-04-08','11'
-'221','val_221','2008-04-08','11'
-'65','val_65','2008-04-08','11'
-'318','val_318','2008-04-08','11'
-'332','val_332','2008-04-08','11'
-'311','val_311','2008-04-08','11'
-'275','val_275','2008-04-08','11'
-'137','val_137','2008-04-08','11'
-'241','val_241','2008-04-08','11'
-'83','val_83','2008-04-08','11'
-'333','val_333','2008-04-08','11'
-'180','val_180','2008-04-08','11'
-'284','val_284','2008-04-08','11'
-'12','val_12','2008-04-08','11'
-'230','val_230','2008-04-08','11'
-'181','val_181','2008-04-08','11'
-'67','val_67','2008-04-08','11'
-'260','val_260','2008-04-08','11'
-'404','val_404','2008-04-08','11'
-'384','val_384','2008-04-08','11'
-'489','val_489','2008-04-08','11'
-'353','val_353','2008-04-08','11'
-'373','val_373','2008-04-08','11'
-'272','val_272','2008-04-08','11'
-'138','val_138','2008-04-08','11'
-'217','val_217','2008-04-08','11'
-'84','val_84','2008-04-08','11'
-'348','val_348','2008-04-08','11'
-'466','val_466','2008-04-08','11'
-'58','val_58','2008-04-08','11'
-'8','val_8','2008-04-08','11'
-'411','val_411','2008-04-08','11'
-'230','val_230','2008-04-08','11'
-'208','val_208','2008-04-08','11'
-'348','val_348','2008-04-08','11'
-'24','val_24','2008-04-08','11'
-'463','val_463','2008-04-08','11'
-'431','val_431','2008-04-08','11'
-'179','val_179','2008-04-08','11'
-'172','val_172','2008-04-08','11'
-'42','val_42','2008-04-08','11'
-'129','val_129','2008-04-08','11'
-'158','val_158','2008-04-08','11'
-'119','val_119','2008-04-08','11'
-'496','val_496','2008-04-08','11'
-'0','val_0','2008-04-08','11'
-'322','val_322','2008-04-08','11'
-'197','val_197','2008-04-08','11'
-'468','val_468','2008-04-08','11'
-'393','val_393','2008-04-08','11'
-'454','val_454','2008-04-08','11'
-'100','val_100','2008-04-08','11'
-'298','val_298','2008-04-08','11'
-'199','val_199','2008-04-08','11'
-'191','val_191','2008-04-08','11'
-'418','val_418','2008-04-08','11'
-'96','val_96','2008-04-08','11'
-'26','val_26','2008-04-08','11'
-'165','val_165','2008-04-08','11'
-'327','val_327','2008-04-08','11'
-'230','val_230','2008-04-08','11'
-'205','val_205','2008-04-08','11'
-'120','val_120','2008-04-08','11'
-'131','val_131','2008-04-08','11'
-'51','val_51','2008-04-08','11'
-'404','val_404','2008-04-08','11'
-'43','val_43','2008-04-08','11'
-'436','val_436','2008-04-08','11'
-'156','val_156','2008-04-08','11'
-'469','val_469','2008-04-08','11'
-'468','val_468','2008-04-08','11'
-'308','val_308','2008-04-08','11'
-'95','val_95','2008-04-08','11'
-'196','val_196','2008-04-08','11'
-'288','val_288','2008-04-08','11'
-'481','val_481','2008-04-08','11'
-'457','val_457','2008-04-08','11'
-'98','val_98','2008-04-08','11'
-'282','val_282','2008-04-08','11'
-'197','val_197','2008-04-08','11'
-'187','val_187','2008-04-08','11'
-'318','val_318','2008-04-08','11'
-'318','val_318','2008-04-08','11'
-'409','val_409','2008-04-08','11'
-'470','val_470','2008-04-08','11'
-'137','val_137','2008-04-08','11'
-'369','val_369','2008-04-08','11'
-'316','val_316','2008-04-08','11'
-'169','val_169','2008-04-08','11'
-'413','val_413','2008-04-08','11'
-'85','val_85','2008-04-08','11'
-'77','val_77','2008-04-08','11'
-'0','val_0','2008-04-08','11'
-'490','val_490','2008-04-08','11'
-'87','val_87','2008-04-08','11'
-'364','val_364','2008-04-08','11'
-'179','val_179','2008-04-08','11'
-'118','val_118','2008-04-08','11'
-'134','val_134','2008-04-08','11'
-'395','val_395','2008-04-08','11'
-'282','val_282','2008-04-08','11'
-'138','val_138','2008-04-08','11'
-'238','val_238','2008-04-08','11'
-'419','val_419','2008-04-08','11'
-'15','val_15','2008-04-08','11'
-'118','val_118','2008-04-08','11'
-'72','val_72','2008-04-08','11'
-'90','val_90','2008-04-08','11'
-'307','val_307','2008-04-08','11'
-'19','val_19','2008-04-08','11'
-'435','val_435','2008-04-08','11'
-'10','val_10','2008-04-08','11'
-'277','val_277','2008-04-08','11'
-'273','val_273','2008-04-08','11'
-'306','val_306','2008-04-08','11'
-'224','val_224','2008-04-08','11'
-'309','val_309','2008-04-08','11'
-'389','val_389','2008-04-08','11'
-'327','val_327','2008-04-08','11'
-'242','val_242','2008-04-08','11'
-'369','val_369','2008-04-08','11'
-'392','val_392','2008-04-08','11'
-'272','val_272','2008-04-08','11'
-'331','val_331','2008-04-08','11'
-'401','val_401','2008-04-08','11'
-'242','val_242','2008-04-08','11'
-'452','val_452','2008-04-08','11'
-'177','val_177','2008-04-08','11'
-'226','val_226','2008-04-08','11'
-'5','val_5','2008-04-08','11'
-'497','val_497','2008-04-08','11'
-'402','val_402','2008-04-08','11'
-'396','val_396','2008-04-08','11'
-'317','val_317','2008-04-08','11'
-'395','val_395','2008-04-08','11'
-'58','val_58','2008-04-08','11'
-'35','val_35','2008-04-08','11'
-'336','val_336','2008-04-08','11'
-'95','val_95','2008-04-08','11'
-'11','val_11','2008-04-08','11'
-'168','val_168','2008-04-08','11'
-'34','val_34','2008-04-08','11'
-'229','val_229','2008-04-08','11'
-'233','val_233','2008-04-08','11'
-'143','val_143','2008-04-08','11'
-'472','val_472','2008-04-08','11'
-'322','val_322','2008-04-08','11'
-'498','val_498','2008-04-08','11'
-'160','val_160','2008-04-08','11'
-'195','val_195','2008-04-08','11'
-'42','val_42','2008-04-08','11'
-'321','val_321','2008-04-08','11'
-'430','val_430','2008-04-08','11'
-'119','val_119','2008-04-08','11'
-'489','val_489','2008-04-08','11'
-'458','val_458','2008-04-08','11'
-'78','val_78','2008-04-08','11'
-'76','val_76','2008-04-08','11'
-'41','val_41','2008-04-08','11'
-'223','val_223','2008-04-08','11'
-'492','val_492','2008-04-08','11'
-'149','val_149','2008-04-08','11'
-'449','val_449','2008-04-08','11'
-'218','val_218','2008-04-08','11'
-'228','val_228','2008-04-08','11'
-'138','val_138','2008-04-08','11'
-'453','val_453','2008-04-08','11'
-'30','val_30','2008-04-08','11'
-'209','val_209','2008-04-08','11'
-'64','val_64','2008-04-08','11'
-'468','val_468','2008-04-08','11'
-'76','val_76','2008-04-08','11'
-'74','val_74','2008-04-08','11'
-'342','val_342','2008-04-08','11'
-'69','val_69','2008-04-08','11'
-'230','val_230','2008-04-08','11'
-'33','val_33','2008-04-08','11'
-'368','val_368','2008-04-08','11'
-'103','val_103','2008-04-08','11'
-'296','val_296','2008-04-08','11'
-'113','val_113','2008-04-08','11'
-'216','val_216','2008-04-08','11'
-'367','val_367','2008-04-08','11'
-'344','val_344','2008-04-08','11'
-'167','val_167','2008-04-08','11'
-'274','val_274','2008-04-08','11'
-'219','val_219','2008-04-08','11'
-'239','val_239','2008-04-08','11'
-'485','val_485','2008-04-08','11'
-'116','val_116','2008-04-08','11'
-'223','val_223','2008-04-08','11'
-'256','val_256','2008-04-08','11'
-'263','val_263','2008-04-08','11'
-'70','val_70','2008-04-08','11'
-'487','val_487','2008-04-08','11'
-'480','val_480','2008-04-08','11'
-'401','val_401','2008-04-08','11'
-'288','val_288','2008-04-08','11'
-'191','val_191','2008-04-08','11'
-'5','val_5','2008-04-08','11'
-'244','val_244','2008-04-08','11'
-'438','val_438','2008-04-08','11'
-'128','val_128','2008-04-08','11'
-'467','val_467','2008-04-08','11'
-'432','val_432','2008-04-08','11'
-'202','val_202','2008-04-08','11'
-'316','val_316','2008-04-08','11'
-'229','val_229','2008-04-08','11'
-'469','val_469','2008-04-08','11'
-'463','val_463','2008-04-08','11'
-'280','val_280','2008-04-08','11'
-'2','val_2','2008-04-08','11'
-'35','val_35','2008-04-08','11'
-'283','val_283','2008-04-08','11'
-'331','val_331','2008-04-08','11'
-'235','val_235','2008-04-08','11'
-'80','val_80','2008-04-08','11'
-'44','val_44','2008-04-08','11'
-'193','val_193','2008-04-08','11'
-'321','val_321','2008-04-08','11'
-'335','val_335','2008-04-08','11'
-'104','val_104','2008-04-08','11'
-'466','val_466','2008-04-08','11'
-'366','val_366','2008-04-08','11'
-'175','val_175','2008-04-08','11'
-'403','val_403','2008-04-08','11'
-'483','val_483','2008-04-08','11'
-'53','val_53','2008-04-08','11'
-'105','val_105','2008-04-08','11'
-'257','val_257','2008-04-08','11'
-'406','val_406','2008-04-08','11'
-'409','val_409','2008-04-08','11'
-'190','val_190','2008-04-08','11'
-'406','val_406','2008-04-08','11'
-'401','val_401','2008-04-08','11'
-'114','val_114','2008-04-08','11'
-'258','val_258','2008-04-08','11'
-'90','val_90','2008-04-08','11'
-'203','val_203','2008-04-08','11'
-'262','val_262','2008-04-08','11'
-'348','val_348','2008-04-08','11'
-'424','val_424','2008-04-08','11'
-'12','val_12','2008-04-08','11'
-'396','val_396','2008-04-08','11'
-'201','val_201','2008-04-08','11'
-'217','val_217','2008-04-08','11'
-'164','val_164','2008-04-08','11'
-'431','val_431','2008-04-08','11'
-'454','val_454','2008-04-08','11'
-'478','val_478','2008-04-08','11'
-'298','val_298','2008-04-08','11'
-'125','val_125','2008-04-08','11'
-'431','val_431','2008-04-08','11'
-'164','val_164','2008-04-08','11'
-'424','val_424','2008-04-08','11'
-'187','val_187','2008-04-08','11'
-'382','val_382','2008-04-08','11'
-'5','val_5','2008-04-08','11'
-'70','val_70','2008-04-08','11'
-'397','val_397','2008-04-08','11'
-'480','val_480','2008-04-08','11'
-'291','val_291','2008-04-08','11'
-'24','val_24','2008-04-08','11'
-'351','val_351','2008-04-08','11'
-'255','val_255','2008-04-08','11'
-'104','val_104','2008-04-08','11'
-'70','val_70','2008-04-08','11'
-'163','val_163','2008-04-08','11'
-'438','val_438','2008-04-08','11'
-'119','val_119','2008-04-08','11'
-'414','val_414','2008-04-08','11'
-'200','val_200','2008-04-08','11'
-'491','val_491','2008-04-08','11'
-'237','val_237','2008-04-08','11'
-'439','val_439','2008-04-08','11'
-'360','val_360','2008-04-08','11'
-'248','val_248','2008-04-08','11'
-'479','val_479','2008-04-08','11'
-'305','val_305','2008-04-08','11'
-'417','val_417','2008-04-08','11'
-'199','val_199','2008-04-08','11'
-'444','val_444','2008-04-08','11'
-'120','val_120','2008-04-08','11'
-'429','val_429','2008-04-08','11'
-'169','val_169','2008-04-08','11'
-'443','val_443','2008-04-08','11'
-'323','val_323','2008-04-08','11'
-'325','val_325','2008-04-08','11'
-'277','val_277','2008-04-08','11'
-'230','val_230','2008-04-08','11'
-'478','val_478','2008-04-08','11'
-'178','val_178','2008-04-08','11'
-'468','val_468','2008-04-08','11'
-'310','val_310','2008-04-08','11'
-'317','val_317','2008-04-08','11'
-'333','val_333','2008-04-08','11'
-'493','val_493','2008-04-08','11'
-'460','val_460','2008-04-08','11'
-'207','val_207','2008-04-08','11'
-'249','val_249','2008-04-08','11'
-'265','val_265','2008-04-08','11'
-'480','val_480','2008-04-08','11'
-'83','val_83','2008-04-08','11'
-'136','val_136','2008-04-08','11'
-'353','val_353','2008-04-08','11'
-'172','val_172','2008-04-08','11'
-'214','val_214','2008-04-08','11'
-'462','val_462','2008-04-08','11'
-'233','val_233','2008-04-08','11'
-'406','val_406','2008-04-08','11'
-'133','val_133','2008-04-08','11'
-'175','val_175','2008-04-08','11'
-'189','val_189','2008-04-08','11'
-'454','val_454','2008-04-08','11'
-'375','val_375','2008-04-08','11'
-'401','val_401','2008-04-08','11'
-'421','val_421','2008-04-08','11'
-'407','val_407','2008-04-08','11'
-'384','val_384','2008-04-08','11'
-'256','val_256','2008-04-08','11'
-'26','val_26','2008-04-08','11'
-'134','val_134','2008-04-08','11'
-'67','val_67','2008-04-08','11'
-'384','val_384','2008-04-08','11'
-'379','val_379','2008-04-08','11'
-'18','val_18','2008-04-08','11'
-'462','val_462','2008-04-08','11'
-'492','val_492','2008-04-08','11'
-'100','val_100','2008-04-08','11'
-'298','val_298','2008-04-08','11'
-'9','val_9','2008-04-08','11'
-'341','val_341','2008-04-08','11'
-'498','val_498','2008-04-08','11'
-'146','val_146','2008-04-08','11'
-'458','val_458','2008-04-08','11'
-'362','val_362','2008-04-08','11'
-'186','val_186','2008-04-08','11'
-'285','val_285','2008-04-08','11'
-'348','val_348','2008-04-08','11'
-'167','val_167','2008-04-08','11'
-'18','val_18','2008-04-08','11'
-'273','val_273','2008-04-08','11'
-'183','val_183','2008-04-08','11'
-'281','val_281','2008-04-08','11'
-'344','val_344','2008-04-08','11'
-'97','val_97','2008-04-08','11'
-'469','val_469','2008-04-08','11'
-'315','val_315','2008-04-08','11'
-'84','val_84','2008-04-08','11'
-'28','val_28','2008-04-08','11'
-'37','val_37','2008-04-08','11'
-'448','val_448','2008-04-08','11'
-'152','val_152','2008-04-08','11'
-'348','val_348','2008-04-08','11'
-'307','val_307','2008-04-08','11'
-'194','val_194','2008-04-08','11'
-'414','val_414','2008-04-08','11'
-'477','val_477','2008-04-08','11'
-'222','val_222','2008-04-08','11'
-'126','val_126','2008-04-08','11'
-'90','val_90','2008-04-08','11'
-'169','val_169','2008-04-08','11'
-'403','val_403','2008-04-08','11'
-'400','val_400','2008-04-08','11'
-'200','val_200','2008-04-08','11'
-'97','val_97','2008-04-08','11'
-500 rows selected 
->>>  !record

http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/input_part4.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/beelinepositive/input_part4.q.out 
b/ql/src/test/results/beelinepositive/input_part4.q.out
deleted file mode 100644
index 0351fe1..0000000
--- a/ql/src/test/results/beelinepositive/input_part4.q.out
+++ /dev/null
@@ -1,42 +0,0 @@
-Saving all output to "!!{outputDirectory}!!/input_part4.q.raw". Enter "record" 
with no arguments to stop it.
->>>  !run !!{qFileDirectory}!!/input_part4.q
->>>  EXPLAIN 
-SELECT x.* FROM SRCPART x WHERE x.ds = '2008-04-08' and x.hr = 15;
-'Explain'
-'ABSTRACT SYNTAX TREE:'
-'  (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME SRCPART) x)) (TOK_INSERT 
(TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR 
(TOK_ALLCOLREF (TOK_TABNAME x)))) (TOK_WHERE (and (= (. (TOK_TABLE_OR_COL x) 
ds) '2008-04-08') (= (. (TOK_TABLE_OR_COL x) hr) 15)))))'
-''
-'STAGE DEPENDENCIES:'
-'  Stage-0 is a root stage'
-''
-'STAGE PLANS:'
-'  Stage: Stage-0'
-'    Fetch Operator'
-'      limit: -1'
-'      Processor Tree:'
-'        TableScan'
-'          alias: x'
-'          Filter Operator'
-'            predicate:'
-'                expr: ((ds = '2008-04-08') and (hr = 15.0))'
-'                type: boolean'
-'            Select Operator'
-'              expressions:'
-'                    expr: key'
-'                    type: string'
-'                    expr: value'
-'                    type: string'
-'                    expr: ds'
-'                    type: string'
-'                    expr: hr'
-'                    type: string'
-'              outputColumnNames: _col0, _col1, _col2, _col3'
-'              ListSink'
-''
-''
-31 rows selected 
->>>  
->>>  SELECT x.* FROM SRCPART x WHERE x.ds = '2008-04-08' and x.hr = 15;
-'key','value','ds','hr'
-No rows selected 
->>>  !record

Reply via email to