Modified: hive/branches/spark/ql/src/test/results/clientpositive/parquet_ctas.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/parquet_ctas.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/parquet_ctas.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/parquet_ctas.q.out Sat Aug 30 06:44:46 2014 @@ -39,9 +39,12 @@ POSTHOOK: Lineage: staging.value SIMPLE PREHOOK: query: create table parquet_ctas stored as parquet as select * from staging PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@staging +PREHOOK: Output: database:default +PREHOOK: Output: default@parquet_ctas POSTHOOK: query: create table parquet_ctas stored as parquet as select * from staging POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@staging +POSTHOOK: Output: database:default POSTHOOK: Output: default@parquet_ctas PREHOOK: query: describe parquet_ctas PREHOOK: type: DESCTABLE @@ -72,9 +75,12 @@ POSTHOOK: Input: default@parquet_ctas PREHOOK: query: create table parquet_ctas_advanced stored as parquet as select key+1,concat(value,"value") from staging PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@staging +PREHOOK: Output: database:default +PREHOOK: Output: default@parquet_ctas_advanced POSTHOOK: query: create table parquet_ctas_advanced stored as parquet as select key+1,concat(value,"value") from staging POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@staging +POSTHOOK: Output: database:default POSTHOOK: Output: default@parquet_ctas_advanced PREHOOK: query: describe parquet_ctas_advanced PREHOOK: type: DESCTABLE @@ -105,9 +111,12 @@ POSTHOOK: Input: default@parquet_ctas_ad PREHOOK: query: create table parquet_ctas_alias stored as parquet as select key+1 as mykey,concat(value,"value") as myvalue from staging PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@staging +PREHOOK: Output: database:default +PREHOOK: Output: default@parquet_ctas_alias POSTHOOK: query: create table parquet_ctas_alias stored as parquet as select key+1 as mykey,concat(value,"value") as myvalue from staging POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@staging +POSTHOOK: Output: database:default POSTHOOK: Output: default@parquet_ctas_alias PREHOOK: query: describe parquet_ctas_alias PREHOOK: type: DESCTABLE @@ -138,9 +147,12 @@ POSTHOOK: Input: default@parquet_ctas_al PREHOOK: query: create table parquet_ctas_mixed stored as parquet as select key,key+1,concat(value,"value") as myvalue from staging PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@staging +PREHOOK: Output: database:default +PREHOOK: Output: default@parquet_ctas_mixed POSTHOOK: query: create table parquet_ctas_mixed stored as parquet as select key,key+1,concat(value,"value") as myvalue from staging POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@staging +POSTHOOK: Output: database:default POSTHOOK: Output: default@parquet_ctas_mixed PREHOOK: query: describe parquet_ctas_mixed PREHOOK: type: DESCTABLE
Modified: hive/branches/spark/ql/src/test/results/clientpositive/parquet_join.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/parquet_join.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/parquet_join.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/parquet_join.q.out Sat Aug 30 06:44:46 2014 @@ -39,16 +39,22 @@ POSTHOOK: Lineage: staging.value SIMPLE PREHOOK: query: create table parquet_jointable1 stored as parquet as select * from staging PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@staging +PREHOOK: Output: database:default +PREHOOK: Output: default@parquet_jointable1 POSTHOOK: query: create table parquet_jointable1 stored as parquet as select * from staging POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@staging +POSTHOOK: Output: database:default POSTHOOK: Output: default@parquet_jointable1 PREHOOK: query: create table parquet_jointable2 stored as parquet as select key,key+1,concat(value,"value") as myvalue from staging PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@staging +PREHOOK: Output: database:default +PREHOOK: Output: default@parquet_jointable2 POSTHOOK: query: create table parquet_jointable2 stored as parquet as select key,key+1,concat(value,"value") as myvalue from staging POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@staging +POSTHOOK: Output: database:default POSTHOOK: Output: default@parquet_jointable2 PREHOOK: query: -- MR join Modified: hive/branches/spark/ql/src/test/results/clientpositive/partition_decode_name.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/partition_decode_name.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/partition_decode_name.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/partition_decode_name.q.out Sat Aug 30 06:44:46 2014 @@ -6,6 +6,8 @@ from (select '2011-01-11', '2011-01-11+1 select '2011-01-11', '2011-01-11+16:18:26' from src tablesample (1 rows) ) s PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@sc POSTHOOK: query: create table sc as select * from (select '2011-01-11', '2011-01-11+14:18:26' from src tablesample (1 rows) union all @@ -14,6 +16,7 @@ from (select '2011-01-11', '2011-01-11+1 select '2011-01-11', '2011-01-11+16:18:26' from src tablesample (1 rows) ) s POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@sc PREHOOK: query: create table sc_part (key string) partitioned by (ts string) stored as rcfile PREHOOK: type: CREATETABLE Modified: hive/branches/spark/ql/src/test/results/clientpositive/partition_special_char.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/partition_special_char.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/partition_special_char.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/partition_special_char.q.out Sat Aug 30 06:44:46 2014 @@ -6,6 +6,8 @@ from (select '2011-01-11', '2011-01-11+1 select '2011-01-11', '2011-01-11+16:18:26' from src tablesample (1 rows) ) s PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@sc POSTHOOK: query: create table sc as select * from (select '2011-01-11', '2011-01-11+14:18:26' from src tablesample (1 rows) union all @@ -14,6 +16,7 @@ from (select '2011-01-11', '2011-01-11+1 select '2011-01-11', '2011-01-11+16:18:26' from src tablesample (1 rows) ) s POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@sc PREHOOK: query: create table sc_part (key string) partitioned by (ts string) stored as rcfile PREHOOK: type: CREATETABLE Modified: hive/branches/spark/ql/src/test/results/clientpositive/ppd_field_garbage.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/ppd_field_garbage.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/ppd_field_garbage.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/ppd_field_garbage.q.out Sat Aug 30 06:44:46 2014 @@ -11,9 +11,12 @@ POSTHOOK: Output: default@test_issue PREHOOK: query: CREATE VIEW v_test_issue AS SELECT fileid, i.user, test_c.user_c.age FROM test_issue LATERAL VIEW explode(infos) info AS i PREHOOK: type: CREATEVIEW PREHOOK: Input: default@test_issue +PREHOOK: Output: database:default +PREHOOK: Output: default@v_test_issue POSTHOOK: query: CREATE VIEW v_test_issue AS SELECT fileid, i.user, test_c.user_c.age FROM test_issue LATERAL VIEW explode(infos) info AS i POSTHOOK: type: CREATEVIEW POSTHOOK: Input: default@test_issue +POSTHOOK: Output: database:default POSTHOOK: Output: default@v_test_issue PREHOOK: query: -- dummy data LOAD DATA LOCAL INPATH '../../data/files/kv1.txt' INTO TABLE test_issue Modified: hive/branches/spark/ql/src/test/results/clientpositive/ppd_union_view.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/ppd_union_view.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== Files hive/branches/spark/ql/src/test/results/clientpositive/ppd_union_view.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/ppd_union_view.q.out Sat Aug 30 06:44:46 2014 differ Modified: hive/branches/spark/ql/src/test/results/clientpositive/ptf.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/ptf.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/ptf.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/ptf.q.out Sat Aug 30 06:44:46 2014 @@ -772,6 +772,8 @@ from part group by p_mfgr, p_brand PREHOOK: type: CREATEVIEW PREHOOK: Input: default@part +PREHOOK: Output: database:default +PREHOOK: Output: default@mfgr_price_view POSTHOOK: query: -- 16. testViewAsTableInputToPTF create view IF NOT EXISTS mfgr_price_view as select p_mfgr, p_brand, @@ -780,6 +782,7 @@ from part group by p_mfgr, p_brand POSTHOOK: type: CREATEVIEW POSTHOOK: Input: default@part +POSTHOOK: Output: database:default POSTHOOK: Output: default@mfgr_price_view PREHOOK: query: select p_mfgr, p_brand, s, sum(s) over w1 as s1 Modified: hive/branches/spark/ql/src/test/results/clientpositive/query_result_fileformat.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/query_result_fileformat.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/query_result_fileformat.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/query_result_fileformat.q.out Sat Aug 30 06:44:46 2014 @@ -4,12 +4,15 @@ PREHOOK: query: create table nzhang_test http://asdf' value from src limit 1 PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@nzhang_test1 POSTHOOK: query: create table nzhang_test1 stored as sequencefile as select 'key1' as key, 'value 1 http://asdf' value from src limit 1 POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@nzhang_test1 PREHOOK: query: select * from nzhang_test1 PREHOOK: type: QUERY Modified: hive/branches/spark/ql/src/test/results/clientpositive/quotedid_basic.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/quotedid_basic.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/quotedid_basic.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/quotedid_basic.q.out Sat Aug 30 06:44:46 2014 @@ -359,12 +359,15 @@ select `x+1```, `y&y` from t4 where `x+1``` < '200' PREHOOK: type: CREATEVIEW PREHOOK: Input: default@t4 +PREHOOK: Output: database:default +PREHOOK: Output: default@v1 POSTHOOK: query: -- view create view v1 as select `x+1```, `y&y` from t4 where `x+1``` < '200' POSTHOOK: type: CREATEVIEW POSTHOOK: Input: default@t4 +POSTHOOK: Output: database:default POSTHOOK: Output: default@v1 PREHOOK: query: select `x+1```, `y&y`, rank() over(partition by `x+1``` order by `y&y`) from v1 Modified: hive/branches/spark/ql/src/test/results/clientpositive/rcfile_createas1.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/rcfile_createas1.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/rcfile_createas1.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/rcfile_createas1.q.out Sat Aug 30 06:44:46 2014 @@ -105,7 +105,7 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat serde name: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: rcfile_createas1b + name: default.rcfile_createas1b Stage: Stage-2 Stats-Aggr Operator @@ -134,6 +134,8 @@ PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@rcfile_createas1a PREHOOK: Input: default@rcfile_createas1a@ds=1 PREHOOK: Input: default@rcfile_createas1a@ds=2 +PREHOOK: Output: database:default +PREHOOK: Output: default@rcfile_createas1b POSTHOOK: query: CREATE TABLE rcfile_createas1b STORED AS RCFILE AS SELECT key, value, PMOD(HASH(key), 50) as part @@ -142,6 +144,7 @@ POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@rcfile_createas1a POSTHOOK: Input: default@rcfile_createas1a@ds=1 POSTHOOK: Input: default@rcfile_createas1a@ds=2 +POSTHOOK: Output: database:default POSTHOOK: Output: default@rcfile_createas1b PREHOOK: query: SELECT SUM(HASH(c)) FROM ( SELECT TRANSFORM(key, value) USING 'tr \t _' AS (c) Modified: hive/branches/spark/ql/src/test/results/clientpositive/rcfile_default_format.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/rcfile_default_format.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/rcfile_default_format.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/rcfile_default_format.q.out Sat Aug 30 06:44:46 2014 @@ -39,9 +39,12 @@ Storage Desc Params: PREHOOK: query: CREATE TABLE rcfile_default_format_ctas AS SELECT key,value FROM src PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@rcfile_default_format_ctas POSTHOOK: query: CREATE TABLE rcfile_default_format_ctas AS SELECT key,value FROM src POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@rcfile_default_format_ctas PREHOOK: query: DESCRIBE FORMATTED rcfile_default_format_ctas PREHOOK: type: DESCTABLE @@ -134,9 +137,12 @@ Storage Desc Params: PREHOOK: query: CREATE TABLE textfile_default_format_ctas AS SELECT key,value FROM rcfile_default_format_ctas PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@rcfile_default_format_ctas +PREHOOK: Output: database:default +PREHOOK: Output: default@textfile_default_format_ctas POSTHOOK: query: CREATE TABLE textfile_default_format_ctas AS SELECT key,value FROM rcfile_default_format_ctas POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@rcfile_default_format_ctas +POSTHOOK: Output: database:default POSTHOOK: Output: default@textfile_default_format_ctas PREHOOK: query: DESCRIBE FORMATTED textfile_default_format_ctas PREHOOK: type: DESCTABLE @@ -177,9 +183,12 @@ Storage Desc Params: PREHOOK: query: CREATE TABLE rcfile_default_format_ctas_default_serde AS SELECT key,value FROM rcfile_default_format_ctas PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@rcfile_default_format_ctas +PREHOOK: Output: database:default +PREHOOK: Output: default@rcfile_default_format_ctas_default_serde POSTHOOK: query: CREATE TABLE rcfile_default_format_ctas_default_serde AS SELECT key,value FROM rcfile_default_format_ctas POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@rcfile_default_format_ctas +POSTHOOK: Output: database:default POSTHOOK: Output: default@rcfile_default_format_ctas_default_serde PREHOOK: query: DESCRIBE FORMATTED rcfile_default_format_ctas_default_serde PREHOOK: type: DESCTABLE @@ -258,9 +267,12 @@ Storage Desc Params: PREHOOK: query: CREATE TABLE rcfile_ctas_default_serde STORED AS rcfile AS SELECT key,value FROM rcfile_default_format_ctas PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@rcfile_default_format_ctas +PREHOOK: Output: database:default +PREHOOK: Output: default@rcfile_ctas_default_serde POSTHOOK: query: CREATE TABLE rcfile_ctas_default_serde STORED AS rcfile AS SELECT key,value FROM rcfile_default_format_ctas POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@rcfile_default_format_ctas +POSTHOOK: Output: database:default POSTHOOK: Output: default@rcfile_ctas_default_serde PREHOOK: query: DESCRIBE FORMATTED rcfile_ctas_default_serde PREHOOK: type: DESCTABLE Modified: hive/branches/spark/ql/src/test/results/clientpositive/sample_islocalmode_hook.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/sample_islocalmode_hook.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/sample_islocalmode_hook.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/sample_islocalmode_hook.q.out Sat Aug 30 06:44:46 2014 @@ -48,19 +48,25 @@ PREHOOK: Input: default@sih_i_part PREHOOK: Input: default@sih_i_part@p=1 PREHOOK: Input: default@sih_i_part@p=2 PREHOOK: Input: default@sih_i_part@p=3 +PREHOOK: Output: database:default +PREHOOK: Output: default@sih_src POSTHOOK: query: create table sih_src as select key, value from sih_i_part order by key, value POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@sih_i_part POSTHOOK: Input: default@sih_i_part@p=1 POSTHOOK: Input: default@sih_i_part@p=2 POSTHOOK: Input: default@sih_i_part@p=3 +POSTHOOK: Output: database:default POSTHOOK: Output: default@sih_src PREHOOK: query: create table sih_src2 as select key, value from sih_src order by key, value PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@sih_src +PREHOOK: Output: database:default +PREHOOK: Output: default@sih_src2 POSTHOOK: query: create table sih_src2 as select key, value from sih_src order by key, value POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@sih_src +POSTHOOK: Output: database:default POSTHOOK: Output: default@sih_src2 PREHOOK: query: -- Relaxing hive.exec.mode.local.auto.input.files.max=1. -- Hadoop20 will not generate more splits than there are files (one). Modified: hive/branches/spark/ql/src/test/results/clientpositive/semijoin.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/semijoin.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/semijoin.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/semijoin.q.out Sat Aug 30 06:44:46 2014 @@ -1,9 +1,12 @@ PREHOOK: query: create table t1 as select cast(key as int) key, value from src where key <= 10 PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@t1 POSTHOOK: query: create table t1 as select cast(key as int) key, value from src where key <= 10 POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@t1 PREHOOK: query: select * from t1 sort by key PREHOOK: type: QUERY @@ -27,9 +30,12 @@ POSTHOOK: Input: default@t1 PREHOOK: query: create table t2 as select cast(2*key as int) key, value from t1 PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@t1 +PREHOOK: Output: database:default +PREHOOK: Output: default@t2 POSTHOOK: query: create table t2 as select cast(2*key as int) key, value from t1 POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@t1 +POSTHOOK: Output: database:default POSTHOOK: Output: default@t2 PREHOOK: query: select * from t2 sort by key PREHOOK: type: QUERY @@ -54,10 +60,13 @@ PREHOOK: query: create table t3 as selec PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@t1 PREHOOK: Input: default@t2 +PREHOOK: Output: database:default +PREHOOK: Output: default@t3 POSTHOOK: query: create table t3 as select * from (select * from t1 union all select * from t2) b POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@t1 POSTHOOK: Input: default@t2 +POSTHOOK: Output: database:default POSTHOOK: Output: default@t3 PREHOOK: query: select * from t3 sort by key, value PREHOOK: type: QUERY Modified: hive/branches/spark/ql/src/test/results/clientpositive/serde_regex.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/serde_regex.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/serde_regex.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/serde_regex.q.out Sat Aug 30 06:44:46 2014 @@ -45,7 +45,7 @@ STAGE PLANS: serde name: org.apache.hadoop.hive.serde2.RegexSerDe serde properties: input.regex ([^ ]*) ([^ ]*) ([^ ]*) (-|\[[^\]]*\]) ([^ "]*|"[^"]*") (-|[0-9]*) (-|[0-9]*)(?: ([^ "]*|"[^"]*") ([^ "]*|"[^"]*"))? - name: serde_regex + name: default.serde_regex PREHOOK: query: CREATE TABLE serde_regex( host STRING, @@ -160,7 +160,7 @@ STAGE PLANS: serde name: org.apache.hadoop.hive.serde2.RegexSerDe serde properties: input.regex ([^ ]*) ([^ ]*) - name: serde_regex1 + name: default.serde_regex1 PREHOOK: query: CREATE TABLE serde_regex1( key decimal(38,18), Modified: hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_db_table.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_db_table.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_db_table.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_db_table.q.out Sat Aug 30 06:44:46 2014 @@ -17,11 +17,10 @@ tmp_feng PREHOOK: query: CREATE TABLE tmp_feng.tmp_showcrt (key string, value int) PREHOOK: type: CREATETABLE PREHOOK: Output: database:tmp_feng -PREHOOK: Output: tmp_feng@tmp_feng.tmp_showcrt +PREHOOK: Output: tmp_feng@tmp_showcrt POSTHOOK: query: CREATE TABLE tmp_feng.tmp_showcrt (key string, value int) POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:tmp_feng -POSTHOOK: Output: tmp_feng@tmp_feng.tmp_showcrt POSTHOOK: Output: tmp_feng@tmp_showcrt PREHOOK: query: USE default PREHOOK: type: SWITCHDATABASE Modified: hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_temp_table.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_temp_table.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_temp_table.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_temp_table.q.out Sat Aug 30 06:44:46 2014 @@ -7,12 +7,11 @@ POSTHOOK: Output: database:tmpdb PREHOOK: query: create temporary table tmpdb.tmp1 (c1 string, c2 string) PREHOOK: type: CREATETABLE PREHOOK: Output: database:tmpdb -PREHOOK: Output: [email protected] +PREHOOK: Output: tmpdb@tmp1 POSTHOOK: query: create temporary table tmpdb.tmp1 (c1 string, c2 string) POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:tmpdb POSTHOOK: Output: tmpdb@tmp1 -POSTHOOK: Output: [email protected] PREHOOK: query: show create table tmpdb.tmp1 PREHOOK: type: SHOW_CREATETABLE PREHOOK: Input: tmpdb@tmp1 Modified: hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_view.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_view.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_view.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/show_create_table_view.q.out Sat Aug 30 06:44:46 2014 @@ -3,11 +3,14 @@ PREHOOK: query: -- Test SHOW CREATE TABL CREATE VIEW tmp_copy_src AS SELECT * FROM src PREHOOK: type: CREATEVIEW PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@tmp_copy_src POSTHOOK: query: -- Test SHOW CREATE TABLE on a view name. CREATE VIEW tmp_copy_src AS SELECT * FROM src POSTHOOK: type: CREATEVIEW POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@tmp_copy_src PREHOOK: query: SHOW CREATE TABLE tmp_copy_src PREHOOK: type: SHOW_CREATETABLE Modified: hive/branches/spark/ql/src/test/results/clientpositive/skewjoin_noskew.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/skewjoin_noskew.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/skewjoin_noskew.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/skewjoin_noskew.q.out Sat Aug 30 06:44:46 2014 @@ -147,7 +147,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: noskew + name: default.noskew Stage: Stage-3 Stats-Aggr Operator @@ -155,9 +155,12 @@ STAGE PLANS: PREHOOK: query: create table noskew as select a.* from src a join src b on a.key=b.key order by a.key limit 30 PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@noskew POSTHOOK: query: create table noskew as select a.* from src a join src b on a.key=b.key order by a.key limit 30 POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@noskew PREHOOK: query: select * from noskew PREHOOK: type: QUERY Modified: hive/branches/spark/ql/src/test/results/clientpositive/smb_mapjoin9.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/smb_mapjoin9.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/smb_mapjoin9.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/smb_mapjoin9.q.out Sat Aug 30 06:44:46 2014 @@ -352,7 +352,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: smb_mapjoin9_results + name: default.smb_mapjoin9_results Stage: Stage-3 Stats-Aggr Operator @@ -397,6 +397,8 @@ PREHOOK: Input: default@hive_test_smb_bu PREHOOK: Input: default@hive_test_smb_bucket1@ds=2010-10-15 PREHOOK: Input: default@hive_test_smb_bucket2 PREHOOK: Input: default@hive_test_smb_bucket2@ds=2010-10-15 +PREHOOK: Output: database:default +PREHOOK: Output: default@smb_mapjoin9_results POSTHOOK: query: create table smb_mapjoin9_results as SELECT /* + MAPJOIN(b) */ b.key as k1, b.value, b.ds, a.key as k2 FROM hive_test_smb_bucket1 a JOIN @@ -407,6 +409,7 @@ POSTHOOK: Input: default@hive_test_smb_b POSTHOOK: Input: default@hive_test_smb_bucket1@ds=2010-10-15 POSTHOOK: Input: default@hive_test_smb_bucket2 POSTHOOK: Input: default@hive_test_smb_bucket2@ds=2010-10-15 +POSTHOOK: Output: database:default POSTHOOK: Output: default@smb_mapjoin9_results PREHOOK: query: drop table smb_mapjoin9_results PREHOOK: type: DROPTABLE Modified: hive/branches/spark/ql/src/test/results/clientpositive/stats5.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/stats5.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/stats5.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/stats5.q.out Sat Aug 30 06:44:46 2014 @@ -1,9 +1,12 @@ PREHOOK: query: create table analyze_src as select * from src PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@analyze_src POSTHOOK: query: create table analyze_src as select * from src POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@analyze_src PREHOOK: query: explain analyze table analyze_src compute statistics PREHOOK: type: QUERY Modified: hive/branches/spark/ql/src/test/results/clientpositive/stats_counter.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/stats_counter.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/stats_counter.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/stats_counter.q.out Sat Aug 30 06:44:46 2014 @@ -2,10 +2,13 @@ PREHOOK: query: -- by analyze create table dummy1 as select * from src PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@dummy1 POSTHOOK: query: -- by analyze create table dummy1 as select * from src POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@dummy1 PREHOOK: query: analyze table dummy1 compute statistics PREHOOK: type: QUERY @@ -55,10 +58,13 @@ PREHOOK: query: -- by autogather create table dummy2 as select * from src PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@dummy2 POSTHOOK: query: -- by autogather create table dummy2 as select * from src POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@dummy2 PREHOOK: query: desc formatted dummy2 PREHOOK: type: DESCTABLE Modified: hive/branches/spark/ql/src/test/results/clientpositive/str_to_map.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/str_to_map.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/str_to_map.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/str_to_map.q.out Sat Aug 30 06:44:46 2014 @@ -187,9 +187,12 @@ POSTHOOK: type: DROPTABLE PREHOOK: query: create table tbl_s2m as select 'ABC=CC_333=444' as t from src tablesample (3 rows) PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@tbl_s2m POSTHOOK: query: create table tbl_s2m as select 'ABC=CC_333=444' as t from src tablesample (3 rows) POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@tbl_s2m PREHOOK: query: select str_to_map(t,'_','=')['333'] from tbl_s2m PREHOOK: type: QUERY Modified: hive/branches/spark/ql/src/test/results/clientpositive/subquery_exists.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/subquery_exists.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/subquery_exists.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/subquery_exists.q.out Sat Aug 30 06:44:46 2014 @@ -125,6 +125,8 @@ where exists where b.value = a.value and a.key = b.key and a.value > 'val_9') PREHOOK: type: CREATEVIEW PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@cv1 POSTHOOK: query: -- view test create view cv1 as select * @@ -135,6 +137,7 @@ where exists where b.value = a.value and a.key = b.key and a.value > 'val_9') POSTHOOK: type: CREATEVIEW POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@cv1 PREHOOK: query: select * from cv1 PREHOOK: type: QUERY Modified: hive/branches/spark/ql/src/test/results/clientpositive/subquery_exists_having.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/subquery_exists_having.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/subquery_exists_having.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/subquery_exists_having.q.out Sat Aug 30 06:44:46 2014 @@ -320,6 +320,8 @@ having exists ) PREHOOK: type: CREATEVIEW PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@cv1 POSTHOOK: query: -- view test create view cv1 as select b.key, count(*) as c @@ -332,6 +334,7 @@ having exists ) POSTHOOK: type: CREATEVIEW POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@cv1 PREHOOK: query: select * from cv1 PREHOOK: type: QUERY Modified: hive/branches/spark/ql/src/test/results/clientpositive/subquery_notin.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/subquery_notin.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/subquery_notin.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/subquery_notin.q.out Sat Aug 30 06:44:46 2014 @@ -1413,22 +1413,28 @@ create view T1_v as select key from src where key <'11' PREHOOK: type: CREATEVIEW PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_v POSTHOOK: query: -- null check create view T1_v as select key from src where key <'11' POSTHOOK: type: CREATEVIEW POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@T1_v PREHOOK: query: create view T2_v as select case when key > '104' then null else key end as key from T1_v PREHOOK: type: CREATEVIEW PREHOOK: Input: default@src PREHOOK: Input: default@t1_v +PREHOOK: Output: database:default +PREHOOK: Output: default@T2_v POSTHOOK: query: create view T2_v as select case when key > '104' then null else key end as key from T1_v POSTHOOK: type: CREATEVIEW POSTHOOK: Input: default@src POSTHOOK: Input: default@t1_v +POSTHOOK: Output: database:default POSTHOOK: Output: default@T2_v Warning: Shuffle Join JOIN[24][tables = [t1_v, sq_1_notin_nullcheck]] in Stage 'Stage-2:MAPRED' is a cross product PREHOOK: query: explain Modified: hive/branches/spark/ql/src/test/results/clientpositive/subquery_views.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/subquery_views.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/subquery_views.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/subquery_views.q.out Sat Aug 30 06:44:46 2014 @@ -10,6 +10,8 @@ where exists where b.value = a.value and a.key = b.key and a.value > 'val_9') PREHOOK: type: CREATEVIEW PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@cv1 POSTHOOK: query: -- SORT_QUERY_RESULTS -- exists test @@ -22,6 +24,7 @@ where exists where b.value = a.value and a.key = b.key and a.value > 'val_9') POSTHOOK: type: CREATEVIEW POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@cv1 PREHOOK: query: select * from cv1 where cv1.key in (select key from cv1 c where c.key > '95') @@ -51,6 +54,8 @@ where b.key not in ) PREHOOK: type: CREATEVIEW PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@cv2 POSTHOOK: query: -- not in test create view cv2 as select * @@ -62,6 +67,7 @@ where b.key not in ) POSTHOOK: type: CREATEVIEW POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@cv2 Warning: Shuffle Join JOIN[42][tables = [b, sq_1_notin_nullcheck]] in Stage 'Stage-5:MAPRED' is a cross product Warning: Shuffle Join JOIN[18][tables = [b, sq_1_notin_nullcheck]] in Stage 'Stage-1:MAPRED' is a cross product @@ -416,6 +422,8 @@ group by key, value having count(*) in (select count(*) from src s1 where s1.key > '9' group by s1.key ) PREHOOK: type: CREATEVIEW PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@cv3 POSTHOOK: query: -- in where + having create view cv3 as select key, value, count(*) @@ -425,6 +433,7 @@ group by key, value having count(*) in (select count(*) from src s1 where s1.key > '9' group by s1.key ) POSTHOOK: type: CREATEVIEW POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@cv3 PREHOOK: query: select * from cv3 PREHOOK: type: QUERY Modified: hive/branches/spark/ql/src/test/results/clientpositive/symlink_text_input_format.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/symlink_text_input_format.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/symlink_text_input_format.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/symlink_text_input_format.q.out Sat Aug 30 06:44:46 2014 @@ -18,7 +18,7 @@ STAGE PLANS: columns: key string, value string input format: org.apache.hadoop.hive.ql.io.SymlinkTextInputFormat output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat - name: symlink_text_input_format + name: default.symlink_text_input_format PREHOOK: query: CREATE TABLE symlink_text_input_format (key STRING, value STRING) STORED AS INPUTFORMAT 'org.apache.hadoop.hive.ql.io.SymlinkTextInputFormat' OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat' PREHOOK: type: CREATETABLE Modified: hive/branches/spark/ql/src/test/results/clientpositive/temp_table.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/temp_table.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/temp_table.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/temp_table.q.out Sat Aug 30 06:44:46 2014 @@ -59,7 +59,7 @@ STAGE PLANS: #### A masked pattern was here #### output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: foo + name: default.foo isTemporary: true Stage: Stage-2 @@ -98,9 +98,12 @@ STAGE PLANS: PREHOOK: query: CREATE TEMPORARY TABLE foo AS SELECT * FROM src WHERE key % 2 = 0 PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@foo POSTHOOK: query: CREATE TEMPORARY TABLE foo AS SELECT * FROM src WHERE key % 2 = 0 POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@foo PREHOOK: query: EXPLAIN CREATE TEMPORARY TABLE bar AS SELECT * FROM src WHERE key % 2 = 1 PREHOOK: type: CREATETABLE_AS_SELECT @@ -163,7 +166,7 @@ STAGE PLANS: #### A masked pattern was here #### output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: bar + name: default.bar isTemporary: true Stage: Stage-2 @@ -202,9 +205,12 @@ STAGE PLANS: PREHOOK: query: CREATE TEMPORARY TABLE bar AS SELECT * FROM src WHERE key % 2 = 1 PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@bar POSTHOOK: query: CREATE TEMPORARY TABLE bar AS SELECT * FROM src WHERE key % 2 = 1 POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@bar PREHOOK: query: DESCRIBE foo PREHOOK: type: DESCTABLE @@ -452,9 +458,12 @@ POSTHOOK: type: SHOWTABLES PREHOOK: query: CREATE TEMPORARY TABLE foo AS SELECT * FROM default.foo PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@foo +PREHOOK: Output: database:two +PREHOOK: Output: two@foo POSTHOOK: query: CREATE TEMPORARY TABLE foo AS SELECT * FROM default.foo POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@foo +POSTHOOK: Output: database:two POSTHOOK: Output: two@foo PREHOOK: query: SHOW TABLES PREHOOK: type: SHOWTABLES Modified: hive/branches/spark/ql/src/test/results/clientpositive/temp_table_gb1.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/temp_table_gb1.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/temp_table_gb1.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/temp_table_gb1.q.out Sat Aug 30 06:44:46 2014 @@ -11,9 +11,12 @@ POSTHOOK: Output: default@dest_g2 PREHOOK: query: CREATE TEMPORARY TABLE src_temp AS SELECT * FROM src PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@src_temp POSTHOOK: query: CREATE TEMPORARY TABLE src_temp AS SELECT * FROM src POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@src_temp PREHOOK: query: FROM src_temp INSERT OVERWRITE TABLE dest_g2 SELECT substr(src_temp.key,1,1), count(DISTINCT substr(src_temp.value,5)), concat(substr(src_temp.key,1,1),sum(substr(src_temp.value,5))) GROUP BY substr(src_temp.key,1,1) Modified: hive/branches/spark/ql/src/test/results/clientpositive/temp_table_join1.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/temp_table_join1.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/temp_table_join1.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/temp_table_join1.q.out Sat Aug 30 06:44:46 2014 @@ -1,16 +1,22 @@ PREHOOK: query: CREATE TABLE src_nontemp AS SELECT * FROM src limit 10 PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@src_nontemp POSTHOOK: query: CREATE TABLE src_nontemp AS SELECT * FROM src limit 10 POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@src_nontemp PREHOOK: query: CREATE TEMPORARY TABLE src_temp AS SELECT * FROM src limit 10 PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@src_temp POSTHOOK: query: CREATE TEMPORARY TABLE src_temp AS SELECT * FROM src limit 10 POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@src_temp PREHOOK: query: -- Non temp table join EXPLAIN Modified: hive/branches/spark/ql/src/test/results/clientpositive/temp_table_names.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/temp_table_names.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/temp_table_names.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/temp_table_names.q.out Sat Aug 30 06:44:46 2014 @@ -1,12 +1,11 @@ PREHOOK: query: -- Test temp tables with upper/lower case names create temporary table Default.Temp_Table_Names (C1 string, c2 string) PREHOOK: type: CREATETABLE -PREHOOK: Output: [email protected]_Table_Names +PREHOOK: Output: Default@Temp_Table_Names PREHOOK: Output: database:default POSTHOOK: query: -- Test temp tables with upper/lower case names create temporary table Default.Temp_Table_Names (C1 string, c2 string) POSTHOOK: type: CREATETABLE -POSTHOOK: Output: [email protected]_Table_Names POSTHOOK: Output: Default@Temp_Table_Names POSTHOOK: Output: database:default PREHOOK: query: show tables 'Temp_Table*' Modified: hive/branches/spark/ql/src/test/results/clientpositive/temp_table_precedence.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/temp_table_precedence.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/temp_table_precedence.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/temp_table_precedence.q.out Sat Aug 30 06:44:46 2014 @@ -8,13 +8,12 @@ PREHOOK: query: -- Create non-temp table create table ttp.tab1 (a1 string, a2 string) PREHOOK: type: CREATETABLE PREHOOK: Output: database:ttp -PREHOOK: Output: [email protected] +PREHOOK: Output: ttp@tab1 POSTHOOK: query: -- Create non-temp tables create table ttp.tab1 (a1 string, a2 string) POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:ttp POSTHOOK: Output: ttp@tab1 -POSTHOOK: Output: [email protected] PREHOOK: query: insert overwrite table ttp.tab1 select * from src where key = 5 limit 5 PREHOOK: type: QUERY PREHOOK: Input: default@src @@ -47,12 +46,11 @@ POSTHOOK: Input: ttp@tab1 PREHOOK: query: create table ttp.tab2 (b1 string, b2 string) PREHOOK: type: CREATETABLE PREHOOK: Output: database:ttp -PREHOOK: Output: [email protected] +PREHOOK: Output: ttp@tab2 POSTHOOK: query: create table ttp.tab2 (b1 string, b2 string) POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:ttp POSTHOOK: Output: ttp@tab2 -POSTHOOK: Output: [email protected] PREHOOK: query: insert overwrite table ttp.tab2 select * from src where key = 2 limit 5 PREHOOK: type: QUERY PREHOOK: Input: default@src @@ -84,13 +82,12 @@ PREHOOK: query: -- Now create temp table create temporary table ttp.tab1 (c1 int, c2 string) PREHOOK: type: CREATETABLE PREHOOK: Output: database:ttp -PREHOOK: Output: [email protected] +PREHOOK: Output: ttp@tab1 POSTHOOK: query: -- Now create temp table with same name create temporary table ttp.tab1 (c1 int, c2 string) POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:ttp POSTHOOK: Output: ttp@tab1 -POSTHOOK: Output: [email protected] PREHOOK: query: insert overwrite table ttp.tab1 select * from src where key = 0 limit 5 PREHOOK: type: QUERY PREHOOK: Input: default@src Modified: hive/branches/spark/ql/src/test/results/clientpositive/temp_table_subquery1.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/temp_table_subquery1.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/temp_table_subquery1.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/temp_table_subquery1.q.out Sat Aug 30 06:44:46 2014 @@ -1,9 +1,12 @@ PREHOOK: query: create temporary table src_temp as select * from src PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@src_temp POSTHOOK: query: create temporary table src_temp as select * from src POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@src_temp PREHOOK: query: -- subquery exists select * Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_1.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_1.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_1.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_1.q.out Sat Aug 30 06:44:46 2014 @@ -2,21 +2,27 @@ PREHOOK: query: create table A as select * from src PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@A POSTHOOK: query: create table A as select * from src POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@A PREHOOK: query: create table B as select * from src limit 10 PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@B POSTHOOK: query: create table B as select * from src limit 10 POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@B Warning: Shuffle Join JOIN[4][tables = [a, b]] in Stage 'Reducer 2' is a cross product PREHOOK: query: explain select * from A join B Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_2.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_2.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_2.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/tez/cross_product_check_2.q.out Sat Aug 30 06:44:46 2014 @@ -2,21 +2,27 @@ PREHOOK: query: create table A as select * from src PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@A POSTHOOK: query: create table A as select * from src POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@A PREHOOK: query: create table B as select * from src limit 10 PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@B POSTHOOK: query: create table B as select * from src limit 10 POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@B Warning: Map Join MAPJOIN[7][bigTable=a] in task 'Map 2' is a cross product PREHOOK: query: explain select * from A join B Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/ctas.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/ctas.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/tez/ctas.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/tez/ctas.q.out Sat Aug 30 06:44:46 2014 @@ -91,7 +91,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: nzhang_CTAS1 + name: default.nzhang_CTAS1 Stage: Stage-3 Stats-Aggr Operator @@ -105,9 +105,12 @@ STAGE PLANS: PREHOOK: query: create table nzhang_CTAS1 as select key k, value from src sort by k, value limit 10 PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@nzhang_CTAS1 POSTHOOK: query: create table nzhang_CTAS1 as select key k, value from src sort by k, value limit 10 POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@nzhang_CTAS1 PREHOOK: query: select * from nzhang_CTAS1 PREHOOK: type: QUERY @@ -236,7 +239,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: nzhang_ctas2 + name: default.nzhang_ctas2 Stage: Stage-3 Stats-Aggr Operator @@ -250,9 +253,12 @@ STAGE PLANS: PREHOOK: query: create table nzhang_ctas2 as select * from src sort by key, value limit 10 PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@nzhang_ctas2 POSTHOOK: query: create table nzhang_ctas2 as select * from src sort by key, value limit 10 POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@nzhang_ctas2 PREHOOK: query: select * from nzhang_ctas2 PREHOOK: type: QUERY @@ -381,7 +387,7 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat serde name: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: nzhang_ctas3 + name: default.nzhang_ctas3 Stage: Stage-3 Stats-Aggr Operator @@ -395,9 +401,12 @@ STAGE PLANS: PREHOOK: query: create table nzhang_ctas3 row format serde "org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe" stored as RCFile as select key/2 half_key, concat(value, "_con") conb from src sort by half_key, conb limit 10 PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@nzhang_ctas3 POSTHOOK: query: create table nzhang_ctas3 row format serde "org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe" stored as RCFile as select key/2 half_key, concat(value, "_con") conb from src sort by half_key, conb limit 10 POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@nzhang_ctas3 PREHOOK: query: select * from nzhang_ctas3 PREHOOK: type: QUERY @@ -592,7 +601,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: nzhang_ctas4 + name: default.nzhang_ctas4 Stage: Stage-3 Stats-Aggr Operator @@ -606,9 +615,12 @@ STAGE PLANS: PREHOOK: query: create table nzhang_ctas4 row format delimited fields terminated by ',' stored as textfile as select key, value from src sort by key, value limit 10 PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@nzhang_ctas4 POSTHOOK: query: create table nzhang_ctas4 row format delimited fields terminated by ',' stored as textfile as select key, value from src sort by key, value limit 10 POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@nzhang_ctas4 PREHOOK: query: select * from nzhang_ctas4 PREHOOK: type: QUERY @@ -854,7 +866,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: nzhang_ctas5 + name: default.nzhang_ctas5 Stage: Stage-3 Stats-Aggr Operator @@ -869,9 +881,12 @@ STAGE PLANS: PREHOOK: query: create table nzhang_ctas5 row format delimited fields terminated by ',' lines terminated by '\012' stored as textfile as select key, value from src sort by key, value limit 10 PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@nzhang_ctas5 POSTHOOK: query: create table nzhang_ctas5 row format delimited fields terminated by ',' lines terminated by '\012' stored as textfile as select key, value from src sort by key, value limit 10 POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@nzhang_ctas5 PREHOOK: query: create table nzhang_ctas6 (key string, `to` string) PREHOOK: type: CREATETABLE @@ -894,7 +909,10 @@ POSTHOOK: Lineage: nzhang_ctas6.to SIMPL PREHOOK: query: create table nzhang_ctas7 as select key, `to` from nzhang_ctas6 PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@nzhang_ctas6 +PREHOOK: Output: database:default +PREHOOK: Output: default@nzhang_ctas7 POSTHOOK: query: create table nzhang_ctas7 as select key, `to` from nzhang_ctas6 POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@nzhang_ctas6 +POSTHOOK: Output: database:default POSTHOOK: Output: default@nzhang_ctas7 Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out Sat Aug 30 06:44:46 2014 @@ -14,6 +14,7 @@ PREHOOK: query: create table over1k( fields terminated by '|' PREHOOK: type: CREATETABLE PREHOOK: Output: database:default +PREHOOK: Output: default@over1k POSTHOOK: query: create table over1k( t tinyint, si smallint, @@ -42,6 +43,7 @@ POSTHOOK: Output: default@over1k PREHOOK: query: create table over1k_orc like over1k PREHOOK: type: CREATETABLE PREHOOK: Output: database:default +PREHOOK: Output: default@over1k_orc POSTHOOK: query: create table over1k_orc like over1k POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default @@ -81,6 +83,7 @@ PREHOOK: query: create table over1k_part partitioned by (ds string, t tinyint) stored as orc PREHOOK: type: CREATETABLE PREHOOK: Output: database:default +PREHOOK: Output: default@over1k_part_orc POSTHOOK: query: create table over1k_part_orc( si smallint, i int, @@ -93,6 +96,7 @@ POSTHOOK: Output: default@over1k_part_or PREHOOK: query: create table over1k_part_limit_orc like over1k_part_orc PREHOOK: type: CREATETABLE PREHOOK: Output: database:default +PREHOOK: Output: default@over1k_part_limit_orc POSTHOOK: query: create table over1k_part_limit_orc like over1k_part_orc POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default @@ -114,6 +118,7 @@ PREHOOK: query: create table over1k_part clustered by (si) into 4 buckets stored as orc PREHOOK: type: CREATETABLE PREHOOK: Output: database:default +PREHOOK: Output: default@over1k_part_buck_orc POSTHOOK: query: create table over1k_part_buck_orc( si smallint, i int, @@ -134,6 +139,7 @@ PREHOOK: query: create table over1k_part sorted by (f) into 4 buckets stored as orc PREHOOK: type: CREATETABLE PREHOOK: Output: database:default +PREHOOK: Output: default@over1k_part_buck_sort_orc POSTHOOK: query: create table over1k_part_buck_sort_orc( si smallint, i int, @@ -195,6 +201,7 @@ STAGE PLANS: Map-reduce partition columns: _col4 (type: tinyint) Statistics: Num rows: 1048 Data size: 310873 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint) + Execution mode: vectorized Reducer 3 Reduce Operator Tree: Extract @@ -280,6 +287,7 @@ STAGE PLANS: Map-reduce partition columns: _col4 (type: tinyint) Statistics: Num rows: 10 Data size: 2960 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint) + Execution mode: vectorized Reducer 3 Reduce Operator Tree: Extract @@ -564,6 +572,7 @@ STAGE PLANS: Map-reduce partition columns: _col4 (type: tinyint) Statistics: Num rows: 1048 Data size: 310873 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint) + Execution mode: vectorized Reducer 3 Reduce Operator Tree: Extract @@ -649,6 +658,7 @@ STAGE PLANS: Map-reduce partition columns: _col4 (type: tinyint) Statistics: Num rows: 10 Data size: 2960 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint) + Execution mode: vectorized Reducer 3 Reduce Operator Tree: Extract @@ -1292,6 +1302,7 @@ create table over1k_part2_orc( partitioned by (ds string, t tinyint) PREHOOK: type: CREATETABLE PREHOOK: Output: database:default +PREHOOK: Output: default@over1k_part2_orc POSTHOOK: query: -- tests for HIVE-6883 create table over1k_part2_orc( si smallint, @@ -1419,6 +1430,7 @@ STAGE PLANS: Map-reduce partition columns: _col4 (type: tinyint) Statistics: Num rows: 1048 Data size: 310873 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: smallint), _col1 (type: int), _col2 (type: bigint), _col3 (type: float), _col4 (type: tinyint) + Execution mode: vectorized Reducer 3 Reduce Operator Tree: Extract @@ -1762,6 +1774,7 @@ create table over1k_part_buck_sort2_orc( sorted by (f) into 1 buckets PREHOOK: type: CREATETABLE PREHOOK: Output: database:default +PREHOOK: Output: default@over1k_part_buck_sort2_orc POSTHOOK: query: -- hadoop-1 does not honor number of reducers in local mode. There is always only 1 reducer irrespective of the number of buckets. -- Hence all records go to one bucket and all other buckets will be empty. Similar to HIVE-6867. However, hadoop-2 honors number -- of reducers and records are spread across all reducers. To avoid this inconsistency we will make number of buckets to 1 for this test. Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/insert1.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/insert1.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/tez/insert1.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/tez/insert1.q.out Sat Aug 30 06:44:46 2014 @@ -143,12 +143,11 @@ POSTHOOK: Output: database:x PREHOOK: query: create table x.insert1(key int, value string) stored as textfile PREHOOK: type: CREATETABLE PREHOOK: Output: database:x -PREHOOK: Output: [email protected] +PREHOOK: Output: x@insert1 POSTHOOK: query: create table x.insert1(key int, value string) stored as textfile POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:x POSTHOOK: Output: x@insert1 -POSTHOOK: Output: [email protected] PREHOOK: query: explain insert into table x.INSERT1 select a.key, a.value from insert2 a WHERE (a.key=-1) PREHOOK: type: QUERY POSTHOOK: query: explain insert into table x.INSERT1 select a.key, a.value from insert2 a WHERE (a.key=-1) @@ -406,11 +405,10 @@ POSTHOOK: Output: database:db1 PREHOOK: query: CREATE TABLE db1.result(col1 STRING) PREHOOK: type: CREATETABLE PREHOOK: Output: database:db1 -PREHOOK: Output: [email protected] +PREHOOK: Output: db1@result POSTHOOK: query: CREATE TABLE db1.result(col1 STRING) POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:db1 -POSTHOOK: Output: [email protected] POSTHOOK: Output: db1@result PREHOOK: query: INSERT OVERWRITE TABLE db1.result SELECT 'db1_insert1' FROM src LIMIT 1 PREHOOK: type: QUERY Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/orc_analyze.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/orc_analyze.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/tez/orc_analyze.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/tez/orc_analyze.q.out Sat Aug 30 06:44:46 2014 @@ -105,7 +105,7 @@ Table Parameters: numFiles 1 numRows 100 rawDataSize 52600 - totalSize 3042 + totalSize 3098 #### A masked pattern was here #### # Storage Information @@ -195,7 +195,7 @@ Table Parameters: numFiles 1 numRows 100 rawDataSize 52600 - totalSize 3042 + totalSize 3098 #### A masked pattern was here #### # Storage Information @@ -309,7 +309,7 @@ Partition Parameters: numFiles 1 numRows 50 rawDataSize 21950 - totalSize 1962 + totalSize 2016 #### A masked pattern was here #### # Storage Information @@ -354,7 +354,7 @@ Partition Parameters: numFiles 1 numRows 50 rawDataSize 22050 - totalSize 1981 + totalSize 2036 #### A masked pattern was here #### # Storage Information @@ -456,7 +456,7 @@ Partition Parameters: numFiles 1 numRows 50 rawDataSize 21950 - totalSize 1962 + totalSize 2016 #### A masked pattern was here #### # Storage Information @@ -501,7 +501,7 @@ Partition Parameters: numFiles 1 numRows 50 rawDataSize 22050 - totalSize 1981 + totalSize 2036 #### A masked pattern was here #### # Storage Information @@ -621,7 +621,7 @@ Partition Parameters: numFiles 4 numRows 50 rawDataSize 21980 - totalSize 4746 + totalSize 4955 #### A masked pattern was here #### # Storage Information @@ -666,7 +666,7 @@ Partition Parameters: numFiles 4 numRows 50 rawDataSize 22048 - totalSize 4829 + totalSize 5046 #### A masked pattern was here #### # Storage Information @@ -774,7 +774,7 @@ Partition Parameters: numFiles 4 numRows 50 rawDataSize 21980 - totalSize 4746 + totalSize 4955 #### A masked pattern was here #### # Storage Information @@ -819,7 +819,7 @@ Partition Parameters: numFiles 4 numRows 50 rawDataSize 22048 - totalSize 4829 + totalSize 5046 #### A masked pattern was here #### # Storage Information @@ -984,7 +984,7 @@ Partition Parameters: numFiles 1 numRows 50 rawDataSize 21950 - totalSize 1962 + totalSize 2016 #### A masked pattern was here #### # Storage Information Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/ptf.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/ptf.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/tez/ptf.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/tez/ptf.q.out Sat Aug 30 06:44:46 2014 @@ -772,6 +772,8 @@ from part group by p_mfgr, p_brand PREHOOK: type: CREATEVIEW PREHOOK: Input: default@part +PREHOOK: Output: database:default +PREHOOK: Output: default@mfgr_price_view POSTHOOK: query: -- 16. testViewAsTableInputToPTF create view IF NOT EXISTS mfgr_price_view as select p_mfgr, p_brand, @@ -780,6 +782,7 @@ from part group by p_mfgr, p_brand POSTHOOK: type: CREATEVIEW POSTHOOK: Input: default@part +POSTHOOK: Output: database:default POSTHOOK: Output: default@mfgr_price_view PREHOOK: query: select p_mfgr, p_brand, s, sum(s) over w1 as s1 Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/stats_counter.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/stats_counter.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/tez/stats_counter.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/tez/stats_counter.q.out Sat Aug 30 06:44:46 2014 @@ -2,10 +2,13 @@ PREHOOK: query: -- by analyze create table dummy1 as select * from src PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@dummy1 POSTHOOK: query: -- by analyze create table dummy1 as select * from src POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@dummy1 PREHOOK: query: analyze table dummy1 compute statistics PREHOOK: type: QUERY @@ -55,10 +58,13 @@ PREHOOK: query: -- by autogather create table dummy2 as select * from src PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@dummy2 POSTHOOK: query: -- by autogather create table dummy2 as select * from src POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@dummy2 PREHOOK: query: desc formatted dummy2 PREHOOK: type: DESCTABLE Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/subquery_exists.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/subquery_exists.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/tez/subquery_exists.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/tez/subquery_exists.q.out Sat Aug 30 06:44:46 2014 @@ -133,6 +133,8 @@ where exists where b.value = a.value and a.key = b.key and a.value > 'val_9') PREHOOK: type: CREATEVIEW PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@cv1 POSTHOOK: query: -- view test create view cv1 as select * @@ -143,6 +145,7 @@ where exists where b.value = a.value and a.key = b.key and a.value > 'val_9') POSTHOOK: type: CREATEVIEW POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@cv1 PREHOOK: query: select * from cv1 PREHOOK: type: QUERY Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/temp_table.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/temp_table.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/tez/temp_table.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/tez/temp_table.q.out Sat Aug 30 06:44:46 2014 @@ -46,7 +46,7 @@ STAGE PLANS: #### A masked pattern was here #### output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: foo + name: default.foo isTemporary: true Stage: Stage-3 @@ -61,9 +61,12 @@ STAGE PLANS: PREHOOK: query: CREATE TEMPORARY TABLE foo AS SELECT * FROM src WHERE key % 2 = 0 PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@foo POSTHOOK: query: CREATE TEMPORARY TABLE foo AS SELECT * FROM src WHERE key % 2 = 0 POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@foo PREHOOK: query: EXPLAIN CREATE TEMPORARY TABLE bar AS SELECT * FROM src WHERE key % 2 = 1 PREHOOK: type: CREATETABLE_AS_SELECT @@ -113,7 +116,7 @@ STAGE PLANS: #### A masked pattern was here #### output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: bar + name: default.bar isTemporary: true Stage: Stage-3 @@ -128,9 +131,12 @@ STAGE PLANS: PREHOOK: query: CREATE TEMPORARY TABLE bar AS SELECT * FROM src WHERE key % 2 = 1 PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@bar POSTHOOK: query: CREATE TEMPORARY TABLE bar AS SELECT * FROM src WHERE key % 2 = 1 POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@bar PREHOOK: query: DESCRIBE foo PREHOOK: type: DESCTABLE @@ -378,9 +384,12 @@ POSTHOOK: type: SHOWTABLES PREHOOK: query: CREATE TEMPORARY TABLE foo AS SELECT * FROM default.foo PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@foo +PREHOOK: Output: database:two +PREHOOK: Output: two@foo POSTHOOK: query: CREATE TEMPORARY TABLE foo AS SELECT * FROM default.foo POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@foo +POSTHOOK: Output: database:two POSTHOOK: Output: two@foo PREHOOK: query: SHOW TABLES PREHOOK: type: SHOWTABLES Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/tez_dml.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/tez_dml.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/tez/tez_dml.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/tez/tez_dml.q.out Sat Aug 30 06:44:46 2014 @@ -82,7 +82,7 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: tmp_src + name: default.tmp_src Stage: Stage-3 Stats-Aggr Operator @@ -96,9 +96,12 @@ STAGE PLANS: PREHOOK: query: CREATE TABLE tmp_src AS SELECT * FROM (SELECT value, count(value) AS cnt FROM src GROUP BY value) f1 ORDER BY cnt PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@tmp_src POSTHOOK: query: CREATE TABLE tmp_src AS SELECT * FROM (SELECT value, count(value) AS cnt FROM src GROUP BY value) f1 ORDER BY cnt POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@tmp_src PREHOOK: query: SELECT * FROM tmp_src PREHOOK: type: QUERY @@ -1478,10 +1481,13 @@ PREHOOK: query: -- create empty table CREATE TABLE empty STORED AS orc AS SELECT * FROM tmp_src_part WHERE d = -1000 PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@tmp_src_part +PREHOOK: Output: database:default +PREHOOK: Output: default@empty POSTHOOK: query: -- create empty table CREATE TABLE empty STORED AS orc AS SELECT * FROM tmp_src_part WHERE d = -1000 POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@tmp_src_part +POSTHOOK: Output: database:default POSTHOOK: Output: default@empty PREHOOK: query: SELECT * FROM empty PREHOOK: type: QUERY Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/tez_union.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/tez_union.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/tez/tez_union.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/tez/tez_union.q.out Sat Aug 30 06:44:46 2014 @@ -92,12 +92,15 @@ UNION ALL select s2.key as key, s2.value as value from src s2 PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@ut POSTHOOK: query: create table ut as select s1.key as key, s1.value as value from src s1 join src s3 on s1.key=s3.key UNION ALL select s2.key as key, s2.value as value from src s2 POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@ut PREHOOK: query: select * from ut order by key, value limit 20 PREHOOK: type: QUERY @@ -268,12 +271,15 @@ select count(*) as cnt from (select u1.k u as u1 join u as u2 on (u1.key = u2.key)) a PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@ut POSTHOOK: query: create table ut as with u as (select * from src union all select * from src) select count(*) as cnt from (select u1.key as k1, u2.key as k2 from u as u1 join u as u2 on (u1.key = u2.key)) a POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@ut PREHOOK: query: select * from ut order by cnt limit 20 PREHOOK: type: QUERY @@ -401,12 +407,15 @@ src s1 join (select * from src union all select * from src) u1 on s1.key = u1.key PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@ut POSTHOOK: query: create table ut as select s1.key as skey, u1.key as ukey from src s1 join (select * from src union all select * from src) u1 on s1.key = u1.key POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@ut PREHOOK: query: select * from ut order by skey, ukey limit 20 PREHOOK: type: QUERY @@ -708,6 +717,8 @@ join src s8 on (u1.key = s8.key) order by lkey PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@ut POSTHOOK: query: create table ut as select s1.key as skey, u1.key as ukey, s8.key as lkey from src s1 @@ -718,6 +729,7 @@ join src s8 on (u1.key = s8.key) order by lkey POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@ut PREHOOK: query: select * from ut order by skey, ukey, lkey limit 100 PREHOOK: type: QUERY @@ -951,11 +963,14 @@ select s2.key as key from src s2 join sr union all select s4.key from src s4 join src s5 on s4.key = s5.key PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@ut POSTHOOK: query: create table ut as select s2.key as key from src s2 join src s3 on s2.key = s3.key union all select s4.key from src s4 join src s5 on s4.key = s5.key POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@ut PREHOOK: query: select * from ut order by key limit 30 PREHOOK: type: QUERY @@ -1190,12 +1205,15 @@ select u.key as ukey, s.key as skey from right outer join src s on u.key = s.key PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@ut POSTHOOK: query: create table ut as select u.key as ukey, s.key as skey from (select * from src union all select * from src) u right outer join src s on u.key = s.key POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src +POSTHOOK: Output: database:default POSTHOOK: Output: default@ut PREHOOK: query: select * from ut order by ukey, skey limit 20 PREHOOK: type: QUERY Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/vector_decimal_aggregate.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/vector_decimal_aggregate.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/tez/vector_decimal_aggregate.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/tez/vector_decimal_aggregate.q.out Sat Aug 30 06:44:46 2014 @@ -5,6 +5,8 @@ PREHOOK: query: CREATE TABLE decimal_vgb FROM alltypesorc PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@alltypesorc +PREHOOK: Output: database:default +PREHOOK: Output: default@decimal_vgby POSTHOOK: query: CREATE TABLE decimal_vgby STORED AS ORC AS SELECT cdouble, CAST (((cdouble*22.1)/37) AS DECIMAL(20,10)) AS cdecimal1, CAST (((cdouble*9.3)/13) AS DECIMAL(23,14)) AS cdecimal2, @@ -12,6 +14,7 @@ POSTHOOK: query: CREATE TABLE decimal_vg FROM alltypesorc POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@alltypesorc +POSTHOOK: Output: database:default POSTHOOK: Output: default@decimal_vgby PREHOOK: query: EXPLAIN SELECT cint, COUNT(cdecimal1), MAX(cdecimal1), MIN(cdecimal1), SUM(cdecimal1), AVG(cdecimal1), STDDEV_POP(cdecimal1), STDDEV_SAMP(cdecimal1), Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/vectorization_part_project.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/vectorization_part_project.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/tez/vectorization_part_project.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/tez/vectorization_part_project.q.out Sat Aug 30 06:44:46 2014 @@ -65,28 +65,28 @@ STAGE PLANS: Map Operator Tree: TableScan alias: alltypesorc_part - Statistics: Num rows: 200 Data size: 4068 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 200 Data size: 41576 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: (cdouble + 2) (type: double) outputColumnNames: _col0 - Statistics: Num rows: 200 Data size: 4068 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 200 Data size: 41576 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: double) sort order: + - Statistics: Num rows: 200 Data size: 4068 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 200 Data size: 41576 Basic stats: COMPLETE Column stats: NONE Execution mode: vectorized Reducer 2 Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: double) outputColumnNames: _col0 - Statistics: Num rows: 200 Data size: 4068 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 200 Data size: 41576 Basic stats: COMPLETE Column stats: NONE Limit Number of rows: 10 - Statistics: Num rows: 10 Data size: 200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 10 Data size: 2070 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 200 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 10 Data size: 2070 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Modified: hive/branches/spark/ql/src/test/results/clientpositive/tez/vectorized_ptf.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/tez/vectorized_ptf.q.out?rev=1621416&r1=1621415&r2=1621416&view=diff ============================================================================== Files hive/branches/spark/ql/src/test/results/clientpositive/tez/vectorized_ptf.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/tez/vectorized_ptf.q.out Sat Aug 30 06:44:46 2014 differ
