http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/auto_join_filters.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/beelinepositive/auto_join_filters.q.out b/ql/src/test/results/beelinepositive/auto_join_filters.q.out deleted file mode 100644 index a1573c2..0000000 --- a/ql/src/test/results/beelinepositive/auto_join_filters.q.out +++ /dev/null @@ -1,254 +0,0 @@ -Saving all output to "!!{outputDirectory}!!/auto_join_filters.q.raw". Enter "record" with no arguments to stop it. ->>> !run !!{qFileDirectory}!!/auto_join_filters.q ->>> set hive.auto.convert.join = true; -No rows affected ->>> ->>> CREATE TABLE myinput1(key int, value int); -No rows affected ->>> LOAD DATA LOCAL INPATH '../data/files/in3.txt' INTO TABLE myinput1; -No rows affected ->>> ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a JOIN myinput1 b on a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b on a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'4937935' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a RIGHT OUTER JOIN myinput1 b on a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3080335' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a FULL OUTER JOIN myinput1 b on a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'19749880' -1 row selected ->>> ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a JOIN myinput1 b ON a.key = b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a JOIN myinput1 b ON a.key = b.key AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a JOIN myinput1 b ON a.value = b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a JOIN myinput1 b ON a.value = b.value and a.key=b.key AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b ON a.key = b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'4937935' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b ON a.value = b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'4937935' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b ON a.key = b.key AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'4937935' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b ON a.key = b.key and a.value=b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'4937935' -1 row selected ->>> ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a RIGHT OUTER JOIN myinput1 b ON a.key = b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3080335' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a RIGHT OUTER JOIN myinput1 b ON a.key = b.key AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3080335' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a RIGHT OUTER JOIN myinput1 b ON a.value = b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3080335' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a RIGHT OUTER JOIN myinput1 b ON a.key=b.key and a.value = b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3080335' -1 row selected ->>> ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a FULL OUTER JOIN myinput1 b ON a.key = b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'4939870' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a FULL OUTER JOIN myinput1 b ON a.key = b.key AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'4939870' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a FULL OUTER JOIN myinput1 b ON a.value = b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'4939870' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a FULL OUTER JOIN myinput1 b ON a.value = b.value and a.key=b.key AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'4939870' -1 row selected ->>> ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) from myinput1 a LEFT OUTER JOIN myinput1 b ON (a.value=b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value) RIGHT OUTER JOIN myinput1 c ON (b.value=c.value AND c.key > 40 AND c.value > 50 AND c.key = c.value AND b.key > 40 AND b.value > 50 AND b.key = b.value); -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) from myinput1 a RIGHT OUTER JOIN myinput1 b ON (a.value=b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value) LEFT OUTER JOIN myinput1 c ON (b.value=c.value AND c.key > 40 AND c.value > 50 AND c.key = c.value AND b.key > 40 AND b.value > 50 AND b.key = b.value); -'_c0' -'3080335' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b RIGHT OUTER JOIN myinput1 c ON a.value = b.value and b.value = c.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value AND c.key > 40 AND c.value > 50 AND c.key = c.value; -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) from myinput1 a LEFT OUTER JOIN myinput1 b ON (a.value=b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value) RIGHT OUTER JOIN myinput1 c ON (b.key=c.key AND c.key > 40 AND c.value > 50 AND c.key = c.value AND b.key > 40 AND b.value > 50 AND b.key = b.value); -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) from myinput1 a RIGHT OUTER JOIN myinput1 b ON (a.value=b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value) LEFT OUTER JOIN myinput1 c ON (b.key=c.key AND c.key > 40 AND c.value > 50 AND c.key = c.value AND b.key > 40 AND b.value > 50 AND b.key = b.value); -'_c0' -'3080335' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b RIGHT OUTER JOIN myinput1 c ON a.value = b.value and b.key = c.key AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value AND c.key > 40 AND c.value > 50 AND c.key = c.value; -'_c0' -'3078400' -1 row selected ->>> ->>> ->>> CREATE TABLE smb_input1(key int, value int) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS; -No rows affected ->>> CREATE TABLE smb_input2(key int, value int) CLUSTERED BY (value) SORTED BY (value) INTO 2 BUCKETS; -No rows affected ->>> LOAD DATA LOCAL INPATH '../data/files/in1.txt' into table smb_input1; -No rows affected ->>> LOAD DATA LOCAL INPATH '../data/files/in2.txt' into table smb_input1; -No rows affected ->>> LOAD DATA LOCAL INPATH '../data/files/in1.txt' into table smb_input2; -No rows affected ->>> LOAD DATA LOCAL INPATH '../data/files/in2.txt' into table smb_input2; -No rows affected ->>> ->>> SET hive.optimize.bucketmapjoin = true; -No rows affected ->>> SET hive.optimize.bucketmapjoin.sortedmerge = true; -No rows affected ->>> SET hive.input.format = org.apache.hadoop.hive.ql.io.BucketizedHiveInputFormat; -No rows affected ->>> ->>> SET hive.outerjoin.supports.filters = false; -No rows affected ->>> ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a JOIN myinput1 b on a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b on a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a RIGHT OUTER JOIN myinput1 b on a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a FULL OUTER JOIN myinput1 b on a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a JOIN myinput1 b ON a.key = b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a JOIN myinput1 b ON a.key = b.key AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a JOIN myinput1 b ON a.value = b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a JOIN myinput1 b ON a.value = b.value and a.key=b.key AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b ON a.key = b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b ON a.value = b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b ON a.key = b.key AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b ON a.key = b.key and a.value=b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a RIGHT OUTER JOIN myinput1 b ON a.key = b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a RIGHT OUTER JOIN myinput1 b ON a.key = b.key AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a RIGHT OUTER JOIN myinput1 b ON a.value = b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a RIGHT OUTER JOIN myinput1 b ON a.key=b.key and a.value = b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a FULL OUTER JOIN myinput1 b ON a.key = b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a FULL OUTER JOIN myinput1 b ON a.key = b.key AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a FULL OUTER JOIN myinput1 b ON a.value = b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a FULL OUTER JOIN myinput1 b ON a.value = b.value and a.key=b.key AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value; -'_c0' -'3078400' -1 row selected ->>> ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) from myinput1 a LEFT OUTER JOIN myinput1 b ON (a.value=b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value) RIGHT OUTER JOIN myinput1 c ON (b.value=c.value AND c.key > 40 AND c.value > 50 AND c.key = c.value AND b.key > 40 AND b.value > 50 AND b.key = b.value); -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) from myinput1 a RIGHT OUTER JOIN myinput1 b ON (a.value=b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value) LEFT OUTER JOIN myinput1 c ON (b.value=c.value AND c.key > 40 AND c.value > 50 AND c.key = c.value AND b.key > 40 AND b.value > 50 AND b.key = b.value); -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b RIGHT OUTER JOIN myinput1 c ON a.value = b.value and b.value = c.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value AND c.key > 40 AND c.value > 50 AND c.key = c.value; -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) from myinput1 a LEFT OUTER JOIN myinput1 b ON (a.value=b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value) RIGHT OUTER JOIN myinput1 c ON (b.key=c.key AND c.key > 40 AND c.value > 50 AND c.key = c.value AND b.key > 40 AND b.value > 50 AND b.key = b.value); -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) from myinput1 a RIGHT OUTER JOIN myinput1 b ON (a.value=b.value AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value) LEFT OUTER JOIN myinput1 c ON (b.key=c.key AND c.key > 40 AND c.value > 50 AND c.key = c.value AND b.key > 40 AND b.value > 50 AND b.key = b.value); -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b RIGHT OUTER JOIN myinput1 c ON a.value = b.value and b.key = c.key AND a.key > 40 AND a.value > 50 AND a.key = a.value AND b.key > 40 AND b.value > 50 AND b.key = b.value AND c.key > 40 AND c.value > 50 AND c.key = c.value; -'_c0' -'3078400' -1 row selected ->>> !record
http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/auto_join_nulls.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/beelinepositive/auto_join_nulls.q.out b/ql/src/test/results/beelinepositive/auto_join_nulls.q.out deleted file mode 100644 index 822fe42..0000000 --- a/ql/src/test/results/beelinepositive/auto_join_nulls.q.out +++ /dev/null @@ -1,101 +0,0 @@ -Saving all output to "!!{outputDirectory}!!/auto_join_nulls.q.raw". Enter "record" with no arguments to stop it. ->>> !run !!{qFileDirectory}!!/auto_join_nulls.q ->>> set hive.auto.convert.join = true; -No rows affected ->>> ->>> CREATE TABLE myinput1(key int, value int); -No rows affected ->>> LOAD DATA LOCAL INPATH '../data/files/in1.txt' INTO TABLE myinput1; -No rows affected ->>> ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a JOIN myinput1 b; -'_c0' -'13630578' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b; -'_c0' -'13630578' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a RIGHT OUTER JOIN myinput1 b; -'_c0' -'13630578' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a JOIN myinput1 b ON a.key = b.value; -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a JOIN myinput1 b ON a.key = b.key; -'_c0' -'4509856' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a JOIN myinput1 b ON a.value = b.value; -'_c0' -'3112070' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a JOIN myinput1 b ON a.value = b.value and a.key=b.key; -'_c0' -'3078400' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b ON a.key = b.value; -'_c0' -'4542003' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b ON a.value = b.value; -'_c0' -'4542038' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b ON a.key = b.key; -'_c0' -'4543491' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b ON a.key = b.key and a.value=b.value; -'_c0' -'4542003' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a RIGHT OUTER JOIN myinput1 b ON a.key = b.value; -'_c0' -'3079923' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a RIGHT OUTER JOIN myinput1 b ON a.key = b.key; -'_c0' -'4509891' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a RIGHT OUTER JOIN myinput1 b ON a.value = b.value; -'_c0' -'3113558' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a RIGHT OUTER JOIN myinput1 b ON a.key=b.key and a.value = b.value; -'_c0' -'3079923' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a FULL OUTER JOIN myinput1 b ON a.key = b.value; -'_c0' -'4543526' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a FULL OUTER JOIN myinput1 b ON a.key = b.key; -'_c0' -'4543526' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a FULL OUTER JOIN myinput1 b ON a.value = b.value; -'_c0' -'4543526' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a FULL OUTER JOIN myinput1 b ON a.value = b.value and a.key=b.key; -'_c0' -'4543526' -1 row selected ->>> ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) from myinput1 a LEFT OUTER JOIN myinput1 b ON (a.value=b.value) RIGHT OUTER JOIN myinput1 c ON (b.value=c.value); -'_c0' -'3112070' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) from myinput1 a RIGHT OUTER JOIN myinput1 b ON (a.value=b.value) LEFT OUTER JOIN myinput1 c ON (b.value=c.value); -'_c0' -'3113558' -1 row selected ->>> SELECT sum(hash(a.key,a.value,b.key,b.value)) FROM myinput1 a LEFT OUTER JOIN myinput1 b RIGHT OUTER JOIN myinput1 c ON a.value = b.value and b.value = c.value; -'_c0' -'3112070' -1 row selected ->>> ->>> !record http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/autogen_colalias.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/beelinepositive/autogen_colalias.q.out b/ql/src/test/results/beelinepositive/autogen_colalias.q.out deleted file mode 100644 index 5c537bd..0000000 --- a/ql/src/test/results/beelinepositive/autogen_colalias.q.out +++ /dev/null @@ -1,70 +0,0 @@ -Saving all output to "!!{outputDirectory}!!/autogen_colalias.q.raw". Enter "record" with no arguments to stop it. ->>> !run !!{qFileDirectory}!!/autogen_colalias.q ->>> CREATE TEMPORARY FUNCTION test_max AS 'org.apache.hadoop.hive.ql.udf.UDAFTestMax'; -No rows affected ->>> ->>> create table dest_grouped_old1 as select 1+1, 2+2 as zz, src.key, test_max(length(src.value)), count(src.value), sin(count(src.value)), count(sin(src.value)), unix_timestamp(), CAST(SUM(IF(value > 10, value, 1)) AS INT), if(src.key > 1, -1, -0) -from src group by src.key; -'_c0','zz','key','_c3','_c4','_c5','_c6','_c7','_c8','_c9' -No rows selected ->>> describe dest_grouped_old1; -'col_name','data_type','comment' -'_c0','int','' -'zz','int','' -'key','string','' -'_c3','int','' -'_c4','bigint','' -'_c5','double','' -'_c6','bigint','' -'_c7','bigint','' -'_c8','int','' -'_c9','int','' -10 rows selected ->>> ->>> create table dest_grouped_old2 as select distinct src.key from src; -'key' -No rows selected ->>> describe dest_grouped_old2; -'col_name','data_type','comment' -'key','string','' -1 row selected ->>> ->>> set hive.autogen.columnalias.prefix.label=column_; -No rows affected ->>> set hive.autogen.columnalias.prefix.includefuncname=true; -No rows affected ->>> ->>> create table dest_grouped_new1 as select 1+1, 2+2 as zz, ((src.key % 2)+2)/2, test_max(length(src.value)), count(src.value), sin(count(src.value)), count(sin(src.value)), unix_timestamp(), CAST(SUM(IF(value > 10, value, 1)) AS INT), if(src.key > 10, -(src.key +5) % 2, -0) -from src group by src.key; -'column_0','zz','column_2','test_max_length_src__3','count_src_value_4','sin_count_src_value_5','count_sin_src_value_6','unix_timestamp_7','sum_if_value_10_valu_8','if_src_key_10_src_ke_9' -No rows selected ->>> describe dest_grouped_new1; -'col_name','data_type','comment' -'column_0','int','' -'zz','int','' -'column_2','double','' -'test_max_length_src__3','int','' -'count_src_value_4','bigint','' -'sin_count_src_value_5','double','' -'count_sin_src_value_6','bigint','' -'unix_timestamp_7','bigint','' -'sum_if_value_10_valu_8','int','' -'if_src_key_10_src_ke_9','double','' -10 rows selected ->>> ->>> create table dest_grouped_new2 as select distinct src.key from src; -'key' -No rows selected ->>> describe dest_grouped_new2; -'col_name','data_type','comment' -'key','string','' -1 row selected ->>> ->>> -- Drop the temporary function at the end till HIVE-3160 gets fixed ->>> DROP TEMPORARY FUNCTION test_max; -No rows affected ->>> !record http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/avro_change_schema.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/beelinepositive/avro_change_schema.q.out b/ql/src/test/results/beelinepositive/avro_change_schema.q.out deleted file mode 100644 index b89ba4a..0000000 --- a/ql/src/test/results/beelinepositive/avro_change_schema.q.out +++ /dev/null @@ -1,42 +0,0 @@ -Saving all output to "!!{outputDirectory}!!/avro_change_schema.q.raw". Enter "record" with no arguments to stop it. ->>> !run !!{qFileDirectory}!!/avro_change_schema.q ->>> -- verify that we can update the table properties ->>> CREATE TABLE avro2 -ROW FORMAT -SERDE 'org.apache.hadoop.hive.serde2.avro.AvroSerDe' -STORED AS -INPUTFORMAT 'org.apache.hadoop.hive.ql.io.avro.AvroContainerInputFormat' -OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.avro.AvroContainerOutputFormat' -TBLPROPERTIES ('avro.schema.literal'='{ "namespace": "org.apache.hive", -"name": "first_schema", -"type": "record", -"fields": [ -{ "name":"string1", "type":"string" }, -{ "name":"string2", "type":"string" } -] }'); -No rows affected ->>> ->>> DESCRIBE avro2; -'col_name','data_type','comment' -'string1','string','from deserializer' -'string2','string','from deserializer' -2 rows selected ->>> ->>> ALTER TABLE avro2 SET TBLPROPERTIES ('avro.schema.literal'='{ "namespace": "org.apache.hive", -"name": "second_schema", -"type": "record", -"fields": [ -{ "name":"int1", "type":"int" }, -{ "name":"float1", "type":"float" }, -{ "name":"double1", "type":"double" } -] }'); -No rows affected ->>> ->>> DESCRIBE avro2; -'col_name','data_type','comment' -'int1','int','from deserializer' -'float1','float','from deserializer' -'double1','double','from deserializer' -3 rows selected ->>> ->>> !record http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/avro_evolved_schemas.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/beelinepositive/avro_evolved_schemas.q.out b/ql/src/test/results/beelinepositive/avro_evolved_schemas.q.out deleted file mode 100644 index bd17fa9..0000000 --- a/ql/src/test/results/beelinepositive/avro_evolved_schemas.q.out +++ /dev/null @@ -1,66 +0,0 @@ -Saving all output to "!!{outputDirectory}!!/avro_evolved_schemas.q.raw". Enter "record" with no arguments to stop it. ->>> !run !!{qFileDirectory}!!/avro_evolved_schemas.q ->>> -- verify that new fields in schema get propagated to table scans ->>> CREATE TABLE doctors_with_new_field -ROW FORMAT -SERDE 'org.apache.hadoop.hive.serde2.avro.AvroSerDe' -STORED AS -INPUTFORMAT 'org.apache.hadoop.hive.ql.io.avro.AvroContainerInputFormat' -OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.avro.AvroContainerOutputFormat' -TBLPROPERTIES ('avro.schema.literal'='{ -"namespace": "testing.hive.avro.serde", -"name": "doctors", -"type": "record", -"fields": [ -{ -"name":"number", -"type":"int", -"doc":"Order of playing the role" -}, -{ -"name":"first_name", -"type":"string", -"doc":"first name of actor playing role" -}, -{ -"name":"last_name", -"type":"string", -"doc":"last name of actor playing role" -}, -{ -"name":"extra_field", -"type":"string", -"doc:":"an extra field not in the original file", -"default":"fishfingers and custard" -} -] -}'); -No rows affected ->>> ->>> DESCRIBE doctors_with_new_field; -'col_name','data_type','comment' -'number','int','from deserializer' -'first_name','string','from deserializer' -'last_name','string','from deserializer' -'extra_field','string','from deserializer' -4 rows selected ->>> ->>> LOAD DATA LOCAL INPATH '../data/files/doctors.avro' INTO TABLE doctors_with_new_field; -No rows affected ->>> ->>> SELECT * FROM doctors_with_new_field ORDER BY first_name; -'number','first_name','last_name','extra_field' -'9','Christopher','Eccleston','fishfingers and custard' -'6','Colin','Baker','fishfingers and custard' -'10','David','Tennant','fishfingers and custard' -'3','Jon','Pertwee','fishfingers and custard' -'11','Matt','Smith','fishfingers and custard' -'2','Patrick','Troughton','fishfingers and custard' -'8','Paul','McGann','fishfingers and custard' -'5','Peter','Davison','fishfingers and custard' -'7','Sylvester','McCoy','fishfingers and custard' -'4','Tom','Baker','fishfingers and custard' -'1','William','Hartnell','fishfingers and custard' -11 rows selected ->>> ->>> !record http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/avro_joins.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/beelinepositive/avro_joins.q.out b/ql/src/test/results/beelinepositive/avro_joins.q.out deleted file mode 100644 index fdced96..0000000 --- a/ql/src/test/results/beelinepositive/avro_joins.q.out +++ /dev/null @@ -1,107 +0,0 @@ -Saving all output to "!!{outputDirectory}!!/avro_joins.q.raw". Enter "record" with no arguments to stop it. ->>> !run !!{qFileDirectory}!!/avro_joins.q ->>> -- verify that new joins bring in correct schemas (including evolved schemas) ->>> ->>> CREATE TABLE doctors4 -ROW FORMAT -SERDE 'org.apache.hadoop.hive.serde2.avro.AvroSerDe' -STORED AS -INPUTFORMAT 'org.apache.hadoop.hive.ql.io.avro.AvroContainerInputFormat' -OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.avro.AvroContainerOutputFormat' -TBLPROPERTIES ('avro.schema.literal'='{ -"namespace": "testing.hive.avro.serde", -"name": "doctors", -"type": "record", -"fields": [ -{ -"name":"number", -"type":"int", -"doc":"Order of playing the role" -}, -{ -"name":"first_name", -"type":"string", -"doc":"first name of actor playing role" -}, -{ -"name":"last_name", -"type":"string", -"doc":"last name of actor playing role" -}, -{ -"name":"extra_field", -"type":"string", -"doc:":"an extra field not in the original file", -"default":"fishfingers and custard" -} -] -}'); -No rows affected ->>> ->>> DESCRIBE doctors4; -'col_name','data_type','comment' -'number','int','from deserializer' -'first_name','string','from deserializer' -'last_name','string','from deserializer' -'extra_field','string','from deserializer' -4 rows selected ->>> ->>> LOAD DATA LOCAL INPATH '../data/files/doctors.avro' INTO TABLE doctors4; -No rows affected ->>> ->>> CREATE TABLE episodes -ROW FORMAT -SERDE 'org.apache.hadoop.hive.serde2.avro.AvroSerDe' -STORED AS -INPUTFORMAT 'org.apache.hadoop.hive.ql.io.avro.AvroContainerInputFormat' -OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.avro.AvroContainerOutputFormat' -TBLPROPERTIES ('avro.schema.literal'='{ -"namespace": "testing.hive.avro.serde", -"name": "episodes", -"type": "record", -"fields": [ -{ -"name":"title", -"type":"string", -"doc":"episode title" -}, -{ -"name":"air_date", -"type":"string", -"doc":"initial date" -}, -{ -"name":"doctor", -"type":"int", -"doc":"main actor playing the Doctor in episode" -} -] -}'); -No rows affected ->>> ->>> DESCRIBE episodes; -'col_name','data_type','comment' -'title','string','from deserializer' -'air_date','string','from deserializer' -'doctor','int','from deserializer' -3 rows selected ->>> ->>> LOAD DATA LOCAL INPATH '../data/files/episodes.avro' INTO TABLE episodes; -No rows affected ->>> ->>> SELECT e.title, e.air_date, d.first_name, d.last_name, d.extra_field, e.air_date -FROM doctors4 d JOIN episodes e ON (d.number=e.doctor) -ORDER BY d.last_name, e.title; -'title','air_date','first_name','last_name','extra_field','air_date' -'Horror of Fang Rock','3 September 1977','Tom','Baker','fishfingers and custard','3 September 1977' -'The Mysterious Planet','6 September 1986','Colin','Baker','fishfingers and custard','6 September 1986' -'Castrolava','4 January 1982','Peter','Davison','fishfingers and custard','4 January 1982' -'Rose','26 March 2005','Christopher','Eccleston','fishfingers and custard','26 March 2005' -'An Unearthly Child','23 November 1963','William','Hartnell','fishfingers and custard','23 November 1963' -'The Doctor's Wife','14 May 2011','Matt','Smith','fishfingers and custard','14 May 2011' -'The Eleventh Hour','3 April 2010','Matt','Smith','fishfingers and custard','3 April 2010' -'The Power of the Daleks','5 November 1966','Patrick','Troughton','fishfingers and custard','5 November 1966' -8 rows selected ->>> ->>> ->>> !record http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/avro_sanity_test.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/beelinepositive/avro_sanity_test.q.out b/ql/src/test/results/beelinepositive/avro_sanity_test.q.out deleted file mode 100644 index 2dd8d7b..0000000 --- a/ql/src/test/results/beelinepositive/avro_sanity_test.q.out +++ /dev/null @@ -1,59 +0,0 @@ -Saving all output to "!!{outputDirectory}!!/avro_sanity_test.q.raw". Enter "record" with no arguments to stop it. ->>> !run !!{qFileDirectory}!!/avro_sanity_test.q ->>> -- verify that we can actually read avro files ->>> CREATE TABLE doctors -ROW FORMAT -SERDE 'org.apache.hadoop.hive.serde2.avro.AvroSerDe' -STORED AS -INPUTFORMAT 'org.apache.hadoop.hive.ql.io.avro.AvroContainerInputFormat' -OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.avro.AvroContainerOutputFormat' -TBLPROPERTIES ('avro.schema.literal'='{ -"namespace": "testing.hive.avro.serde", -"name": "doctors", -"type": "record", -"fields": [ -{ -"name":"number", -"type":"int", -"doc":"Order of playing the role" -}, -{ -"name":"first_name", -"type":"string", -"doc":"first name of actor playing role" -}, -{ -"name":"last_name", -"type":"string", -"doc":"last name of actor playing role" -} -] -}'); -No rows affected ->>> ->>> DESCRIBE doctors; -'col_name','data_type','comment' -'number','int','from deserializer' -'first_name','string','from deserializer' -'last_name','string','from deserializer' -3 rows selected ->>> ->>> LOAD DATA LOCAL INPATH '../data/files/doctors.avro' INTO TABLE doctors; -No rows affected ->>> ->>> SELECT * FROM doctors ORDER BY number; -'number','first_name','last_name' -'1','William','Hartnell' -'2','Patrick','Troughton' -'3','Jon','Pertwee' -'4','Tom','Baker' -'5','Peter','Davison' -'6','Colin','Baker' -'7','Sylvester','McCoy' -'8','Paul','McGann' -'9','Christopher','Eccleston' -'10','David','Tennant' -'11','Matt','Smith' -11 rows selected ->>> ->>> !record http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/avro_schema_literal.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/beelinepositive/avro_schema_literal.q.out b/ql/src/test/results/beelinepositive/avro_schema_literal.q.out deleted file mode 100644 index 45ae129..0000000 --- a/ql/src/test/results/beelinepositive/avro_schema_literal.q.out +++ /dev/null @@ -1,54 +0,0 @@ -Saving all output to "!!{outputDirectory}!!/avro_schema_literal.q.raw". Enter "record" with no arguments to stop it. ->>> !run !!{qFileDirectory}!!/avro_schema_literal.q ->>> CREATE TABLE avro1 -ROW FORMAT -SERDE 'org.apache.hadoop.hive.serde2.avro.AvroSerDe' -STORED AS -INPUTFORMAT 'org.apache.hadoop.hive.ql.io.avro.AvroContainerInputFormat' -OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.avro.AvroContainerOutputFormat' -TBLPROPERTIES ('avro.schema.literal'='{ -"namespace": "org.apache.hive", -"name": "big_old_schema", -"type": "record", -"fields": [ -{ "name":"string1", "type":"string" }, -{ "name":"int1", "type":"int" }, -{ "name":"tinyint1", "type":"int" }, -{ "name":"smallint1", "type":"int" }, -{ "name":"bigint1", "type":"long" }, -{ "name":"boolean1", "type":"boolean" }, -{ "name":"float1", "type":"float" }, -{ "name":"double1", "type":"double" }, -{ "name":"list1", "type":{"type":"array", "items":"string"} }, -{ "name":"map1", "type":{"type":"map", "values":"int"} }, -{ "name":"struct1", "type":{"type":"record", "name":"struct1_name", "fields": [ -{ "name":"sInt", "type":"int" }, { "name":"sBoolean", "type":"boolean" }, { "name":"sString", "type":"string" } ] } }, -{ "name":"union1", "type":["float", "boolean", "string"] }, -{ "name":"enum1", "type":{"type":"enum", "name":"enum1_values", "symbols":["BLUE","RED", "GREEN"]} }, -{ "name":"nullableint", "type":["int", "null"] }, -{ "name":"bytes1", "type":"bytes" }, -{ "name":"fixed1", "type":{"type":"fixed", "name":"threebytes", "size":3} } -] }'); -No rows affected ->>> ->>> DESCRIBE avro1; -'col_name','data_type','comment' -'string1','string','from deserializer' -'int1','int','from deserializer' -'tinyint1','int','from deserializer' -'smallint1','int','from deserializer' -'bigint1','bigint','from deserializer' -'boolean1','boolean','from deserializer' -'float1','float','from deserializer' -'double1','double','from deserializer' -'list1','array<string>','from deserializer' -'map1','map<string,int>','from deserializer' -'struct1','struct<sint:int,sboolean:boolean,sstring:string>','from deserializer' -'union1','uniontype<float,boolean,string>','from deserializer' -'enum1','string','from deserializer' -'nullableint','int','from deserializer' -'bytes1','array<tinyint>','from deserializer' -'fixed1','array<tinyint>','from deserializer' -16 rows selected ->>> ->>> !record http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/ba_table_union.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/beelinepositive/ba_table_union.q.out b/ql/src/test/results/beelinepositive/ba_table_union.q.out deleted file mode 100644 index b8b9c6a..0000000 --- a/ql/src/test/results/beelinepositive/ba_table_union.q.out +++ /dev/null @@ -1,40 +0,0 @@ -Saving all output to "!!{outputDirectory}!!/ba_table_union.q.raw". Enter "record" with no arguments to stop it. ->>> !run !!{qFileDirectory}!!/ba_table_union.q ->>> drop table ba_test; -No rows affected ->>> ->>> -- this query tests ba_table1.q + nested queries with multiple operations on binary data types + union on binary types ->>> create table ba_test (ba_key binary, ba_val binary) ; -No rows affected ->>> ->>> describe extended ba_test; -'col_name','data_type','comment' -'ba_key','binary','' -'ba_val','binary','' -'','','' -'Detailed Table Information','Table(tableName:ba_test, dbName:ba_table_union, owner:!!{user.name}!!, createTime:!!UNIXTIME!!, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:ba_key, type:binary, comment:null), FieldSchema(name:ba_val, type:binary, comment:null)], location:!!{hive.metastore.warehouse.dir}!!/ba_table_union.db/ba_test, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[], parameters:{transient_lastDdlTime=!!UNIXTIME!!}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)','' -4 rows selected ->>> ->>> from src insert overwrite table ba_test select cast (src.key as binary), cast (src.value as binary); -'_c0','_c1' -No rows selected ->>> ->>> select * from ( select key from src where key < 50 union all select cast(ba_key as string) as key from ba_test limit 50) unioned order by key limit 10; -'key' -'0' -'0' -'0' -'10' -'11' -'12' -'12' -'128' -'145' -'146' -10 rows selected ->>> ->>> drop table ba_test; -No rows affected ->>> ->>> ->>> !record http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/binary_constant.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/beelinepositive/binary_constant.q.out b/ql/src/test/results/beelinepositive/binary_constant.q.out deleted file mode 100644 index 63b3046..0000000 --- a/ql/src/test/results/beelinepositive/binary_constant.q.out +++ /dev/null @@ -1,7 +0,0 @@ -Saving all output to "!!{outputDirectory}!!/binary_constant.q.raw". Enter "record" with no arguments to stop it. ->>> !run !!{qFileDirectory}!!/binary_constant.q ->>> select cast(cast('a' as binary) as string) from src limit 1; -'_c0' -'a' -1 row selected ->>> !record http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/binary_output_format.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/beelinepositive/binary_output_format.q.out b/ql/src/test/results/beelinepositive/binary_output_format.q.out deleted file mode 100644 index 64257ca..0000000 --- a/ql/src/test/results/beelinepositive/binary_output_format.q.out +++ /dev/null @@ -1,859 +0,0 @@ -Saving all output to "!!{outputDirectory}!!/binary_output_format.q.raw". Enter "record" with no arguments to stop it. ->>> !run !!{qFileDirectory}!!/binary_output_format.q ->>> -- Create a table with binary output format ->>> CREATE TABLE dest1(mydata STRING) -ROW FORMAT SERDE -'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -WITH SERDEPROPERTIES ( -'serialization.last.column.takes.rest'='true' -) -STORED AS -INPUTFORMAT 'org.apache.hadoop.mapred.TextInputFormat' -OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat'; -No rows affected ->>> ->>> -- Insert into that table using transform ->>> EXPLAIN EXTENDED -INSERT OVERWRITE TABLE dest1 -SELECT TRANSFORM(*) -USING 'cat' -AS mydata STRING -ROW FORMAT SERDE -'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -WITH SERDEPROPERTIES ( -'serialization.last.column.takes.rest'='true' -) -RECORDREADER 'org.apache.hadoop.hive.ql.exec.BinaryRecordReader' -FROM src; -'Explain' -'ABSTRACT SYNTAX TREE:' -' (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME dest1))) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST TOK_ALLCOLREF) TOK_SERDE TOK_RECORDWRITER 'cat' (TOK_SERDE (TOK_SERDENAME 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' (TOK_TABLEPROPERTIES (TOK_TABLEPROPLIST (TOK_TABLEPROPERTY 'serialization.last.column.takes.rest' 'true'))))) (TOK_RECORDREADER 'org.apache.hadoop.hive.ql.exec.BinaryRecordReader') (TOK_TABCOLLIST (TOK_TABCOL mydata TOK_STRING)))))))' -'' -'STAGE DEPENDENCIES:' -' Stage-1 is a root stage' -' Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5' -' Stage-4' -' Stage-0 depends on stages: Stage-4, Stage-3, Stage-6' -' Stage-2 depends on stages: Stage-0' -' Stage-3' -' Stage-5' -' Stage-6 depends on stages: Stage-5' -'' -'STAGE PLANS:' -' Stage: Stage-1' -' Map Reduce' -' Alias -> Map Operator Tree:' -' src ' -' TableScan' -' alias: src' -' GatherStats: false' -' Select Operator' -' expressions:' -' expr: key' -' type: string' -' expr: value' -' type: string' -' outputColumnNames: _col0, _col1' -' Transform Operator' -' command: cat' -' output info:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' properties:' -' columns _col0' -' columns.types string' -' field.delim 9' -' serialization.format 9' -' serialization.last.column.takes.rest true' -' File Output Operator' -' compressed: false' -' GlobalTableId: 1' -' directory: pfile:!!{hive.exec.scratchdir}!!' -' NumFilesPerFileSink: 1' -' Stats Publishing Key Prefix: pfile:!!{hive.exec.scratchdir}!!' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat' -' properties:' -' bucket_count -1' -' columns mydata' -' columns.types string' -' file.inputformat org.apache.hadoop.mapred.TextInputFormat' -' file.outputformat org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat' -' location !!{hive.metastore.warehouse.dir}!!/binary_output_format.db/dest1' -' name binary_output_format.dest1' -' serialization.ddl struct dest1 { string mydata}' -' serialization.format 1' -' serialization.last.column.takes.rest true' -' serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' transient_lastDdlTime !!UNIXTIME!!' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' name: binary_output_format.dest1' -' TotalFiles: 1' -' GatherStats: true' -' MultiFileSpray: false' -' Needs Tagging: false' -' Path -> Alias:' -' !!{hive.metastore.warehouse.dir}!!/binary_output_format.db/src [src]' -' Path -> Partition:' -' !!{hive.metastore.warehouse.dir}!!/binary_output_format.db/src ' -' Partition' -' base file name: src' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' properties:' -' bucket_count -1' -' columns key,value' -' columns.types string:string' -' file.inputformat org.apache.hadoop.mapred.TextInputFormat' -' file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' location !!{hive.metastore.warehouse.dir}!!/binary_output_format.db/src' -' name binary_output_format.src' -' numFiles 1' -' numPartitions 0' -' numRows 0' -' rawDataSize 0' -' serialization.ddl struct src { string key, string value}' -' serialization.format 1' -' serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' totalSize 5812' -' transient_lastDdlTime !!UNIXTIME!!' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' ' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' properties:' -' bucket_count -1' -' columns key,value' -' columns.types string:string' -' file.inputformat org.apache.hadoop.mapred.TextInputFormat' -' file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' location !!{hive.metastore.warehouse.dir}!!/binary_output_format.db/src' -' name binary_output_format.src' -' numFiles 1' -' numPartitions 0' -' numRows 0' -' rawDataSize 0' -' serialization.ddl struct src { string key, string value}' -' serialization.format 1' -' serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' totalSize 5812' -' transient_lastDdlTime !!UNIXTIME!!' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' name: binary_output_format.src' -' name: binary_output_format.src' -'' -' Stage: Stage-7' -' Conditional Operator' -'' -' Stage: Stage-4' -' Move Operator' -' files:' -' hdfs directory: true' -' source: pfile:!!{hive.exec.scratchdir}!!' -' destination: pfile:!!{hive.exec.scratchdir}!!' -'' -' Stage: Stage-0' -' Move Operator' -' tables:' -' replace: true' -' source: pfile:!!{hive.exec.scratchdir}!!' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat' -' properties:' -' bucket_count -1' -' columns mydata' -' columns.types string' -' file.inputformat org.apache.hadoop.mapred.TextInputFormat' -' file.outputformat org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat' -' location !!{hive.metastore.warehouse.dir}!!/binary_output_format.db/dest1' -' name binary_output_format.dest1' -' serialization.ddl struct dest1 { string mydata}' -' serialization.format 1' -' serialization.last.column.takes.rest true' -' serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' transient_lastDdlTime !!UNIXTIME!!' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' name: binary_output_format.dest1' -' tmp directory: pfile:!!{hive.exec.scratchdir}!!' -'' -' Stage: Stage-2' -' Stats-Aggr Operator' -' Stats Aggregation Key Prefix: pfile:!!{hive.exec.scratchdir}!!' -'' -' Stage: Stage-3' -' Map Reduce' -' Alias -> Map Operator Tree:' -' pfile:!!{hive.exec.scratchdir}!! ' -' File Output Operator' -' compressed: false' -' GlobalTableId: 0' -' directory: pfile:!!{hive.exec.scratchdir}!!' -' NumFilesPerFileSink: 1' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat' -' properties:' -' bucket_count -1' -' columns mydata' -' columns.types string' -' file.inputformat org.apache.hadoop.mapred.TextInputFormat' -' file.outputformat org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat' -' location !!{hive.metastore.warehouse.dir}!!/binary_output_format.db/dest1' -' name binary_output_format.dest1' -' serialization.ddl struct dest1 { string mydata}' -' serialization.format 1' -' serialization.last.column.takes.rest true' -' serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' transient_lastDdlTime !!UNIXTIME!!' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' name: binary_output_format.dest1' -' TotalFiles: 1' -' GatherStats: false' -' MultiFileSpray: false' -' Needs Tagging: false' -' Path -> Alias:' -' pfile:!!{hive.exec.scratchdir}!! [pfile:!!{hive.exec.scratchdir}!!]' -' Path -> Partition:' -' pfile:!!{hive.exec.scratchdir}!! ' -' Partition' -' base file name: -ext-10002' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat' -' properties:' -' bucket_count -1' -' columns mydata' -' columns.types string' -' file.inputformat org.apache.hadoop.mapred.TextInputFormat' -' file.outputformat org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat' -' location !!{hive.metastore.warehouse.dir}!!/binary_output_format.db/dest1' -' name binary_output_format.dest1' -' serialization.ddl struct dest1 { string mydata}' -' serialization.format 1' -' serialization.last.column.takes.rest true' -' serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' transient_lastDdlTime !!UNIXTIME!!' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' ' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat' -' properties:' -' bucket_count -1' -' columns mydata' -' columns.types string' -' file.inputformat org.apache.hadoop.mapred.TextInputFormat' -' file.outputformat org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat' -' location !!{hive.metastore.warehouse.dir}!!/binary_output_format.db/dest1' -' name binary_output_format.dest1' -' serialization.ddl struct dest1 { string mydata}' -' serialization.format 1' -' serialization.last.column.takes.rest true' -' serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' transient_lastDdlTime !!UNIXTIME!!' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' name: binary_output_format.dest1' -' name: binary_output_format.dest1' -'' -' Stage: Stage-5' -' Map Reduce' -' Alias -> Map Operator Tree:' -' pfile:!!{hive.exec.scratchdir}!! ' -' File Output Operator' -' compressed: false' -' GlobalTableId: 0' -' directory: pfile:!!{hive.exec.scratchdir}!!' -' NumFilesPerFileSink: 1' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat' -' properties:' -' bucket_count -1' -' columns mydata' -' columns.types string' -' file.inputformat org.apache.hadoop.mapred.TextInputFormat' -' file.outputformat org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat' -' location !!{hive.metastore.warehouse.dir}!!/binary_output_format.db/dest1' -' name binary_output_format.dest1' -' serialization.ddl struct dest1 { string mydata}' -' serialization.format 1' -' serialization.last.column.takes.rest true' -' serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' transient_lastDdlTime !!UNIXTIME!!' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' name: binary_output_format.dest1' -' TotalFiles: 1' -' GatherStats: false' -' MultiFileSpray: false' -' Needs Tagging: false' -' Path -> Alias:' -' pfile:!!{hive.exec.scratchdir}!! [pfile:!!{hive.exec.scratchdir}!!]' -' Path -> Partition:' -' pfile:!!{hive.exec.scratchdir}!! ' -' Partition' -' base file name: -ext-10002' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat' -' properties:' -' bucket_count -1' -' columns mydata' -' columns.types string' -' file.inputformat org.apache.hadoop.mapred.TextInputFormat' -' file.outputformat org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat' -' location !!{hive.metastore.warehouse.dir}!!/binary_output_format.db/dest1' -' name binary_output_format.dest1' -' serialization.ddl struct dest1 { string mydata}' -' serialization.format 1' -' serialization.last.column.takes.rest true' -' serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' transient_lastDdlTime !!UNIXTIME!!' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' ' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat' -' properties:' -' bucket_count -1' -' columns mydata' -' columns.types string' -' file.inputformat org.apache.hadoop.mapred.TextInputFormat' -' file.outputformat org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat' -' location !!{hive.metastore.warehouse.dir}!!/binary_output_format.db/dest1' -' name binary_output_format.dest1' -' serialization.ddl struct dest1 { string mydata}' -' serialization.format 1' -' serialization.last.column.takes.rest true' -' serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' transient_lastDdlTime !!UNIXTIME!!' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' name: binary_output_format.dest1' -' name: binary_output_format.dest1' -'' -' Stage: Stage-6' -' Move Operator' -' files:' -' hdfs directory: true' -' source: pfile:!!{hive.exec.scratchdir}!!' -' destination: pfile:!!{hive.exec.scratchdir}!!' -'' -'' -310 rows selected ->>> ->>> INSERT OVERWRITE TABLE dest1 -SELECT TRANSFORM(*) -USING 'cat' -AS mydata STRING -ROW FORMAT SERDE -'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -WITH SERDEPROPERTIES ( -'serialization.last.column.takes.rest'='true' -) -RECORDREADER 'org.apache.hadoop.hive.ql.exec.BinaryRecordReader' -FROM src; -'mydata' -No rows selected ->>> ->>> -- Test the result ->>> SELECT * FROM dest1; -'mydata' -'238 val_238' -'86 val_86' -'311 val_311' -'27 val_27' -'165 val_165' -'409 val_409' -'255 val_255' -'278 val_278' -'98 val_98' -'484 val_484' -'265 val_265' -'193 val_193' -'401 val_401' -'150 val_150' -'273 val_273' -'224 val_224' -'369 val_369' -'66 val_66' -'128 val_128' -'213 val_213' -'146 val_146' -'406 val_406' -'429 val_429' -'374 val_374' -'152 val_152' -'469 val_469' -'145 val_145' -'495 val_495' -'37 val_37' -'327 val_327' -'281 val_281' -'277 val_277' -'209 val_209' -'15 val_15' -'82 val_82' -'403 val_403' -'166 val_166' -'417 val_417' -'430 val_430' -'252 val_252' -'292 val_292' -'219 val_219' -'287 val_287' -'153 val_153' -'193 val_193' -'338 val_338' -'446 val_446' -'459 val_459' -'394 val_394' -'237 val_237' -'482 val_482' -'174 val_174' -'413 val_413' -'494 val_494' -'207 val_207' -'199 val_199' -'466 val_466' -'208 val_208' -'174 val_174' -'399 val_399' -'396 val_396' -'247 val_247' -'417 val_417' -'489 val_489' -'162 val_162' -'377 val_377' -'397 val_397' -'309 val_309' -'365 val_365' -'266 val_266' -'439 val_439' -'342 val_342' -'367 val_367' -'325 val_325' -'167 val_167' -'195 val_195' -'475 val_475' -'17 val_17' -'113 val_113' -'155 val_155' -'203 val_203' -'339 val_339' -'0 val_0' -'455 val_455' -'128 val_128' -'311 val_311' -'316 val_316' -'57 val_57' -'302 val_302' -'205 val_205' -'149 val_149' -'438 val_438' -'345 val_345' -'129 val_129' -'170 val_170' -'20 val_20' -'489 val_489' -'157 val_157' -'378 val_378' -'221 val_221' -'92 val_92' -'111 val_111' -'47 val_47' -'72 val_72' -'4 val_4' -'280 val_280' -'35 val_35' -'427 val_427' -'277 val_277' -'208 val_208' -'356 val_356' -'399 val_399' -'169 val_169' -'382 val_382' -'498 val_498' -'125 val_125' -'386 val_386' -'437 val_437' -'469 val_469' -'192 val_192' -'286 val_286' -'187 val_187' -'176 val_176' -'54 val_54' -'459 val_459' -'51 val_51' -'138 val_138' -'103 val_103' -'239 val_239' -'213 val_213' -'216 val_216' -'430 val_430' -'278 val_278' -'176 val_176' -'289 val_289' -'221 val_221' -'65 val_65' -'318 val_318' -'332 val_332' -'311 val_311' -'275 val_275' -'137 val_137' -'241 val_241' -'83 val_83' -'333 val_333' -'180 val_180' -'284 val_284' -'12 val_12' -'230 val_230' -'181 val_181' -'67 val_67' -'260 val_260' -'404 val_404' -'384 val_384' -'489 val_489' -'353 val_353' -'373 val_373' -'272 val_272' -'138 val_138' -'217 val_217' -'84 val_84' -'348 val_348' -'466 val_466' -'58 val_58' -'8 val_8' -'411 val_411' -'230 val_230' -'208 val_208' -'348 val_348' -'24 val_24' -'463 val_463' -'431 val_431' -'179 val_179' -'172 val_172' -'42 val_42' -'129 val_129' -'158 val_158' -'119 val_119' -'496 val_496' -'0 val_0' -'322 val_322' -'197 val_197' -'468 val_468' -'393 val_393' -'454 val_454' -'100 val_100' -'298 val_298' -'199 val_199' -'191 val_191' -'418 val_418' -'96 val_96' -'26 val_26' -'165 val_165' -'327 val_327' -'230 val_230' -'205 val_205' -'120 val_120' -'131 val_131' -'51 val_51' -'404 val_404' -'43 val_43' -'436 val_436' -'156 val_156' -'469 val_469' -'468 val_468' -'308 val_308' -'95 val_95' -'196 val_196' -'288 val_288' -'481 val_481' -'457 val_457' -'98 val_98' -'282 val_282' -'197 val_197' -'187 val_187' -'318 val_318' -'318 val_318' -'409 val_409' -'470 val_470' -'137 val_137' -'369 val_369' -'316 val_316' -'169 val_169' -'413 val_413' -'85 val_85' -'77 val_77' -'0 val_0' -'490 val_490' -'87 val_87' -'364 val_364' -'179 val_179' -'118 val_118' -'134 val_134' -'395 val_395' -'282 val_282' -'138 val_138' -'238 val_238' -'419 val_419' -'15 val_15' -'118 val_118' -'72 val_72' -'90 val_90' -'307 val_307' -'19 val_19' -'435 val_435' -'10 val_10' -'277 val_277' -'273 val_273' -'306 val_306' -'224 val_224' -'309 val_309' -'389 val_389' -'327 val_327' -'242 val_242' -'369 val_369' -'392 val_392' -'272 val_272' -'331 val_331' -'401 val_401' -'242 val_242' -'452 val_452' -'177 val_177' -'226 val_226' -'5 val_5' -'497 val_497' -'402 val_402' -'396 val_396' -'317 val_317' -'395 val_395' -'58 val_58' -'35 val_35' -'336 val_336' -'95 val_95' -'11 val_11' -'168 val_168' -'34 val_34' -'229 val_229' -'233 val_233' -'143 val_143' -'472 val_472' -'322 val_322' -'498 val_498' -'160 val_160' -'195 val_195' -'42 val_42' -'321 val_321' -'430 val_430' -'119 val_119' -'489 val_489' -'458 val_458' -'78 val_78' -'76 val_76' -'41 val_41' -'223 val_223' -'492 val_492' -'149 val_149' -'449 val_449' -'218 val_218' -'228 val_228' -'138 val_138' -'453 val_453' -'30 val_30' -'209 val_209' -'64 val_64' -'468 val_468' -'76 val_76' -'74 val_74' -'342 val_342' -'69 val_69' -'230 val_230' -'33 val_33' -'368 val_368' -'103 val_103' -'296 val_296' -'113 val_113' -'216 val_216' -'367 val_367' -'344 val_344' -'167 val_167' -'274 val_274' -'219 val_219' -'239 val_239' -'485 val_485' -'116 val_116' -'223 val_223' -'256 val_256' -'263 val_263' -'70 val_70' -'487 val_487' -'480 val_480' -'401 val_401' -'288 val_288' -'191 val_191' -'5 val_5' -'244 val_244' -'438 val_438' -'128 val_128' -'467 val_467' -'432 val_432' -'202 val_202' -'316 val_316' -'229 val_229' -'469 val_469' -'463 val_463' -'280 val_280' -'2 val_2' -'35 val_35' -'283 val_283' -'331 val_331' -'235 val_235' -'80 val_80' -'44 val_44' -'193 val_193' -'321 val_321' -'335 val_335' -'104 val_104' -'466 val_466' -'366 val_366' -'175 val_175' -'403 val_403' -'483 val_483' -'53 val_53' -'105 val_105' -'257 val_257' -'406 val_406' -'409 val_409' -'190 val_190' -'406 val_406' -'401 val_401' -'114 val_114' -'258 val_258' -'90 val_90' -'203 val_203' -'262 val_262' -'348 val_348' -'424 val_424' -'12 val_12' -'396 val_396' -'201 val_201' -'217 val_217' -'164 val_164' -'431 val_431' -'454 val_454' -'478 val_478' -'298 val_298' -'125 val_125' -'431 val_431' -'164 val_164' -'424 val_424' -'187 val_187' -'382 val_382' -'5 val_5' -'70 val_70' -'397 val_397' -'480 val_480' -'291 val_291' -'24 val_24' -'351 val_351' -'255 val_255' -'104 val_104' -'70 val_70' -'163 val_163' -'438 val_438' -'119 val_119' -'414 val_414' -'200 val_200' -'491 val_491' -'237 val_237' -'439 val_439' -'360 val_360' -'248 val_248' -'479 val_479' -'305 val_305' -'417 val_417' -'199 val_199' -'444 val_444' -'120 val_120' -'429 val_429' -'169 val_169' -'443 val_443' -'323 val_323' -'325 val_325' -'277 val_277' -'230 val_230' -'478 val_478' -'178 val_178' -'468 val_468' -'310 val_310' -'317 val_317' -'333 val_333' -'493 val_493' -'460 val_460' -'207 val_207' -'249 val_249' -'265 val_265' -'480 val_480' -'83 val_83' -'136 val_136' -'353 val_353' -'172 val_172' -'214 val_214' -'462 val_462' -'233 val_233' -'406 val_406' -'133 val_133' -'175 val_175' -'189 val_189' -'454 val_454' -'375 val_375' -'401 val_401' -'421 val_421' -'407 val_407' -'384 val_384' -'256 val_256' -'26 val_26' -'134 val_134' -'67 val_67' -'384 val_384' -'379 val_379' -'18 val_18' -'462 val_462' -'492 val_492' -'100 val_100' -'298 val_298' -'9 val_9' -'341 val_341' -'498 val_498' -'146 val_146' -'458 val_458' -'362 val_362' -'186 val_186' -'285 val_285' -'348 val_348' -'167 val_167' -'18 val_18' -'273 val_273' -'183 val_183' -'281 val_281' -'344 val_344' -'97 val_97' -'469 val_469' -'315 val_315' -'84 val_84' -'28 val_28' -'37 val_37' -'448 val_448' -'152 val_152' -'348 val_348' -'307 val_307' -'194 val_194' -'414 val_414' -'477 val_477' -'222 val_222' -'126 val_126' -'90 val_90' -'169 val_169' -'403 val_403' -'400 val_400' -'200 val_200' -'97 val_97' -500 rows selected ->>> !record http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/binarysortable_1.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/beelinepositive/binarysortable_1.q.out b/ql/src/test/results/beelinepositive/binarysortable_1.q.out deleted file mode 100644 index a08d35d..0000000 --- a/ql/src/test/results/beelinepositive/binarysortable_1.q.out +++ /dev/null @@ -1,118 +0,0 @@ -Saving all output to "!!{outputDirectory}!!/binarysortable_1.q.raw". Enter "record" with no arguments to stop it. ->>> !run !!{qFileDirectory}!!/binarysortable_1.q ->>> CREATE TABLE mytable(key STRING, value STRING) -ROW FORMAT DELIMITED -FIELDS TERMINATED BY '9' -STORED AS TEXTFILE; -No rows affected ->>> ->>> LOAD DATA LOCAL INPATH '../data/files/string.txt' INTO TABLE mytable; -No rows affected ->>> ->>> EXPLAIN -SELECT REGEXP_REPLACE(REGEXP_REPLACE(REGEXP_REPLACE(key, '\001', '^A'), '\0', '^@'), '\002', '^B'), value -FROM ( -SELECT key, sum(value) as value -FROM mytable -GROUP BY key -) a; -'Explain' -'ABSTRACT SYNTAX TREE:' -' (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME mytable))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_FUNCTION sum (TOK_TABLE_OR_COL value)) value)) (TOK_GROUPBY (TOK_TABLE_OR_COL key)))) a)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION REGEXP_REPLACE (TOK_FUNCTION REGEXP_REPLACE (TOK_FUNCTION REGEXP_REPLACE (TOK_TABLE_OR_COL key) '\001' '^A') '\0' '^@') '\002' '^B')) (TOK_SELEXPR (TOK_TABLE_OR_COL value)))))' -'' -'STAGE DEPENDENCIES:' -' Stage-1 is a root stage' -' Stage-0 is a root stage' -'' -'STAGE PLANS:' -' Stage: Stage-1' -' Map Reduce' -' Alias -> Map Operator Tree:' -' a:mytable ' -' TableScan' -' alias: mytable' -' Select Operator' -' expressions:' -' expr: key' -' type: string' -' expr: value' -' type: string' -' outputColumnNames: key, value' -' Group By Operator' -' aggregations:' -' expr: sum(value)' -' bucketGroup: false' -' keys:' -' expr: key' -' type: string' -' mode: hash' -' outputColumnNames: _col0, _col1' -' Reduce Output Operator' -' key expressions:' -' expr: _col0' -' type: string' -' sort order: +' -' Map-reduce partition columns:' -' expr: _col0' -' type: string' -' tag: -1' -' value expressions:' -' expr: _col1' -' type: double' -' Reduce Operator Tree:' -' Group By Operator' -' aggregations:' -' expr: sum(VALUE._col0)' -' bucketGroup: false' -' keys:' -' expr: KEY._col0' -' type: string' -' mode: mergepartial' -' outputColumnNames: _col0, _col1' -' Select Operator' -' expressions:' -' expr: _col0' -' type: string' -' expr: _col1' -' type: double' -' outputColumnNames: _col0, _col1' -' Select Operator' -' expressions:' -' expr: regexp_replace(regexp_replace(regexp_replace(_col0, '' -' type: string' -' expr: _col1' -' type: double' -' outputColumnNames: _col0, _col1' -' File Output Operator' -' compressed: false' -' GlobalTableId: 0' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -'' -' Stage: Stage-0' -' Fetch Operator' -' limit: -1' -'' -'' -78 rows selected ->>> ->>> SELECT REGEXP_REPLACE(REGEXP_REPLACE(REGEXP_REPLACE(key, '\001', '^A'), '\0', '^@'), '\002', '^B'), value -FROM ( -SELECT key, sum(value) as value -FROM mytable -GROUP BY key -) a; -'_c0','value' -'^@^@^@','7.0' -'^@^A^@','9.0' -'^@test^@','2.0' -'^A^@^A','10.0' -'^A^A^A','8.0' -'^Atest^A','3.0' -'a^@bc^A^B^A^@','1.0' -'test^@^@^A^Atest','6.0' -'test^@test','4.0' -'test^Atest','5.0' -10 rows selected ->>> !record http://git-wip-us.apache.org/repos/asf/hive/blob/3890ed65/ql/src/test/results/beelinepositive/bucket1.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/beelinepositive/bucket1.q.out b/ql/src/test/results/beelinepositive/bucket1.q.out deleted file mode 100644 index 295d786..0000000 --- a/ql/src/test/results/beelinepositive/bucket1.q.out +++ /dev/null @@ -1,675 +0,0 @@ -Saving all output to "!!{outputDirectory}!!/bucket1.q.raw". Enter "record" with no arguments to stop it. ->>> !run !!{qFileDirectory}!!/bucket1.q ->>> set hive.enforce.bucketing = true; -No rows affected ->>> set hive.exec.reducers.max = 200; -No rows affected ->>> ->>> CREATE TABLE bucket1_1(key int, value string) CLUSTERED BY (key) INTO 100 BUCKETS; -No rows affected ->>> ->>> explain extended -insert overwrite table bucket1_1 -select * from src; -'Explain' -'ABSTRACT SYNTAX TREE:' -' (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME bucket1_1))) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF))))' -'' -'STAGE DEPENDENCIES:' -' Stage-1 is a root stage' -' Stage-0 depends on stages: Stage-1' -' Stage-2 depends on stages: Stage-0' -'' -'STAGE PLANS:' -' Stage: Stage-1' -' Map Reduce' -' Alias -> Map Operator Tree:' -' src ' -' TableScan' -' alias: src' -' GatherStats: false' -' Select Operator' -' expressions:' -' expr: key' -' type: string' -' expr: value' -' type: string' -' outputColumnNames: _col0, _col1' -' Reduce Output Operator' -' sort order: ' -' Map-reduce partition columns:' -' expr: UDFToInteger(_col0)' -' type: int' -' tag: -1' -' value expressions:' -' expr: _col0' -' type: string' -' expr: _col1' -' type: string' -' Needs Tagging: false' -' Path -> Alias:' -' !!{hive.metastore.warehouse.dir}!!/bucket1.db/src [src]' -' Path -> Partition:' -' !!{hive.metastore.warehouse.dir}!!/bucket1.db/src ' -' Partition' -' base file name: src' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' properties:' -' bucket_count -1' -' columns key,value' -' columns.types string:string' -' file.inputformat org.apache.hadoop.mapred.TextInputFormat' -' file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' location !!{hive.metastore.warehouse.dir}!!/bucket1.db/src' -' name bucket1.src' -' numFiles 1' -' numPartitions 0' -' numRows 0' -' rawDataSize 0' -' serialization.ddl struct src { string key, string value}' -' serialization.format 1' -' serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' totalSize 5812' -' transient_lastDdlTime !!UNIXTIME!!' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' ' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' properties:' -' bucket_count -1' -' columns key,value' -' columns.types string:string' -' file.inputformat org.apache.hadoop.mapred.TextInputFormat' -' file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' location !!{hive.metastore.warehouse.dir}!!/bucket1.db/src' -' name bucket1.src' -' numFiles 1' -' numPartitions 0' -' numRows 0' -' rawDataSize 0' -' serialization.ddl struct src { string key, string value}' -' serialization.format 1' -' serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' totalSize 5812' -' transient_lastDdlTime !!UNIXTIME!!' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' name: bucket1.src' -' name: bucket1.src' -' Reduce Operator Tree:' -' Extract' -' Select Operator' -' expressions:' -' expr: UDFToInteger(_col0)' -' type: int' -' expr: _col1' -' type: string' -' outputColumnNames: _col0, _col1' -' File Output Operator' -' compressed: false' -' GlobalTableId: 1' -' directory: pfile:!!{hive.exec.scratchdir}!!' -' NumFilesPerFileSink: 1' -' Stats Publishing Key Prefix: pfile:!!{hive.exec.scratchdir}!!' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' properties:' -' bucket_count 100' -' bucket_field_name key' -' columns key,value' -' columns.types int:string' -' file.inputformat org.apache.hadoop.mapred.TextInputFormat' -' file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' location !!{hive.metastore.warehouse.dir}!!/bucket1.db/bucket1_1' -' name bucket1.bucket1_1' -' serialization.ddl struct bucket1_1 { i32 key, string value}' -' serialization.format 1' -' serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' transient_lastDdlTime !!UNIXTIME!!' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' name: bucket1.bucket1_1' -' TotalFiles: 1' -' GatherStats: true' -' MultiFileSpray: false' -'' -' Stage: Stage-0' -' Move Operator' -' tables:' -' replace: true' -' source: pfile:!!{hive.exec.scratchdir}!!' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' properties:' -' bucket_count 100' -' bucket_field_name key' -' columns key,value' -' columns.types int:string' -' file.inputformat org.apache.hadoop.mapred.TextInputFormat' -' file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' location !!{hive.metastore.warehouse.dir}!!/bucket1.db/bucket1_1' -' name bucket1.bucket1_1' -' serialization.ddl struct bucket1_1 { i32 key, string value}' -' serialization.format 1' -' serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' transient_lastDdlTime !!UNIXTIME!!' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' name: bucket1.bucket1_1' -' tmp directory: pfile:!!{hive.exec.scratchdir}!!' -'' -' Stage: Stage-2' -' Stats-Aggr Operator' -' Stats Aggregation Key Prefix: pfile:!!{hive.exec.scratchdir}!!' -'' -150 rows selected ->>> ->>> insert overwrite table bucket1_1 -select * from src; -'_col0','_col1' -No rows selected ->>> ->>> select * from bucket1_1 order by key; -'key','value' -'0','val_0' -'0','val_0' -'0','val_0' -'2','val_2' -'4','val_4' -'5','val_5' -'5','val_5' -'5','val_5' -'8','val_8' -'9','val_9' -'10','val_10' -'11','val_11' -'12','val_12' -'12','val_12' -'15','val_15' -'15','val_15' -'17','val_17' -'18','val_18' -'18','val_18' -'19','val_19' -'20','val_20' -'24','val_24' -'24','val_24' -'26','val_26' -'26','val_26' -'27','val_27' -'28','val_28' -'30','val_30' -'33','val_33' -'34','val_34' -'35','val_35' -'35','val_35' -'35','val_35' -'37','val_37' -'37','val_37' -'41','val_41' -'42','val_42' -'42','val_42' -'43','val_43' -'44','val_44' -'47','val_47' -'51','val_51' -'51','val_51' -'53','val_53' -'54','val_54' -'57','val_57' -'58','val_58' -'58','val_58' -'64','val_64' -'65','val_65' -'66','val_66' -'67','val_67' -'67','val_67' -'69','val_69' -'70','val_70' -'70','val_70' -'70','val_70' -'72','val_72' -'72','val_72' -'74','val_74' -'76','val_76' -'76','val_76' -'77','val_77' -'78','val_78' -'80','val_80' -'82','val_82' -'83','val_83' -'83','val_83' -'84','val_84' -'84','val_84' -'85','val_85' -'86','val_86' -'87','val_87' -'90','val_90' -'90','val_90' -'90','val_90' -'92','val_92' -'95','val_95' -'95','val_95' -'96','val_96' -'97','val_97' -'97','val_97' -'98','val_98' -'98','val_98' -'100','val_100' -'100','val_100' -'103','val_103' -'103','val_103' -'104','val_104' -'104','val_104' -'105','val_105' -'111','val_111' -'113','val_113' -'113','val_113' -'114','val_114' -'116','val_116' -'118','val_118' -'118','val_118' -'119','val_119' -'119','val_119' -'119','val_119' -'120','val_120' -'120','val_120' -'125','val_125' -'125','val_125' -'126','val_126' -'128','val_128' -'128','val_128' -'128','val_128' -'129','val_129' -'129','val_129' -'131','val_131' -'133','val_133' -'134','val_134' -'134','val_134' -'136','val_136' -'137','val_137' -'137','val_137' -'138','val_138' -'138','val_138' -'138','val_138' -'138','val_138' -'143','val_143' -'145','val_145' -'146','val_146' -'146','val_146' -'149','val_149' -'149','val_149' -'150','val_150' -'152','val_152' -'152','val_152' -'153','val_153' -'155','val_155' -'156','val_156' -'157','val_157' -'158','val_158' -'160','val_160' -'162','val_162' -'163','val_163' -'164','val_164' -'164','val_164' -'165','val_165' -'165','val_165' -'166','val_166' -'167','val_167' -'167','val_167' -'167','val_167' -'168','val_168' -'169','val_169' -'169','val_169' -'169','val_169' -'169','val_169' -'170','val_170' -'172','val_172' -'172','val_172' -'174','val_174' -'174','val_174' -'175','val_175' -'175','val_175' -'176','val_176' -'176','val_176' -'177','val_177' -'178','val_178' -'179','val_179' -'179','val_179' -'180','val_180' -'181','val_181' -'183','val_183' -'186','val_186' -'187','val_187' -'187','val_187' -'187','val_187' -'189','val_189' -'190','val_190' -'191','val_191' -'191','val_191' -'192','val_192' -'193','val_193' -'193','val_193' -'193','val_193' -'194','val_194' -'195','val_195' -'195','val_195' -'196','val_196' -'197','val_197' -'197','val_197' -'199','val_199' -'199','val_199' -'199','val_199' -'200','val_200' -'200','val_200' -'201','val_201' -'202','val_202' -'203','val_203' -'203','val_203' -'205','val_205' -'205','val_205' -'207','val_207' -'207','val_207' -'208','val_208' -'208','val_208' -'208','val_208' -'209','val_209' -'209','val_209' -'213','val_213' -'213','val_213' -'214','val_214' -'216','val_216' -'216','val_216' -'217','val_217' -'217','val_217' -'218','val_218' -'219','val_219' -'219','val_219' -'221','val_221' -'221','val_221' -'222','val_222' -'223','val_223' -'223','val_223' -'224','val_224' -'224','val_224' -'226','val_226' -'228','val_228' -'229','val_229' -'229','val_229' -'230','val_230' -'230','val_230' -'230','val_230' -'230','val_230' -'230','val_230' -'233','val_233' -'233','val_233' -'235','val_235' -'237','val_237' -'237','val_237' -'238','val_238' -'238','val_238' -'239','val_239' -'239','val_239' -'241','val_241' -'242','val_242' -'242','val_242' -'244','val_244' -'247','val_247' -'248','val_248' -'249','val_249' -'252','val_252' -'255','val_255' -'255','val_255' -'256','val_256' -'256','val_256' -'257','val_257' -'258','val_258' -'260','val_260' -'262','val_262' -'263','val_263' -'265','val_265' -'265','val_265' -'266','val_266' -'272','val_272' -'272','val_272' -'273','val_273' -'273','val_273' -'273','val_273' -'274','val_274' -'275','val_275' -'277','val_277' -'277','val_277' -'277','val_277' -'277','val_277' -'278','val_278' -'278','val_278' -'280','val_280' -'280','val_280' -'281','val_281' -'281','val_281' -'282','val_282' -'282','val_282' -'283','val_283' -'284','val_284' -'285','val_285' -'286','val_286' -'287','val_287' -'288','val_288' -'288','val_288' -'289','val_289' -'291','val_291' -'292','val_292' -'296','val_296' -'298','val_298' -'298','val_298' -'298','val_298' -'302','val_302' -'305','val_305' -'306','val_306' -'307','val_307' -'307','val_307' -'308','val_308' -'309','val_309' -'309','val_309' -'310','val_310' -'311','val_311' -'311','val_311' -'311','val_311' -'315','val_315' -'316','val_316' -'316','val_316' -'316','val_316' -'317','val_317' -'317','val_317' -'318','val_318' -'318','val_318' -'318','val_318' -'321','val_321' -'321','val_321' -'322','val_322' -'322','val_322' -'323','val_323' -'325','val_325' -'325','val_325' -'327','val_327' -'327','val_327' -'327','val_327' -'331','val_331' -'331','val_331' -'332','val_332' -'333','val_333' -'333','val_333' -'335','val_335' -'336','val_336' -'338','val_338' -'339','val_339' -'341','val_341' -'342','val_342' -'342','val_342' -'344','val_344' -'344','val_344' -'345','val_345' -'348','val_348' -'348','val_348' -'348','val_348' -'348','val_348' -'348','val_348' -'351','val_351' -'353','val_353' -'353','val_353' -'356','val_356' -'360','val_360' -'362','val_362' -'364','val_364' -'365','val_365' -'366','val_366' -'367','val_367' -'367','val_367' -'368','val_368' -'369','val_369' -'369','val_369' -'369','val_369' -'373','val_373' -'374','val_374' -'375','val_375' -'377','val_377' -'378','val_378' -'379','val_379' -'382','val_382' -'382','val_382' -'384','val_384' -'384','val_384' -'384','val_384' -'386','val_386' -'389','val_389' -'392','val_392' -'393','val_393' -'394','val_394' -'395','val_395' -'395','val_395' -'396','val_396' -'396','val_396' -'396','val_396' -'397','val_397' -'397','val_397' -'399','val_399' -'399','val_399' -'400','val_400' -'401','val_401' -'401','val_401' -'401','val_401' -'401','val_401' -'401','val_401' -'402','val_402' -'403','val_403' -'403','val_403' -'403','val_403' -'404','val_404' -'404','val_404' -'406','val_406' -'406','val_406' -'406','val_406' -'406','val_406' -'407','val_407' -'409','val_409' -'409','val_409' -'409','val_409' -'411','val_411' -'413','val_413' -'413','val_413' -'414','val_414' -'414','val_414' -'417','val_417' -'417','val_417' -'417','val_417' -'418','val_418' -'419','val_419' -'421','val_421' -'424','val_424' -'424','val_424' -'427','val_427' -'429','val_429' -'429','val_429' -'430','val_430' -'430','val_430' -'430','val_430' -'431','val_431' -'431','val_431' -'431','val_431' -'432','val_432' -'435','val_435' -'436','val_436' -'437','val_437' -'438','val_438' -'438','val_438' -'438','val_438' -'439','val_439' -'439','val_439' -'443','val_443' -'444','val_444' -'446','val_446' -'448','val_448' -'449','val_449' -'452','val_452' -'453','val_453' -'454','val_454' -'454','val_454' -'454','val_454' -'455','val_455' -'457','val_457' -'458','val_458' -'458','val_458' -'459','val_459' -'459','val_459' -'460','val_460' -'462','val_462' -'462','val_462' -'463','val_463' -'463','val_463' -'466','val_466' -'466','val_466' -'466','val_466' -'467','val_467' -'468','val_468' -'468','val_468' -'468','val_468' -'468','val_468' -'469','val_469' -'469','val_469' -'469','val_469' -'469','val_469' -'469','val_469' -'470','val_470' -'472','val_472' -'475','val_475' -'477','val_477' -'478','val_478' -'478','val_478' -'479','val_479' -'480','val_480' -'480','val_480' -'480','val_480' -'481','val_481' -'482','val_482' -'483','val_483' -'484','val_484' -'485','val_485' -'487','val_487' -'489','val_489' -'489','val_489' -'489','val_489' -'489','val_489' -'490','val_490' -'491','val_491' -'492','val_492' -'492','val_492' -'493','val_493' -'494','val_494' -'495','val_495' -'496','val_496' -'497','val_497' -'498','val_498' -'498','val_498' -'498','val_498' -500 rows selected ->>> !record