Modified: hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/util/JavaDataModel.java URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/util/JavaDataModel.java?rev=1527858&r1=1527857&r2=1527858&view=diff ============================================================================== --- hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/util/JavaDataModel.java (original) +++ hive/trunk/ql/src/java/org/apache/hadoop/hive/ql/util/JavaDataModel.java Tue Oct 1 03:12:23 2013 @@ -27,95 +27,133 @@ import org.apache.hadoop.hive.ql.udf.gen public enum JavaDataModel { JAVA32 { + @Override public int object() { return JAVA32_OBJECT; } + @Override public int array() { return JAVA32_ARRAY; } + @Override public int ref() { return JAVA32_REF; } + @Override public int hashMap(int entry) { // base = JAVA32_OBJECT + PRIMITIVES1 * 4 + JAVA32_FIELDREF * 3 + JAVA32_ARRAY; // entry = JAVA32_OBJECT + JAVA32_FIELDREF + PRIMITIVES1 return 64 + 24 * entry; } + @Override + public int hashMapEntry() { + return 24; + } + + @Override public int hashSet(int entry) { // hashMap += JAVA32_OBJECT return 80 + 24 * entry; } + @Override public int linkedHashMap(int entry) { // hashMap += JAVA32_FIELDREF + PRIMITIVES1 // hashMap.entry += JAVA32_FIELDREF * 2 return 72 + 32 * entry; } + @Override public int linkedList(int entry) { // base = JAVA32_OBJECT + PRIMITIVES1 * 2 + JAVA32_FIELDREF; // entry = JAVA32_OBJECT + JAVA32_FIELDREF * 2 return 28 + 24 * entry; } + @Override public int arrayList() { // JAVA32_OBJECT + PRIMITIVES1 * 2 + JAVA32_ARRAY; return 44; } + + @Override + public int memoryAlign() { + return 8; + } }, JAVA64 { + @Override public int object() { return JAVA64_OBJECT; } + @Override public int array() { return JAVA64_ARRAY; } + @Override public int ref() { return JAVA64_REF; } + @Override public int hashMap(int entry) { // base = JAVA64_OBJECT + PRIMITIVES1 * 4 + JAVA64_FIELDREF * 3 + JAVA64_ARRAY; // entry = JAVA64_OBJECT + JAVA64_FIELDREF + PRIMITIVES1 return 112 + 44 * entry; } + @Override + public int hashMapEntry() { + return 44; + } + + @Override public int hashSet(int entry) { // hashMap += JAVA64_OBJECT return 144 + 44 * entry; } + @Override public int linkedHashMap(int entry) { // hashMap += JAVA64_FIELDREF + PRIMITIVES1 // hashMap.entry += JAVA64_FIELDREF * 2 return 128 + 60 * entry; } + @Override public int linkedList(int entry) { // base = JAVA64_OBJECT + PRIMITIVES1 * 2 + JAVA64_FIELDREF; // entry = JAVA64_OBJECT + JAVA64_FIELDREF * 2 return 48 + 48 * entry; } + @Override public int arrayList() { // JAVA64_OBJECT + PRIMITIVES1 * 2 + JAVA64_ARRAY; return 80; } + + @Override + public int memoryAlign() { + return 8; + } }; public abstract int object(); public abstract int array(); public abstract int ref(); public abstract int hashMap(int entry); + public abstract int hashMapEntry(); public abstract int hashSet(int entry); public abstract int linkedHashMap(int entry); public abstract int linkedList(int entry); public abstract int arrayList(); + public abstract int memoryAlign(); // ascii string public int lengthFor(String string) { @@ -161,6 +199,10 @@ public enum JavaDataModel { return PRIMITIVES2; } + public static int alignUp(int value, int align) { + return (value + align - 1) & ~(align - 1); + } + public static final int JAVA32_META = 12; public static final int JAVA32_ARRAY_META = 16; public static final int JAVA32_REF = 4; @@ -176,6 +218,8 @@ public enum JavaDataModel { public static final int PRIMITIVES1 = 4; // void, boolean, byte, short, int, float public static final int PRIMITIVES2 = 8; // long, double + public static final int PRIMITIVE_BYTE = 1; // byte + private static JavaDataModel current; public static JavaDataModel get() { @@ -200,4 +244,27 @@ public enum JavaDataModel { } return ((size + 8) >> 3) << 3; } + + private int lengthForPrimitiveArrayOfSize(int primitiveSize, int length) { + return alignUp(array() + primitiveSize*length, memoryAlign()); + } + + public int lengthForByteArrayOfSize(int length) { + return lengthForPrimitiveArrayOfSize(PRIMITIVE_BYTE, length); + } + public int lengthForObjectArrayOfSize(int length) { + return lengthForPrimitiveArrayOfSize(ref(), length); + } + public int lengthForLongArrayOfSize(int length) { + return lengthForPrimitiveArrayOfSize(primitive2(), length); + } + public int lengthForDoubleArrayOfSize(int length) { + return lengthForPrimitiveArrayOfSize(primitive2(), length); + } + public int lengthForIntArrayOfSize(int length) { + return lengthForPrimitiveArrayOfSize(primitive1(), length); + } + public int lengthForBooleanArrayOfSize(int length) { + return lengthForPrimitiveArrayOfSize(PRIMITIVE_BYTE, length); + } }
Modified: hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/QTestUtil.java URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/QTestUtil.java?rev=1527858&r1=1527857&r2=1527858&view=diff ============================================================================== --- hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/QTestUtil.java (original) +++ hive/trunk/ql/src/test/org/apache/hadoop/hive/ql/QTestUtil.java Tue Oct 1 03:12:23 2013 @@ -59,12 +59,15 @@ import org.apache.hadoop.hive.cli.CliDri import org.apache.hadoop.hive.cli.CliSessionState; import org.apache.hadoop.hive.common.io.CachingPrintStream; import org.apache.hadoop.hive.conf.HiveConf; +import org.apache.hadoop.hive.conf.HiveConf.ConfVars; import org.apache.hadoop.hive.metastore.MetaStoreUtils; import org.apache.hadoop.hive.metastore.api.Index; import org.apache.hadoop.hive.ql.exec.FunctionRegistry; import org.apache.hadoop.hive.ql.exec.Task; import org.apache.hadoop.hive.ql.exec.Utilities; import org.apache.hadoop.hive.ql.exec.Utilities.StreamPrinter; +import org.apache.hadoop.hive.ql.exec.vector.util.AllVectorTypesRecord; +import org.apache.hadoop.hive.ql.exec.vector.util.OrcFileGenerator; import org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat; import org.apache.hadoop.hive.ql.lockmgr.zookeeper.ZooKeeperHiveLockManager; import org.apache.hadoop.hive.ql.metadata.Hive; @@ -108,7 +111,7 @@ public class QTestUtil { public static final HashSet<String> srcTables = new HashSet<String> (Arrays.asList(new String [] { "src", "src1", "srcbucket", "srcbucket2", "src_json", "src_thrift", - "src_sequencefile", "srcpart" + "src_sequencefile", "srcpart", "alltypesorc" })); private ParseDriver pd; @@ -221,6 +224,11 @@ public class QTestUtil { convertPathsFromWindowsToHdfs(); } + String vectorizationEnabled = System.getProperty("test.vectorization.enabled"); + if(vectorizationEnabled != null && vectorizationEnabled.equalsIgnoreCase("true")) { + conf.setBoolVar(ConfVars.HIVE_VECTORIZATION_ENABLED, true); + } + // Plug verifying metastore in for testing. conf.setVar(HiveConf.ConfVars.METASTORE_RAW_STORE_IMPL, "org.apache.hadoop.hive.metastore.VerifyingObjectStore"); @@ -521,7 +529,8 @@ public class QTestUtil { for (String s : new String[] {"src", "src1", "src_json", "src_thrift", "src_sequencefile", "srcpart", "srcbucket", "srcbucket2", "dest1", "dest2", "dest3", "dest4", "dest4_sequencefile", "dest_j1", "dest_j2", - "dest_g1", "dest_g2", "fetchtask_ioexception"}) { + "dest_g1", "dest_g2", "fetchtask_ioexception", + AllVectorTypesRecord.TABLE_NAME}) { db.dropTable(MetaStoreUtils.DEFAULT_DATABASE_NAME, s); } @@ -663,7 +672,17 @@ public class QTestUtil { fpath = new Path(testFiles, "json.txt"); runLoadCmd("LOAD DATA LOCAL INPATH '" + fpath.toUri().getPath() + "' INTO TABLE src_json"); + + FileSystem localFs = FileSystem.getLocal(conf); + // create and load data into orc table + fpath = new Path(testFiles, AllVectorTypesRecord.TABLE_NAME); + + runCreateTableCmd(AllVectorTypesRecord.TABLE_CREATE_COMMAND); + runLoadCmd("LOAD DATA LOCAL INPATH '" + fpath.toUri().getPath() + + "' INTO TABLE "+AllVectorTypesRecord.TABLE_NAME); + conf.setBoolean("hive.test.init.phase", false); + } public void init() throws Exception { Modified: hive/trunk/ql/src/test/results/clientpositive/add_part_exist.q.out URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/add_part_exist.q.out?rev=1527858&r1=1527857&r2=1527858&view=diff ============================================================================== --- hive/trunk/ql/src/test/results/clientpositive/add_part_exist.q.out (original) +++ hive/trunk/ql/src/test/results/clientpositive/add_part_exist.q.out Tue Oct 1 03:12:23 2013 @@ -75,6 +75,7 @@ PREHOOK: query: SHOW TABLES PREHOOK: type: SHOWTABLES POSTHOOK: query: SHOW TABLES POSTHOOK: type: SHOWTABLES +alltypesorc src src1 src_json Modified: hive/trunk/ql/src/test/results/clientpositive/alter1.q.out URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/alter1.q.out?rev=1527858&r1=1527857&r2=1527858&view=diff ============================================================================== --- hive/trunk/ql/src/test/results/clientpositive/alter1.q.out (original) +++ hive/trunk/ql/src/test/results/clientpositive/alter1.q.out Tue Oct 1 03:12:23 2013 @@ -168,6 +168,7 @@ PREHOOK: query: SHOW TABLES PREHOOK: type: SHOWTABLES POSTHOOK: query: SHOW TABLES POSTHOOK: type: SHOWTABLES +alltypesorc src src1 src_json Modified: hive/trunk/ql/src/test/results/clientpositive/alter2.q.out URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/alter2.q.out?rev=1527858&r1=1527857&r2=1527858&view=diff ============================================================================== --- hive/trunk/ql/src/test/results/clientpositive/alter2.q.out (original) +++ hive/trunk/ql/src/test/results/clientpositive/alter2.q.out Tue Oct 1 03:12:23 2013 @@ -172,6 +172,7 @@ PREHOOK: query: SHOW TABLES PREHOOK: type: SHOWTABLES POSTHOOK: query: SHOW TABLES POSTHOOK: type: SHOWTABLES +alltypesorc src src1 src_json Modified: hive/trunk/ql/src/test/results/clientpositive/alter3.q.out URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/alter3.q.out?rev=1527858&r1=1527857&r2=1527858&view=diff ============================================================================== --- hive/trunk/ql/src/test/results/clientpositive/alter3.q.out (original) +++ hive/trunk/ql/src/test/results/clientpositive/alter3.q.out Tue Oct 1 03:12:23 2013 @@ -184,6 +184,7 @@ POSTHOOK: query: SHOW TABLES POSTHOOK: type: SHOWTABLES POSTHOOK: Lineage: alter3 PARTITION(pcol1=test_part:,pcol2=test_part:).col1 SIMPLE [(alter3_src)alter3_src.FieldSchema(name:col1, type:string, comment:null), ] POSTHOOK: Lineage: alter3_like PARTITION(pcol1=test_part:,pcol2=test_part:).col1 SIMPLE [(alter3_src)alter3_src.FieldSchema(name:col1, type:string, comment:null), ] +alltypesorc src src1 src_json Modified: hive/trunk/ql/src/test/results/clientpositive/alter4.q.out URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/alter4.q.out?rev=1527858&r1=1527857&r2=1527858&view=diff ============================================================================== --- hive/trunk/ql/src/test/results/clientpositive/alter4.q.out (original) +++ hive/trunk/ql/src/test/results/clientpositive/alter4.q.out Tue Oct 1 03:12:23 2013 @@ -41,6 +41,7 @@ PREHOOK: query: SHOW TABLES PREHOOK: type: SHOWTABLES POSTHOOK: query: SHOW TABLES POSTHOOK: type: SHOWTABLES +alltypesorc src src1 src_json Modified: hive/trunk/ql/src/test/results/clientpositive/alter5.q.out URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/alter5.q.out?rev=1527858&r1=1527857&r2=1527858&view=diff ============================================================================== --- hive/trunk/ql/src/test/results/clientpositive/alter5.q.out (original) +++ hive/trunk/ql/src/test/results/clientpositive/alter5.q.out Tue Oct 1 03:12:23 2013 @@ -115,6 +115,7 @@ PREHOOK: type: SHOWTABLES POSTHOOK: query: SHOW TABLES POSTHOOK: type: SHOWTABLES POSTHOOK: Lineage: alter5 PARTITION(dt=a).col1 SIMPLE [(alter5_src)alter5_src.FieldSchema(name:col1, type:string, comment:null), ] +alltypesorc src src1 src_json Modified: hive/trunk/ql/src/test/results/clientpositive/alter_index.q.out URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/alter_index.q.out?rev=1527858&r1=1527857&r2=1527858&view=diff ============================================================================== --- hive/trunk/ql/src/test/results/clientpositive/alter_index.q.out (original) +++ hive/trunk/ql/src/test/results/clientpositive/alter_index.q.out Tue Oct 1 03:12:23 2013 @@ -37,6 +37,7 @@ PREHOOK: query: show tables PREHOOK: type: SHOWTABLES POSTHOOK: query: show tables POSTHOOK: type: SHOWTABLES +alltypesorc src src1 src_json Modified: hive/trunk/ql/src/test/results/clientpositive/alter_rename_partition.q.out URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/alter_rename_partition.q.out?rev=1527858&r1=1527857&r2=1527858&view=diff ============================================================================== --- hive/trunk/ql/src/test/results/clientpositive/alter_rename_partition.q.out (original) +++ hive/trunk/ql/src/test/results/clientpositive/alter_rename_partition.q.out Tue Oct 1 03:12:23 2013 @@ -12,6 +12,7 @@ PREHOOK: query: SHOW TABLES PREHOOK: type: SHOWTABLES POSTHOOK: query: SHOW TABLES POSTHOOK: type: SHOWTABLES +alltypesorc src src1 src_json @@ -130,6 +131,7 @@ PREHOOK: type: SHOWTABLES POSTHOOK: query: SHOW TABLES POSTHOOK: type: SHOWTABLES POSTHOOK: Lineage: alter_rename_partition PARTITION(pcol1=old_part1:,pcol2=old_part2:).col1 SIMPLE [(alter_rename_partition_src)alter_rename_partition_src.FieldSchema(name:col1, type:string, comment:null), ] +alltypesorc src src1 src_json Modified: hive/trunk/ql/src/test/results/clientpositive/describe_table_json.q.out URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/describe_table_json.q.out?rev=1527858&r1=1527857&r2=1527858&view=diff ============================================================================== --- hive/trunk/ql/src/test/results/clientpositive/describe_table_json.q.out (original) +++ hive/trunk/ql/src/test/results/clientpositive/describe_table_json.q.out Tue Oct 1 03:12:23 2013 @@ -7,7 +7,7 @@ PREHOOK: query: SHOW TABLES PREHOOK: type: SHOWTABLES POSTHOOK: query: SHOW TABLES POSTHOOK: type: SHOWTABLES -{"tables":["jsontable","src","src1","src_json","src_sequencefile","src_thrift","srcbucket","srcbucket2","srcpart"]} +{"tables":["alltypesorc","jsontable","src","src1","src_json","src_sequencefile","src_thrift","srcbucket","srcbucket2","srcpart"]} PREHOOK: query: SHOW TABLES LIKE 'json*' PREHOOK: type: SHOWTABLES POSTHOOK: query: SHOW TABLES LIKE 'json*' Modified: hive/trunk/ql/src/test/results/clientpositive/index_creation.q.out URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/index_creation.q.out?rev=1527858&r1=1527857&r2=1527858&view=diff ============================================================================== --- hive/trunk/ql/src/test/results/clientpositive/index_creation.q.out (original) +++ hive/trunk/ql/src/test/results/clientpositive/index_creation.q.out Tue Oct 1 03:12:23 2013 @@ -306,6 +306,7 @@ POSTHOOK: Lineage: default___t_x2__._off POSTHOOK: Lineage: default___t_x__._bucketname SIMPLE [(_t)_t.FieldSchema(name:INPUT__FILE__NAME, type:string, comment:), ] POSTHOOK: Lineage: default___t_x__._j SIMPLE [(_t)_t.FieldSchema(name:_j, type:int, comment:null), ] POSTHOOK: Lineage: default___t_x__._offsets EXPRESSION [(_t)_t.FieldSchema(name:BLOCK__OFFSET__INSIDE__FILE, type:bigint, comment:), ] +alltypesorc src src1 src_json Modified: hive/trunk/ql/src/test/results/clientpositive/input2.q.out URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/input2.q.out?rev=1527858&r1=1527857&r2=1527858&view=diff ============================================================================== --- hive/trunk/ql/src/test/results/clientpositive/input2.q.out (original) +++ hive/trunk/ql/src/test/results/clientpositive/input2.q.out Tue Oct 1 03:12:23 2013 @@ -31,6 +31,7 @@ PREHOOK: query: SHOW TABLES PREHOOK: type: SHOWTABLES POSTHOOK: query: SHOW TABLES POSTHOOK: type: SHOWTABLES +alltypesorc src src1 src_json @@ -53,6 +54,7 @@ PREHOOK: query: SHOW TABLES PREHOOK: type: SHOWTABLES POSTHOOK: query: SHOW TABLES POSTHOOK: type: SHOWTABLES +alltypesorc src src1 src_json Modified: hive/trunk/ql/src/test/results/clientpositive/input3.q.out URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/input3.q.out?rev=1527858&r1=1527857&r2=1527858&view=diff ============================================================================== --- hive/trunk/ql/src/test/results/clientpositive/input3.q.out (original) +++ hive/trunk/ql/src/test/results/clientpositive/input3.q.out Tue Oct 1 03:12:23 2013 @@ -25,6 +25,7 @@ PREHOOK: query: SHOW TABLES PREHOOK: type: SHOWTABLES POSTHOOK: query: SHOW TABLES POSTHOOK: type: SHOWTABLES +alltypesorc src src1 src_json @@ -114,6 +115,7 @@ PREHOOK: query: SHOW TABLES PREHOOK: type: SHOWTABLES POSTHOOK: query: SHOW TABLES POSTHOOK: type: SHOWTABLES +alltypesorc src src1 src_json Modified: hive/trunk/ql/src/test/results/clientpositive/input4.q.out URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/input4.q.out?rev=1527858&r1=1527857&r2=1527858&view=diff ============================================================================== --- hive/trunk/ql/src/test/results/clientpositive/input4.q.out (original) +++ hive/trunk/ql/src/test/results/clientpositive/input4.q.out Tue Oct 1 03:12:23 2013 @@ -48,7 +48,7 @@ PREHOOK: type: QUERY POSTHOOK: query: EXPLAIN FORMATTED SELECT Input4Alias.VALUE, Input4Alias.KEY FROM INPUT4 AS Input4Alias POSTHOOK: type: QUERY -{"STAGE PLANS":{"Stage-1":{"Map Reduce":{"Map:":{"Split Sample:":{},"Alias -> Map Operator Tree:":{"input4alias":{"TS_0":{"SEL_1":{"FS_2":{"File Output Operator":{"GlobalTableId:":"0","compressed:":"false","table:":{"serde:":"org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe","input format:":"org.apache.hadoop.mapred.TextInputFormat","output format:":"org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat"}}}}}}}}}},"Stage-0":{"Fetch Operator":{"limit:":"-1"}}},"STAGE DEPENDENCIES":{"Stage-1":{"ROOT STAGE":"TRUE"},"Stage-0":{"ROOT STAGE":"TRUE"}},"ABSTRACT SYNTAX TREE":"(TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME INPUT4) Input4Alias)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL Input4Alias) VALUE)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL Input4Alias) KEY)))))"} +{"STAGE PLANS":{"Stage-1":{"Map Reduce":{"Vectorized execution:":"false","Map:":{"Vectorized execution:":"false","Split Sample:":{},"Alias -> Map Operator Tree:":{"input4alias":{"TS_0":{"SEL_1":{"FS_2":{"File Output Operator":{"Vectorized execution:":"false","GlobalTableId:":"0","compressed:":"false","table:":{"serde:":"org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe","input format:":"org.apache.hadoop.mapred.TextInputFormat","output format:":"org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat"}}}}}}}}}},"Stage-0":{"Fetch Operator":{"limit:":"-1"}}},"STAGE DEPENDENCIES":{"Stage-1":{"ROOT STAGE":"TRUE"},"Stage-0":{"ROOT STAGE":"TRUE"}},"ABSTRACT SYNTAX TREE":"(TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME INPUT4) Input4Alias)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL Input4Alias) VALUE)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL Input4Alias) KEY)))))"} PREHOOK: query: SELECT Input4Alias.VALUE, Input4Alias.KEY FROM INPUT4 AS Input4Alias PREHOOK: type: QUERY PREHOOK: Input: default@input4 Modified: hive/trunk/ql/src/test/results/clientpositive/plan_json.q.out URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/plan_json.q.out?rev=1527858&r1=1527857&r2=1527858&view=diff ============================================================================== --- hive/trunk/ql/src/test/results/clientpositive/plan_json.q.out (original) +++ hive/trunk/ql/src/test/results/clientpositive/plan_json.q.out Tue Oct 1 03:12:23 2013 @@ -6,4 +6,4 @@ POSTHOOK: query: -- explain plan json: EXPLAIN FORMATTED SELECT count(1) FROM src POSTHOOK: type: QUERY -{"STAGE PLANS":{"Stage-1":{"Map Reduce":{"Map:":{"Split Sample:":{},"Alias -> Map Operator Tree:":{"src":{"TS_0":{"SEL_1":{"GBY_2":{"RS_3":{"Reduce Output Operator":{"Map-reduce partition columns:":[],"sort order:":"","tag:":"-1","value expressions:":[{"type:":"bigint","expr:":"_col0"}],"key expressions:":[]}}}}}}}},"Reduce:":{"Reduce Operator Tree:":{"GBY_4":{"SEL_5":{"FS_6":{"File Output Operator":{"GlobalTableId:":"0","compressed:":"false","table:":{"serde:":"org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe","input format:":"org.apache.hadoop.mapred.TextInputFormat","output format:":"org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat"}}}}}}}}},"Stage-0":{"Fetch Operator":{"limit:":"-1"}}},"STAGE DEPENDENCIES":{"Stage-1":{"ROOT STAGE":"TRUE"},"Stage-0":{"ROOT STAGE":"TRUE"}},"ABSTRACT SYNTAX TREE":"(TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION count 1)))))"} +{"STAGE PLANS":{"Stage-1":{"Map Reduce":{"Vectorized execution:":"false","Map:":{"Vectorized execution:":"false","Split Sample:":{},"Alias -> Map Operator Tree:":{"src":{"TS_0":{"SEL_1":{"GBY_2":{"RS_3":{"Reduce Output Operator":{"Vectorized execution:":"false","Map-reduce partition columns:":[],"sort order:":"","tag:":"-1","value expressions:":[{"type:":"bigint","expr:":"_col0"}],"key expressions:":[]}}}}}}}},"Reduce:":{"Vectorized execution:":"false","Reduce Operator Tree:":{"GBY_4":{"SEL_5":{"FS_6":{"File Output Operator":{"Vectorized execution:":"false","GlobalTableId:":"0","compressed:":"false","table:":{"serde:":"org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe","input format:":"org.apache.hadoop.mapred.TextInputFormat","output format:":"org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat"}}}}}}}}},"Stage-0":{"Fetch Operator":{"limit:":"-1"}}},"STAGE DEPENDENCIES":{"Stage-1":{"ROOT STAGE":"TRUE"},"Stage-0":{"ROOT STAGE":"TRUE"}},"ABSTRACT SYNTAX TREE":"(TOK_QUERY (TOK _FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION count 1)))))"} Modified: hive/trunk/ql/src/test/results/clientpositive/rename_column.q.out URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/rename_column.q.out?rev=1527858&r1=1527857&r2=1527858&view=diff ============================================================================== --- hive/trunk/ql/src/test/results/clientpositive/rename_column.q.out (original) +++ hive/trunk/ql/src/test/results/clientpositive/rename_column.q.out Tue Oct 1 03:12:23 2013 @@ -127,6 +127,7 @@ PREHOOK: query: SHOW TABLES PREHOOK: type: SHOWTABLES POSTHOOK: query: SHOW TABLES POSTHOOK: type: SHOWTABLES +alltypesorc src src1 src_json Modified: hive/trunk/ql/src/test/results/clientpositive/show_tables.q.out URL: http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/show_tables.q.out?rev=1527858&r1=1527857&r2=1527858&view=diff ============================================================================== --- hive/trunk/ql/src/test/results/clientpositive/show_tables.q.out (original) +++ hive/trunk/ql/src/test/results/clientpositive/show_tables.q.out Tue Oct 1 03:12:23 2013 @@ -112,6 +112,7 @@ PREHOOK: query: SHOW TABLES FROM default PREHOOK: type: SHOWTABLES POSTHOOK: query: SHOW TABLES FROM default POSTHOOK: type: SHOWTABLES +alltypesorc shtb_test1 shtb_test2 src @@ -133,6 +134,7 @@ PREHOOK: query: SHOW TABLES IN default PREHOOK: type: SHOWTABLES POSTHOOK: query: SHOW TABLES IN default POSTHOOK: type: SHOWTABLES +alltypesorc shtb_test1 shtb_test2 src Modified: hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/columnar/ColumnarSerDe.java URL: http://svn.apache.org/viewvc/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/columnar/ColumnarSerDe.java?rev=1527858&r1=1527857&r2=1527858&view=diff ============================================================================== --- hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/columnar/ColumnarSerDe.java (original) +++ hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/columnar/ColumnarSerDe.java Tue Oct 1 03:12:23 2013 @@ -71,7 +71,7 @@ public class ColumnarSerDe extends Colum public ColumnarSerDe() throws SerDeException { } - SerDeParameters serdeParams = null; + protected SerDeParameters serdeParams = null; /** * Initialize the SerDe given the parameters. Modified: hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyUtils.java URL: http://svn.apache.org/viewvc/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyUtils.java?rev=1527858&r1=1527857&r2=1527858&view=diff ============================================================================== --- hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyUtils.java (original) +++ hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/lazy/LazyUtils.java Tue Oct 1 03:12:23 2013 @@ -34,12 +34,12 @@ import org.apache.hadoop.hive.serde2.Ser import org.apache.hadoop.hive.serde2.io.HiveVarcharWritable; import org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe.SerDeParameters; import org.apache.hadoop.hive.serde2.objectinspector.PrimitiveObjectInspector; -import org.apache.hadoop.hive.serde2.objectinspector.primitive.HiveDecimalObjectInspector; import org.apache.hadoop.hive.serde2.objectinspector.primitive.BinaryObjectInspector; import org.apache.hadoop.hive.serde2.objectinspector.primitive.BooleanObjectInspector; import org.apache.hadoop.hive.serde2.objectinspector.primitive.ByteObjectInspector; import org.apache.hadoop.hive.serde2.objectinspector.primitive.DoubleObjectInspector; import org.apache.hadoop.hive.serde2.objectinspector.primitive.FloatObjectInspector; +import org.apache.hadoop.hive.serde2.objectinspector.primitive.HiveDecimalObjectInspector; import org.apache.hadoop.hive.serde2.objectinspector.primitive.HiveVarcharObjectInspector; import org.apache.hadoop.hive.serde2.objectinspector.primitive.IntObjectInspector; import org.apache.hadoop.hive.serde2.objectinspector.primitive.LongObjectInspector; @@ -130,8 +130,8 @@ public final class LazyUtils { } } - private static byte[] trueBytes = {(byte) 't', 'r', 'u', 'e'}; - private static byte[] falseBytes = {(byte) 'f', 'a', 'l', 's', 'e'}; + public static byte[] trueBytes = {(byte) 't', 'r', 'u', 'e'}; + public static byte[] falseBytes = {(byte) 'f', 'a', 'l', 's', 'e'}; /** * Write the bytes with special characters escaped. @@ -144,7 +144,7 @@ public final class LazyUtils { * if escaped, whether a specific character needs escaping. This * array should have size of 128. */ - private static void writeEscaped(OutputStream out, byte[] bytes, int start, + public static void writeEscaped(OutputStream out, byte[] bytes, int start, int len, boolean escaped, byte escapeChar, boolean[] needsEscape) throws IOException { if (escaped) {
