Modified: hive/trunk/ql/src/test/results/clientpositive/sample10.q.out
URL: 
http://svn.apache.org/viewvc/hive/trunk/ql/src/test/results/clientpositive/sample10.q.out?rev=1089938&r1=1089937&r2=1089938&view=diff
==============================================================================
--- hive/trunk/ql/src/test/results/clientpositive/sample10.q.out (original)
+++ hive/trunk/ql/src/test/results/clientpositive/sample10.q.out Thu Apr  7 
16:32:18 2011
@@ -104,12 +104,12 @@ STAGE PLANS:
                               type: bigint
       Needs Tagging: false
       Path -> Alias:
-        
pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-08/hr=11/000000_0
 [srcpartbucket]
-        
pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-08/hr=12/000000_0
 [srcpartbucket]
-        
pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-09/hr=11/000000_0
 [srcpartbucket]
-        
pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-09/hr=12/000000_0
 [srcpartbucket]
+        
pfile:/data/users/nzhang/work/3/apache-hive/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-08/hr=11/000000_0
 [srcpartbucket]
+        
pfile:/data/users/nzhang/work/3/apache-hive/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-08/hr=12/000000_0
 [srcpartbucket]
+        
pfile:/data/users/nzhang/work/3/apache-hive/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-09/hr=11/000000_0
 [srcpartbucket]
+        
pfile:/data/users/nzhang/work/3/apache-hive/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-09/hr=12/000000_0
 [srcpartbucket]
       Path -> Partition:
-        
pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-08/hr=11/000000_0
 
+        
pfile:/data/users/nzhang/work/3/apache-hive/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-08/hr=11/000000_0
 
           Partition
             base file name: 000000_0
             input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
@@ -124,17 +124,17 @@ STAGE PLANS:
               columns.types string:string
               file.inputformat org.apache.hadoop.hive.ql.io.RCFileInputFormat
               file.outputformat org.apache.hadoop.hive.ql.io.RCFileOutputFormat
-              location 
pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-08/hr=11
+              location 
pfile:/data/users/nzhang/work/3/apache-hive/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-08/hr=11
               name default.srcpartbucket
-              numFiles 16
+              numFiles 4
               numPartitions 4
-              numRows 40
+              numRows 10
               partition_columns ds/hr
               serialization.ddl struct srcpartbucket { string key, string 
value}
               serialization.format 1
               serialization.lib 
org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
-              totalSize 2748
-              transient_lastDdlTime 1297386150
+              totalSize 687
+              transient_lastDdlTime 1302062777
             serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
           
               input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
@@ -146,7 +146,7 @@ STAGE PLANS:
                 columns.types string:string
                 file.inputformat org.apache.hadoop.hive.ql.io.RCFileInputFormat
                 file.outputformat 
org.apache.hadoop.hive.ql.io.RCFileOutputFormat
-                location 
pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcpartbucket
+                location 
pfile:/data/users/nzhang/work/3/apache-hive/build/ql/test/data/warehouse/srcpartbucket
                 name default.srcpartbucket
                 numFiles 16
                 numPartitions 4
@@ -156,11 +156,11 @@ STAGE PLANS:
                 serialization.format 1
                 serialization.lib 
org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
                 totalSize 2748
-                transient_lastDdlTime 1297386150
+                transient_lastDdlTime 1302062777
               serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
               name: default.srcpartbucket
             name: default.srcpartbucket
-        
pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-08/hr=12/000000_0
 
+        
pfile:/data/users/nzhang/work/3/apache-hive/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-08/hr=12/000000_0
 
           Partition
             base file name: 000000_0
             input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
@@ -175,17 +175,17 @@ STAGE PLANS:
               columns.types string:string
               file.inputformat org.apache.hadoop.hive.ql.io.RCFileInputFormat
               file.outputformat org.apache.hadoop.hive.ql.io.RCFileOutputFormat
-              location 
pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-08/hr=12
+              location 
pfile:/data/users/nzhang/work/3/apache-hive/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-08/hr=12
               name default.srcpartbucket
-              numFiles 16
+              numFiles 4
               numPartitions 4
-              numRows 40
+              numRows 10
               partition_columns ds/hr
               serialization.ddl struct srcpartbucket { string key, string 
value}
               serialization.format 1
               serialization.lib 
org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
-              totalSize 2748
-              transient_lastDdlTime 1297386150
+              totalSize 687
+              transient_lastDdlTime 1302062777
             serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
           
               input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
@@ -197,7 +197,7 @@ STAGE PLANS:
                 columns.types string:string
                 file.inputformat org.apache.hadoop.hive.ql.io.RCFileInputFormat
                 file.outputformat 
org.apache.hadoop.hive.ql.io.RCFileOutputFormat
-                location 
pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcpartbucket
+                location 
pfile:/data/users/nzhang/work/3/apache-hive/build/ql/test/data/warehouse/srcpartbucket
                 name default.srcpartbucket
                 numFiles 16
                 numPartitions 4
@@ -207,11 +207,11 @@ STAGE PLANS:
                 serialization.format 1
                 serialization.lib 
org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
                 totalSize 2748
-                transient_lastDdlTime 1297386150
+                transient_lastDdlTime 1302062777
               serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
               name: default.srcpartbucket
             name: default.srcpartbucket
-        
pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-09/hr=11/000000_0
 
+        
pfile:/data/users/nzhang/work/3/apache-hive/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-09/hr=11/000000_0
 
           Partition
             base file name: 000000_0
             input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
@@ -226,17 +226,17 @@ STAGE PLANS:
               columns.types string:string
               file.inputformat org.apache.hadoop.hive.ql.io.RCFileInputFormat
               file.outputformat org.apache.hadoop.hive.ql.io.RCFileOutputFormat
-              location 
pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-09/hr=11
+              location 
pfile:/data/users/nzhang/work/3/apache-hive/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-09/hr=11
               name default.srcpartbucket
-              numFiles 16
+              numFiles 4
               numPartitions 4
-              numRows 40
+              numRows 10
               partition_columns ds/hr
               serialization.ddl struct srcpartbucket { string key, string 
value}
               serialization.format 1
               serialization.lib 
org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
-              totalSize 2748
-              transient_lastDdlTime 1297386150
+              totalSize 687
+              transient_lastDdlTime 1302062777
             serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
           
               input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
@@ -248,7 +248,7 @@ STAGE PLANS:
                 columns.types string:string
                 file.inputformat org.apache.hadoop.hive.ql.io.RCFileInputFormat
                 file.outputformat 
org.apache.hadoop.hive.ql.io.RCFileOutputFormat
-                location 
pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcpartbucket
+                location 
pfile:/data/users/nzhang/work/3/apache-hive/build/ql/test/data/warehouse/srcpartbucket
                 name default.srcpartbucket
                 numFiles 16
                 numPartitions 4
@@ -258,11 +258,11 @@ STAGE PLANS:
                 serialization.format 1
                 serialization.lib 
org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
                 totalSize 2748
-                transient_lastDdlTime 1297386150
+                transient_lastDdlTime 1302062777
               serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
               name: default.srcpartbucket
             name: default.srcpartbucket
-        
pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-09/hr=12/000000_0
 
+        
pfile:/data/users/nzhang/work/3/apache-hive/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-09/hr=12/000000_0
 
           Partition
             base file name: 000000_0
             input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
@@ -277,17 +277,17 @@ STAGE PLANS:
               columns.types string:string
               file.inputformat org.apache.hadoop.hive.ql.io.RCFileInputFormat
               file.outputformat org.apache.hadoop.hive.ql.io.RCFileOutputFormat
-              location 
pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-09/hr=12
+              location 
pfile:/data/users/nzhang/work/3/apache-hive/build/ql/test/data/warehouse/srcpartbucket/ds=2008-04-09/hr=12
               name default.srcpartbucket
-              numFiles 16
+              numFiles 4
               numPartitions 4
-              numRows 40
+              numRows 10
               partition_columns ds/hr
               serialization.ddl struct srcpartbucket { string key, string 
value}
               serialization.format 1
               serialization.lib 
org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
-              totalSize 2748
-              transient_lastDdlTime 1297386150
+              totalSize 687
+              transient_lastDdlTime 1302062777
             serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
           
               input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat
@@ -299,7 +299,7 @@ STAGE PLANS:
                 columns.types string:string
                 file.inputformat org.apache.hadoop.hive.ql.io.RCFileInputFormat
                 file.outputformat 
org.apache.hadoop.hive.ql.io.RCFileOutputFormat
-                location 
pfile:/data/users/sdong/www/open-source-hive1/build/ql/test/data/warehouse/srcpartbucket
+                location 
pfile:/data/users/nzhang/work/3/apache-hive/build/ql/test/data/warehouse/srcpartbucket
                 name default.srcpartbucket
                 numFiles 16
                 numPartitions 4
@@ -309,7 +309,7 @@ STAGE PLANS:
                 serialization.format 1
                 serialization.lib 
org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
                 totalSize 2748
-                transient_lastDdlTime 1297386150
+                transient_lastDdlTime 1302062777
               serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe
               name: default.srcpartbucket
             name: default.srcpartbucket
@@ -333,9 +333,9 @@ STAGE PLANS:
             File Output Operator
               compressed: false
               GlobalTableId: 0
-              directory: 
file:/tmp/sdong/hive_2011-02-10_17-02-30_127_2856183230982405624/-ext-10001
+              directory: 
file:/tmp/nzhang/hive_2011-04-05_21-06-18_053_445935565751534994/-ext-10001
               NumFilesPerFileSink: 1
-              Stats Publishing Key Prefix: 
file:/tmp/sdong/hive_2011-02-10_17-02-30_127_2856183230982405624/-ext-10001/
+              Stats Publishing Key Prefix: 
file:/tmp/nzhang/hive_2011-04-05_21-06-18_053_445935565751534994/-ext-10001/
               table:
                   input format: org.apache.hadoop.mapred.TextInputFormat
                   output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -358,14 +358,14 @@ PREHOOK: Input: default@srcpartbucket@ds
 PREHOOK: Input: default@srcpartbucket@ds=2008-04-08/hr=12
 PREHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=11
 PREHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=12
-PREHOOK: Output: 
file:/tmp/sdong/hive_2011-02-10_17-02-31_052_5780136652498214851/-mr-10000
+PREHOOK: Output: 
file:/tmp/nzhang/hive_2011-04-05_21-06-18_723_3947433659053651214/-mr-10000
 POSTHOOK: query: select ds, count(1) from srcpartbucket tablesample (bucket 1 
out of 4 on key) where ds is not null group by ds
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@srcpartbucket@ds=2008-04-08/hr=11
 POSTHOOK: Input: default@srcpartbucket@ds=2008-04-08/hr=12
 POSTHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=11
 POSTHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=12
-POSTHOOK: Output: 
file:/tmp/sdong/hive_2011-02-10_17-02-31_052_5780136652498214851/-mr-10000
+POSTHOOK: Output: 
file:/tmp/nzhang/hive_2011-04-05_21-06-18_723_3947433659053651214/-mr-10000
 POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-08,hr=11).key SIMPLE 
[(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
 POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-08,hr=11).value SIMPLE 
[(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
 POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-08,hr=12).key SIMPLE 
[(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
@@ -382,14 +382,14 @@ PREHOOK: Input: default@srcpartbucket@ds
 PREHOOK: Input: default@srcpartbucket@ds=2008-04-08/hr=12
 PREHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=11
 PREHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=12
-PREHOOK: Output: 
file:/tmp/sdong/hive_2011-02-10_17-02-35_963_7688294481780262172/-mr-10000
+PREHOOK: Output: 
file:/tmp/nzhang/hive_2011-04-05_21-06-23_117_8234913390986277646/-mr-10000
 POSTHOOK: query: select ds, count(1) from srcpartbucket tablesample (bucket 1 
out of 2 on key) where ds is not null group by ds
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@srcpartbucket@ds=2008-04-08/hr=11
 POSTHOOK: Input: default@srcpartbucket@ds=2008-04-08/hr=12
 POSTHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=11
 POSTHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=12
-POSTHOOK: Output: 
file:/tmp/sdong/hive_2011-02-10_17-02-35_963_7688294481780262172/-mr-10000
+POSTHOOK: Output: 
file:/tmp/nzhang/hive_2011-04-05_21-06-23_117_8234913390986277646/-mr-10000
 POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-08,hr=11).key SIMPLE 
[(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
 POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-08,hr=11).value SIMPLE 
[(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
 POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-08,hr=12).key SIMPLE 
[(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
@@ -406,14 +406,14 @@ PREHOOK: Input: default@srcpartbucket@ds
 PREHOOK: Input: default@srcpartbucket@ds=2008-04-08/hr=12
 PREHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=11
 PREHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=12
-PREHOOK: Output: 
file:/tmp/sdong/hive_2011-02-10_17-02-42_219_2445297351151986459/-mr-10000
+PREHOOK: Output: 
file:/tmp/nzhang/hive_2011-04-05_21-06-28_386_4655661597112299228/-mr-10000
 POSTHOOK: query: select * from srcpartbucket where ds is not null
 POSTHOOK: type: QUERY
 POSTHOOK: Input: default@srcpartbucket@ds=2008-04-08/hr=11
 POSTHOOK: Input: default@srcpartbucket@ds=2008-04-08/hr=12
 POSTHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=11
 POSTHOOK: Input: default@srcpartbucket@ds=2008-04-09/hr=12
-POSTHOOK: Output: 
file:/tmp/sdong/hive_2011-02-10_17-02-42_219_2445297351151986459/-mr-10000
+POSTHOOK: Output: 
file:/tmp/nzhang/hive_2011-04-05_21-06-28_386_4655661597112299228/-mr-10000
 POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-08,hr=11).key SIMPLE 
[(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
 POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-08,hr=11).value SIMPLE 
[(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
 POSTHOOK: Lineage: srcpartbucket PARTITION(ds=2008-04-08,hr=12).key SIMPLE 
[(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]

Modified: 
hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/SerDeUtils.java
URL: 
http://svn.apache.org/viewvc/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/SerDeUtils.java?rev=1089938&r1=1089937&r2=1089938&view=diff
==============================================================================
--- hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/SerDeUtils.java 
(original)
+++ hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/SerDeUtils.java Thu 
Apr  7 16:32:18 2011
@@ -19,9 +19,9 @@
 package org.apache.hadoop.hive.serde2;
 
 import java.util.ArrayList;
-import java.util.HashMap;
 import java.util.List;
 import java.util.Map;
+import java.util.concurrent.ConcurrentHashMap;
 
 import org.apache.hadoop.hive.common.JavaUtils;
 import org.apache.hadoop.hive.serde2.objectinspector.ListObjectInspector;
@@ -54,7 +54,8 @@ public final class SerDeUtils {
   public static final String LBRACE = "{";
   public static final String RBRACE = "}";
 
-  private static HashMap<String, Class<?>> serdes = new HashMap<String, 
Class<?>>();
+  private static ConcurrentHashMap<String, Class<?>> serdes =
+    new ConcurrentHashMap<String, Class<?>>();
 
   public static void registerSerDe(String name, Class<?> serde) {
     if (serdes.containsKey(name)) {


Reply via email to