Forgot about this one

hive> describe extended aggregations;
OK
value   int
account string
application     string
dataset string
hour    int
aggregation     string
aggregated_by   string
Detailed Table Information:
Table(tableName:aggregations,dbName:default,owner:Josh,createTime: 1228463486,lastAccessTime:0,retention:0,sd:StorageDescriptor(cols: [FieldSchema(name:value,type:int,comment:null)],location:/user/hive/ warehouse/ aggregations ,inputFormat:org .apache .hadoop .mapred .TextInputFormat ,outputFormat:org .apache .hadoop .hive .ql .io .IgnoreKeyTextOutputFormat ,compressed:false ,numBuckets :-1 ,serdeInfo:SerDeInfo (name:null ,serializationLib:org .apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe,parameters: {colelction .delim = 44 ,mapkey .delim = 58 ,serialization .format = org .apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol}),bucketCols: [],sortCols:[],parameters:{}),partitionKeys: [FieldSchema(name:account,type:string,comment:null), FieldSchema(name:application,type:string,comment:null), FieldSchema(name:dataset,type:string,comment:null), FieldSchema(name:hour,type:int,comment:null), FieldSchema(name:aggregation,type:string,comment:null), FieldSchema(name:aggregated_by,type:string,comment:null)],parameters:{})
Time taken: 2.884 seconds

On Dec 10, 2008, at 8:31 PM, Josh Ferguson wrote:

hive> SELECT COUNT(actor_id) AS value FROM activities WHERE (
> account='80c27664-b047-4c0a-86f3-342c0cdf36c7' AND application='myproduct'
   > AND dataset='purchase' AND hour=341165 );

This actually works fine, it's writing to the new table that is broken.

hive> describe extended activities;
OK
actor_id        string
actee_id        string
properties      map<string,string>
account string
application     string
dataset string
hour    int
Detailed Table Information:
Table(tableName:activities,dbName:default,owner:Josh,createTime: 1228208598,lastAccessTime:0,retention:0,sd:StorageDescriptor(cols: [FieldSchema(name:actor_id,type:string,comment:null), FieldSchema(name:actee_id,type:string,comment:null), FieldSchema (name:properties,type:map<string,string>,comment:null)],location:/ user/hive/warehouse/ activities ,inputFormat:org .apache .hadoop .mapred .TextInputFormat ,outputFormat:org .apache .hadoop .hive.ql.io.IgnoreKeyTextOutputFormat,compressed:false,numBuckets: 32 ,serdeInfo:SerDeInfo (name:null ,serializationLib:org .apache.hadoop.hive.serde2.dynamic_type.DynamicSerDe,parameters: {colelction .delim = 44 ,mapkey .delim = 58 ,serialization .format = org .apache.hadoop.hive.serde2.thrift.TCTLSeparatedProtocol}),bucketCols: [actor_id, actee_id],sortCols:[],parameters:{}),partitionKeys: [FieldSchema(name:account,type:string,comment:null), FieldSchema(name:application,type:string,comment:null), FieldSchema(name:dataset,type:string,comment:null), FieldSchema(name:hour,type:int,comment:null)],parameters:{})
Time taken: 2.656 seconds

I' not sure what describe extended partition is, doesn't work for me

Josh

On Dec 10, 2008, at 2:31 PM, Raghu Murthy wrote:

Can you check the output of the query without the insert clause?

SELECT COUNT(actor_id) AS value FROM activities WHERE (
account='80c27664-b047-4c0a-86f3-342c0cdf36c7' AND application='myproduct'
AND dataset='purchase' AND hour=341165 )

Is it empty?


On 12/10/08 1:08 AM, "Josh Ferguson" <[EMAIL PROTECTED]> wrote:

I should say this is during the last reduction step that occures in this query

Josh

On Dec 10, 2008, at 1:06 AM, Josh Ferguson wrote:

This Query

INSERT OVERWRITE TABLE aggregations PARTITION (
account='80c27664-b047-4c0a-86f3-342c0cdf36c7', application='myproduct', dataset='purchase', hour=341165, aggregation='count', aggregated_by='all' )
SELECT COUNT(actor_id) AS value FROM activities WHERE (
account='80c27664-b047-4c0a-86f3-342c0cdf36c7' AND application='myproduct'
AND
dataset='purchase' AND hour=341165 )

Generates this message

java.lang.RuntimeException: Error while closing operators:
org.apache.hadoop.hive.ql.metadata.HiveException:
org.apache.hadoop.hive.serde2.SerDeException:
org.apache.hadoop.hive.serde2.SerDeException: Trying to serialize 0 fields
into a struct with 1
at org.apache.hadoop.hive.ql.exec.ExecReducer.close(ExecReducer.java: 202)
at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:440)
at org.apache.hadoop.mapred.Child.main(Child.java:155)


Josh Ferguson




Reply via email to