Hi Bennie,
Can you post your hadoop version and hive version?

Thanks
Yongqiang


On 2/5/10 10:05 AM, "Zheng Shao" <[email protected]> wrote:

> That seems to be a bug.
> Are you using hive trunk or any release?
> 
> 
> On 2/5/10, Bennie Schut <[email protected]> wrote:
>> I have a tab separated files I have loaded it with "load data inpath"
>> then I do a
>> 
>> SET hive.exec.compress.output=true;
>> SET mapred.output.compression.codec=com.hadoop.compression.lzo.LzoCodec;
>> SET mapred.map.output.compression.codec=com.hadoop.compression.lzo.LzoCodec;
>> select distinct login_cldr_id as cldr_id from chatsessions_load;
>> 
>> Ended Job = job_201001151039_1641
>> OK
>> NULL
>> NULL
>> NULL
>> Time taken: 49.06 seconds
>> 
>> however if I start it without the set commands I get this:
>> Ended Job = job_201001151039_1642
>> OK
>> 2283
>> Time taken: 45.308 seconds
>> 
>> Which is the correct result.
>> 
>> When I do a "insert overwrite" on a rcfile table it will actually
>> compress the data correctly.
>> When I disable compression and query this new table the result is correct.
>> When I enable compression it's wrong again.
>> I see no errors in the logs.
>> 
>> Any idea's why this might happen?
>> 
>> 
>> 


Reply via email to