That seems to be a bug. Are you using hive trunk or any release?
On 2/5/10, Bennie Schut <[email protected]> wrote: > I have a tab separated files I have loaded it with "load data inpath" > then I do a > > SET hive.exec.compress.output=true; > SET mapred.output.compression.codec=com.hadoop.compression.lzo.LzoCodec; > SET mapred.map.output.compression.codec=com.hadoop.compression.lzo.LzoCodec; > select distinct login_cldr_id as cldr_id from chatsessions_load; > > Ended Job = job_201001151039_1641 > OK > NULL > NULL > NULL > Time taken: 49.06 seconds > > however if I start it without the set commands I get this: > Ended Job = job_201001151039_1642 > OK > 2283 > Time taken: 45.308 seconds > > Which is the correct result. > > When I do a "insert overwrite" on a rcfile table it will actually > compress the data correctly. > When I disable compression and query this new table the result is correct. > When I enable compression it's wrong again. > I see no errors in the logs. > > Any idea's why this might happen? > > > -- Sent from my mobile device Yours, Zheng
