Lets just start with simple steps to reproduce this. If that does not work then we can look at the java stuff...
Ashish -----Original Message----- From: Saurabh Nanda [mailto:[email protected]] Sent: Thursday, August 20, 2009 10:34 PM To: [email protected] Subject: Re: Output compression not working on hive-trunk (r802989) How do i do that? Just steps to reproduce the error or a formal java based test case? On 8/21/09, Ashish Thusoo <[email protected]> wrote: > Hi Saurabh, > > Can you give a simple reproducible test case for this (unless you have > already done so) ? > > Thanks, > Ashish > > ________________________________ > From: Saurabh Nanda [mailto:[email protected]] > Sent: Thursday, August 20, 2009 12:43 PM > To: [email protected] > Subject: Re: Output compression not working on hive-trunk (r802989) > > Is anyone else facing this issue? > > On Wed, Aug 19, 2009 at 11:16 AM, Saurabh Nanda > <[email protected]<mailto:[email protected]>> wrote: > Any clue? Has anyone else tried to replicate this? Is this really a > bug or am I doing something obviously stupid? > > Saurabh. > > > On Tue, Aug 18, 2009 at 3:51 PM, Saurabh Nanda > <[email protected]<mailto:[email protected]>> wrote: > Is this what you're talking about -- http://pastebin.ca/1533627 ? > Seems like compression is on. > > Is there any difference in how CLI queries and JDBC queries are treated? > > Saurabh. > > > On Tue, Aug 18, 2009 at 11:19 AM, Zheng Shao > <[email protected]<mailto:[email protected]>> wrote: > Hi Saurabh, > > So the compression flag is correct when the plan is generated. > When you run the query, you should see "plan = xxx.xml" in the log > file. Can you open that file (in HDFS) and see whether the compression > flag is on or not? > > Zheng > > On Mon, Aug 17, 2009 at 5:17 AM, Saurabh > Nanda<[email protected]<mailto:[email protected]>> wrote: >> Hey Zheng, any clues as to what the bug is? Or what I'm doing wrong? >> I can do some more digging and logging if required. >> >> Saurabh. >> >> On Mon, Aug 17, 2009 at 1:28 PM, Saurabh Nanda >> <[email protected]<mailto:[email protected]>> >> wrote: >>> >>> Here's the log output: >>> >>> 2009-08-17 13:26:42,183 INFO parse.SemanticAnalyzer >>> (SemanticAnalyzer.java:genFileSinkPlan(2575)) - Created FileSink >>> Plan for >>> clause: insclause-0dest_path: >>> hdfs://master-hadoop/user/hive/warehouse/raw/dt=2009-04-07 row schema: >>> {(_col0,_col0: string)(_col1,_col1: string)(_col2,_col2: >>> string)(_col3,_col3: string)(_col4,_col4: string)(_col5,_col5: >>> string)(_col6,_col6: string)(_col7,_col7: string)(_col8,_col8: >>> string)(_col9,_col9: string)(_col10,_col10: int)} . >>> HiveConf.ConfVars.COMPRESSRESULT=true >>> >>> Is the SemanticAnalyszer run more than once in the lifetime of a job? >>> Should I be looking for another log entry like this one? >>> >>> Saurabh. >>> >>> On Mon, Aug 17, 2009 at 1:26 PM, Saurabh Nanda >>> <[email protected]<mailto:[email protected]>> >>> wrote: >>>> >>>> Strange. The compression configuration log entry was also info but >>>> I could see it in the task logs: >>>> >>>> LOG.info("Compression configuration is:" + isCompressed); >>>> >>>> Saurabh. >>>> >>>> On Mon, Aug 17, 2009 at 12:56 PM, Zheng Shao >>>> <[email protected]<mailto:[email protected]>> wrote: >>>>> >>>>> The default log level is WARN. Please change it to INFO. >>>>> >>>>> hive.root.logger=INFO,DRFA >>>>> >>>>> Of course you can also use LOG.warn() in your test code. >>>>> >>>>> Zheng >>>>> >>>> -- >>>> http://nandz.blogspot.com >>>> http://foodieforlife.blogspot.com >>> >>> >>> >>> -- >>> http://nandz.blogspot.com >>> http://foodieforlife.blogspot.com >> >> >> >> -- >> http://nandz.blogspot.com >> http://foodieforlife.blogspot.com >> > > > > -- > Yours, > Zheng > > > > -- > http://nandz.blogspot.com > http://foodieforlife.blogspot.com > > > > -- > http://nandz.blogspot.com > http://foodieforlife.blogspot.com > > > > -- > http://nandz.blogspot.com > http://foodieforlife.blogspot.com > -- http://nandz.blogspot.com http://foodieforlife.blogspot.com
