This is is happening on PIG 0.7 as well. When I use PigServer to launch pig jobs from Java lots of tmp files (e.g. directory /tmp/temp-203305867) gets created that are not deleted at the end of the Job.
"hadoop.tmp.dir" points to some other directory "/tmp/hadoop-userA". Any ideas on how to get path to these tmp folders so that I can explicitly delete them at the end of the Job ? Thanks, -Rakesh > Date: Mon, 25 Oct 2010 20:30:24 +0200 > From: [email protected] > To: [email protected] > Subject: Re: How to make PIG delete its temporary files ? > > Thanks for your answer, I will investigate on that. > > Le 25/10/10 08:26, Rekha Joshi a écrit : > > Hi Vincent, > > > > I recall it happens under some specific condition and is resolved in later > > versions of pig/hadoop by adding an explicit cleanup before final exit. > > As a workaround in 0.6 you can code finally{} in your java file to remove > > files from your user specific folder under hadoop.tmp.dir.HTH! > > > > Thanks& Regards, > > /Rekha. > > > > On 10/24/10 5:24 AM, "Vincent Barat"<[email protected]> wrote: > > > > I forgot to mention that I use PIG 0.6.0, it may help ! > > > > Le 23/10/10 13:29, Vincent Barat a écrit : > >> Hello, > >> > >> I face an issue with PIG temporary files: they are not deleted > >> once a job is terminated. > >> I got my HDFS storage full of PIG temporary files. > >> > >> I use PIG from Java using a PigServer object. > >> > >> Is there is way to ask PIG to cleanup all that it created to > >> perform a job ? > >> > >> Thanks a lot > >
