Re: why does hadoop creates /tmp/hadoop-user/hadoop-unjar-xxxx/ dir and unjar my fat jar?

2014-10-25 Thread Yang
I thought this might be because that hadoop wants to pack everything (including the -files dfs cache files) into one single jar, so I removed the -files commands I have. but it still extracts the jar. this is rather confusing On Fri, Oct 24, 2014 at 11:51 AM, Yang tedd...@gmail.com wrote:

Re: why does hadoop creates /tmp/hadoop-user/hadoop-unjar-xxxx/ dir and unjar my fat jar?

2014-10-25 Thread Harsh J
If you use 'hadoop jar' to invoke your application, this is the default behaviour. The reason it is done is that the utility supports use of jars-within-jar feature, that lets one pack additional dependency jars into an application as a lib/ subdirectory under the root of the main jar. It is not

why does hadoop creates /tmp/hadoop-user/hadoop-unjar-xxxx/ dir and unjar my fat jar?

2014-10-24 Thread Yang
I just noticed that when I run a hadoop jar my-fat-jar-with-all-dependencies.jar , it unjars the job jar in /tmp/hadoop-username/hadoop-unjar-/ and extracts all the classes in there. the fat jar is pretty big, so it took up a lot of space (particularly inodes ) and ran out of quota. I wonder