I thought this might be because that hadoop wants to pack everything
(including the -files dfs cache files) into one single jar, so I removed
the -files commands I have.
but it still extracts the jar. this is rather confusing
On Fri, Oct 24, 2014 at 11:51 AM, Yang tedd...@gmail.com wrote:
If you use 'hadoop jar' to invoke your application, this is the default
behaviour. The reason it is done is that the utility supports use of
jars-within-jar feature, that lets one pack additional dependency jars into
an application as a lib/ subdirectory under the root of the main jar.
It is not
I just noticed that when I run a hadoop jar
my-fat-jar-with-all-dependencies.jar , it unjars the job jar in
/tmp/hadoop-username/hadoop-unjar-/ and extracts all the classes in
there.
the fat jar is pretty big, so it took up a lot of space (particularly
inodes ) and ran out of quota.
I wonder