Prashant,

You may be able to salvage some free temp-data space (if used up by
earlier maps or reduces) from shared mapred.local.dir directories by
restarting the TaskTracker. Also clear out unwanted items from /tmp
(which is what may be blocking your JVMs). Or if you have another free
disk, set HADOOP_OPTS in hadoop-env.sh as
"-Djava.io.tmpdir=/path/to/dir/on/free/disk" and make sure the chosen
path has permissions 777, and then once HDFS is up delete away some
unwanted files.

On Tue, May 22, 2012 at 6:13 PM, shan s <mysub...@gmail.com> wrote:
> My cluster went out of disk during a job execution and now shows occupied
> disk space = 100%. When I attempt hadoop fs -ls input, I get below.
>
> Java HotSpot(TM) 64-Bit Server VM warning: Insufficient space for shared
> memory file:   /tmp/hsperfdata_user/32062
> Try using the -Djava.io.tmpdir= option to select an alternate temp location.
> Error: tmp.h sperfdata Could not find or load main class ___._user.32074
> Is there an easy way to come out of this and be able to delete hdfs files,
> to get it back to working.
> Thanks, Prashant.



-- 
Harsh J

Reply via email to