I run "hadoop fs -rmr .." immediately after start-all.sh Does the namenode always start in safemode and after sometime switches to normal mode ? If that is the problem then your suggestion of waiting might work. Lemme check.
-Thanks for the pointer. Prasen On Tue, Jan 19, 2010 at 10:47 AM, Amogh Vasekar <[email protected]> wrote: > Hi, > When NN is in safe mode, you get a read-only view of the hadoop file system. > ( since NN is reconstructing its image of FS ) > Use "hadoop dfsadmin -safemode get" to check if in safe mode. > "hadoop dfsadmin -safemode leave" to leave safe mode forcefully. Or use > "hadoop dfsadmin -safemode wait" to block till NN leaves by itself. > > Amogh > > > On 1/19/10 10:31 AM, "prasenjit mukherjee" <[email protected]> wrote: > > Hmmm. I am actually running it from a batch file. Is "hadoop fs -rmr" > not that stable compared to pig's rm OR hadoop's FileSystem ? > > Let me try your suggestion by writing a cleanup script in pig. > > -Thanks, > Prasen > > On Tue, Jan 19, 2010 at 10:25 AM, Rekha Joshi <[email protected]> wrote: >> Can you try with dfs/ without quotes?If using pig to run jobs you can use >> rmf within your script(again w/o quotes) to force remove and avoid error if >> file/dir not present.Or if doing this inside hadoop job, you can use >> FileSystem/FileStatus to delete directories.HTH. >> Cheers, >> /R >> >> On 1/19/10 10:15 AM, "prasenjit mukherjee" <[email protected]> wrote: >> >> "hadoop fs -rmr /op" >> >> That command always fails. I am trying to run sequential hadoop jobs. >> After the first run all subsequent runs fail while cleaning up ( aka >> removing the hadoop dir created by previous run ). What can I do to >> avoid this ? >> >> here is my hadoop version : >> # hadoop version >> Hadoop 0.20.0 >> Subversion https://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.20 >> -r 763504 >> Compiled by ndaley on Thu Apr 9 05:18:40 UTC 2009 >> >> Any help is greatly appreciated. >> >> -Prasen >> >> > >
