Fantastic, I had a look at the hdfs-site.xml and realised the port was 54310.
I was able to turn OFF the safe mode, and have just deleted the file. Thank you very much indeed Wang. Rob. 2010/1/14 Wang Xu <[email protected]> > Hi Rob, > > How about the "fs.default.name" in your configuration files, and can > you find NameNode process with jps? > > On Thu, Jan 14, 2010 at 10:16 PM, Rob Stewart > <[email protected]> wrote: > > Hi Wang, > > > >> hadoop dfsadmin -report > > > > returns nothing at all. > > > > I have tried: > >> hadoop dfsadmin -fs hdfs://hadoopmaster/ -safemode leave > > > > 10/01/14 14:11:24 INFO ipc.Client: Retrying connect to server: > hadoopmaster/ > > 137.195.143.132:8020. Already tried 0 time(s). > > 10/01/14 14:11:25 INFO ipc.Client: Retrying connect to server: > hadoopmaster/ > > 137.195.143.132:8020. Already tried 1 time(s). > > etc...... > > > > ?? > > > > thanks.. > > > > > > > > 2010/1/14 Wang Xu <[email protected]> > > > >> Hi Rob, > >> > >> What do you get if you use > >> hadoop dfsadmin -report > >> ? > >> > >> And you can specify the fs in command line when you call dfsadmin. > >> > >> On Thu, Jan 14, 2010 at 9:41 PM, Rob Stewart > >> <[email protected]> wrote: > >> > Hi, > >> > > >> > I'm having a slight issue with my Hadoop cluster. There are 32 nodes. > I > >> > have: > >> > /usr/lib/hadoop/bin/stop-mapred.sh > >> > /usr/lib/hadoop/bin/stop-dfs.sh > >> > /usr/lib/hadoop/bin/start-dfs.sh > >> > /usr/lib/hadoop/bin/start-mapred.sh > >> > > >> > All worked perfectly, no errors. > >> > > >> > I try and remove a file: hadoop dfs -rmr wordCountOutput.jaql > >> > > >> > This returns: Cannot delete [the file]. Name node is in safe mode. > >> > > >> > I tried to exit safemode manually. I get: > >> >> hadoop dfsadmin -safemode leave > >> > FileSystem is file:/// > >> > > >> > The code generating this is: > >> > > >> > > >> > public void setSafeMode(String[] argv, int idx) throws IOException { > >> > final String safeModeUsage = "Usage: java DFSAdmin -safemode " > >> > + "[enter | leave | get]"; > >> > if (!(fs instanceof DistributedFileSystem)) { > >> > System.out.println("FileSystem is " + fs.getName()); > >> > return; > >> > } > >> > > >> > > >> > What seems to be the issue? the HDFS is running, and I can browse the > >> > filesystem, both via the command line and the Hadoop web interface. > >> > > >> > > >> > Thanks... > >> > > >> > >> > >> > >> -- > >> Wang Xu > >> Stephen Leacock - "I detest life-insurance agents: they always argue > >> that I shall some day die, which is not so." - > >> http://www.brainyquote.com/quotes/authors/s/stephen_leacock.html > >> > > > > > > -- > Wang Xu > Mike Ditka - "If God had wanted man to play soccer, he wouldn't have > given us arms." - > http://www.brainyquote.com/quotes/authors/m/mike_ditka.html >
