Hi, I think the problem was not hadoop related. After a reboot of the nodes, everything is fine.
Thanks, Juan On Mon, Jul 16, 2012 at 4:38 PM, Juan Pino <juancitomiguel...@gmail.com>wrote: > Thanks for your help. I ended up reformatting everything after backing up > the data. > After reinstalling the cluster, I still have problems with hadoop commands > hanging forever before returning the command line prompt. > Also, if I ctrl-Z and bg a hadoop command, then it finishes very quickly, > if I leave running in the foreground, then it hangs before returning the > prompt. > Same thing happens for a pseudo-distributed cluster installation on one > machine. > > Juan > > On Sat, Jul 14, 2012 at 4:23 PM, Edward Capriolo <edlinuxg...@gmail.com>wrote: > >> If the files are gone forever you should run: >> >> hadoop fsck -delete / >> >> To acknowledge they have moved on from existence. Otherwise things >> that attempt to read this files will, to put it in a technical way, >> BARF. >> >> On Fri, Jul 13, 2012 at 12:22 PM, Juan Pino <juancitomiguel...@gmail.com> >> wrote: >> > Thank you for your reply. I ran that command before and it works fine >> but >> > hadoop fs -ls diplays the list of files in the user's directory but then >> > hangs for quite a while (~ 10 minutes) before >> > handing the command line prompt back, then if I rerun the same command >> > there is no problem. That is why I would like to be able to leave safe >> mode >> > automatically (at least I think it's related). >> > Also, in the hdfs web page, clicking on the Live Nodes or Dead Nodes >> links >> > hangs forever but I am able to browse the file >> > system without any problem with the browser. >> > There is no error in the logs. >> > Please let me know what sort of details I can provide to help resolve >> this >> > issue. >> > >> > Best, >> > >> > Juan >> > >> > On Fri, Jul 13, 2012 at 4:10 PM, Edward Capriolo <edlinuxg...@gmail.com >> >wrote: >> > >> >> If the datanode is not coming back you have to explicitly tell hadoop >> >> to leave safemode. >> >> >> >> >> http://hadoop.apache.org/common/docs/r0.17.2/hdfs_user_guide.html#Safemode >> >> >> >> hadoop dfsadmin -safemode leave >> >> >> >> >> >> On Fri, Jul 13, 2012 at 9:35 AM, Juan Pino < >> juancitomiguel...@gmail.com> >> >> wrote: >> >> > Hi, >> >> > >> >> > I can't get HDFS to leave safe mode automatically. Here is what I >> did: >> >> > >> >> > -- there was a dead node >> >> > -- I stopped dfs >> >> > -- I restarted dfs >> >> > -- Safe mode wouldn't leave automatically >> >> > >> >> > I am using hadoop-1.0.2 >> >> > >> >> > Here are the logs: >> >> > >> >> > end of hadoop-hadoop-namenode.log (attached): >> >> > >> >> > 2012-07-13 13:22:29,372 INFO org.apache.hadoop.hdfs.StateChange: >> STATE* >> >> Safe >> >> > mode ON. >> >> > The ratio of reported blocks 0.9795 has not reached the threshold >> 0.9990. >> >> > Safe mode will be turned off automatically. >> >> > 2012-07-13 13:22:29,375 INFO org.apache.hadoop.hdfs.StateChange: >> STATE* >> >> Safe >> >> > mode extension entered. >> >> > The ratio of reported blocks 0.9990 has reached the threshold 0.9990. >> >> Safe >> >> > mode will be turned off automatically in 29 seconds. >> >> > 2012-07-13 13:22:29,375 INFO org.apache.hadoop.hdfs.StateChange: >> *BLOCK* >> >> > NameSystem.processReport: from , blocks: 3128, processing time: 4 >> msecs >> >> > 2012-07-13 13:31:29,201 INFO org.apache.hadoop.hdfs.StateChange: >> BLOCK* >> >> > NameSystem.processReport: discarded non-initial block report from >> because >> >> > namenode still in startup phase >> >> > >> >> > Any help would be greatly appreciated. >> >> > >> >> > Best, >> >> > >> >> > Juan >> >> > >> >> >> > >