Thanks for your help. I ended up reformatting everything after backing up
the data.
After reinstalling the cluster, I still have problems with hadoop commands
hanging forever before returning the command line prompt.
Also, if I ctrl-Z and bg a hadoop command, then it finishes very quickly,
if I leave running in the foreground, then it hangs before returning the
prompt.
Same thing happens for a pseudo-distributed cluster installation on one
machine.

Juan

On Sat, Jul 14, 2012 at 4:23 PM, Edward Capriolo <edlinuxg...@gmail.com>wrote:

> If the files are gone forever you should run:
>
> hadoop fsck -delete /
>
> To acknowledge they have moved on from existence. Otherwise things
> that attempt to read this files will, to put it in a technical way,
> BARF.
>
> On Fri, Jul 13, 2012 at 12:22 PM, Juan Pino <juancitomiguel...@gmail.com>
> wrote:
> > Thank you for your reply. I ran that command before and it works fine but
> > hadoop fs -ls diplays the list of files in the user's directory but then
> > hangs for quite a while (~ 10 minutes) before
> > handing the command line prompt back, then if I rerun the same command
> > there is no problem. That is why I would like to be able to leave safe
> mode
> > automatically (at least I think it's related).
> > Also, in the hdfs web page, clicking on the Live Nodes or Dead Nodes
> links
> > hangs forever but I am able to browse the file
> > system without any problem with the browser.
> > There is no error in the logs.
> > Please let me know what sort of details I can provide to help resolve
> this
> > issue.
> >
> > Best,
> >
> > Juan
> >
> > On Fri, Jul 13, 2012 at 4:10 PM, Edward Capriolo <edlinuxg...@gmail.com
> >wrote:
> >
> >> If the datanode is not coming back you have to explicitly tell hadoop
> >> to leave safemode.
> >>
> >>
> http://hadoop.apache.org/common/docs/r0.17.2/hdfs_user_guide.html#Safemode
> >>
> >> hadoop dfsadmin -safemode leave
> >>
> >>
> >> On Fri, Jul 13, 2012 at 9:35 AM, Juan Pino <juancitomiguel...@gmail.com
> >
> >> wrote:
> >> > Hi,
> >> >
> >> > I can't get HDFS to leave safe mode automatically. Here is what I did:
> >> >
> >> > -- there was a dead node
> >> > -- I stopped dfs
> >> > -- I restarted dfs
> >> > -- Safe mode wouldn't leave automatically
> >> >
> >> > I am using hadoop-1.0.2
> >> >
> >> > Here are the logs:
> >> >
> >> > end of hadoop-hadoop-namenode.log (attached):
> >> >
> >> > 2012-07-13 13:22:29,372 INFO org.apache.hadoop.hdfs.StateChange:
> STATE*
> >> Safe
> >> > mode ON.
> >> > The ratio of reported blocks 0.9795 has not reached the threshold
> 0.9990.
> >> > Safe mode will be turned off automatically.
> >> > 2012-07-13 13:22:29,375 INFO org.apache.hadoop.hdfs.StateChange:
> STATE*
> >> Safe
> >> > mode extension entered.
> >> > The ratio of reported blocks 0.9990 has reached the threshold 0.9990.
> >> Safe
> >> > mode will be turned off automatically in 29 seconds.
> >> > 2012-07-13 13:22:29,375 INFO org.apache.hadoop.hdfs.StateChange:
> *BLOCK*
> >> > NameSystem.processReport: from , blocks: 3128, processing time: 4
> msecs
> >> > 2012-07-13 13:31:29,201 INFO org.apache.hadoop.hdfs.StateChange:
> BLOCK*
> >> > NameSystem.processReport: discarded non-initial block report from
> because
> >> > namenode still in startup phase
> >> >
> >> > Any help would be greatly appreciated.
> >> >
> >> > Best,
> >> >
> >> > Juan
> >> >
> >>
>

Reply via email to