Moving to hdfs-user@, bcc mapreduce-user@. 

You NameNode isn't coming out of safemode since all the datanodes haven't 
rejoined the cluster...

> The ratio of reported blocks 0.2915 has not reached the threshold 0.9990. 
> Safe mode will be turned off automatically.

Can you check why your datanodes aren't joining?

Arun

On Jul 27, 2011, at 12:28 AM, 周俊清 wrote:

> hello everyone,
>     I got an exception from my jobtracker's log file as follow:
> 2011-07-27 01:58:04,197 INFO org.apache.hadoop.mapred.JobTracker: Cleaning up 
> the system directory
> 2011-07-27 01:58:04,230 INFO org.apache.hadoop.mapred.JobTracker: problem 
> cleaning system directory: 
> hdfs://dn224.pengyun.org:56900/home/hadoop/hadoop-tmp203/mapred/system
> org.apache.hadoop.ipc.RemoteException: 
> org.apache.hadoop.hdfs.server.namenode.SafeModeException: Cannot delete 
> /home/hadoop/hadoop-tmp203/mapred/system. Name node is in safe mode.
> The ratio of reported blocks 0.2915 has not reached the threshold 0.9990. 
> Safe mode will be turned off automatically.
>     at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInternal(FSNamesystem.java:1851)
>     at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.delete(FSNamesystem.java:1831)
>     at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.delete(NameNode.java:691)
>     ……
> and 
>    the log message of namenode:
> 2011-07-27 00:00:00,219 INFO org.apache.hadoop.ipc.Server: IPC Server handler 
> 1 on 56900, call delete(/home/hadoop/hadoop-tmp203/mapred/system, true) from 
> 192.168.1.224:5131
> 2: error: org.apache.hadoop.hdfs.server.namenode.SafeModeException: Cannot 
> delete /home/hadoop/hadoop-tmp203/mapred/system. Name node is in safe mode.
> The ratio of reported blocks 0.2915 has not reached the threshold 0.9990. 
> Safe mode will be turned off automatically.
> org.apache.hadoop.hdfs.server.namenode.SafeModeException: Cannot delete 
> /home/hadoop/hadoop-tmp203/mapred/system. Name node is in safe mode.
> The ratio of reported blocks 0.2915 has not reached the threshold 0.9990. 
> Safe mode will be turned off automatically.
>     at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInternal(FSNamesystem.java:1851)
>     at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.delete(FSNamesystem.java:1831)
>     at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.delete(NameNode.java:691)
>     at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source)
>     at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>     at java.lang.reflect.Method.invoke(Method.java:597)
>     at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:523)
>     at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1383)
>     at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1379)
>     at java.security.AccessController.doPrivileged(Native Method)
>     at javax.security.auth.Subject.doAs(Subject.java:396)
>     at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1059)
>     at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1377)
> 
>  It means,I think,that the namenode is always being in the safe mode,so what 
> can i do about these exception.Anyone who can tell me why?I don't find the 
> file "/home/hadoop/hadoop-tmp203/mapred/system" in my system.The exception 
> upon which appearing in the log file are repeated,even when I restart my 
> hadoop.
>    thanks for your concern.
> 
> 
> ----------------------------
> Junqing Zhou
> 2ho...@163.com
> 
> 
> 
> 

Reply via email to