Moving to hdfs-user@, bcc mapreduce-user@. 

You NameNode isn't coming out of safemode since all the datanodes haven't rejoined the cluster...

The ratio of reported blocks 0.2915 has not reached the threshold 0.9990. Safe mode will be turned off automatically.

Can you check why your datanodes aren't joining?

Arun

On Jul 27, 2011, at 12:28 AM, 周俊清 wrote:

hello everyone,
    I got an exception from my jobtracker's log file as follow:
2011-07-27 01:58:04,197 INFO org.apache.hadoop.mapred.JobTracker: Cleaning up the system directory
2011-07-27 01:58:04,230 INFO org.apache.hadoop.mapred.JobTracker: problem cleaning system directory: hdfs://dn224.pengyun.org:56900/home/hadoop/hadoop-tmp203/mapred/system
org.apache.hadoop.ipc.RemoteException: org.apache.hadoop.hdfs.server.namenode.SafeModeException: Cannot delete /home/hadoop/hadoop-tmp203/mapred/system. Name node is in safe mode.
The ratio of reported blocks 0.2915 has not reached the threshold 0.9990. Safe mode will be turned off automatically.
    at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInternal(FSNamesystem.java:1851)
    at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.delete(FSNamesystem.java:1831)
    at org.apache.hadoop.hdfs.server.namenode.NameNode.delete(NameNode.java:691)
    ……
and
   the log message of namenode:
2011-07-27 00:00:00,219 INFO org.apache.hadoop.ipc.Server: IPC Server handler 1 on 56900, call delete(/home/hadoop/hadoop-tmp203/mapred/system, true) from 192.168.1.224:5131
2: error: org.apache.hadoop.hdfs.server.namenode.SafeModeException: Cannot delete /home/hadoop/hadoop-tmp203/mapred/system. Name node is in safe mode.
The ratio of reported blocks 0.2915 has not reached the threshold 0.9990. Safe mode will be turned off automatically.
org.apache.hadoop.hdfs.server.namenode.SafeModeException: Cannot delete /home/hadoop/hadoop-tmp203/mapred/system. Name node is in safe mode.
The ratio of reported blocks 0.2915 has not reached the threshold 0.9990. Safe mode will be turned off automatically.
    at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInternal(FSNamesystem.java:1851)
    at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.delete(FSNamesystem.java:1831)
    at org.apache.hadoop.hdfs.server.namenode.NameNode.delete(NameNode.java:691)
    at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:523)
    at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1383)
    at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1379)
    at java.security.AccessController.doPrivileged(Native Method)
    at javax.security.auth.Subject.doAs(Subject.java:396)
    at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1059)
    at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1377)

 It means,I think,that the namenode is always being in the safe mode,so what can i do about these exception.Anyone who can tell me why?I don't find the file "/home/hadoop/hadoop-tmp203/mapred/system" in my system.The exception upon which appearing in the log file are repeated,even when I restart my hadoop.
   thanks for your concern.



----------------------------
Junqing Zhou