Hi everyone.
I installed hadoop among three pcs. When I ran the command 'start-all.sh', I 
only could start the jobtracker and tasktrackers. I use 192.*.*.x as master and 
use 192.*.*.y and 192.*.*.z as slaves.

The namenode log from the master 192.*.*.x is following like this:

2009-08-18 10:48:44,543 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* 
NameSystem.registerDatanode: node 192.*.*.y:50010 is replaced by 
192.*.*.x:50010 with the same storageID 
DS-1120429845-127.0.0.1-50010-1246697164684
2009-08-18 10:48:44,543 INFO org.apache.hadoop.net.NetworkTopology: Removing a 
node: /default-rack/192.*.*.y:50010
2009-08-18 10:48:44,543 INFO org.apache.hadoop.net.NetworkTopology: Adding a 
new node: /default-rack/192.*.*.x:50010
2009-08-18 10:48:45,932 FATAL org.apache.hadoop.hdfs.StateChange: BLOCK* 
NameSystem.getDatanode: Data node 192.*.*.z:50010 is attempting to report 
storage ID DS-1120429845-127.0.0.1-50010-1246697164684. Node 192.*.*.x:50010 is 
expected to serve this storage.
2009-08-18 10:48:45,932 INFO org.apache.hadoop.ipc.Server: IPC Server handler 8 
on 9000, call blockReport(DatanodeRegistration(192.*.*.z:50010, 
storageID=DS-1120429845-127.0.0.1-50010-1246697164684, infoPort=50075, 
ipcPort=50020), [...@1b8ebe3) from 192.*.*.z:33177: error: 
org.apache.hadoop.hdfs.protocol.UnregisteredDatanodeException: Data node 
192.*.*.z:50010 is attempting to report storage ID 
DS-1120429845-127.0.0.1-50010-1246697164684. Node 192.*.*.x:50010 is expected 
to serve this storage.
org.apache.hadoop.hdfs.protocol.UnregisteredDatanodeException: Data node 
192.*.*.z:50010 is attempting to report storage ID 
DS-1120429845-127.0.0.1-50010-1246697164684. Node 192.*.*.x:50010 is expected 
to serve this storage.
        at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getDatanode(FSNamesystem.java:3800)
        at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.processReport(FSNamesystem.java:2771)
        at 
org.apache.hadoop.hdfs.server.namenode.NameNode.blockReport(NameNode.java:636)
        at sun.reflect.GeneratedMethodAccessor14.invoke(Unknown Source)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:597)
        at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:452)
        at org.apache.hadoop.ipc.Server$Handler.run(Server.java:892)
2009-08-18 10:48:46,398 FATAL org.apache.hadoop.hdfs.StateChange: BLOCK* 
NameSystem.getDatanode: Data node 192.*.*.y:50010 is attempting to report 
storage ID DS-1120429845-127.0.0.1-50010-1246697164684. Node 192.*.*.x:50010 is 
expected to serve this storage.
2009-08-18 10:48:46,398 INFO org.apache.hadoop.ipc.Server: IPC Server handler 0 
on 9000, call blockReport(DatanodeRegistration(192.9.200.y:50010, 
storageID=DS-1120429845-127.0.0.1-50010-1246697164684, infoPort=50075, 
ipcPort=50020), [...@186b634) from 192.*.*.y:47367: error: 
org.apache.hadoop.hdfs.protocol.UnregisteredDatanodeException: Data node 
192.*.*.y:50010 is attempting to report storage ID 
DS-1120429845-127.0.0.1-50010-1246697164684. Node 192.*.*.x:50010 is expected 
to serve this storage.
org.apache.hadoop.hdfs.protocol.UnregisteredDatanodeException: Data node 
192.*.*.y:50010 is attempting to report storage ID 
DS-1120429845-127.0.0.1-50010-1246697164684. Node 192.*.*.x:50010 is expected 
to serve this storage.
        at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getDatanode(FSNamesystem.java:3800)
        at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.processReport(FSNamesystem.java:2771)
        at 
org.apache.hadoop.hdfs.server.namenode.NameNode.blockReport(NameNode.java:636)
        at sun.reflect.GeneratedMethodAccessor14.invoke(Unknown Source)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:597)
        at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:452)
        at org.apache.hadoop.ipc.Server$Handler.run(Server.java:892)
2009-08-18 10:48:47,000 INFO 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Roll Edit Log from 
192.*.*.x
~        

The message on the shell looks like this:
192.*.*.x: Exception in thread "main" java.io.IOException: Cannot lock storage 
/home/gaojun/HadoopInstall/tmp/dfs/namesecondary. The directory is already 
locked.
192.*.*.x:     at 
org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:510)
192.*.*.x:     at 
org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:363)
192.*.*.x:     at 
org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode$CheckpointStorage.recoverCreate(SecondaryNameNode.java:517)
192.*.*.x:     at 
org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.initialize(SecondaryNameNode.java:145)
192.*.*.x:     at 
org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.<init>(SecondaryNameNode.java:115)
192.*.*.x:     at 
org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.main(SecondaryNameNode.java:469)
                                                                   
I could not find the reason. Can someone help me?
Thanks!

yan                                                   



      ___________________________________________________________ 
  好玩贺卡等你发,邮箱贺卡全新上线! 
http://card.mail.cn.yahoo.com/

Reply via email to