[ 
https://issues.apache.org/jira/browse/HDFS-1106?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo Nicholas Sze resolved HDFS-1106.
---------------------------------------

    Resolution: Not a Problem

Resolving as Not a Problem.

> Datanode throwing UnregisteredDatanodeException -- expects itself to serve 
> storage!
> -----------------------------------------------------------------------------------
>
>                 Key: HDFS-1106
>                 URL: https://issues.apache.org/jira/browse/HDFS-1106
>             Project: Hadoop HDFS
>          Issue Type: Bug
>    Affects Versions: 0.20.1
>            Reporter: Eugene Hung
>
> We run a large Hadoop cluster used by many different universities.  When some 
> DataNodes went down recently, they came back up and then generated this error
> message in their datanode logs:
> 2010-04-22 16:58:37,314 ERROR 
> org.apache.hadoop.hdfs.server.datanode.DataNode: 
> org.apache.hadoop.ipc.RemoteException: 
> org.apache.hadoop.hdfs.protocol.UnregisteredDatanodeException: Data node 
> vm-10-160-4-109:50010 is attempting to report storage ID 
> DS-1884904520-10.160.4.109-50010-1255720271773. Node 10.160.4.109:50010 is 
> expected to serve this storage.
>       at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getDatanode(FSNamesystem.java:3972)
>       at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.verifyNodeRegistration(FSNamesystem.java:3937)
>       at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.registerDatanode(FSNamesystem.java:2052)
>       at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.register(NameNode.java:735)
>       at sun.reflect.GeneratedMethodAccessor8.invoke(Unknown Source)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>       at java.lang.reflect.Method.invoke(Method.java:597)
>       at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
>       at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:966)
>       at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:962)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:396)
>       at org.apache.hadoop.ipc.Server$Handler.run(Server.java:960)
>       at org.apache.hadoop.ipc.Client.call(Client.java:740)
>       at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
>       at $Proxy4.register(Unknown Source)
>       at 
> org.apache.hadoop.hdfs.server.datanode.DataNode.register(DataNode.java:544)
>       at 
> org.apache.hadoop.hdfs.server.datanode.DataNode.runDatanodeDaemon(DataNode.java:1230)
>       at 
> org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1273)
>       at 
> org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1394)
> Note it is correctly expecting itself to serve the data, but throwing an 
> UnregisteredDatanodeException for some reason.  This is causing these 
> datanodes to remain "dead" to the namenode.  Does anyone know why this is 
> occuring and what we can do to fix it?



--
This message was sent by Atlassian JIRA
(v6.2#6252)

Reply via email to