It is right. When datanodes try to connect the namenode hadoop01,in fact it connect the 127.0.0.1. This error happened.
On Thu, Sep 9, 2010 at 1:34 PM, Adarsh Sharma <adarsh.sha...@orkash.com>wrote: > Mark just write down the following lines in /etc/hosts file > > Ip-address hostname > > e. g. 192.168.0.111 ws-test > 192.1685.0.165 rahul > > samr for all nodes > > > > > Mark wrote: > >> I am getting the following errors from my datanodes when I start the >> namenode. >> >> 2010-09-08 14:17:40,690 INFO org.apache.hadoop.ipc.RPC: Server at >> hadoop1/10.XXX.XXX.XX:9000 not available yet, Zzzzz... >> 2010-09-08 14:17:42,690 INFO org.apache.hadoop.ipc.Client: Retrying >> connect to server: hadoop1/10.XXX.XXX.XX:9000. Already tried 0 time(s). >> 2010-09-08 14:17:43,691 INFO org.apache.hadoop.ipc.Client: Retrying >> connect to server: hadoop1/10.XXX.XXX.XX:9000. Already tried 1 time(s). >> 2010-09-08 14:17:44,691 INFO org.apache.hadoop.ipc.Client: Retrying >> connect to server: hadoop1/10.XXX.XXX.XX:9000. Already tried 2 time(s). >> 2010-09-08 14:17:45,692 INFO org.apache.hadoop.ipc.Client: Retrying >> connect to server: hadoop1/10.XXX.XXX.XX:9000. Already tried 3 time(s). >> .... >> >> If I remove the following line from /etc/hosts then the datanodes will be >> able to connect: >> >> 127.0.0.1 localhost.localdomain localhost hadoop1 >> >> Can someone explain why this is happening? I feel like I shouldn't have to >> alter my hosts file to get this up and running. >> >> Here is the log from the NameNode server >> >> 2010-09-08 14:16:45,179 INFO >> org.apache.hadoop.hdfs.server.namenode.NameNode: STARTUP_MSG: >> /************************************************************ >> STARTUP_MSG: Starting NameNode >> STARTUP_MSG: host = hadoop1.mycompany.com/10.XXX.XXX.XX >> STARTUP_MSG: args = [] >> STARTUP_MSG: version = 0.20.2 >> STARTUP_MSG: build = >> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20 -r >> 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010 >> ************************************************************/ >> 2010-09-08 14:16:45,309 INFO org.apache.hadoop.ipc.metrics.RpcMetrics: >> Initializing RPC Metrics with hostName=NameNode, port=9000 >> 2010-09-08 14:16:45,315 INFO >> org.apache.hadoop.hdfs.server.namenode.NameNode: Namenode up at: >> localhost.localdomain/127.0.0.1:9000 >> 2010-09-08 14:16:45,318 INFO org.apache.hadoop.metrics.jvm.JvmMetrics: >> Initializing JVM Metrics with processName=NameNode, sessionId=null >> 2010-09-08 14:16:45,319 INFO >> org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics: Initializing >> NameNodeMeterics using context >> object:org.apache.hadoop.metrics.spi.NullContext >> 2010-09-08 14:16:45,386 INFO >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: >> fsOwner=root,root,bin,daemon,sys,adm,disk,wheel >> 2010-09-08 14:16:45,387 INFO >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: supergroup=supergroup >> 2010-09-08 14:16:45,387 INFO >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: >> isPermissionEnabled=true >> 2010-09-08 14:16:45,396 INFO >> org.apache.hadoop.hdfs.server.namenode.metrics.FSNamesystemMetrics: >> Initializing FSNamesystemMetrics using context >> object:org.apache.hadoop.metrics.spi.NullContext >> 2010-09-08 14:16:45,398 INFO >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered >> FSNamesystemStatusMBean >> 2010-09-08 14:16:45,440 INFO org.apache.hadoop.hdfs.server.common.Storage: >> Number of files = 5 >> 2010-09-08 14:16:45,448 INFO org.apache.hadoop.hdfs.server.common.Storage: >> Number of files under construction = 0 >> 2010-09-08 14:16:45,449 INFO org.apache.hadoop.hdfs.server.common.Storage: >> Image file of size 419 loaded in 0 seconds. >> 2010-09-08 14:16:45,449 INFO org.apache.hadoop.hdfs.server.common.Storage: >> Edits file /home/hadoop/dfs/name/current/edits of size 4 edits # 0 loaded in >> 0 seconds. >> 2010-09-08 14:16:45,488 INFO org.apache.hadoop.hdfs.server.common.Storage: >> Image file of size 419 saved in 0 seconds. >> 2010-09-08 14:16:45,606 INFO >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Finished loading >> FSImage in 248 msecs >> 2010-09-08 14:16:45,612 INFO org.apache.hadoop.hdfs.StateChange: STATE* >> Safe mode ON. >> The ratio of reported blocks 0.0000 has not reached the threshold 0.9990. >> Safe mode will be turned off automatically. >> 2010-09-08 14:16:45,767 INFO org.mortbay.log: Logging to >> org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via >> org.mortbay.log.Slf4jLog >> 2010-09-08 14:16:45,847 INFO org.apache.hadoop.http.HttpServer: Port >> returned by webServer.getConnectors()[0].getLocalPort() before open() is -1. >> Opening the listener on 50070 >> 2010-09-08 14:16:45,848 INFO org.apache.hadoop.http.HttpServer: >> listener.getLocalPort() returned 50070 >> webServer.getConnectors()[0].getLocalPort() returned 50070 >> 2010-09-08 14:16:45,848 INFO org.apache.hadoop.http.HttpServer: Jetty >> bound to port 50070 >> 2010-09-08 14:16:45,848 INFO org.mortbay.log: jetty-6.1.14 >> 2010-09-08 14:17:15,274 INFO org.mortbay.log: Started >> selectchannelconnec...@0.0.0.0:50070 >> 2010-09-08 14:17:15,274 INFO >> org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server up at: >> 0.0.0.0:50070 >> 2010-09-08 14:17:15,275 INFO org.apache.hadoop.ipc.Server: IPC Server >> Responder: starting >> 2010-09-08 14:17:15,275 INFO org.apache.hadoop.ipc.Server: IPC Server >> listener on 9000: starting >> 2010-09-08 14:17:15,276 INFO org.apache.hadoop.ipc.Server: IPC Server >> handler 0 on 9000: starting >> 2010-09-08 14:17:15,276 INFO org.apache.hadoop.ipc.Server: IPC Server >> handler 1 on 9000: starting >> 2010-09-08 14:17:15,277 INFO org.apache.hadoop.ipc.Server: IPC Server >> handler 2 on 9000: starting >> 2010-09-08 14:17:15,277 INFO org.apache.hadoop.ipc.Server: IPC Server >> handler 3 on 9000: starting >> 2010-09-08 14:17:15,277 INFO org.apache.hadoop.ipc.Server: IPC Server >> handler 4 on 9000: starting >> 2010-09-08 14:17:15,277 INFO org.apache.hadoop.ipc.Server: IPC Server >> handler 5 on 9000: starting >> 2010-09-08 14:17:15,277 INFO org.apache.hadoop.ipc.Server: IPC Server >> handler 6 on 9000: starting >> 2010-09-08 14:17:15,278 INFO org.apache.hadoop.ipc.Server: IPC Server >> handler 7 on 9000: starting >> 2010-09-08 14:17:15,278 INFO org.apache.hadoop.ipc.Server: IPC Server >> handler 8 on 9000: starting >> 2010-09-08 14:17:15,278 INFO org.apache.hadoop.ipc.Server: IPC Server >> handler 9 on 9000: starting >> 2010-09-08 14:17:15,323 INFO org.apache.hadoop.ipc.Server: Error register >> getProtocolVersion >> java.lang.IllegalArgumentException: Duplicate >> metricsName:getProtocolVersion >> at >> org.apache.hadoop.metrics.util.MetricsRegistry.add(MetricsRegistry.java:53) >> at >> org.apache.hadoop.metrics.util.MetricsTimeVaryingRate.<init>(MetricsTimeVaryingRate.java:89) >> >> at >> org.apache.hadoop.metrics.util.MetricsTimeVaryingRate.<init>(MetricsTimeVaryingRate.java:99) >> >> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:523) >> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959) >> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955) >> at java.security.AccessController.doPrivileged(Native Method) >> at javax.security.auth.Subject.doAs(Subject.java:416) >> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953) >> >> >> Thanks >> >> >