Hello Starry,
Thank you for your reply. As you said, last time we changed the
cluster, removing a datanode form the cluster. And today we have built a
whole new cluster. But when we ran "start-all.sh", it took several minutes
to startup the hadoop system. Isn't it too long? Before the startup is
finished, if we run "hadoop/bin/hadoop dfs -ls /", it will be waiting
forever. Following is the log on namenode server, is it normal? Thanks.
2009-09-27 11:52:38,190 INFO
org.apache.hadoop.hdfs.server.namenode.NameNode: STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting NameNode
STARTUP_MSG: host = ubuntu6/192.168.33.7
STARTUP_MSG: args = []
STARTUP_MSG: version = 0.20.0
STARTUP_MSG: build =
https://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.20 -r 763504;
compiled by 'ndaley' on Thu Apr 9 05:18:40 UTC 2009
************************************************************/
2009-09-27 11:52:38,284 INFO org.apache.hadoop.ipc.metrics.RpcMetrics:
Initializing RPC Metrics with hostName=NameNode, port=9000
2009-09-27 11:52:38,288 INFO
org.apache.hadoop.hdfs.server.namenode.NameNode: Namenode up at: ubuntu6/
192.168.33.7:9000
2009-09-27 11:52:38,290 INFO org.apache.hadoop.metrics.jvm.JvmMetrics:
Initializing JVM Metrics with processName=NameNode, sessionId=null
2009-09-27 11:52:38,291 INFO
org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics: Initializing
NameNodeMeterics using context object:org.apache.had
oop.metrics.spi.NullContext
2009-09-27 11:52:38,373 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
fsOwner=cyd,cyd,adm,dialout,cdrom,plugdev,lpadmin,sambashare,admin
2009-09-27 11:52:38,373 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: supergroup=supergroup
2009-09-27 11:52:38,373 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
isPermissionEnabled=true
2009-09-27 11:52:38,380 INFO
org.apache.hadoop.hdfs.server.namenode.metrics.FSNamesystemMetrics:
Initializing FSNamesystemMetrics using context object:org.apa
che.hadoop.metrics.spi.NullContext
2009-09-27 11:52:38,381 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered
FSNamesystemStatusMBean
2009-09-27 11:52:38,407 INFO org.apache.hadoop.hdfs.server.common.Storage:
Number of files = 1
2009-09-27 11:52:38,411 INFO org.apache.hadoop.hdfs.server.common.Storage:
Number of files under construction = 0
2009-09-27 11:52:38,411 INFO org.apache.hadoop.hdfs.server.common.Storage:
Image file of size 93 loaded in 0 seconds.
2009-09-27 11:52:38,417 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Invalid opcode, reached
end of edit log Number of transactions found 9
2009-09-27 11:52:38,417 INFO org.apache.hadoop.hdfs.server.common.Storage:
Edits file /home/cyd/hdfs2/name/current/edits of size 1049092 edits # 9
loaded in 0
seconds.
2009-09-27 11:52:38,462 INFO org.apache.hadoop.hdfs.server.common.Storage:
Image file of size 541 saved in 0 seconds.
2009-09-27 11:52:38,587 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Finished loading
FSImage in 269 msecs
2009-09-27 11:52:38,602 INFO org.apache.hadoop.hdfs.StateChange: STATE* Safe
mode ON.
The ratio of reported blocks 0.0000 has not reached the threshold 0.9990.
Safe mode will be turned off automatically.
2009-09-27 11:52:43,972 INFO org.mortbay.log: Logging to
org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
org.mortbay.log.Slf4jLog
2009-09-27 11:52:44,138 INFO org.apache.hadoop.http.HttpServer: Jetty bound
to port 50070
2009-09-27 11:52:44,138 INFO org.mortbay.log: jetty-6.1.14
2009-09-27 11:53:44,318 INFO org.mortbay.log: Started
[email protected]:50070
2009-09-27 11:53:44,319 INFO
org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server up at:
0.0.0.0:50070
2009-09-27 11:53:44,320 INFO org.apache.hadoop.ipc.Server: IPC Server
Responder: starting
2009-09-27 11:53:44,321 INFO org.apache.hadoop.ipc.Server: IPC Server
listener on 9000: starting
2009-09-27 11:53:44,350 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 0 on 9000: starting
2009-09-27 11:53:44,366 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 1 on 9000: starting
2009-09-27 11:53:44,366 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 2 on 9000: starting
2009-09-27 11:53:44,367 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 3 on 9000: starting
2009-09-27 11:53:44,367 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 4 on 9000: starting
2009-09-27 11:53:44,368 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 5 on 9000: starting
2009-09-27 11:53:44,368 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 6 on 9000: starting
2009-09-27 11:53:44,369 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 7 on 9000: starting
2009-09-27 11:53:44,369 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 8 on 9000: starting
2009-09-27 11:53:44,370 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 9 on 9000: starting
2009-09-27 11:56:32,036 INFO org.apache.hadoop.hdfs.StateChange: BLOCK*
NameSystem.registerDatanode: node registration from
192.168.33.6:50010storage DS-3923
30531-192.168.33.6-50010-1254023159672
2009-09-27 11:56:32,043 INFO org.apache.hadoop.net.NetworkTopology: Adding a
new node: /default-rack/192.168.33.6:50010
2009-09-27 11:56:32,086 INFO org.apache.hadoop.hdfs.StateChange: STATE* Safe
mode extension entered.
The ratio of reported blocks 1.0000 has reached the threshold 0.9990. Safe
mode will be turned off automatically in 29 seconds.
2009-09-27 11:56:52,090 INFO org.apache.hadoop.hdfs.StateChange: STATE* Safe
mode ON.
The ratio of reported blocks 1.0000 has reached the threshold 0.9990. Safe
mode will be turned off automatically in 9 seconds.
2009-09-27 11:57:02,092 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Total number of blocks
= 1
2009-09-27 11:57:02,092 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of invalid
blocks = 0
2009-09-27 11:57:02,092 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
under-replicated blocks = 1
2009-09-27 11:57:02,092 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
over-replicated blocks = 0
2009-09-27 11:57:02,093 INFO org.apache.hadoop.hdfs.StateChange: STATE*
Leaving safe mode after 263 secs.
2009-09-27 11:57:02,093 INFO org.apache.hadoop.hdfs.StateChange: STATE* Safe
mode is OFF.
2009-09-27 11:57:02,093 INFO org.apache.hadoop.hdfs.StateChange: STATE*
Network topology has 1 racks and 1 datanodes
2009-09-27 11:57:02,093 INFO org.apache.hadoop.hdfs.StateChange: STATE*
UnderReplicatedBlocks has 1 blocks
2009-09-27 11:59:00,157 INFO org.apache.hadoop.hdfs.StateChange: BLOCK*
NameSystem.registerDatanode: node registration from
192.168.33.8:50010storage DS-1578
942809-192.168.33.8-50010-1254023408249
2009-09-27 11:59:00,158 INFO org.apache.hadoop.net.NetworkTopology: Adding a
new node: /default-rack/192.168.33.8:50010
2009-09-27 11:59:00,192 INFO org.apache.hadoop.hdfs.StateChange: BLOCK*
NameSystem.addStoredBlock: blockMap updated: 192.168.33.8:50010 is added to
blk_234235
2795116225569_1001 size 4
2009-09-27 11:59:20,745 INFO org.apache.hadoop.hdfs.StateChange: BLOCK*
NameSystem.registerDatanode: node registration from
192.168.33.5:50010storage DS-7945
10456-192.168.33.5-50010-1254023267768
2009-09-27 11:59:20,746 INFO org.apache.hadoop.net.NetworkTopology: Adding a
new node: /default-rack/192.168.33.5:50010
2009-09-27 12:00:02,336 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
ugi=cyd,cyd,adm,dialout,cdrom,plugdev,lpadmin,sambashare,admin ip=/19
2.168.33.7 cmd=listStatus src=/tmp/hadoop-cyd/mapred/system
dst=null perm=null
2009-09-27 12:00:02,346 INFO org.apache.hadoop.hdfs.StateChange: BLOCK*
NameSystem.addToInvalidates: blk_2342352795116225569 is added to invalidSet
of 192.168
.33.6:50010
2009-09-27 12:00:02,346 INFO org.apache.hadoop.hdfs.StateChange: BLOCK*
NameSystem.addToInvalidates: blk_2342352795116225569 is added to invalidSet
of 192.168
.33.8:50010
2009-09-27 12:00:02,349 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of transactions:
1 Total time for transactions(ms): 1Number of transa
ctions batched in Syncs: 0 Number of syncs: 0 SyncTimes(ms): 0
2009-09-27 12:00:02,384 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
ugi=cyd,cyd,adm,dialout,cdrom,plugdev,lpadmin,sambashare,admin ip=/19
2.168.33.7 cmd=delete src=/tmp/hadoop-cyd/mapred/system
dst=null perm=null
2009-09-27 12:00:02,400 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
ugi=cyd,cyd,adm,dialout,cdrom,plugdev,lpadmin,sambashare,admin ip=/19
2.168.33.7 cmd=mkdirs src=/tmp/hadoop-cyd/mapred/system
dst=null perm=cyd:supergroup:rwxr-xr-x
2009-09-27 12:00:02,408 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
ugi=cyd,cyd,adm,dialout,cdrom,plugdev,lpadmin,sambashare,admin ip=/19
2.168.33.7 cmd=setPermission
src=/tmp/hadoop-cyd/mapred/system dst=null
perm=cyd:supergroup:rwx-wx-wx
2009-09-27 12:00:02,432 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
ugi=cyd,cyd,adm,dialout,cdrom,plugdev,lpadmin,sambashare,admin ip=/19
2.168.33.7 cmd=create src=/tmp/hadoop-cyd/mapred/system/
jobtracker.info dst=null perm=cyd:supergroup:rw-r--r--
2009-09-27 12:00:02,440 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
ugi=cyd,cyd,adm,dialout,cdrom,plugdev,lpadmin,sambashare,admin ip=/19
2.168.33.7 cmd=setPermission src=/tmp/hadoop-cyd/mapred/system/
jobtracker.info dst=null perm=cyd:supergroup:rw-------
2009-09-27 12:00:02,445 INFO org.apache.hadoop.hdfs.StateChange: BLOCK*
NameSystem.allocateBlock: /tmp/hadoop-cyd/mapred/system/jobtracker.info.
blk_-69940863
0358636044_1002
2009-09-27 12:00:02,593 INFO org.apache.hadoop.hdfs.StateChange: BLOCK*
NameSystem.addStoredBlock: blockMap updated: 192.168.33.6:50010 is added to
blk_-69940
8630358636044_1002 size 4
2009-09-27 12:00:02,600 INFO org.apache.hadoop.hdfs.StateChange: BLOCK*
NameSystem.addStoredBlock: blockMap updated: 192.168.33.5:50010 is added to
blk_-69940
8630358636044_1002 size 4
2009-09-27 12:00:02,644 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* ask
192.168.33.8:50010 to delete blk_2342352795116225569_1001
2009-09-27 12:00:05,644 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* ask
192.168.33.6:50010 to delete blk_2342352795116225569_1001
2009-09-27 12:03:49,176 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
ugi=cyd,cyd,adm,dialout,cdrom,plugdev,lpadmin,sambashare,admin ip=/19
2.168.33.7 cmd=listStatus src=/ dst=null perm=null
2009/9/26 Starry SHI <[email protected]>
> Is that the first time you start your cluster? My experience is that, when
> you start the cluster once, then change the conf (say, add another slave),
> and restart your cluster, it sometimes generate some IPC issues (like the
> timeout in the namenode log). This change will cause the filesystem into
> safe mode so there will be the exceptions.
>
> The exception in your NN log is caused when the cluster was trying to add a
> new node but reached a timeout. If you have changed your cluster conf, you
> can use "hadoop namenode format" to format the filesystem and then start,
> then your cluster should be started.
>
> Hope this can help.
>
> Best regards,
> Starry
>
> /* Tomorrow is another day. So is today. */
>
>
> On Sat, Sep 26, 2009 at 09:03, Zheng Lv <[email protected]> wrote:
>
> > No one can give some suggestions?
> >
> > 2009/9/25 Zheng Lv <[email protected]>
> >
> > > Hello Everyone,
> > > We have a cluster with one namenode and three datanodes. And we got
> > > these logs when starting hadoop0.20. Is it normal?
> > > 2009-09-25 10:45:00,616 INFO
> > > org.apache.hadoop.hdfs.server.namenode.NameNode: STARTUP_MSG:
> > > /************************************************************
> > > STARTUP_MSG: Starting NameNode
> > > STARTUP_MSG: host = ubuntu6/192.168.33.7
> > > STARTUP_MSG: args = []
> > > STARTUP_MSG: version = 0.20.1-dev
> > > STARTUP_MSG: build = -r ; compiled by 'cyd' on 2009?ê 08?? 07??
> ??????
> > > 15:20:16 CST
> > > ************************************************************/
> > > 2009-09-25 10:45:00,721 INFO org.apache.hadoop.ipc.metrics.RpcMetrics:
> > > Initializing RPC Metrics with hostName=NameNode, port=9000
> > > 2009-09-25 10:45:00,767 INFO
> > > org.apache.hadoop.hdfs.server.namenode.NameNode: Namenode up at:
> ubuntu6/
> > > 192.168.33.7:9000
> > > 2009-09-25 10:45:00,769 INFO org.apache.hadoop.metrics.jvm.JvmMetrics:
> > > Initializing JVM Metrics with processName=NameNode, sessionId=null
> > > 2009-09-25 10:45:00,771 INFO
> > > org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics:
> > Initializing
> > > NameNodeMeterics using context
> > > object:org.apache.hadoop.metrics.spi.NullContext
> > > 2009-09-25 10:45:00,835 INFO
> > > org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
> > > fsOwner=cyd,cyd,adm,dialout,cdrom,plugdev,lpadmin,sambashare,admin
> > > 2009-09-25 10:45:00,835 INFO
> > > org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
> > supergroup=supergroup
> > > 2009-09-25 10:45:00,835 INFO
> > > org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
> > > isPermissionEnabled=true
> > > 2009-09-25 10:45:00,842 INFO
> > > org.apache.hadoop.hdfs.server.namenode.metrics.FSNamesystemMetrics:
> > > Initializing FSNamesystemMetrics using context
> > > object:org.apache.hadoop.metrics.spi.NullContext
> > > 2009-09-25 10:45:00,843 INFO
> > > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered
> > > FSNamesystemStatusMBean
> > > 2009-09-25 10:45:00,871 INFO
> > org.apache.hadoop.hdfs.server.common.Storage:
> > > Number of files = 2834
> > > 2009-09-25 10:45:01,412 INFO
> > org.apache.hadoop.hdfs.server.common.Storage:
> > > Number of files under construction = 1
> > > 2009-09-25 10:45:01,416 INFO
> > org.apache.hadoop.hdfs.server.common.Storage:
> > > Image file of size 329240 loaded in 0 seconds.
> > > 2009-09-25 10:45:01,417 INFO
> > org.apache.hadoop.hdfs.server.common.Storage:
> > > Edits file /home/cyd/hdfs2/name/current/edits of size 4 edits # 0
> loaded
> > in
> > > 0 seconds.
> > > 2009-09-25 10:45:01,683 INFO
> > org.apache.hadoop.hdfs.server.common.Storage:
> > > Image file of size 329240 saved in 0 seconds.
> > > 2009-09-25 10:45:01,805 INFO
> > > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Finished loading
> > > FSImage in 1007 msecs
> > > 2009-09-25 10:45:01,821 INFO org.apache.hadoop.hdfs.StateChange: STATE*
> > > Safe mode ON.
> > > The ratio of reported blocks 0.0000 has not reached the threshold
> 0.9990.
> > > Safe mode will be turned off automatically.
> > > 2009-09-25 10:45:07,162 INFO org.mortbay.log: Logging to
> > > org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
> > > org.mortbay.log.Slf4jLog
> > > 2009-09-25 10:45:07,228 INFO org.apache.hadoop.http.HttpServer: Jetty
> > bound
> > > to port 50070
> > > 2009-09-25 10:45:07,229 INFO org.mortbay.log: jetty-6.1.14
> > > 2009-09-25 10:47:34,305 INFO org.mortbay.log: Started
> > > [email protected]:50070
> > > 2009-09-25 10:47:34,319 INFO
> > > org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server up at:
> > > 0.0.0.0:50070
> > > 2009-09-25 10:47:34,320 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > Responder: starting
> > > 2009-09-25 10:47:34,321 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > listener on 9000: starting
> > > 2009-09-25 10:47:34,355 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 0 on 9000: starting
> > > 2009-09-25 10:47:34,355 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 1 on 9000: starting
> > > 2009-09-25 10:47:34,356 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 2 on 9000: starting
> > > 2009-09-25 10:47:34,356 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 3 on 9000: starting
> > > 2009-09-25 10:47:34,384 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 5 on 9000: starting
> > > 2009-09-25 10:47:34,397 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 4 on 9000: starting
> > > 2009-09-25 10:47:34,398 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 6 on 9000: starting
> > > 2009-09-25 10:47:34,400 WARN org.apache.hadoop.ipc.Server: IPC Server
> > > Responder, call
> > > getProtocolVersion(org.apache.hadoop.hdfs.protocol.ClientProtocol, 41)
> > from
> > > 192.168.33.7:54977: output error
> > > 2009-09-25 10:47:34,401 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 7 on 9000: starting
> > > 2009-09-25 10:47:34,402 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 8 on 9000: starting
> > > 2009-09-25 10:47:34,404 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 9 on 9000: starting
> > > 2009-09-25 10:47:34,433 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 1 on 9000 caught: java.nio.channels.ClosedChannelException
> > > at
> > >
> sun.nio.ch.SocketChannelImpl.ensureWriteOpen(SocketChannelImpl.java:126)
> > > at
> sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:324)
> > > at org.apache.hadoop.ipc.Server.channelWrite(Server.java:1195)
> > > at org.apache.hadoop.ipc.Server.access$1900(Server.java:77)
> > > at
> > > org.apache.hadoop.ipc.Server$Responder.processResponse(Server.java:613)
> > > at
> > > org.apache.hadoop.ipc.Server$Responder.doRespond(Server.java:677)
> > > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:981)
> > > 2009-09-25 10:48:09,304 INFO org.apache.hadoop.hdfs.StateChange: BLOCK*
> > > NameSystem.registerDatanode: node registration from 192.168.33.5:50010
> storage
> > DS-1782592326-192.168.33.5-50010-1253784633678
> > > 2009-09-25 10:48:09,311 INFO org.apache.hadoop.net.NetworkTopology:
> > Adding
> > > a new node: /default-rack/192.168.33.5:50010
> > > 2009-09-25 10:48:09,920 INFO org.apache.hadoop.hdfs.StateChange: STATE*
> > > Safe mode ON.
> > > The ratio of reported blocks 0.0005 has not reached the threshold
> 0.9990.
> > > Safe mode will be turned off automatically.
> > > 2009-09-25 10:48:09,985 INFO org.apache.hadoop.hdfs.StateChange: STATE*
> > > Safe mode extension entered.
> > > The ratio of reported blocks 0.9990 has reached the threshold 0.9990.
> > Safe
> > > mode will be turned off automatically in 29 seconds.
> > > 2009-09-25 10:48:29,989 INFO org.apache.hadoop.hdfs.StateChange: STATE*
> > > Safe mode ON.
> > > The ratio of reported blocks 1.0000 has reached the threshold 0.9990.
> > Safe
> > > mode will be turned off automatically in 9 seconds.
> > > 2009-09-25 10:48:32,412 INFO org.apache.hadoop.hdfs.StateChange: BLOCK*
> > > NameSystem.registerDatanode: node registration from 192.168.33.6:50010
> storage
> > DS-119886468-192.168.33.6-50010-1253784635966
> > > 2009-09-25 10:48:32,412 INFO org.apache.hadoop.net.NetworkTopology:
> > Adding
> > > a new node: /default-rack/192.168.33.6:50010
> > > 2009-09-25 10:48:40,147 INFO
> > > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Total number of
> > blocks
> > > = 2052
> > > 2009-09-25 10:48:40,147 INFO
> > > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of invalid
> > > blocks = 0
> > > 2009-09-25 10:48:40,147 INFO
> > > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
> > > under-replicated blocks = 2051
> > > 2009-09-25 10:48:40,147 INFO
> > > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
> > > over-replicated blocks = 0
> > > 2009-09-25 10:48:40,147 INFO org.apache.hadoop.hdfs.StateChange: STATE*
> > > Leaving safe mode after 219 secs.
> > > 2009-09-25 10:48:40,147 INFO org.apache.hadoop.hdfs.StateChange: STATE*
> > > Safe mode is OFF.
> > > 2009-09-25 10:48:40,148 INFO org.apache.hadoop.hdfs.StateChange: STATE*
> > > Network topology has 1 racks and 2 datanodes
> > > 2009-09-25 10:48:40,148 INFO org.apache.hadoop.hdfs.StateChange: STATE*
> > > UnderReplicatedBlocks has 2051 blocks
> > >
> > >
> > > I noticed the exceptions and "org.apache.hadoop.hdfs.StateChange:
> > > STATE* Network topology has 1 racks and 2 datanodes".Sometimes it tells
> 1
> > > datanodes, and sometimes 2, but we have 3.Please give some suggestions,
> > > thank you.
> > >
> > >
> > >
> > > Following is logs in one of datanodes:
> > > 2009-09-25 10:47:22,110 INFO
> > > org.apache.hadoop.hdfs.server.datanode.DataNode: Registered
> > > FSDatasetStatusMBean
> > > 2009-09-25 10:47:22,112 INFO
> > > org.apache.hadoop.hdfs.server.datanode.DataNode: Opened info server at
> > 50010
> > > 2009-09-25 10:47:22,114 INFO
> > > org.apache.hadoop.hdfs.server.datanode.DataNode: Balancing bandwith is
> > > 1048576 bytes/s
> > > 2009-09-25 10:47:27,257 INFO org.mortbay.log: Logging to
> > > org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
> > > org.mortbay.log.Slf4jLog
> > > 2009-09-25 10:47:27,434 INFO org.apache.hadoop.http.HttpServer: Jetty
> > bound
> > > to port 50075
> > > 2009-09-25 10:47:27,434 INFO org.mortbay.log: jetty-6.1.14
> > > 2009-09-25 10:48:14,324 INFO org.mortbay.log: Started
> > > [email protected]:50075
> > > 2009-09-25 10:48:14,334 INFO org.apache.hadoop.metrics.jvm.JvmMetrics:
> > > Initializing JVM Metrics with processName=DataNode, sessionId=null
> > > 2009-09-25 10:48:19,369 INFO org.apache.hadoop.ipc.metrics.RpcMetrics:
> > > Initializing RPC Metrics with hostName=DataNode, port=50020
> > > 2009-09-25 10:48:19,376 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > Responder: starting
> > > 2009-09-25 10:48:19,376 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > listener on 50020: starting
> > > 2009-09-25 10:48:19,378 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 0 on 50020: starting
> > > 2009-09-25 10:48:19,380 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 1 on 50020: starting
> > > 2009-09-25 10:48:19,380 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 2 on 50020: starting
> > > 2009-09-25 10:48:19,381 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 3 on 50020: starting
> > > 2009-09-25 10:48:19,381 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 4 on 50020: starting
> > > 2009-09-25 10:48:19,381 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 5 on 50020: starting
> > > 2009-09-25 10:48:19,382 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 6 on 50020: starting
> > > 2009-09-25 10:48:19,382 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 7 on 50020: starting
> > > 2009-09-25 10:48:19,382 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 8 on 50020: starting
> > > 2009-09-25 10:48:19,383 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 9 on 50020: starting
> > > 2009-09-25 10:48:19,383 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 10 on 50020: starting
> > > 2009-09-25 10:48:19,383 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 11 on 50020: starting
> > > 2009-09-25 10:48:19,384 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 12 on 50020: starting
> > > 2009-09-25 10:48:19,384 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 13 on 50020: starting
> > > 2009-09-25 10:48:19,385 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 14 on 50020: starting
> > > 2009-09-25 10:48:19,385 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 15 on 50020: starting
> > > 2009-09-25 10:48:19,385 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 16 on 50020: starting
> > > 2009-09-25 10:48:19,386 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 17 on 50020: starting
> > > 2009-09-25 10:48:19,386 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 18 on 50020: starting
> > > 2009-09-25 10:48:19,386 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 19 on 50020: starting
> > > 2009-09-25 10:48:19,387 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 20 on 50020: starting
> > > 2009-09-25 10:48:19,387 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 21 on 50020: starting
> > > 2009-09-25 10:48:19,388 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 22 on 50020: starting
> > > 2009-09-25 10:48:19,388 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 23 on 50020: starting
> > > 2009-09-25 10:48:19,388 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 24 on 50020: starting
> > > 2009-09-25 10:48:19,388 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 25 on 50020: starting
> > > 2009-09-25 10:48:19,389 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 26 on 50020: starting
> > > 2009-09-25 10:48:19,389 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 27 on 50020: starting
> > > 2009-09-25 10:48:19,390 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 28 on 50020: starting
> > > 2009-09-25 10:48:19,390 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 29 on 50020: starting
> > > 2009-09-25 10:48:19,390 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 30 on 50020: starting
> > > 2009-09-25 10:48:19,391 INFO org.apache.hadoop.ipc.Server: IPC Server
> > > handler 31 on 50020: starting
> > > 2009-09-25 10:48:19,391 INFO
> > > org.apache.hadoop.hdfs.server.datanode.DataNode: dnRegistration =
> > > DatanodeRegistration(ubuntu7:50010,
> > > storageID=DS-119886468-192.168.33.6-50010-1253784635966,
> infoPort=50075,
> > > ipcPort=50020)
> > > 2009-09-25 10:48:19,434 INFO
> > > org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(
> > > 192.168.33.6:50010,
> > > storageID=DS-119886468-192.168.33.6-50010-1253784635966,
> infoPort=50075,
> > > ipcPort=50020)In DataNode.run, data =
> > > FSDataset{dirpath='/home/cyd/hdfs2/data/current'}
> > > 2009-09-25 10:48:19,456 INFO
> > > org.apache.hadoop.hdfs.server.datanode.DataNode: using
> > BLOCKREPORT_INTERVAL
> > > of 3600000msec Initial delay: 0msec
> > > 2009-09-25 10:48:19,950 INFO
> > > org.apache.hadoop.hdfs.server.datanode.DataNode: BlockReport of 2052
> > blocks
> > > got processed in 458 msecs
> > > 2009-09-25 10:48:19,951 INFO
> > > org.apache.hadoop.hdfs.server.datanode.DataNode: Starting Periodic
> block
> > > scanner.
> > > 2009-09-25 10:48:20,601 INFO
> > > org.apache.hadoop.hdfs.server.datanode.DataBlockScanner: Verification
> > > succeeded for blk_2990263655324635865_9556
> > > 2009-09-25 10:48:58,692 INFO
> > > org.apache.hadoop.hdfs.server.datanode.DataNode: BlockReport of 2052
> > blocks
> > > got processed in 224 msecs
> > > 2009-09-25 10:51:07,955 INFO
> > > org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving block
> > > blk_-1955286602120745744_18039 src: /192.168.33.5:43889 dest: /
> > > 192.168.33.6:50010
> > > 2009-09-25 10:51:07,982 INFO
> > > org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /
> > > 192.168.33.5:43889, dest: /192.168.33.6:50010, bytes: 4, op:
> HDFS_WRITE,
> > > cliID: DFSClient_2050853952, srvID:
> > > DS-119886468-192.168.33.6-50010-1253784635966, blockid:
> > > blk_-1955286602120745744_18039
> > > 2009-09-25 10:51:07,983 INFO
> > > org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder 0 for
> > block
> > > blk_-1955286602120745744_18039 terminating
> > > 2009-09-25 10:51:10,742 INFO
> > > org.apache.hadoop.hdfs.server.datanode.DataNode: Deleting block
> > > blk_-6733538557527582040_18038 file
> > > /home/cyd/hdfs2/data/current/blk_-6733538557527582040
> > > 2009-09-25 10:56:45,523 INFO
> > > org.apache.hadoop.hdfs.server.datanode.DataBlockScanner: Verification
> > > succeeded for blk_-8204684875565980176_3413
> > > 2009-09-25 11:03:06,572 INFO
> > > org.apache.hadoop.hdfs.server.datanode.DataBlockScanner: Verification
> > > succeeded for blk_8093352211012570299_17220
> > > 2009-09-25 11:10:28,649 INFO
> > > org.apache.hadoop.hdfs.server.datanode.DataBlockScanner: Verification
> > > succeeded for blk_-6978381366409301329_8279
> > >
> > > Thanks a lot,
> > > LvZheng
> > >
> >
>