I'd say that this is your problem right here:

Heap Size is 972.5 MB / 972.5 MB (100%)

As suspected, your very small block size + many files has completely
filled the NameNode heap. All bets are off as to what Hadoop will do
at this point.

Potential solutions:
1) increase the HADOOP_HEAPSIZE parameter in
$HADOOP_HOME/conf/hadoop-env.sh on the NameNode to give your NameNode
more space

2) increase your block size to something more reasonable. You are
running Hadoop drastically outside the normal configuration. I don't
know anyone who uses blocks less than 64MB; and certainly can't
imagine that there's too many people running it at 8 or fewer
MB/block. You may be uncovering new bugs by running this far outside
of the normal configuration. Note that increasing the configured block
size will not do anything to your existing data; you'll need to
reimport it somehow.

3) Delete a lot of files til your NameNode can cope.

- Aaron




On Mon, Jul 20, 2009 at 8:51 PM, mingyang<[email protected]> wrote:
> I have carefully examined the issue of versions to ensure that the hadoop
> are tar from the same package, on the issue of block size,   The system is
> because I do a test, when I set the block size to 64MB, the front-end apache
> can only read the document to reach 300-400kb, and I see my network traffic
> to port as much as 40-60Mb, so I trying to become a smaller file to solve
> the problem, so I through the process, I would like to store the file into
> 2mB, and then in the store, would be better to see the results.
>  In the document shredding, I encountered a problem, is to expand the number
> of documents two orders of magnitude, reaching 200 million, initially, my
> hadoop only two machines, one running datanode, a running namenode, when the
> document After more than one million, when I put data to hadoop when 10
> minutes will be initially normal, and then I will see datanode error,  
> 09/07/19
> 18:46:13 INFO hdfs.DFSClient: Waiting to find target node:
> 192.168.1.97:50010
> 09/07/19 18:46:13 INFO hdfs.DFSClient: Exception in createBlockOutputStream
> java.net.SocketTimeoutException: 66000 millis timeout while waiting for
> channel to be ready for read. Ch: java.nio.channels.SocketChannel [connected
> local = / 192.168.1.94:16153 remote = / 192.168.1.97:50010]
> 2009-07-19 18:47:21,367 INFO org.apache.hadoop.hdfs.server.datanode.DataNode:
> writeBlock blk_-8134427742713358791_1418899 received exception
> java.io.EOFException: while trying to read 65557 bytes
> 2009-07-19 18:47:21,367 ERROR
> org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(
> 192.168.1.97:50010, storageID=DS-148204255-192.168.1.97-50010-1247630449653,
> infoPort=50075, ipcPort=50020):DataXceiver
> java.io.EOFException: while trying to read 65557 bytes
>
>
>  At the same time does not log the error namenode will wait about 10 minutes
> to receive the information datanode loss of heartbeat, the host at this time
> in the run datanode jps still can see the process of datanode, telnet 50010
> ports and at the same time, the problem seems solved, At this point, At this
> point, i run lsof can see the port seems to have been waiting forjava 9233
> root 67u IPv4 5508710 TCP master: 50010 -> web01: 10630 (ESTABLISHED)
> java 9233 root 69u IPv4 5509810 TCP master: 50010 -> web01: 10633
> (ESTABLISHED)
> I am also not clear why these occur, I can only add a 2 datanode, and to
> ensure that each file does not datanode of more than one million The
> following is my namenode log
>
>
> [r...@named logs]# free
>             total       used       free     shared    buffers     cached
> Mem:       4051552    3583876     467676          0     113360     944980
> -/+ buffers/cache:    2525536    1526016
> Swap:      4192924        120    4192804
>
> Cluster Summary***2027545 files and directories, 2008837 blocks = 4036382
> total. Heap Size is 972.5 MB / 972.5 MB (100%)
> *   Configured Capacity : 12.09 TB DFS Used : 2.88 TB Non DFS Used : 632.67
> GB DFS Remaining : 8.59 TB DFS Used% : 23.85 % DFS Remaining% : 71.03 % Live
> Nodes <http://dfsnodelist.jsp?whatNodes=LIVE> : 3 Dead
> Nodes<http://dfsnodelist.jsp?whatNodes=DEAD> :
> 0
>
> ------------------------------
> NameNode Storage:   *Storage Directory* *Type* *State*  /hadoop/name
> IMAGE_AND_EDITS Active
>
>
> 2009/7/21 Aaron Kimball <[email protected]>
>
>> A VersionMismatch occurs because you're using different builds of Hadoop on
>> your different nodes. All DataNodes and the NameNode must be running the
>> exact same compilation of Hadoop (It's very strict).
>>
>> One thing I noticed in your config is that you set dfs.block.size to
>> 2560000. That's certainly not the default value. Note that this value is in
>> bytes. That's 2.5 MB/block. If you've got a million or more files in there,
>> of any reasonable size, then you might have several million blocks.
>> Depending on how powerful your NameNode's hardware is, that could take up a
>> lot of RAM. Each block is 192 bytes, and each file is 124 bytes + the
>> filename length.
>>
>> When you run 'bin/hadoop dfsadmin -report' or look at the status JSP at
>> http://named:50070/, how much RAM free does it report? You might be using
>> up
>> a lot of the NameNode's available RAM, which could cause various problems
>> involving spending all its time in GC.
>>
>> In the future I'd recommend setting your block size 5 or 10x higher. Note
>> that this won't affect existing blocks/files though. You may need to
>> somehow
>> read and rewrite those files to take advantage of a larger block size to
>> alleviate block pressure.
>>
>> - Aaron
>>
>>
>> On Sun, Jul 19, 2009 at 3:59 PM, mingyang <[email protected]> wrote:
>>
>> > time find / b / c / d / e / f / g / h / i / j / k / l-type f-print> / dev
>> /
>> > null
>> >
>> > real 0m4.608s
>> > user 0m1.489s
>> > sys 0m3.068s
>> >
>> > I run this, the result is far less than eight minutes, is not where I set
>> > up
>> > right? I just changed the fs.default.name, other settings are default
>> > values
>> >
>> >   <name> fs.default.name </ name>
>> >   <value> hdfs: / / named: 9000 </ value>
>> >
>> >
>> >
>> >   <name> mapred.job.tracker </ name>
>> >   <value> hdfs: / / named: 9001 </ value>
>> >
>> >
>> >
>> >   <name> hadoop.tmp.dir </ name>
>> >   <value> / hadoop / tmp </ value>
>> >
>> >   <name> dfs.name.dir </ name>
>> >   <value> / hadoop / name </ value>
>> >
>> >
>> >   <name> dfs.data.dir </ name>
>> >   <value> / b, / c, / d, / e, / f, / g, / h, / i, / j, / k, / l </ value>
>> >
>> >
>> >
>> >   <name> dfs.replication </ name>
>> >   <value> 1 </ value>
>> >
>> >
>> >   <name> dfs.block.size </ name>
>> >   <value> 2560000 </ value>
>> >
>> >
>> >
>> > 2009/7/20 Jason Venner <[email protected]>
>> >
>> > > Did you run this command on the datanode that not responding?
>> > >
>> > > On Sun, Jul 19, 2009 at 3:59 AM, mingyang <[email protected]>
>> wrote:
>> > >
>> > > > in datanode logs , I found a new error message.
>> > > > Would like to help solve the problem
>> > > >
>> > > > 2009-07-19 18:40:43,464 ERROR
>> > > > org.apache.hadoop.hdfs.server.datanode.DataNode:
>> DatanodeRegistration(
>> > > > 192.168.1.97:50010,
>> > > > storageID=DS-148204255-192.168.1.97-50010-1247630449653,
>> > > > infoPort=50075, ipcPort=50020):DataXceiver
>> > > > java.io.IOException: Version Mismatch
>> > > >        at
>> > > >
>> > >
>> >
>> org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:81)
>> > > >        at java.lang.Thread.run(Thread.java:619)
>> > > > 2009-07-19 18:40:43,464 INFO
>> > > > org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving block
>> > > > blk_9198148243909908835_1418899 src: /192.168.1.94:21486 dest: /
>> > > > 192.168.1.97:50010
>> > > > 2009-07-19 18:41:43,363 INFO
>> > > > org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving block
>> > > > blk_4464496078754951058_1418901 src: /192.168.1.105:17434 dest: /
>> > > > 192.168.1.97:50010
>> > > > 2009-07-19 18:41:55,318 INFO
>> > > > org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving block
>> > > > blk_-998769390764365324_1418903 src: /192.168.1.105:17436 dest: /
>> > > > 192.168.1.97:50010
>> > > > 2009-07-19 18:42:51,983 INFO
>> > > > org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving block
>> > > > blk_7357805261059813870_1418903 src: /192.168.1.94:16148 dest: /
>> > > > 192.168.1.97:50010
>> > > > 2009-07-19 18:43:00,936 INFO
>> > > > org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving block
>> > > > blk_6766332640664853809_1418903 src: /192.168.1.105:60392 dest: /
>> > > > 192.168.1.97:50010
>> > > > 2009-07-19 18:44:03,112 INFO
>> > > > org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving block
>> > > > blk_8833189579369772797_1418903 src: /192.168.1.94:16152 dest: /
>> > > > 192.168.1.97:50010
>> > > > 2009-07-19 18:44:07,105 INFO
>> > > > org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving block
>> > > > blk_-92272670489114760_1418903 src: /192.168.1.94:16153 dest: /
>> > > > 192.168.1.97:50010
>> > > > 2009-07-19 18:45:15,301 INFO
>> > > > org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving block
>> > > > blk_-3189060218922169017_1418903 src: /192.168.1.94:16157 dest: /
>> > > > 192.168.1.97:50010
>> > > > 2009-07-19 18:45:19,299 INFO
>> > > > org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving block
>> > > > blk_2093621896162975184_1418903 src: /192.168.1.105:60400 dest: /
>> > > > 192.168.1.97:50010
>> > > > 2009-07-19 18:46:27,602 INFO
>> > > > org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving block
>> > > > blk_-2879021713701012781_1418905 src: /192.168.1.105:60404 dest: /
>> > > > 192.168.1.97:50010
>> > > > 2009-07-19 18:46:27,602 INFO
>> > > > org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving block
>> > > > blk_356680792598474797_1418907 src: /192.168.1.105:60406 dest: /
>> > > > 192.168.1.97:50010
>> > > > 2009-07-19 18:47:21,366 INFO
>> > > > org.apache.hadoop.hdfs.server.datanode.DataNode: Exception in
>> > > receiveBlock
>> > > > for block blk_-8134427742713358791_1418899 java.io.EOFException:
>> while
>> > > > trying to read 65557 bytes
>> > > > 2009-07-19 18:47:21,367 INFO
>> > > > org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder 0
>> for
>> > > > block
>> > > > blk_-8134427742713358791_1418899 Interrupted.
>> > > > 2009-07-19 18:47:21,367 INFO
>> > > > org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder 0
>> for
>> > > > block
>> > > > blk_-8134427742713358791_1418899 terminating
>> > > > 2009-07-19 18:47:21,367 INFO
>> > > > org.apache.hadoop.hdfs.server.datanode.DataNode: writeBlock
>> > > > blk_-8134427742713358791_1418899 received exception
>> > java.io.EOFException:
>> > > > while trying to read 65557 bytes
>> > > > 2009-07-19 18:47:21,367 ERROR
>> > > > org.apache.hadoop.hdfs.server.datanode.DataNode:
>> DatanodeRegistration(
>> > > > 192.168.1.97:50010,
>> > > > storageID=DS-148204255-192.168.1.97-50010-1247630449653,
>> > > > infoPort=50075, ipcPort=50020):DataXceiver
>> > > > java.io.EOFException: while trying to read 65557 bytes
>> > > >        at
>> > > >
>> > > >
>> > >
>> >
>> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.readToBuf(BlockReceiver.java:264)
>> > > >        at
>> > > >
>> > > >
>> > >
>> >
>> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.readNextPacket(BlockReceiver.java:308)
>> > > >        at
>> > > >
>> > > >
>> > >
>> >
>> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:372)
>> > > >        at
>> > > >
>> > > >
>> > >
>> >
>> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:524)
>> > > >        at
>> > > >
>> > > >
>> > >
>> >
>> org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:357)
>> > > >        at
>> > > >
>> > > >
>> > >
>> >
>> org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:103)
>> > > >        at java.lang.Thread.run(Thread.java:619)
>> > > > 2009-07-19 18:47:21,368 INFO
>> > > > org.apache.hadoop.hdfs.server.datanode.DataNode: Exception in
>> > > receiveBlock
>> > > > for block blk_-346503319966521195_1418897 java.io.EOFException: while
>> > > > trying
>> > > > to read 65557 bytes
>> > > > 2009-07-19 18:47:21,371 INFO
>> > > > org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder 0
>> for
>> > > > block
>> > > > blk_-346503319966521195_1418897 Interrupted.
>> > > > 2009-07-19 18:47:21,373 INFO
>> > > > org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder 0
>> for
>> > > > block
>> > > > blk_-346503319966521195_1418897 terminating
>> > > >
>> > > > 2009/7/19 mingyang <[email protected]>
>> > > >
>> > > > > But I added datanode still appear in this phenomenon, and I added a
>> > new
>> > > > > datanode, the above data only 10g
>> > > > > Strangely enough, wait about 10 minutes, hadoop will resume, I do
>> not
>> > > > know
>> > > > > what the problem is this, ask for help
>> > > > >
>> > > > >
>> > > > > it's log
>> > > > > 09/07/19 18:46:13 INFO hdfs.DFSClient: Waiting to find target node:
>> > > > > 192.168.1.97:50010
>> > > > > 09/07/19 18:46:13 INFO hdfs.DFSClient: Exception in
>> > > > createBlockOutputStream
>> > > > > java.net.SocketTimeoutException: 66000 millis timeout wh
>> > > > > ile waiting for channel to be ready for read. ch :
>> > > > > java.nio.channels.SocketChannel[connected local=/
>> 192.168.1.94:16153
>> > > > remote=/192.1
>> > > > > 68.1.97:50010]
>> > > > > 09/07/19 18:46:08 INFO hdfs.DFSClient: Waiting to find target node:
>> > > > > 192.168.1.97:50010
>> > > > > 09/07/19 18:46:08 INFO hdfs.DFSClient: Exception in
>> > > > createBlockOutputStream
>> > > > > java.net.SocketTimeoutException: 66000 millis timeout wh
>> > > > > ile waiting for channel to be ready for read. ch :
>> > > > > java.nio.channels.SocketChannel[connected local=/
>> 192.168.1.94:16152
>> > > > remote=/192.1
>> > > > > 68.1.97:50010]
>> > > > > 09/07/19 18:45:01 INFO hdfs.DFSClient: Waiting to find target node:
>> > > > > 192.168.1.105:50010
>> > > > > 09/07/19 18:45:01 INFO hdfs.DFSClient: Exception in
>> > > > createBlockOutputStream
>> > > > > java.io.IOException: Bad connect ack with firstBadLink 1
>> > > > > 92.168.1.97:50010
>> > > > > 09/07/19 18:44:56 INFO hdfs.DFSClient: Waiting to find target node:
>> > > > > 192.168.1.97:50010
>> > > > > 09/07/19 18:44:56 INFO hdfs.DFSClient: Exception in
>> > > > createBlockOutputStream
>> > > > > java.net.SocketTimeoutException: 66000 millis timeout wh
>> > > > > ile waiting for channel to be ready for read. ch :
>> > > > > java.nio.channels.SocketChannel[connected local=/
>> 192.168.1.94:16148
>> > > > remote=/192.1
>> > > > > 68.1.97:50010]
>> > > > >
>> > > > >
>> > > > >
>> > > > >
>> > > > >
>> > > > > 2009/7/17 mingyang <[email protected]>
>> > > > >
>> > > > > I run this, the result is far less than eight minutes, is not where
>> I
>> > > set
>> > > > >> up right? I just changed the fs.default.name, other settings are
>> > > > default
>> > > > >> values
>> > > > >>    <name> fs.default.name </ name>
>> > > > >>    <value> hdfs: / / named: 9000 </ value>
>> > > > >>
>> > > > >>
>> > > > >>
>> > > > >>    <name> mapred.job.tracker </ name>
>> > > > >>    <value> hdfs: / / named: 9001 </ value>
>> > > > >>
>> > > > >>
>> > > > >>
>> > > > >>    <name> hadoop.tmp.dir </ name>
>> > > > >>    <value> / hadoop / tmp </ value>
>> > > > >>
>> > > > >>    <name> dfs.name.dir </ name>
>> > > > >>    <value> / hadoop / name </ value>
>> > > > >>
>> > > > >>
>> > > > >>    <name> dfs.data.dir </ name>
>> > > > >>    <value> / b, / c, / d, / e, / f, / g, / h, / i, / j, / k, / l
>> </
>> > > > value>
>> > > > >>
>> > > > >>
>> > > > >>
>> > > > >>
>> > > > >>    <name> dfs.replication </ name>
>> > > > >>    <value> 1 </ value>
>> > > > >>
>> > > > >>
>> > > > >>    <name> dfs.block.size </ name>
>> > > > >>    <value> 2560000 </ value>
>> > > > >>
>> > > > >>
>> > > > >>
>> > > > >> time find / b / c / d / e / f / g / h / i / j / k / l-type
>> f-print>
>> > /
>> > > > dev
>> > > > >> / null
>> > > > >>
>> > > > >> real 0m4.608s
>> > > > >> user 0m1.489s
>> > > > >> sys 0m3.068s
>> > > > >>
>> > > > >>
>> > > > >> 2009/7/17 jason hadoop <[email protected]>
>> > > > >>
>> > > > >>> From the shell,
>> > > > >>> time find /b /c /d /e /f /g /h /i /j /k /l -type f -print >
>> > /dev/null
>> > > > >>>
>> > > > >>> Also, unless your distribution is magic, white space is not
>> ignored
>> > > in
>> > > > >>> value
>> > > > >>> statements. Hopefully your actual value is
>> > > > >>> <value>/b,/c,/d,/e,/f,/g,/h,/i,/j,/k,/l</value>
>> > > > >>>
>> > > > >>> On Fri, Jul 17, 2009 at 6:29 AM, mingyang <[email protected]>
>> > > > wrote:
>> > > > >>>
>> > > > >>> > Thank you for your reply, I am novice, I do not quite
>> understand
>> > > how
>> > > > to
>> > > > >>> > check dfs.data.dir more than eight minutes?
>> > > > >>> > My settings are dfs.data.dir
>> > > > >>> > <property>
>> > > > >>> >   <name> dfs.data.dir </ name>
>> > > > >>> >   <value> / b, / c, / d, / e, / f, / g, / h, / i, / j, / k, / l
>> > </
>> > > > >>> value>
>> > > > >>> >   </ property>
>> > > > >>> >
>> > > > >>> > Each directory has a disk mount
>> > > > >>> >
>> > > > >>> > 2009/7/17 jason hadoop <[email protected]>
>> > > > >>> >
>> > > > >>> > > Most likely problem is that the block report is taking more
>> > than
>> > > 10
>> > > > >>> > > minutes.
>> > > > >>> > > Due to the placement of the sync blocks in the core Datanode
>> > > code,
>> > > > >>> the
>> > > > >>> > > block
>> > > > >>> > > report locks out the heartbeat.
>> > > > >>> > > This can cause the namenode to think the datanode has
>> vanished.
>> > > > >>> > >
>> > > > >>> > > A simple way to check, is to run a find on the directory set
>> > > > >>> specified
>> > > > >>> > for
>> > > > >>> > > dfs.data.dir. If this find takes more than 8 minutes or so,
>> you
>> > > are
>> > > > >>> in
>> > > > >>> > > trouble.
>> > > > >>> > > The only solutions are to add more datanodes, reducing the
>> > block
>> > > > >>> count,
>> > > > >>> > or
>> > > > >>> > > increase your system io speed so that the block report may
>> > > complete
>> > > > >>> in
>> > > > >>> > > time.
>> > > > >>> > >
>> > > > >>> > > On Fri, Jul 17, 2009 at 6:12 AM, mingyang <
>> > [email protected]>
>> > > > >>> wrote:
>> > > > >>> > >
>> > > > >>> > > > i using hadoop storage my media files,, but when the number
>> > of
>> > > > >>> > documents
>> > > > >>> > > > when more than one million,
>> > > > >>> > > > Hadoop start about 10-20 minutes, my datanode automatically
>> > > down,
>> > > > >>> > > > namenode log shows that the loss of heart, but I see my
>> > normal
>> > > > >>> > datanode,
>> > > > >>> > > > port 50010 can be a normal telnet, use jps to see can see
>> > > > datanode
>> > > > >>> > still
>> > > > >>> > > > running, but at this time have been unable to put data to a
>> > > > hadoop,
>> > > > >>> I
>> > > > >>> > > guess
>> > > > >>> > > > datanode services is dead, hadoop does not support more
>> than
>> > > one
>> > > > >>> > million
>> > > > >>> > > > documents? How do I adjust those parameters? I have already
>> > set
>> > > > up
>> > > > >>> at
>> > > > >>> > the
>> > > > >>> > > > same time the number of open file 65535
>> > > > >>> > > >
>> > > > >>> > > >
>> > > > >>> > > > namenode log
>> > > > >>> > > >
>> > > > >>> > > > 2009-07-17 18:14:29,330 INFO
>> > > > >>> > > > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> > > > >>> > > >
>> > > > >>> > > > ugi=root,root,bin,daemon,sys,adm,disk,wheelip=/
>> 192.168.1.96
>> > > > >>> > > > cmd=setPermission
>> > > > >>> > > >
>> > > > >>> > > > src=/hadoop/tmp/mapred/system/jobtracker.info   dst=null
>> > > > >>> > > > perm=root:supergroup
>> > > > >>> > > > :rw-------
>> > > > >>> > > > 2009-07-17 18:14:29,336 INFO
>> > > org.apache.hadoop.hdfs.StateChange:
>> > > > >>> BLOCK*
>> > > > >>> > > > NameSystem.allocateBlock:
>> > > > >>> > > >
>> > > > >>> > > > /hadoop/tmp/mapred/system/jobtrack
>> > > > >>> > > > er.info. blk_-2148480138731090754_1403179
>> > > > >>> > > > 2009-07-17 18:14:32,958 INFO
>> > > org.apache.hadoop.hdfs.StateChange:
>> > > > >>> BLOCK*
>> > > > >>> > > > NameSystem.addStoredBlock: blockMap updated:
>> > > > >>> > > >
>> > > > >>> > > > 192.168.1.97:50
>> > > > >>> > > > 010 is added to blk_-2148480138731090754_1403179 size 4
>> > > > >>> > > > 2009-07-17 18:14:33,340 INFO
>> > > org.apache.hadoop.hdfs.StateChange:
>> > > > >>> DIR*
>> > > > >>> > > > NameSystem.completeFile: file
>> > > > >>> > > >
>> > > > >>> > > > /hadoop/tmp/mapred/system/jobtra
>> > > > >>> > > > cker.info is closed by DFSClient_1037557306
>> > > > >>> > > > 2009-07-17 18:16:21,349 INFO
>> > > > >>> > > > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Roll
>> > Edit
>> > > > Log
>> > > > >>> from
>> > > > >>> > > > 192.168.1.96
>> > > > >>> > > > 2009-07-17 18:16:21,349 INFO
>> > > > >>> > > > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number
>> > of
>> > > > >>> > > > transactions:
>> > > > >>> > > > 7 Total time for
>> > > > >>> > > >
>> > > > >>> > > > transacti
>> > > > >>> > > > ons(ms): 1Number of transactions batched in Syncs: 1 Number
>> > of
>> > > > >>> syncs: 6
>> > > > >>> > > > SyncTimes(ms): 9
>> > > > >>> > > > 2009-07-17 18:17:12,171 INFO
>> > > > >>> > > > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Roll
>> > > FSImage
>> > > > >>> from
>> > > > >>> > > > 192.168.1.96
>> > > > >>> > > > 2009-07-17 18:17:12,171 INFO
>> > > > >>> > > > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number
>> > of
>> > > > >>> > > > transactions:
>> > > > >>> > > > 0 Total time for
>> > > > >>> > > >
>> > > > >>> > > > transacti
>> > > > >>> > > > ons(ms): 0Number of transactions batched in Syncs: 0 Number
>> > of
>> > > > >>> syncs: 1
>> > > > >>> > > > SyncTimes(ms): 0
>> > > > >>> > > > 2009-07-17 18:51:00,566 INFO
>> > > org.apache.hadoop.hdfs.StateChange:
>> > > > >>> BLOCK*
>> > > > >>> > > > NameSystem.heartbeatCheck: lost heartbeat from
>> > > > >>> > > >
>> > > > >>> > > > 192.168.1.97:
>> > > > >>> > > > 50010
>> > > > >>> > > > 2009-07-17 18:51:25,383 INFO
>> > > > org.apache.hadoop.net.NetworkTopology:
>> > > > >>> > > > Removing
>> > > > >>> > > > a node: /default-rack/192.168.1.97:50010
>> > > > >>> > > > 2009-07-17 19:10:48,564 INFO
>> > > > >>> > > > org.apache.hadoop.hdfs.server.namenode.LeaseManager: Lease
>> > > > [Lease.
>> > > > >>> > >  Holder:
>> > > > >>> > > > DFSClient_-
>> > > > >>> > > >
>> > > > >>> > > > 1624377199, pend
>> > > > >>> > > > ingcreates: 69] has expired hard limit
>> > > > >>> > > > 2009-07-17 19:10:48,564 INFO
>> > > > >>> > > > org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>> > Recovering
>> > > > >>> > > > lease=[Lease.  Holder:
>> > > > >>> > > >
>> > > > >>> > > > DFSClient_-16243
>> > > > >>> > > > 77199, pendingcreates: 69],
>> > > > >>> > > >
>> > > > >>>
>> > > src=/unp/01/video/B3/94/{B394EDB2-0302-34B9-5357-4904FFFEFF36}_100.unp
>> > > > >>> > > >
>> > > > >>> > > > datanode log
>> > > > >>> > > >
>> > > > >>> > > > 2009-07-17 18:52:40,719 INFO
>> > > > >>> > > > org.apache.hadoop.hdfs.server.datanode.DataBlockScanner:
>> > > > >>> Verification
>> > > > >>> > > > succeeded for blk_664105388033641
>> > > > >>> > > > 1514_601647
>> > > > >>> > > > 2009-07-17 18:52:12,421 INFO
>> > > > >>> > > > org.apache.hadoop.hdfs.server.datanode.DataBlockScanner:
>> > > > >>> Verification
>> > > > >>> > > > succeeded for blk_-10747966535898
>> > > > >>> > > > 19594_1392025
>> > > > >>> > > > 2009-07-17 18:51:44,074 INFO
>> > > > >>> > > > org.apache.hadoop.hdfs.server.datanode.DataBlockScanner:
>> > > > >>> Verification
>> > > > >>> > > > succeeded for blk_-63504301593802
>> > > > >>> > > > 31402_155334
>> > > > >>> > > > 2009-07-17 18:51:12,760 INFO
>> > > > >>> > > > org.apache.hadoop.hdfs.server.datanode.DataBlockScanner:
>> > > > >>> Verification
>> > > > >>> > > > succeeded for blk_460729998775184
>> > > > >>> > > > 5359_395290
>> > > > >>> > > > 2009-07-17 18:50:39,977 INFO
>> > > > >>> > > > org.apache.hadoop.hdfs.server.datanode.DataBlockScanner:
>> > > > >>> Verification
>> > > > >>> > > > succeeded for blk_802918354954113
>> > > > >>> > > > 9011_474989
>> > > > >>> > > > 2009-07-17 18:50:11,707 INFO
>> > > > >>> > > > org.apache.hadoop.hdfs.server.datanode.DataBlockScanner:
>> > > > >>> Verification
>> > > > >>> > > > succeeded for blk_846865664811904
>> > > > >>> > > > 9754_1065465
>> > > > >>> > > > 2009-07-17 18:49:39,421 INFO
>> > > > >>> > > > org.apache.hadoop.hdfs.server.datanode.DataBlockScanner:
>> > > > >>> Verification
>> > > > >>> > > > succeeded for blk_473953565994615
>> > > > >>> > > > 8302_532204
>> > > > >>> > > > 2009-07-17 18:49:11,213 INFO
>> > > > >>> > > > org.apache.hadoop.hdfs.server.datanode.DataBlockScanner:
>> > > > >>> Verification
>> > > > >>> > > > succeeded for blk_-14950858387931
>> > > > >>> > > > 09024_354553
>> > > > >>> > > >
>> > > > >>> > > > 09/07/17 18:02:09 WARN hdfs.DFSClient: DFSOutputStream
>> > > > >>> > ResponseProcessor
>> > > > >>> > > > exception  for block
>> > > > >>> > > >
>> > > blk_-2536746364442878375_1403164java.net.SocketTimeoutException:
>> > > > >>> 63000
>> > > > >>> > > > millis timeout while waiting for channel to be ready for
>> > read.
>> > > ch
>> > > > :
>> > > > >>> > > > java.nio.channels.SocketChannel[connected local=/
>> > > > >>> 192.168.1.94:54783
>> > > > >>> > > remote=/
>> > > > >>> > > > 192.168.1.97:50010]
>> > > > >>> > > > 11096473        09/07/17 18:02:12 INFO hdfs.DFSClient:
>> > > Exception
>> > > > in
>> > > > >>> > > > createBlockOutputStream java.net.SocketTimeoutException:
>> > 63000
>> > > > >>> millis
>> > > > >>> > > > timeout while waiting for channel to be ready for read. ch
>> :
>> > > > >>> > > > java.nio.channels.SocketChannel[connected local=/
>> > > > >>> 192.168.1.94:54790
>> > > > >>> > > remote=/
>> > > > >>> > > > 192.168.1.97:50010]
>> > > > >>> > > > 11096475        09/07/17 18:02:12 INFO hdfs.DFSClient:
>> > > Exception
>> > > > in
>> > > > >>> > > > createBlockOutputStream java.net.SocketTimeoutException:
>> > 63000
>> > > > >>> millis
>> > > > >>> > > > timeout while waiting for channel to be ready for read. ch
>> :
>> > > > >>> > > > java.nio.channels.SocketChannel[connected local=/
>> > > > >>> 192.168.1.94:54791
>> > > > >>> > > remote=/
>> > > > >>> > > > 192.168.1.97:50010]
>> > > > >>> > > >
>> > > > >>> > >
>> > > > >>> > >
>> > > > >>> > >
>> > > > >>> > > --
>> > > > >>> > > Pro Hadoop, a book to guide you from beginner to hadoop
>> > mastery,
>> > > > >>> > > http://www.amazon.com/dp/1430219424?tag=jewlerymall
>> > > > >>> > > www.prohadoopbook.com a community for Hadoop Professionals
>> > > > >>> > >
>> > > > >>> >
>> > > > >>> >
>> > > > >>> >
>> > > > >>> > --
>> > > > >>> > 致
>> > > > >>> > 礼!
>> > > > >>> >
>> > > > >>> >
>> > > > >>> > 王明阳
>> > > > >>> >
>> > > > >>>
>> > > > >>>
>> > > > >>>
>> > > > >>> --
>> > > > >>> Pro Hadoop, a book to guide you from beginner to hadoop mastery,
>> > > > >>> http://www.amazon.com/dp/1430219424?tag=jewlerymall
>> > > > >>> www.prohadoopbook.com a community for Hadoop Professionals
>> > > > >>>
>> > > > >>
>> > > > >>
>> > > > >>
>> > > > >> --
>> > > > >> 致
>> > > > >> 礼!
>> > > > >>
>> > > > >>
>> > > > >> 王明阳
>> > > > >>
>> > > > >
>> > > > >
>> > > > >
>> > > > > --
>> > > > > 致
>> > > > > 礼!
>> > > > >
>> > > > >
>> > > > > 王明阳
>> > > > >
>> > > >
>> > > >
>> > > >
>> > > > --
>> > > > 致
>> > > > 礼!
>> > > >
>> > > >
>> > > > 王明阳
>> > > >
>> > >
>> > >
>> > >
>> > > --
>> > > Pro Hadoop, a book to guide you from beginner to hadoop mastery,
>> > > http://www.amazon.com/dp/1430219424?tag=jewlerymall
>> > > www.prohadoopbook.com a community for Hadoop Professionals
>> > >
>> >
>> >
>> >
>> > --
>> > 致
>> > 礼!
>> >
>> >
>> > 王明阳
>> >
>>
>
>
>
> --
> 致
> 礼!
>
>
> 王明阳
>

Reply via email to