[
https://issues.apache.org/jira/browse/HDFS-7480?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14266045#comment-14266045
]
Frode Halvorsen commented on HDFS-7480:
---------------------------------------
2.6.1 is not out yet, but one thought; This fix might resolve the issue when
namenodes are started with a lot of incoming information about 'loose'
data-blokcs, but it probably won't resolve the issue that causes the namenodes
to be killed by zookeeper when I delete a lot of files.
Athe the delete-moment, I don't think that the logging is that problematic.
The logging-issue, I believe, is secondary. I believe that the active namenode
gets busy calculating/distributing delete-orders to datanodes when I drop
500.000 files at once, and that this is the causer fo the zookeeper-shutdown.
When the namenode gets overloaded with caclulating/distributing those
delete-orders, it doesn't keep up with responses to zoo-keeper, which the kills
the namenode in order to failover to NN2.
> Namenodes loops on 'block does not belong to any file' after deleting many
> files
> --------------------------------------------------------------------------------
>
> Key: HDFS-7480
> URL: https://issues.apache.org/jira/browse/HDFS-7480
> Project: Hadoop HDFS
> Issue Type: Bug
> Affects Versions: 2.5.0
> Environment: CentOS - HDFS-HA (journal), zookeeper
> Reporter: Frode Halvorsen
>
> A small cluster has 8 servers with 32 G RAM.
> Two is namenodes (HA-configured), six is Datanodes (8x3 TB disks configured
> with RAID as one 21 TB drive).
> The cluster recieves avg 400.000 small files each day. I started archiving
> (HAR) each day as separate archives. After deleting the orinigal files for
> one month, the namenodes stared acting up really bad.
> When restaring those, both active and passive nodes seems to work OK for some
> time, but then starts to report a lot of blocks belonging to no files, and
> the name-node just spins those messages in a massive loop. If the passive
> node is first, it also influences the active node in susch a way that it's no
> longer possible to archive new files. If the active node also starts in this
> loop, it suddenly dies without any error-message.
> The only way I'm able to get rid of the problem, is to start decommission
> nodes, watching the cluster closely to avoid downtime, and make sure every
> datanode gets a 'clean' start. After all datanodes has been decommisioned (in
> turns), and restarted with clean disks, the problem is gone. But if I then
> delete a lot of files in a short time, the problem starts again...
> The main problem (I think), is that the recieving and reporting of those
> blocks takes so many resources, that the namenodes is too busy to tell the
> datanodes to delete those blocks..
> If the active name-node starts on the loop, it does the 'right' thing by
> telling the datanode to invalidate the block, But the amount of blocks is so
> massive, that the namenode doesn't do anything else. Just now, I have about
> 1200-1400 log-entries pr second in the passive node.
> update :
> Just got the active namenode in the loop - it logs 1000 lines pr second.
> 500 'BlockStateChange: BLOCK* processReport: blk_1080796332_7056241 on
> x.x.x.x:50010 size 1742 does not belong to any file'
> and
> 500 ' BlockStateChange: BLOCK* InvalidateBlocks: add blk_1080796332_7056241
> to x.x.x.x:50010'
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)