[ 
https://issues.apache.org/jira/browse/HDFS-7480?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14266045#comment-14266045
 ] 

Frode Halvorsen commented on HDFS-7480:
---------------------------------------

2.6.1 is not out yet, but one thought; This fix might resolve the issue when 
namenodes are started with a lot of incoming information about 'loose' 
data-blokcs, but it probably won't resolve the issue that causes the namenodes 
to be killed by zookeeper when I delete a lot of files.
Athe the delete-moment, I don't think that the logging is that problematic.
The logging-issue, I believe, is secondary. I believe that the active namenode 
gets busy calculating/distributing delete-orders to datanodes when I drop 
500.000 files at once, and that this is the causer fo the zookeeper-shutdown. 
When the namenode gets overloaded with caclulating/distributing those 
delete-orders, it doesn't keep up with responses to zoo-keeper, which the kills 
the namenode in order to failover to NN2.

> Namenodes loops on 'block does not belong to any file' after deleting many 
> files
> --------------------------------------------------------------------------------
>
>                 Key: HDFS-7480
>                 URL: https://issues.apache.org/jira/browse/HDFS-7480
>             Project: Hadoop HDFS
>          Issue Type: Bug
>    Affects Versions: 2.5.0
>         Environment: CentOS - HDFS-HA (journal), zookeeper
>            Reporter: Frode Halvorsen
>
> A small cluster has 8 servers with 32 G RAM.
> Two is namenodes (HA-configured), six is Datanodes (8x3 TB disks configured 
> with RAID as one 21 TB drive).
> The cluster recieves avg 400.000 small files each day. I started archiving 
> (HAR) each day as separate archives. After deleting the orinigal files for 
> one month, the namenodes stared acting up really bad.
> When restaring those, both active and passive nodes seems to work OK for some 
> time, but then starts to report a lot of blocks belonging to no files, and 
> the name-node just spins those messages in a massive loop. If the passive 
> node is first, it also influences the active node in susch a way that it's no 
> longer possible to archive new files. If the active node also starts in this 
> loop, it suddenly dies without any error-message.
> The only way I'm able to get rid of the problem, is to start decommission 
> nodes, watching the cluster closely to avoid downtime, and make sure every 
> datanode gets a 'clean' start. After all datanodes has been decommisioned (in 
> turns), and restarted with clean disks, the problem is gone. But if I then 
> delete a lot of files in a short time, the problem starts again...  
> The main problem (I think), is that the recieving and reporting of those 
> blocks takes so many resources, that the namenodes is too busy to tell the 
> datanodes to delete those blocks.. 
> If the active name-node starts on the loop, it does the 'right' thing by 
> telling the datanode to invalidate the block, But the amount of blocks is so 
> massive, that the namenode doesn't do anything else. Just now, I have about 
> 1200-1400 log-entries pr second in the passive node.
> update :
> Just got the active namenode in the loop - it logs 1000 lines pr second. 
> 500 'BlockStateChange: BLOCK* processReport: blk_1080796332_7056241 on 
> x.x.x.x:50010 size 1742 does not belong to any file'
> and 
> 500 ' BlockStateChange: BLOCK* InvalidateBlocks: add blk_1080796332_7056241 
> to x.x.x.x:50010'



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to