[ 
https://issues.apache.org/jira/browse/HDFS-13314?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16405556#comment-16405556
 ] 

Arpit Agarwal commented on HDFS-13314:
--------------------------------------

Hi Rushabh,

bq. You need to change the namenode code and build again to bypass
Correct.

bq. On top of that if we don't purge the old images in case of corruption, we 
will run out of disk space in 3-4 weeks.
Impossible, as you will need to restart the standby to get any new checkpoints. 
And it will fail to restart because the latest fsimage is bad.

bq. IMO instead of putting a hack in namenode, we should actively chase the 
bugs and root cause them.
This is not a hack - it is the equivalent of an {{assert}}. We have been 
chasing this bug for a few weeks without success. Other community members have 
been doing the same for a while (please see HDFS-13101 and comments on 
HDFS-9406). To root cause the problem, we need (1) a good image and (2) the 
sequence of write transactions causing the corruption.

> NameNode should optionally exit if it detects FsImage corruption
> ----------------------------------------------------------------
>
>                 Key: HDFS-13314
>                 URL: https://issues.apache.org/jira/browse/HDFS-13314
>             Project: Hadoop HDFS
>          Issue Type: Improvement
>          Components: namenode
>            Reporter: Arpit Agarwal
>            Assignee: Arpit Agarwal
>            Priority: Major
>         Attachments: HDFS-13314.01.patch, HDFS-13314.02.patch
>
>
> The NameNode should optionally exit after writing an FsImage if it detects 
> the following kinds of corruptions:
> # INodeReference pointing to non-existent INode
> # Duplicate entries in snapshot deleted diff list.
> This behavior is controlled via an undocumented configuration setting, and 
> disabled by default.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to