[
https://issues.apache.org/jira/browse/HDFS-13314?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16407196#comment-16407196
]
Arpit Agarwal commented on HDFS-13314:
--------------------------------------
bq. How is the "safe" choice to knowingly write a corrupt image? One that
renders the NN incapable of starting up?
Once we get to this point, the metadata is already corrupt. Writing out a new
FsImage doesn't make it any worse because replaying the prior image and edits
would lead to the same state.
bq. We need to immediately do a full stop anytime data structures are known to
be corrupt.
That is my intention. Looks like you've gone over the patch. Is there anything
you suggest doing differently?
> NameNode should optionally exit if it detects FsImage corruption
> ----------------------------------------------------------------
>
> Key: HDFS-13314
> URL: https://issues.apache.org/jira/browse/HDFS-13314
> Project: Hadoop HDFS
> Issue Type: Improvement
> Components: namenode
> Reporter: Arpit Agarwal
> Assignee: Arpit Agarwal
> Priority: Major
> Attachments: HDFS-13314.01.patch, HDFS-13314.02.patch,
> HDFS-13314.03.patch
>
>
> The NameNode should optionally exit after writing an FsImage if it detects
> the following kinds of corruptions:
> # INodeReference pointing to non-existent INode
> # Duplicate entries in snapshot deleted diff list.
> This behavior is controlled via an undocumented configuration setting, and
> disabled by default.
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]