[ 
https://issues.apache.org/jira/browse/HDFS-13314?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16407196#comment-16407196
 ] 

Arpit Agarwal commented on HDFS-13314:
--------------------------------------

bq. How is the "safe" choice to knowingly write a corrupt image? One that 
renders the NN incapable of starting up?
Once we get to this point, the metadata is already corrupt. Writing out a new 
FsImage doesn't make it any worse because replaying the prior image and edits 
would lead to the same state. 

bq. We need to immediately do a full stop anytime data structures are known to 
be corrupt.
That is my intention. Looks like you've gone over the patch. Is there anything 
you suggest doing differently?

> NameNode should optionally exit if it detects FsImage corruption
> ----------------------------------------------------------------
>
>                 Key: HDFS-13314
>                 URL: https://issues.apache.org/jira/browse/HDFS-13314
>             Project: Hadoop HDFS
>          Issue Type: Improvement
>          Components: namenode
>            Reporter: Arpit Agarwal
>            Assignee: Arpit Agarwal
>            Priority: Major
>         Attachments: HDFS-13314.01.patch, HDFS-13314.02.patch, 
> HDFS-13314.03.patch
>
>
> The NameNode should optionally exit after writing an FsImage if it detects 
> the following kinds of corruptions:
> # INodeReference pointing to non-existent INode
> # Duplicate entries in snapshot deleted diff list.
> This behavior is controlled via an undocumented configuration setting, and 
> disabled by default.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to