[ 
https://issues.apache.org/jira/browse/HDFS-12996?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16326698#comment-16326698
 ] 

Allen Wittenauer commented on HDFS-12996:
-----------------------------------------

bq. Also the design looks very similar to Checkpointing/Snapshots.

The fact that this JIRA even exists points that snapshots are/were a failure.  
On other file systems, snapshots are exactly the recovery model for these types 
of deletes.

...

Reading through the doc, there are handful of spots where I see the use cases 
are extremely limited.  But I'm really left with a basic question:

Why isn't there an option to just have the NN automatically do a snapshot for 
deletes over a certain size instead and then automatically delete these 
snapshots after X amount time?  Wouldn't that add the protection that is being 
requested while avoiding the requirement to restart the NN? 



> DataNode Replica Trash
> ----------------------
>
>                 Key: HDFS-12996
>                 URL: https://issues.apache.org/jira/browse/HDFS-12996
>             Project: Hadoop HDFS
>          Issue Type: New Feature
>            Reporter: Hanisha Koneru
>            Assignee: Hanisha Koneru
>            Priority: Major
>         Attachments: DataNode_Replica_Trash_Design_Doc.pdf
>
>
> DataNode Replica Trash will allow administrators to recover from a recent 
> delete request that resulted in catastrophic loss of user data. This is 
> achieved by placing all invalidated blocks in a replica trash on the datanode 
> before completely purging them from the system. The design doc is attached 
> here.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to