[ 
https://issues.apache.org/jira/browse/HDFS-12996?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16336172#comment-16336172
 ] 

Hanisha Koneru commented on HDFS-12996:
---------------------------------------

Thanks [~aw] for reviewing the doc.

We considered something along similar lines. However we don’t want to take 
automatic snapshots on behalf of the user as it would not relieve the NN 
memory, as expected by the user. For example, say the user deletes 10M files 
and expects the NN memory usage to go down. Automatic snapshots would prevent 
this.

> DataNode Replica Trash
> ----------------------
>
>                 Key: HDFS-12996
>                 URL: https://issues.apache.org/jira/browse/HDFS-12996
>             Project: Hadoop HDFS
>          Issue Type: New Feature
>            Reporter: Hanisha Koneru
>            Assignee: Hanisha Koneru
>            Priority: Major
>         Attachments: DataNode_Replica_Trash_Design_Doc.pdf
>
>
> DataNode Replica Trash will allow administrators to recover from a recent 
> delete request that resulted in catastrophic loss of user data. This is 
> achieved by placing all invalidated blocks in a replica trash on the datanode 
> before completely purging them from the system. The design doc is attached 
> here.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to