[ 
https://issues.apache.org/jira/browse/HDFS-9819?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15151684#comment-15151684
 ] 

Vinayakumar B commented on HDFS-9819:
-------------------------------------

Since HDFS-8845, deleting some of the blocks inside <bpDir>/finalized will not 
cause any harm by DiskChecker, as it will not check recursively.
But if you delete <bpDir> itself, as [~kihwal] mentioned it looks really 
strange. This level of accidental deletion should not be ignored.

> FsVolume should tolerate few times check-dir failed due to deletion by mistake
> ------------------------------------------------------------------------------
>
>                 Key: HDFS-9819
>                 URL: https://issues.apache.org/jira/browse/HDFS-9819
>             Project: Hadoop HDFS
>          Issue Type: Bug
>    Affects Versions: 2.7.1
>            Reporter: Lin Yiqun
>            Assignee: Lin Yiqun
>         Attachments: HDFS-9819.001.patch
>
>
> FsVolume should tolerate few times check-dir failed because sometimes we will 
> do a delete dir/file operation by mistake in datanode data-dirs. Then the 
> {{DataNode#startCheckDiskErrorThread}} will invoking checkDir method 
> periodicity and find dir not existed, throw exception. The checked volume 
> will be added to failed volume list. The blocks on this volume will be 
> replicated again. But actually, this is not needed to do. We should let 
> volume can be tolerated few times check-dir failed like config 
> {{dfs.datanode.failed.volumes.tolerated}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to