[
https://issues.apache.org/jira/browse/HDFS-7686?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14317262#comment-14317262
]
Andrew Wang commented on HDFS-7686:
-----------------------------------
Thanks for finding this Rushabh, thanks Colin for the patch. A few light review
comments:
* Unused Iterator import in the test, LoadingCache import in VolumeScanner
* I like the cache since it's a nice way of preventing scanning the same blocks
over and over again, but it'd be good to also use a LinkedHashMap instead of
the LinkedList and also check existence in there before adding. That way we
never have dupes in the suspect queue. It seems possible to have a working set
bigger than the 1000 element cache size, like if an entire disk goes bad.
Otherwise looks good!
> Re-add rapid rescan of possibly corrupt block feature to the block scanner
> --------------------------------------------------------------------------
>
> Key: HDFS-7686
> URL: https://issues.apache.org/jira/browse/HDFS-7686
> Project: Hadoop HDFS
> Issue Type: Bug
> Affects Versions: 3.0.0
> Reporter: Rushabh S Shah
> Assignee: Colin Patrick McCabe
> Priority: Blocker
> Attachments: HDFS-7686.002.patch
>
>
> When doing a transferTo (aka sendfile operation) from the DataNode to a
> client, we may hit an I/O error from the disk. If we believe this is the
> case, we should be able to tell the block scanner to rescan that block soon.
> The feature was originally implemented in HDFS-7548 but was removed by
> HDFS-7430. We should re-add it.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)