[ 
https://issues.apache.org/jira/browse/HDFS-8617?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14596748#comment-14596748
 ] 

Andrew Wang commented on HDFS-8617:
-----------------------------------

bq. As you mentioned in the evaluation, there are adverse effects on 
throughputs...The proposed solution looks like lowering the utilization of the 
cluster (at the cost of making checkDir() really slow) to meet the SLOs.

I'd like to turn this question around and ask: is there a downside to 
throttling checkDisk throughput? We might end up taking longer to detect a bad 
disk, but this is not a performance-critical workload.

Here's also another idea for a throttle: spend at most x% of time doing 
checkDisk work. Maybe we say it can only run for 250ms of every 1000ms 
interval. Timeslicing like this automatically tunes for faster vs. slower IO 
rates.

> Throttle DiskChecker#checkDirs() speed.
> ---------------------------------------
>
>                 Key: HDFS-8617
>                 URL: https://issues.apache.org/jira/browse/HDFS-8617
>             Project: Hadoop HDFS
>          Issue Type: Improvement
>          Components: HDFS
>    Affects Versions: 2.7.0
>            Reporter: Lei (Eddy) Xu
>            Assignee: Lei (Eddy) Xu
>         Attachments: HDFS-8617.000.patch
>
>
> As described in HDFS-8564,  {{DiskChecker.checkDirs(finalizedDir)}} is 
> causing excessive I/Os because {{finalizedDirs}} might have up to 64K 
> sub-directories (HDFS-6482).
> This patch proposes to limit the rate of IO operations in 
> {{DiskChecker.checkDirs()}}. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to