[
https://issues.apache.org/jira/browse/HDFS-13277?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16410549#comment-16410549
]
Ajay Kumar commented on HDFS-13277:
-----------------------------------
[~bharatviswa] thanks for working on this. Patch 2 looks good. Some suggestions:
* FsDatasetAsyncDiskService L346: Do we need this else if check as previous
else checks if it is equal.
* javadoc for ReplicaTrashInfo & FsDatasetAsyncDiskService#replicaTrashInfoMap
* Rename ReplicaTrashInfo to ReplicaTrashCurDirInfo ?
* FsDatasetAsyncDiskService L102: rename replicaTrashSubDirMax to
replicaTrashSubDirMaxBlocks or replicaTrashSubDirMaxFiles ?
> Improve move to account for usage (number of files) to limit trash dir size
> ---------------------------------------------------------------------------
>
> Key: HDFS-13277
> URL: https://issues.apache.org/jira/browse/HDFS-13277
> Project: Hadoop HDFS
> Issue Type: Sub-task
> Reporter: Bharat Viswanadham
> Assignee: Bharat Viswanadham
> Priority: Major
> Attachments: HDFS-13277-HDFS-12996.00.patch,
> HDFS-13277-HDFS-12996.01.patch, HDFS-13277-HDFS-12996.02.patch
>
>
> The trash subdirectory maximum entries. This puts an upper limit on the size
> of subdirectories in replica-trash. Set this default value to
> blockinvalidateLimit.
>
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]