[
https://issues.apache.org/jira/browse/HDFS-1403?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12913908#action_12913908
]
sam rash commented on HDFS-1403:
--------------------------------
can you elaborate?
also, this truncate option will have to work on open files. I think
-list-corruptfiles only works on closed ones. we have to handle the missing
last block problem (the main reason I filed this)
> add -truncate option to fsck
> ----------------------------
>
> Key: HDFS-1403
> URL: https://issues.apache.org/jira/browse/HDFS-1403
> Project: Hadoop HDFS
> Issue Type: New Feature
> Components: hdfs client, name-node
> Reporter: sam rash
>
> When running fsck, it would be useful to be able to tell hdfs to truncate any
> corrupt file to the last valid position in the latest block. Then, when
> running hadoop fsck, an admin can cleanup the filesystem.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.