[
https://issues.apache.org/jira/browse/HADOOP-2063?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12563213#action_12563213
]
Marco Nicosia commented on HADOOP-2063:
---------------------------------------
I'm disappointed this didn't go in for Hadoop 0.16, despite having been set as
target for 0.16. I understand that this is a pretty big change, but I really
want to be sure we get something for Hadoop 0.17.
While we wait for this, any old Hadoop DFS' with corrupted files will need to
sit, waiting for their owners to have a way to retrieve the files. For that
time fsck will always return corrupt. The inability to do anything with these
files (except delete them) could be masking us from being able to detect other
hadoop issues.
> Command to pull corrupted files
> -------------------------------
>
> Key: HADOOP-2063
> URL: https://issues.apache.org/jira/browse/HADOOP-2063
> Project: Hadoop Core
> Issue Type: New Feature
> Components: fs
> Affects Versions: 0.14.1
> Reporter: Koji Noguchi
> Priority: Blocker
> Fix For: 0.17.0
>
>
> Before 0.14, dfs -get didn't perform checksum checking.
> Users were able to download the corrupted files to see if they want to delete
> them.
> After 0.14, dfs -get also does the checksumming.
> Requesting a command for no-checksum-get command.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.