[
https://issues.apache.org/jira/browse/HADOOP-2797?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12915086#action_12915086
]
Tanping Wang commented on HADOOP-2797:
--------------------------------------
The test file images used , hadoop-14-dfs-dir.tgz as for now are generated and
uploaded only. Based on the comments in hadoop-1629, these FSimages suppose
to contain the various categories
* zero length files
* file with replication set higher than number of datanodes
* file with no .crc file
* file with corrupt .crc file
* file with multiple blocks (will need to set dfs.block.size to a small
value)
* file with multiple checksum blocks
* empty directory
* all of the above again but with a different io.bytes.per.checksum setting
As FSImage structure could be changed in future releases, developers need to
generate new FS images. There should be at least test program that can
generate these FSImages contained in version control. Otherwise, this test is
not very sustainable.
> Withdraw CRC upgrade from HDFS
> ------------------------------
>
> Key: HADOOP-2797
> URL: https://issues.apache.org/jira/browse/HADOOP-2797
> Project: Hadoop Common
> Issue Type: New Feature
> Affects Versions: 0.13.0
> Reporter: Robert Chansler
> Assignee: Raghu Angadi
> Priority: Critical
> Fix For: 0.18.0
>
> Attachments: hadoop-14-dfs-dir.tgz, HADOOP-2797.patch,
> HADOOP-2797.patch, HADOOP-2797.patch, HADOOP-2797.patch
>
>
> HDFS will no longer support upgrades from versions without CRCs for block
> data. Users upgrading from version 0.13 or earlier must first upgrade to an
> intermediate (0.14, 0.15, 0.16, 0.17) version before upgrade to 0.18 or later.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.