[ https://issues.apache.org/jira/browse/HADOOP-1629?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12519494 ]
Raghu Angadi commented on HADOOP-1629: -------------------------------------- This conflicts with HADOOP-1621 on the trunk. Will submit different patches for 14 and trunk. > Block CRC Unit Tests: upgrade test > ---------------------------------- > > Key: HADOOP-1629 > URL: https://issues.apache.org/jira/browse/HADOOP-1629 > Project: Hadoop > Issue Type: Test > Components: dfs > Affects Versions: 0.14.0 > Reporter: Nigel Daley > Assignee: Raghu Angadi > Priority: Blocker > Fix For: 0.14.0 > > Attachments: hadoop-12-dfs-dir.tgz, HADOOP-1629.patch, > HADOOP-1629.patch, HADOOP-1629.patch > > > HADOOP-1286 introduced a distributed upgrade framework. 1 or more unit tests > should be developed that start with a zipped up Hadoop 0.12 file system (that > is included in Hadoop's src/test directory under version controlled) and > attempts to upgrade it to the current version of Hadoop (ie the version that > the tests are running against). The zipped up file system should include > some "interesting" files, such as: > - zero length files > - file with replication set higher than number of datanodes > - file with no .crc file > - file with corrupt .crc file > - file with multiple blocks (will need to set dfs.block.size to a small value) > - file with multiple checksum blocks > - empty directory > - all of the above again but with a different io.bytes.per.checksum setting > The class that generates the zipped up file system should also be included in > this patch. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.