[ https://issues.apache.org/jira/browse/HADOOP-5598?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12719234#action_12719234 ]
Todd Lipcon commented on HADOOP-5598: ------------------------------------- The other solution that would solve this whole issue is to simply not checksum the data in FSOutputSummer until an entire chunk has been buffered up. This would bring the "write size" up to io.bytes.per.checksum, 512 by default, which performs better in java.util.zip.CRC32 than in the Java implementation. Doug expressed some concern over that idea in HADOOP-5318, I guess since data can sit in memory for an arbitrarily long amount of time before getting checksummed, but I would imagine most people are using ECC RAM anyway :) > Implement a pure Java CRC32 calculator > -------------------------------------- > > Key: HADOOP-5598 > URL: https://issues.apache.org/jira/browse/HADOOP-5598 > Project: Hadoop Core > Issue Type: Improvement > Components: dfs > Reporter: Owen O'Malley > Assignee: Todd Lipcon > Attachments: crc32-results.txt, hadoop-5598-hybrid.txt, > hadoop-5598.txt, TestCrc32Performance.java, TestCrc32Performance.java > > > We've seen a reducer writing 200MB to HDFS with replication = 1 spending a > long time in crc calculation. In particular, it was spending 5 seconds in crc > calculation out of a total of 6 for the write. I suspect that it is the > java-jni border that is causing us grief. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.