[ https://issues.apache.org/jira/browse/HDFS-3788?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13434634#comment-13434634 ]
Tsz Wo (Nicholas), SZE commented on HDFS-3788: ---------------------------------------------- > I believe it's legitimate to send a content-length (if known) with a chunked > response. ... I believe it is not. See below from http://www.w3.org/Protocols/rfc2616/rfc2616-sec4.html#sec4.4 bq. The Content-Length header field MUST NOT be sent if these two lengths are different (i.e., if a Transfer-Encoding header field is present) > I think a test case would be invaluable since the file size issue has reared > itself a few times. Could you add a test that uses a mock? I did have a mock test but it requires changing DatanodeWebHdfsMethods. I don't see an easy way to have mock tests without changing the main code. Do you have any idea? If yes, could you add the tests? > distcp can't copy large files using webhdfs due to missing Content-Length > header > -------------------------------------------------------------------------------- > > Key: HDFS-3788 > URL: https://issues.apache.org/jira/browse/HDFS-3788 > Project: Hadoop HDFS > Issue Type: Bug > Components: webhdfs > Affects Versions: 0.23.3, 2.0.0-alpha > Reporter: Eli Collins > Assignee: Tsz Wo (Nicholas), SZE > Priority: Critical > Attachments: distcp-webhdfs-errors.txt, h3788_20120813.patch, > h3788_20120814.patch > > > The following command fails when data1 contains a 3gb file. It passes when > using hftp or when the directory just contains smaller (<2gb) files, so looks > like a webhdfs issue with large files. > {{hadoop distcp webhdfs://eli-thinkpad:50070/user/eli/data1 > hdfs://localhost:8020/user/eli/data2}} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira