[
https://issues.apache.org/jira/browse/COMPRESS-16?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13218088#comment-13218088
]
Stefan Bodewig commented on COMPRESS-16:
----------------------------------------
Our code is wrong.
to_chars in src/create.c in GNU tar only uses the remaining bytes and sets the
first one to 255 or 128 for negative/positive numbers. Negative numbers only
occur in time fields where we don't support anything non-octal ATM anyway, so
this isn't a real problem right now. It becomes one if we support star/GNU
tar/POSIX dialects for the other numeric fields as well. This would be
required for COMPRESS-177.
I suggest to broaden and reopen COMPRESS-177 to something like "extend
STAR/POSIX support to all numeric fields" or alternatively create a new issue
and close this one again.
> unable to extract a TAR file that contains an entry which is 10 GB in size
> --------------------------------------------------------------------------
>
> Key: COMPRESS-16
> URL: https://issues.apache.org/jira/browse/COMPRESS-16
> Project: Commons Compress
> Issue Type: Bug
> Components: Archivers
> Environment: I am using win xp sp3, but this should be platform
> independent.
> Reporter: Sam Smith
> Fix For: 1.4
>
> Attachments:
> 0001-Accept-GNU-tar-files-with-entries-over-8GB-in-size.patch,
> 0002-Allow-creating-tar-archives-with-files-over-8GB.patch,
> 0004-Prefer-octal-over-binary-size-representation.patch, ant-8GB-tar.patch,
> patch-for-compress.txt
>
>
> I made a TAR file which contains a file entry where the file is 10 GB in size.
> When I attempt to extract the file using TarInputStream, it fails with the
> following stack trace:
> java.io.IOException: unexpected EOF with 24064 bytes unread
> at
> org.apache.commons.compress.archivers.tar.TarInputStream.read(TarInputStream.java:348)
> at
> org.apache.commons.compress.archivers.tar.TarInputStream.copyEntryContents(TarInputStream.java:388)
> So, TarInputStream does not seem to support large (> 8 GB?) files.
> Here is something else to note: I created that TAR file using TarOutputStream
> , which did not complain when asked to write a 10 GB file into the TAR file,
> so I assume that TarOutputStream has no file size limits? That, or does it
> silently create corrupted TAR files (which would be the worst situation of
> all...)?
--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators:
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira