[ https://issues.apache.org/jira/browse/HADOOP-8615?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13422383#comment-13422383 ]
Tim Broberg commented on HADOOP-8615: ------------------------------------- I think I like thomastechs' idea. It has the advantage that it makes the codec figure out whether the problem is in the stream or not. It does mean patching every codec at once and then each client one by one. Really, in any scheme, we would need to be clear that the stream is the problem in any case, so touching the codecs seems unavoidable. Are there issues for clients where they have problems knowing the stream source too? Does this ripple through a bunch of layers? > EOFException in DecompressorStream.java needs to be more verbose > ---------------------------------------------------------------- > > Key: HADOOP-8615 > URL: https://issues.apache.org/jira/browse/HADOOP-8615 > Project: Hadoop Common > Issue Type: Bug > Components: io > Affects Versions: 0.20.2 > Reporter: Jeff Lord > > In ./src/core/org/apache/hadoop/io/compress/DecompressorStream.java > The following exception should at least pass back the file that it encounters > this error in relation to: > protected void getCompressedData() throws IOException { > checkStream(); > int n = in.read(buffer, 0, buffer.length); > if (n == -1) { > throw new EOFException("Unexpected end of input stream"); > } > This would help greatly to debug bad/corrupt files. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira