[ https://issues.apache.org/jira/browse/HBASE-14307?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14712608#comment-14712608 ]
Shradha Revankar commented on HBASE-14307: ------------------------------------------ Yes, but it is not guaranteed that it will even read as much as the 'size' using the positional read api, shouldn't there be a loop to read until at least 'size'. We tried running hbase with WebhdfsFilesystem (with server implementation that sets http header for Transfer-encoding as chunked encoding, there is no content-length present), the positional read api reads only the first chunk which is far less than the size. Unless there is a loop, the rest of the bytes are not read. We ended up getting errors like this : Caused by: java.io.IOException: Positional read of 16425 bytes failed at offset 4132767 (returned 26) at org.apache.hadoop.hbase.io.hfile.HFileBlock$AbstractFSReader.readAtOffset(HFileBlock.java:1322) > Incorrect use of positional read api in HFileBlock > -------------------------------------------------- > > Key: HBASE-14307 > URL: https://issues.apache.org/jira/browse/HBASE-14307 > Project: HBase > Issue Type: Bug > Reporter: Shradha Revankar > Priority: Minor > > Considering that {{read()}} is not guaranteed to read all bytes, > I'm interested to understand this particular piece of code and why is partial > read treated as an error : > https://github.com/apache/hbase/blob/master/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java#L1446-L1450 > Particularly, if hbase were to use a different filesystem, say > WebhdfsFileSystem, this would not work, please also see > https://issues.apache.org/jira/browse/HDFS-8943 for discussion around this. -- This message was sent by Atlassian JIRA (v6.3.4#6332)