[
https://issues.apache.org/jira/browse/HBASE-10052?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13838409#comment-13838409
]
Enis Soztutar commented on HBASE-10052:
---------------------------------------
Sounds good. Any API we can use to direct hdfs to NOT cache our hfiles in the
first place. We are already caching the blocks, and it is better to have hdfs
to not double cache, but instead spend the memory elsewhere. Since only the
HBase layer knows about block boundaries, block access patterns and block types
(index, data), HBase cache should always be better in theory.
> use HDFS advisory caching to avoid caching HFiles that are not going to be
> read again (because they are being compacted)
> ------------------------------------------------------------------------------------------------------------------------
>
> Key: HBASE-10052
> URL: https://issues.apache.org/jira/browse/HBASE-10052
> Project: HBase
> Issue Type: Improvement
> Reporter: Colin Patrick McCabe
> Fix For: 0.98.0
>
>
> HBase can benefit from doing dropbehind during compaction since compacted
> files are not read again. HDFS advisory caching, introduced in HDFS-4817,
> can help here. The right API here is {{DataInputStream#setDropBehind}}.
--
This message was sent by Atlassian JIRA
(v6.1#6144)