[ 
https://issues.apache.org/jira/browse/HBASE-18203?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16045581#comment-16045581
 ] 

Allan Yang commented on HBASE-18203:
------------------------------------

It is a very useful feature. We often encounter "open too many files" exception 
if storefile number is too many on a single RS. the RS process holds too many 
handlers(tcp connections) and grows unboundedly with storefile numbers.

> Intelligently manage a pool of open references to store files
> -------------------------------------------------------------
>
>                 Key: HBASE-18203
>                 URL: https://issues.apache.org/jira/browse/HBASE-18203
>             Project: HBase
>          Issue Type: Improvement
>          Components: regionserver
>    Affects Versions: 2.0.0
>            Reporter: Andrew Purtell
>
> When bringing a region online we open every store file and keep the file 
> open, to avoid further round trips to the HDFS namenode during reads. Naively 
> keeping open every store file we encounter is a bad idea. There should be an 
> upper bound. We should close and reopen files as needed once we are above the 
> upper bound. We should choose candidates to close on a LRU basis. Otherwise 
> we can (and some users have in production) overrun high (~64k) open file 
> handle limits on the server if the aggregate number of store files is too 
> large. 
> Note the 'open files' here refers to open/active references to files at the 
> HDFS level. How this maps to active file descriptors at the OS level depends 
> on concurrency of access (block transfers, short circuit reads). The more 
> open files we have at the HDFS level the higher number of OS level file 
> handles we can expect to consume.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

Reply via email to