[ 
https://issues.apache.org/jira/browse/HBASE-3649?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13007563#comment-13007563
 ] 

stack commented on HBASE-3649:
------------------------------

I thought the problem was that compression slowed the flush. If problem is 
rather the count of files, yeah, compression doesn't factor.

bq. I think the better solution would be "merging flushes"?

Its about time we did this (Its only 5 years since it was described in BT 
paper).  I made HBASE-3656.

> Separate compression setting for flush files
> --------------------------------------------
>
>                 Key: HBASE-3649
>                 URL: https://issues.apache.org/jira/browse/HBASE-3649
>             Project: HBase
>          Issue Type: Improvement
>            Reporter: Andrew Purtell
>            Assignee: Andrew Purtell
>             Fix For: 0.90.2, 0.92.0
>
>
> In this thread on user@hbase: http://search-hadoop.com/m/WUnLM6ojHm1 J-D 
> conjectures that compressing flush files leads to a suboptimal situation 
> where "the puts are sometimes blocked on the memstores which are blocked by 
> the flusher thread which is blocked because there's too many files to compact 
> because the compactor is given too many small files to compact and has to 
> compact the same data a bunch of times."
> We have a separate compression setting already for major compaction vs store 
> files written during minor compaction, for background/archival apps. Add a 
> separate compression setting for flush files, default to none, to avoid the 
> above condition.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira

Reply via email to