[ https://issues.apache.org/jira/browse/HBASE-69?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12568027#action_12568027 ]
Billy Pearson commented on HBASE-69: ------------------------------------ I thank we are still missing something when I talk about hlog build up I mean there is never a removal of the hlogs unless there is a split or a shutdown. I have over 106 hlog on one server hbase.regionserver.maxlogentries = 250000 The modification time of the oldest to the newest logs files are much greater then the 30 min option flush time set in hbase-default.xml > [hbase] Make cache flush triggering less simplistic > --------------------------------------------------- > > Key: HBASE-69 > URL: https://issues.apache.org/jira/browse/HBASE-69 > Project: Hadoop HBase > Issue Type: Improvement > Components: regionserver > Reporter: stack > Assignee: Jim Kellerman > Fix For: 0.2.0 > > Attachments: patch.txt, patch.txt, patch.txt, patch.txt, patch.txt, > patch.txt, patch.txt, patch.txt, patch.txt, patch.txt, patch.txt, patch.txt > > > When flusher runs -- its triggered when the sum of all Stores in a Region > a > configurable max size -- we flush all Stores though a Store memcache might > have but a few bytes. > I would think Stores should only dump their memcache disk if they have some > substance. > The problem becomes more acute, the more families you have in a Region. > Possible behaviors would be to dump the biggest Store only, or only those > Stores > 50% of max memcache size. Behavior would vary dependent on the > prompt that provoked the flush. Would also log why the flush is running: > optional or > max size. > This issue comes out of HADOOP-2621. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.