[ https://issues.apache.org/jira/browse/HBASE-69?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12568200#action_12568200 ]
Billy Pearson commented on HBASE-69: ------------------------------------ sorry my last post I am talking about while the cluster is busy. The hlogs have > 4 hours between them that would mean that I am rolling the hlog many times after several optional memcash flushes while jobs are running. Might check to make sure the option flush is updating the sequence id even if there is 0 entries and the flush is skipped. The problem could be from the code that removed the code before I used to see 0 logs removed while having debug turned on and hlog rolling but I do not see these messages after this patches. > [hbase] Make cache flush triggering less simplistic > --------------------------------------------------- > > Key: HBASE-69 > URL: https://issues.apache.org/jira/browse/HBASE-69 > Project: Hadoop HBase > Issue Type: Improvement > Components: regionserver > Reporter: stack > Assignee: Jim Kellerman > Fix For: 0.2.0 > > Attachments: patch.txt, patch.txt, patch.txt, patch.txt, patch.txt, > patch.txt, patch.txt, patch.txt, patch.txt, patch.txt, patch.txt, patch.txt > > > When flusher runs -- its triggered when the sum of all Stores in a Region > a > configurable max size -- we flush all Stores though a Store memcache might > have but a few bytes. > I would think Stores should only dump their memcache disk if they have some > substance. > The problem becomes more acute, the more families you have in a Region. > Possible behaviors would be to dump the biggest Store only, or only those > Stores > 50% of max memcache size. Behavior would vary dependent on the > prompt that provoked the flush. Would also log why the flush is running: > optional or > max size. > This issue comes out of HADOOP-2621. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.