[ 
https://issues.apache.org/jira/browse/HBASE-3421?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12978016#action_12978016
 ] 

Nicolas Spiegelberg commented on HBASE-3421:
--------------------------------------------

Note that you can limit the number of StoreFiles that can be compacted at one 
time...

Store.java#204:  this.maxFilesToCompact =
conf.getInt("hbase.hstore.compaction.max", 10)

30M * 10 SF == 300MB.  What is your RAM capacity?  You are likely stuck on an 
merging outlier that exists in every SF.  I would run:

bin/hbase org.apache.hadoop.hbase.io.hfile.HFile -f <FILE_NAME> -p |sed 
's/V:.*$//g'|less

on the HFiles in that Store to see what your high watermark is.

> Very wide rows -- 30M plus -- cause us OOME
> -------------------------------------------
>
>                 Key: HBASE-3421
>                 URL: https://issues.apache.org/jira/browse/HBASE-3421
>             Project: HBase
>          Issue Type: Bug
>    Affects Versions: 0.90.0
>            Reporter: stack
>
> From the list, see 'jvm oom' in 
> http://mail-archives.apache.org/mod_mbox/hbase-user/201101.mbox/browser, it 
> looks like wide rows -- 30M or so -- causes OOME during compaction.  We 
> should check it out. Can the scanner used during compactions use the 'limit' 
> when nexting?  If so, this should save our OOME'ing (or, we need to add to 
> the next a max size rather than count of KVs).

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to