[ 
https://issues.apache.org/jira/browse/HBASE-3421?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13040082#comment-13040082
 ] 

Todd Lipcon commented on HBASE-3421:
------------------------------------

I think I'm seeing this on one workload as well - I added a "stats" command for 
the HFile tool, then ran it on a problematic 1.5GB gzip-compressed HFile that 
won't compact due to OOME.

Key length: count: 272358037    min: 59 max: 2259       mean: 59.620136732003246
Val length: count: 272358037    min: 0  max: 0  mean: 0.0
Row size (bytes): count: 1287745        min: 67 max: 15615127736        mean: 
14301.657317248368
Row size (columns): count: 1287745      min: 1  max: 233061608  mean: 
211.49997631518661

so in this case there's actually a row with 233 million columns which takes up 
15GB uncompressed!

> Very wide rows -- 30M plus -- cause us OOME
> -------------------------------------------
>
>                 Key: HBASE-3421
>                 URL: https://issues.apache.org/jira/browse/HBASE-3421
>             Project: HBase
>          Issue Type: Bug
>    Affects Versions: 0.90.0
>            Reporter: stack
>
> From the list, see 'jvm oom' in 
> http://mail-archives.apache.org/mod_mbox/hbase-user/201101.mbox/browser, it 
> looks like wide rows -- 30M or so -- causes OOME during compaction.  We 
> should check it out. Can the scanner used during compactions use the 'limit' 
> when nexting?  If so, this should save our OOME'ing (or, we need to add to 
> the next a max size rather than count of KVs).

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira

Reply via email to