[ https://issues.apache.org/jira/browse/HBASE-2251?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12837327#action_12837327 ]
stack commented on HBASE-2251: ------------------------------ Whats being described sounds like the yahoo tool thats supposed to be open sourced any time soon. While I think these additions to PE would be sweet, before this, before each release, we need to run perf tests so we find these slow downs before release -- even if its only PE (though as Dan Washuen pointed out -- PE currently clears memstore so its not factored in PE evals -- that needs fixing). > PE defaults to 1k rows - uncommon use case, and easy to hit benchmarks > ---------------------------------------------------------------------- > > Key: HBASE-2251 > URL: https://issues.apache.org/jira/browse/HBASE-2251 > Project: Hadoop HBase > Issue Type: Bug > Reporter: ryan rawson > Fix For: 0.20.4, 0.21.0 > > > The PerformanceEvaluation uses 1k rows, which I would argue is uncommon, and > also provides an easy to hit performance goal. Most of the harder > performance issues happens at the low and high side of cell size. In our own > application, our key sizes range from 4 bytes to maybe 100 bytes. Very > rarely 1000 bytes. If we have large values, they are VERY large, like > multiple k sizes. > Recently a change went into HBase that ran well with PE because the overhead > of 1k rows is very low in memory, but under small rows, the expected > performance would be hit much more. This is because the per-value overhead > (eg: node objects of the skip list/memstore) is amortized more with 1k > values. > We should make this a tunable setting, and have a low default. I would argue > for a 10-30 byte default. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.