It is not jvm version dependent. 

Stack



On Feb 19, 2011, at 6:43, Wayne <[email protected]> wrote:

> What JVM is recommended for the new memstore allocator? We swtiched from u23
> back to u17 which helped a lot. Is this optimized for a specific JVM or does
> it not matter?
> 
> On Fri, Feb 18, 2011 at 5:46 PM, Todd Lipcon <[email protected]> wrote:
> 
>> On Fri, Feb 18, 2011 at 12:10 PM, Jean-Daniel Cryans <[email protected]
>>> wrote:
>> 
>>> The bigger the heap the longer the GC pause of the world when
>> 
>> fragmentation requires it, 8GB is "safer".
>>> 
>> 
>> On my boxes, a stop-the-world on 8G heap is already around 80 seconds...
>> pretty catastrophic. Of course we've bumped the ZK timeout up to several
>> minutes these days, but it's just a bandaid.
>> 
>> 
>>> 
>>> In 0.90.1 you can try enabling the new memstore allocator that seems
>>> to do a really good job, checkout the jira first:
>>> https://issues.apache.org/jira/browse/HBASE-3455
>>> 
>>> 
>> Yep. Hopefully will have time to do a blog post this weekend about it as
>> well. In my testing, try as I might, I can't get my region servers to do a
>> full GC anymore when this is enabled.
>> 
>> -Todd
>> 
>> 
>>> On Fri, Feb 18, 2011 at 12:05 PM, Chris Tarnas <[email protected]> wrote:
>>>> Thank you , ad that bring me to my next question...
>>>> 
>>>> What is the current recommendation on the max heap size for Hbase if
>> RAM
>>> on the server is not an issue? Right now I am at 8GB and have no issues,
>> can
>>> I safely do 12GB? The servers have plenty of RAM (48GB) so that should
>> not
>>> be an issue - I just want to minimize the risk that GC will cause
>> problems.
>>>> 
>>>> thanks again.
>>>> -chris
>>>> 
>>>> On Feb 18, 2011, at 11:59 AM, Jean-Daniel Cryans wrote:
>>>> 
>>>>> That's what I usually recommend, the bigger the flushed files the
>>>>> better. On the other hand, you only have so much memory to dedicate to
>>>>> the MemStore...
>>>>> 
>>>>> J-D
>>>>> 
>>>>> On Fri, Feb 18, 2011 at 11:50 AM, Chris Tarnas <[email protected]> wrote:
>>>>>> Would it be a good idea to raise the
>> hbase.hregion.memstore.flush.size
>>> if you have really large regions?
>>>>>> 
>>>>>> -chris
>>>>>> 
>>>>>> On Feb 18, 2011, at 11:43 AM, Jean-Daniel Cryans wrote:
>>>>>> 
>>>>>>> Less regions, but it's often a good thing if you have a lot of data
>> :)
>>>>>>> 
>>>>>>> It's probably a good thing to bump the HDFS block size to 128 or
>> 256MB
>>>>>>> since you know you're going to have huge-ish files.
>>>>>>> 
>>>>>>> But anyway regarding penalties, I can't think of one that clearly
>>>>>>> comes out (unless you use a very small heap). The IO usage patterns
>>>>>>> will change, but unless you flush very small files all the time and
>>>>>>> need to recompact them into much bigger ones, then it shouldn't
>> really
>>>>>>> be an issue.
>>>>>>> 
>>>>>>> J-D
>>>>>>> 
>>>>>>> On Fri, Feb 18, 2011 at 11:36 AM, Jason Rutherglen
>>>>>>> <[email protected]> wrote:
>>>>>>>>> We are also using a 5Gb region size to keep our region
>>>>>>>>> counts in the 100-200 range/node per Jonathan Grey's
>> recommendation.
>>>>>>>> 
>>>>>>>> So there isn't a penalty incurred from increasing the max region
>> size
>>>>>>>> from 256MB to 5GB?
>>>>>>>> 
>>>>>> 
>>>>>> 
>>>> 
>>>> 
>>> 
>> 
>> 
>> 
>> --
>> Todd Lipcon
>> Software Engineer, Cloudera
>> 

Reply via email to