you will need to time how long it takes to read all that state back in
and adjust the initTime accordingly. it will probably take a while to
pull all that data into memory.
ben
On 10/05/2010 11:36 AM, Avinash Lakshman wrote:
I have run it over 5 GB of heap with over 10M znodes. We will definitely run
it with over 64 GB of heap. Technically I do not see any limitiation.
However I will the experts chime in.
Avinash
On Tue, Oct 5, 2010 at 11:14 AM, Mahadev Konar<maha...@yahoo-inc.com>wrote:
Hi Maarteen,
I definitely know of a group which uses around 3GB of memory heap for
zookeeper but never heard of someone with such huge requirements. I would
say it definitely would be a learning experience with such high memory
which
I definitely think would be very very useful for others in the community as
well.
Thanks
mahadev
On 10/5/10 11:03 AM, "Maarten Koopmans"<maar...@vrijheid.net> wrote:
Hi,
I just wondered: has anybody ever ran zookeeper "to the max" on a 68GB
quadruple extra large high memory EC2 instance? With, say, 60GB allocated
or
so?
Because EC2 with EBS is a nice way to grow your zookeeper cluster (data
on the
ebs columes, upgrade as your memory utilization grows....) - I just
wonder
what the limits are there, or if I am foing where angels fear to tread...
--Maarten