Yes, and syncing after a crash will be interesting as well. Off note; I am running it with a 6GB heap now, but it's not filled yet. I do have smoke tests thoug, so maybe I'll give it a try.
Op 5 okt. 2010 om 21:13 heeft Benjamin Reed <br...@yahoo-inc.com> het volgende geschreven: > > you will need to time how long it takes to read all that state back in and > adjust the initTime accordingly. it will probably take a while to pull all > that data into memory. > > ben > > On 10/05/2010 11:36 AM, Avinash Lakshman wrote: >> I have run it over 5 GB of heap with over 10M znodes. We will definitely run >> it with over 64 GB of heap. Technically I do not see any limitiation. >> However I will the experts chime in. >> >> Avinash >> >> On Tue, Oct 5, 2010 at 11:14 AM, Mahadev Konar<maha...@yahoo-inc.com>wrote: >> >>> Hi Maarteen, >>> I definitely know of a group which uses around 3GB of memory heap for >>> zookeeper but never heard of someone with such huge requirements. I would >>> say it definitely would be a learning experience with such high memory >>> which >>> I definitely think would be very very useful for others in the community as >>> well. >>> >>> Thanks >>> mahadev >>> >>> >>> On 10/5/10 11:03 AM, "Maarten Koopmans"<maar...@vrijheid.net> wrote: >>> >>>> Hi, >>>> >>>> I just wondered: has anybody ever ran zookeeper "to the max" on a 68GB >>>> quadruple extra large high memory EC2 instance? With, say, 60GB allocated >>> or >>>> so? >>>> >>>> Because EC2 with EBS is a nice way to grow your zookeeper cluster (data >>> on the >>>> ebs columes, upgrade as your memory utilization grows....) - I just >>> wonder >>>> what the limits are there, or if I am foing where angels fear to tread... >>>> >>>> --Maarten >>>> >>> > >