Our compactions/repairs have already become nightmares and we have not 
approached the levels of data you describe here (~200 GB). Have any 
pointers/case studies for optimizing this?


On Nov 5, 2012, at 12:00 PM, Michael Kjellman wrote:

> Rule of thumb is to try to keep nodes under 400GB.
> Compactions/Repairs/Move operations etc become a nightmare otherwise. How
> much data do you expect to have on each node? Also depends on caches,
> bloom filters etc
> 
> On 11/5/12 8:57 AM, "Oleg Dulin" <oleg.du...@gmail.com> wrote:
> 
>> I have 4 nodes at my disposal.
>> 
>> I can configure them like this:
>> 
>> 1) RF=1, each node has 25% of the data. On random-reads, how big is the
>> performance penalty if a node needs to look for data on another replica
>> ?
>> 
>> 2) RF=2, each node has 50% of the data. Same question ?
>> 
>> 
>> 
>> -- 
>> Regards,
>> Oleg Dulin
>> NYC Java Big Data Engineer
>> http://www.olegdulin.com/
>> 
>> 
> 
> 
> 'Like' us on Facebook for exclusive content and other resources on all 
> Barracuda Networks solutions.
> Visit http://barracudanetworks.com/facebook
> 
> 

Reply via email to