cassandra-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Bryan <br...@appssavvy.com>
Subject Re: Replication factor and performance questions
Date Mon, 05 Nov 2012 17:12:03 GMT
Our compactions/repairs have already become nightmares and we have not approached the levels
of data you describe here (~200 GB). Have any pointers/case studies for optimizing this?


On Nov 5, 2012, at 12:00 PM, Michael Kjellman wrote:

> Rule of thumb is to try to keep nodes under 400GB.
> Compactions/Repairs/Move operations etc become a nightmare otherwise. How
> much data do you expect to have on each node? Also depends on caches,
> bloom filters etc
> 
> On 11/5/12 8:57 AM, "Oleg Dulin" <oleg.dulin@gmail.com> wrote:
> 
>> I have 4 nodes at my disposal.
>> 
>> I can configure them like this:
>> 
>> 1) RF=1, each node has 25% of the data. On random-reads, how big is the
>> performance penalty if a node needs to look for data on another replica
>> ?
>> 
>> 2) RF=2, each node has 50% of the data. Same question ?
>> 
>> 
>> 
>> -- 
>> Regards,
>> Oleg Dulin
>> NYC Java Big Data Engineer
>> http://www.olegdulin.com/
>> 
>> 
> 
> 
> 'Like' us on Facebook for exclusive content and other resources on all Barracuda Networks
solutions.
> Visit http://barracudanetworks.com/facebook
> 
> 


Mime
View raw message