incubator-cassandra-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From aaron morton <aa...@thelastpickle.com>
Subject Re: maximum storage per node
Date Mon, 29 Jul 2013 08:30:03 GMT
> Does anyone have opinions on the maximum amount of data reasonable to store on one Cassandra
node?
spinning disk and 1Gbe networking, rule of thumb was 300Gb to 500GB. 

SSD or very fast local disk, 10Gbe networking, optionally JBOD, cassandra 1.2 and vnodes people
are talking about multiple TB's per node. 

>  If there are limitations, what are the reasons for it?   

The main issues were:

* As discussed potentially very long compaction
* As discussed repair taking a very long time to calculate the merkle trees. 
* Potentially taking a very long time to rebuild a new node after one totally fails. vNodes
address this by increasing the number of nodes that can stream data to one bootstrapping.
This is really something that has to fit into your operations. 

Hope that helps. 
 
-----------------
Aaron Morton
Cassandra Consultant
New Zealand

@aaronmorton
http://www.thelastpickle.com

On 27/07/2013, at 5:00 AM, Robert Coli <rcoli@eventbrite.com> wrote:

> On Fri, Jul 26, 2013 at 4:23 AM, Romain HARDOUIN <romain.hardouin@urssaf.fr> wrote:
> Do you have some fairly complex queries to run against your data? 
> Or your need is just to store large pieces of data? (In which case Object Storage like
OpenStack Swift could be more appropriate IMHO)  
> 
> Or distributed blob storage like MogileFS.
> 
> https://code.google.com/p/mogilefs/
> 
> =Rob


Mime
View raw message