incubator-cassandra-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Ryan King <r...@twitter.com>
Subject Re: question on capacity planning
Date Wed, 29 Jun 2011 16:17:23 GMT
On Wed, Jun 29, 2011 at 5:36 AM, Jacob, Arun <Arun.Jacob@disney.com> wrote:
> if I'm planning to store 20TB of new data per week, and expire all data
> every 2 weeks, with a replication factor of 3, do I only need approximately
> 120 TB of disk? I'm going to use ttl in my column values to automatically
> expire data. Or would I need more capacity to handle sstable merges? Given
> this amount of data, would you recommend node storage at 2TB per node or
> more? This application will have a heavy write /moderate read use profile.

You'll need extra space for both compaction and the overhead in the
storage format.

As to the amount of storage per node, that depends on your latency and
throughput requirements.

-ryan

Mime
View raw message