Return-Path: X-Original-To: apmail-cassandra-user-archive@www.apache.org Delivered-To: apmail-cassandra-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id C83768069 for ; Wed, 7 Sep 2011 17:10:00 +0000 (UTC) Received: (qmail 27292 invoked by uid 500); 7 Sep 2011 17:09:58 -0000 Delivered-To: apmail-cassandra-user-archive@cassandra.apache.org Received: (qmail 27105 invoked by uid 500); 7 Sep 2011 17:09:58 -0000 Mailing-List: contact user-help@cassandra.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@cassandra.apache.org Delivered-To: mailing list user@cassandra.apache.org Received: (qmail 27097 invoked by uid 99); 7 Sep 2011 17:09:57 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 07 Sep 2011 17:09:57 +0000 X-ASF-Spam-Status: No, hits=2.5 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_SOFTFAIL X-Spam-Check-By: apache.org Received-SPF: softfail (nike.apache.org: transitioning domain of hfyuan@rhapsody.com does not designate 209.85.216.179 as permitted sender) Received: from [209.85.216.179] (HELO mail-qy0-f179.google.com) (209.85.216.179) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 07 Sep 2011 17:09:50 +0000 Received: by qyk35 with SMTP id 35so4762651qyk.10 for ; Wed, 07 Sep 2011 10:09:28 -0700 (PDT) Received: by 10.68.6.33 with SMTP id x1mr11931997pbx.268.1315415368337; Wed, 07 Sep 2011 10:09:28 -0700 (PDT) Received: from cafebabe.sea1.rhapsody.com (rev-66-150-171-4.rhapsody.com. [66.150.171.4]) by mx.google.com with ESMTPS id f6sm6987102pbp.2.2011.09.07.10.09.27 (version=TLSv1/SSLv3 cipher=OTHER); Wed, 07 Sep 2011 10:09:27 -0700 (PDT) From: Hefeng Yuan Mime-Version: 1.0 (Apple Message framework v1244.3) Content-Type: multipart/alternative; boundary="Apple-Mail=_80BDB883-AAE0-4C67-9D36-464A4B29AC59" Subject: Re: Calculate number of nodes required based on data Date: Wed, 7 Sep 2011 10:09:42 -0700 In-Reply-To: To: user@cassandra.apache.org References: <18A410A9-951F-4002-B188-DFA5C421D919@rhapsody.com> Message-Id: <51889EA5-9480-4565-88EE-2F9BD7B1ED36@rhapsody.com> X-Mailer: Apple Mail (2.1244.3) X-Virus-Checked: Checked by ClamAV on apache.org --Apple-Mail=_80BDB883-AAE0-4C67-9D36-464A4B29AC59 Content-Transfer-Encoding: quoted-printable Content-Type: text/plain; charset=iso-8859-1 Adi, The reason we're attempting to add more nodes is trying to solve the = long/simultaneous compactions, i.e. the performance issue, not the = storage issue yet. We have RF 5 and CL QUORUM for read and write, we have currently 6 = nodes, and when 4 nodes doing compaction at the same period, we're = screwed, especially on read, since it'll cover one of the compaction = node anyways.=20 My assumption is that if we add more nodes, each node will have less = load, and therefore need less compaction, and probably will compact = faster, eternally avoid 4+ nodes doing compaction simultaneously. Any suggestion on how to calculate how many more nodes to add? Or, = generally how to plan for number of nodes required, from a performance = perspective? Thanks, Hefeng On Sep 7, 2011, at 9:56 AM, Adi wrote: > On Tue, Sep 6, 2011 at 3:53 PM, Hefeng Yuan = wrote: > Hi, >=20 > Is there any suggested way of calculating number of nodes needed based = on data? > =20 > We currently have 6 nodes (each has 8G memory) with RF5 (because we = want to be able to survive loss of 2 nodes). > The flush of memtable happens around every 30 min (while not doing = compaction), with ~9m serialized bytes. >=20 > The problem is that we see more than 3 nodes doing compaction at the = same time, which slows down the application. > (tried to increase/decrease compaction_throughput_mb_per_sec, not = helping much) >=20 > So I'm thinking probably we should add more nodes, but not sure how = many more to add. > Based on the data rate, is there any suggested way of calculating = number of nodes required? >=20 > Thanks, > Hefeng >=20 >=20 > What is the total amount of data? > What is the total amount in the biggest column family? >=20 > There is no hard limit per node. Cassandra gurus like more nodes :-). = One number for 'happy cassandra users' I have seen mentioned in = discussions is around 250-300 GB per node. But you could store more per = node by having multiple column families each storing around 250-300 GB = per column family. The main problem being repair/compactions and such = operations taking longer and requiring much more spare disk space. >=20 > As for slow down in application during compaction I was wondering=20 > what is the CL you are using for read and writes? > Make sure it is not a client issue - Is your client hitting all nodes = in round-robin or some other fashion? >=20 > -Adi --Apple-Mail=_80BDB883-AAE0-4C67-9D36-464A4B29AC59 Content-Transfer-Encoding: quoted-printable Content-Type: text/html; charset=iso-8859-1
On Tue, Sep 6, 2011 at 3:53 PM, Hefeng Yuan <hfyuan@rhapsody.com> = wrote:
Hi,

Is there any suggested way of calculating number of nodes needed based = on data?
 
We currently have 6 nodes (each has 8G memory) with RF5 (because we want = to be able to survive loss of 2 nodes).
The flush of memtable happens around every 30 min (while not doing = compaction), with ~9m serialized bytes.

The problem is that we see more than 3 nodes doing compaction at the = same time, which slows down the application.
(tried to increase/decrease compaction_throughput_mb_per_sec, not = helping much)

So I'm thinking probably we should add more nodes, but not sure how many = more to add.
Based on the data rate, is there any suggested way of calculating number = of nodes required?

Thanks,
Hefeng


=
What is the total  amount of data?
What is the total = amount in the biggest column family?

There is = no hard limit per node. Cassandra gurus like more nodes :-). One number = for 'happy cassandra users'  I have seen mentioned in discussions = is around 250-300 GB per node. But you could store more per node by = having multiple column families each storing around 250-300 GB per = column family. The main problem being repair/compactions and such = operations taking longer and requiring much more spare disk space.

As for slow down in application during = compaction I was wondering 
what is the CL you are using = for read and writes?
Make sure it is not a client issue - Is = your client hitting all nodes in round-robin or some other = fashion?

-Adi

= --Apple-Mail=_80BDB883-AAE0-4C67-9D36-464A4B29AC59--