Return-Path: X-Original-To: apmail-cassandra-user-archive@www.apache.org Delivered-To: apmail-cassandra-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id EA4AB6B2E for ; Wed, 22 Jun 2011 17:03:50 +0000 (UTC) Received: (qmail 73559 invoked by uid 500); 22 Jun 2011 17:03:48 -0000 Delivered-To: apmail-cassandra-user-archive@cassandra.apache.org Received: (qmail 73530 invoked by uid 500); 22 Jun 2011 17:03:48 -0000 Mailing-List: contact user-help@cassandra.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@cassandra.apache.org Delivered-To: mailing list user@cassandra.apache.org Received: (qmail 73522 invoked by uid 99); 22 Jun 2011 17:03:48 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 22 Jun 2011 17:03:48 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=FREEMAIL_FROM,HTML_MESSAGE,RCVD_IN_DNSWL_LOW,RFC_ABUSE_POST,SPF_PASS,T_TO_NO_BRKTS_FREEMAIL X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of jonathan.colby@gmail.com designates 74.125.82.44 as permitted sender) Received: from [74.125.82.44] (HELO mail-ww0-f44.google.com) (74.125.82.44) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 22 Jun 2011 17:03:39 +0000 Received: by wwe5 with SMTP id 5so919632wwe.25 for ; Wed, 22 Jun 2011 10:03:19 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:from:mime-version:content-type:subject:date :in-reply-to:to:references:message-id:x-mailer; bh=Lk0jIMl/W8EYF2bbtXZbVXNg+tJXXtUdNcXyCxX6TJs=; b=touDmnlkTE5UV01MCmp+cOZgz4iqg7970B8pYYfdmq7sqAOqqC5Yvtuu+GrZGwRqva Ea9e0taUZxqzw54yE8a8qArxFtjsIM8exNlo43+/66gHnLKCfdC8fdDMGI7C7msth2PI KX6DpK1YX3kEhZnnsJfCONr4Fc5gBMgSnBtQg= DomainKey-Signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=from:mime-version:content-type:subject:date:in-reply-to:to :references:message-id:x-mailer; b=foV+YT2qxhSjamQt0KKiM+oM+C2wqVEA+eggusnCJYkyZIbdERreMxFFdMYFE7PASu tY4gdEj2DvhgBUEwwAVY2xzuwSPRMGgg79AbJNdiqYkxIXdafkdtF9R/KxtpDBHD7x46 1jHtz0N93qjk9X1MVBy9CKVwSrSWIKV/3v+yI= Received: by 10.216.79.7 with SMTP id h7mr911090wee.105.1308762199040; Wed, 22 Jun 2011 10:03:19 -0700 (PDT) Received: from [10.250.128.161] ([195.234.136.12]) by mx.google.com with ESMTPS id g2sm422161weg.14.2011.06.22.10.03.17 (version=TLSv1/SSLv3 cipher=OTHER); Wed, 22 Jun 2011 10:03:18 -0700 (PDT) From: Jonathan Colby Mime-Version: 1.0 (Apple Message framework v1084) Content-Type: multipart/alternative; boundary=Apple-Mail-3--257485195 Subject: Re: simple question about merged SSTable sizes Date: Wed, 22 Jun 2011 19:03:16 +0200 In-Reply-To: To: user@cassandra.apache.org References: Message-Id: <16FEB70B-119E-4377-B993-8D8F39712444@gmail.com> X-Mailer: Apple Mail (2.1084) X-Virus-Checked: Checked by ClamAV on apache.org --Apple-Mail-3--257485195 Content-Transfer-Encoding: quoted-printable Content-Type: text/plain; charset=us-ascii So the take-away is try to avoid major compactions at all costs! = Thanks Ed and Eric. On Jun 22, 2011, at 7:00 PM, Edward Capriolo wrote: > Yes, if you are not deleting fast enough they will grow. This is not = specifically a cassandra problem /var/log/messages has the same issue.=20= >=20 > There is a JIRA ticket about having a maximum size for SSTables, so = they always stay manageable >=20 > You fall into a small trap when you force major compaction in that = many small tables turn into one big one, from their it is hard to get = back to many smaller ones again, the other side of the coin if you do = not major compact you can end up with much more disk usage then live = data (IE large % of disk is overwrites and tombstones). >=20 > You can tune the compaction rate now so compaction does not kill your = IO. Generally I think avoiding really large SSTables is the best way to = do. Scale out and avoid very large SSTables/node if possible. >=20 > Edward >=20 >=20 > On Wed, Jun 22, 2011 at 12:35 PM, Jonathan Colby = wrote: >=20 > The way compaction works, "x" same-sized files are merged into a new = SSTable. This repeats itself and the SSTable get bigger and bigger. >=20 > So what is the upper limit?? If you are not deleting stuff fast = enough, wouldn't the SSTable sizes grow indefinitely? >=20 > I ask because we have some rather large SSTable files (80-100 GB) and = I'm starting to worry about future compactions. >=20 > Second, compacting such large files is an IO killer. What can be = tuned other than compaction_threshold to help optimize this and prevent = the files from getting too big? >=20 > Thanks! >=20 --Apple-Mail-3--257485195 Content-Transfer-Encoding: 7bit Content-Type: text/html; charset=us-ascii So the take-away is try to avoid major compactions at all costs!   Thanks Ed and Eric.

On Jun 22, 2011, at 7:00 PM, Edward Capriolo wrote:

Yes, if you are not deleting fast enough they will grow. This is not specifically a cassandra problem /var/log/messages has the same issue.

There is a JIRA ticket about having a maximum size for SSTables, so they always stay manageable

You fall into a small trap when you force major compaction in that many small tables turn into one big one, from their it is hard to get back to many smaller ones again, the other side of the coin if you do not major compact you can end up with much more disk usage then live data (IE large % of disk is overwrites and tombstones).

You can tune the compaction rate now so compaction does not kill your IO. Generally I think avoiding really large SSTables is the best way to do. Scale out and avoid very large SSTables/node if possible.

Edward


On Wed, Jun 22, 2011 at 12:35 PM, Jonathan Colby <jonathan.colby@gmail.com> wrote:

The way compaction works,  "x" same-sized files are merged into a new SSTable.  This repeats itself and the SSTable get bigger and bigger.

So what is the upper limit??     If you are not deleting stuff fast enough, wouldn't the SSTable sizes grow indefinitely?

I ask because we have some rather large SSTable files (80-100 GB) and I'm starting to worry about future compactions.

Second, compacting such large files is an IO killer.    What can be tuned other than compaction_threshold to help optimize this and prevent the files from getting too big?

Thanks!


--Apple-Mail-3--257485195--