Return-Path: X-Original-To: apmail-cassandra-user-archive@www.apache.org Delivered-To: apmail-cassandra-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id E29D6112A9 for ; Fri, 10 May 2013 08:34:00 +0000 (UTC) Received: (qmail 70982 invoked by uid 500); 10 May 2013 08:33:58 -0000 Delivered-To: apmail-cassandra-user-archive@cassandra.apache.org Received: (qmail 70845 invoked by uid 500); 10 May 2013 08:33:56 -0000 Mailing-List: contact user-help@cassandra.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@cassandra.apache.org Delivered-To: mailing list user@cassandra.apache.org Received: (qmail 70832 invoked by uid 99); 10 May 2013 08:33:55 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 10 May 2013 08:33:55 +0000 X-ASF-Spam-Status: No, hits=-0.7 required=5.0 tests=RCVD_IN_DNSWL_LOW X-Spam-Check-By: apache.org Received-SPF: error (athena.apache.org: local policy) Received: from [74.125.82.181] (HELO mail-we0-f181.google.com) (74.125.82.181) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 10 May 2013 08:33:50 +0000 Received: by mail-we0-f181.google.com with SMTP id q55so3620659wes.26 for ; Fri, 10 May 2013 01:33:09 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20120113; h=x-received:content-type:mime-version:subject:from:in-reply-to:date :content-transfer-encoding:message-id:references:to:x-mailer :x-gm-message-state; bh=IOJBIyLIUdJuil/9lhQtqJdvAuFIUXvUEqPVxEN1330=; b=O6F1Z656HJsWyS4W2CiUt922oWgKj64wvubmVHINkFBFTFnBfXNP+nHjEzbyqN8kvd TTwILkUZYZkhUH+gezrRMnOYYEWTSTYWkWT+dN2rK5vfMg8pehsLr4y99qFWvV+9Y8Y0 2mdSKvY82LOxlTbydynaUz9ffym47sFjp06lsDh7iIwH1irIXGyk57cNtg9fFRQ2WQEI Vu8Dl6a+gFwjQAj/4s8Clz0ZC1VCN2zhn7BOtdCviWDIDdLaZRh5GEnDxqDUMenN8Czm acqqbRhf9HI6uTplcar8UGdVjgwh9tj7+4RrSG73zJU4Zw+pfjcist8IHNbnVa5mhBUk mo7Q== X-Received: by 10.180.210.225 with SMTP id mx1mr2140194wic.15.1368174788930; Fri, 10 May 2013 01:33:08 -0700 (PDT) Received: from [10.0.10.243] ([195.184.103.10]) by mx.google.com with ESMTPSA id bs20sm1918681wib.0.2013.05.10.01.33.06 for (version=TLSv1 cipher=ECDHE-RSA-RC4-SHA bits=128/128); Fri, 10 May 2013 01:33:07 -0700 (PDT) Content-Type: text/plain; charset=iso-8859-1 Mime-Version: 1.0 (Mac OS X Mail 6.3 \(1503\)) Subject: Re: Sudden increase in diskspace usage From: Nicolai Gylling In-Reply-To: Date: Fri, 10 May 2013 10:33:11 +0200 Content-Transfer-Encoding: quoted-printable Message-Id: References: To: user@cassandra.apache.org X-Mailer: Apple Mail (2.1503) X-Gm-Message-State: ALoCoQkjaoYKyFd8uS7W5/KG+l3I48or3lzRuHu/sZhokdKo0WM7EEz9rBrURT8vmnvNvMLGYRED X-Virus-Checked: Checked by ClamAV on apache.org > On Wed, May 8, 2013 at 10:43 PM, Nicolai Gylling wrote: >> At the time of normal operation there was 800 gb free space on each = node. >> After the crash, C* started using a lot more, resulting in an >> out-of-diskspace situation on 2 nodes, eg. C* used up the 800 gb in = just 2 >> days, giving us very little time to do anything about it, since >> repairs/joins takes a considerable amount of time. >=20 > Did someone do a repair? Repair very frequently results in (usually > temporary) >2x disk consumption. >=20 Repairs is running regularly once a week, and normally doesn't take up = much space, as we're using Leveled Compaction Strategy.=20 >> What can make C* suddenly use this amount of disk-space? We did see a = lot of >> pending compactions on one node (7k). >=20 > Mostly repair. >=20 >> Any tips on recovering from an out-of-diskspace on multiple nodes, >> situation? I've tried moving some SStables away, but C* seems to use >> whatever space I free up in no time. I'm not sure if any of the nodes = is >> fully updated as 'nodetool status' reports 3 different loads >=20 > A relevant note here is that moving sstables out of the full partition > while cassandra is running will not result in any space recovery, > because Cassandra still has an open filehandle to that sstable. In > order to deal with out of disk space condition you need to stop > Cassandra. Unfortunately the JVM stops responding to clean shutdown > request when the disk is full, you will have to kill -KILL the > process. >=20 > If you have a lot of overwrites/fragmentation, you could attempt to > clear enough space to do a major compaction of remaining data, do that > major compaction, split your One Huge sstable with the (experimental) > sstable_split tool and then copy temporarily moved sstables back onto > the node. You could also attempt to use user defined compaction (via > JMX endpoint) to strategically compact such data. If you grep for > compaction in your logs, do you see compactions resulting in smaller > output file sizes? (compacted to X% of original messages) >=20 > I agree with Alexis Rodriguez that Cassandra 1.2.0 is not a version > anyone should run, it contains significant bugs. >=20 > =3DRob We're storing timeseries, so we don't have any overwrites and hardly any = reduction in sizes during compaction. I'll try to upgrade and see if = that can help get some diskspace back. Let's say we're seing some bug in C*, and SSTables doesn't get deleted = during compaction (which I guess is the only reason for this consumption = of diskspace). Will C* 1.2.4 be able to fix this? Or would it be a = better solution to replace one node at a time, so we're sure to only = have the data, that C* knows about?