Return-Path: X-Original-To: apmail-cassandra-user-archive@www.apache.org Delivered-To: apmail-cassandra-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 127EA7E0C for ; Fri, 19 Aug 2011 18:28:17 +0000 (UTC) Received: (qmail 68775 invoked by uid 500); 19 Aug 2011 18:28:14 -0000 Delivered-To: apmail-cassandra-user-archive@cassandra.apache.org Received: (qmail 68707 invoked by uid 500); 19 Aug 2011 18:28:14 -0000 Mailing-List: contact user-help@cassandra.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@cassandra.apache.org Delivered-To: mailing list user@cassandra.apache.org Received: (qmail 68699 invoked by uid 99); 19 Aug 2011 18:28:14 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 19 Aug 2011 18:28:14 +0000 X-ASF-Spam-Status: No, hits=-0.0 required=5.0 tests=RCVD_IN_DNSWL_LOW,SPF_NEUTRAL X-Spam-Check-By: apache.org Received-SPF: neutral (athena.apache.org: local policy) Received: from [209.85.212.44] (HELO mail-vw0-f44.google.com) (209.85.212.44) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 19 Aug 2011 18:28:06 +0000 Received: by vws12 with SMTP id 12so3272762vws.31 for ; Fri, 19 Aug 2011 11:26:23 -0700 (PDT) MIME-Version: 1.0 Received: by 10.52.20.228 with SMTP id q4mr37410vde.498.1313778383456; Fri, 19 Aug 2011 11:26:23 -0700 (PDT) Sender: scode@scode.org Received: by 10.52.164.131 with HTTP; Fri, 19 Aug 2011 11:26:23 -0700 (PDT) X-Originating-IP: [213.114.154.128] In-Reply-To: References: Date: Fri, 19 Aug 2011 20:26:23 +0200 X-Google-Sender-Auth: mH1P913RTEZq4irGtUK57-S9jLI Message-ID: Subject: Re: nodetool repair caused high disk space usage From: Peter Schuller To: user@cassandra.apache.org Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: quoted-printable > After upgrading to cass 0.8.4 from cass 0.6.11.=C2=A0 I ran scrub.=C2=A0 = That worked > fine.=C2=A0 Then I ran nodetool repair on one of the nodes.=C2=A0 The dis= k usage on > data directory increased from 40GB to 480GB, and it's still growing. If you check your data directory, does it contain a lot of "*Compacted" files? It sounds like you're churning sstables from a combination of compactions/flushes (including triggered by repair) and the old ones aren't being deleted. I wonder if there is still some issue causing sstable retention Since you're on 0.8.4, I'm a bit suspicious. I'd have to re-check each JIRA but I think the major known repair problems should be fixed except for CASSANDRA-2280 which is not your problem since you're going form a total load of 40 gig to hundreds of gigs (so even with all cf:s streaming, that's unexpected). Do you have any old left-over streams active on the nodes? "nodetool netstats". If there are "stuck" streams, they might be causing sstable retention beyond what you'd expect. --=20 / Peter Schuller (@scode on twitter)