Return-Path: X-Original-To: apmail-cassandra-user-archive@www.apache.org Delivered-To: apmail-cassandra-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 107D571A4 for ; Fri, 19 Aug 2011 18:53:51 +0000 (UTC) Received: (qmail 46910 invoked by uid 500); 19 Aug 2011 18:53:49 -0000 Delivered-To: apmail-cassandra-user-archive@cassandra.apache.org Received: (qmail 46848 invoked by uid 500); 19 Aug 2011 18:53:48 -0000 Mailing-List: contact user-help@cassandra.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@cassandra.apache.org Delivered-To: mailing list user@cassandra.apache.org Received: (qmail 46837 invoked by uid 99); 19 Aug 2011 18:53:48 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 19 Aug 2011 18:53:48 +0000 X-ASF-Spam-Status: No, hits=-0.0 required=5.0 tests=RCVD_IN_DNSWL_LOW,SPF_NEUTRAL X-Spam-Check-By: apache.org Received-SPF: neutral (nike.apache.org: local policy) Received: from [209.85.220.172] (HELO mail-vx0-f172.google.com) (209.85.220.172) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 19 Aug 2011 18:53:39 +0000 Received: by vxi29 with SMTP id 29so3539162vxi.31 for ; Fri, 19 Aug 2011 11:53:18 -0700 (PDT) MIME-Version: 1.0 Received: by 10.52.184.10 with SMTP id eq10mr113310vdc.96.1313779943962; Fri, 19 Aug 2011 11:52:23 -0700 (PDT) Sender: scode@scode.org Received: by 10.52.164.131 with HTTP; Fri, 19 Aug 2011 11:52:23 -0700 (PDT) X-Originating-IP: [213.114.154.128] In-Reply-To: References: Date: Fri, 19 Aug 2011 20:52:23 +0200 X-Google-Sender-Auth: VJgzLr1u-xxnF1f7JFStNGgltkY Message-ID: Subject: Re: nodetool repair caused high disk space usage From: Peter Schuller To: user@cassandra.apache.org Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: quoted-printable X-Virus-Checked: Checked by ClamAV on apache.org > There were few Compacted files.=C2=A0 I thought that might have been the = cause, > but it wasn't it.=C2=A0 We have a CF that is 23GB, and while repair is ru= nning, > there are multiple instances of that CF created along with other CFs. To confirm - are you saying the data directory size is huge, but the live size as reported by nodetool ring and nodetool info does NOT reflect this inflated size? What files *do* you have in the data directory? Any left-over *tmp* files for example? Are you sure you're only running a single repair at a time? (Sorry if this was covered, I did a quick swipe through thread history because I was unsure whether I was confusing two different threads, and I don't think so.) The question is what's taking the space. If it's sstables, they really should be either compacted onces that are marked for deletion but being retained, or "live" sstables in which case they should show up as load in nodetool. What else... maybe streams are being re-tried from the source nodes and the disk space is coming from a bunch of half-finished streams of the same data. But if so, those should be *tmp* files IIRC. I'm just wildly speculation, but it would be nice to get to the bottom of t= his. --=20 / Peter Schuller (@scode on twitter)