Return-Path: Delivered-To: apmail-incubator-cassandra-user-archive@minotaur.apache.org Received: (qmail 40131 invoked from network); 4 Dec 2009 20:45:41 -0000 Received: from hermes.apache.org (HELO mail.apache.org) (140.211.11.3) by minotaur.apache.org with SMTP; 4 Dec 2009 20:45:41 -0000 Received: (qmail 19334 invoked by uid 500); 4 Dec 2009 20:45:41 -0000 Delivered-To: apmail-incubator-cassandra-user-archive@incubator.apache.org Received: (qmail 19311 invoked by uid 500); 4 Dec 2009 20:45:41 -0000 Mailing-List: contact cassandra-user-help@incubator.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: cassandra-user@incubator.apache.org Delivered-To: mailing list cassandra-user@incubator.apache.org Received: (qmail 19301 invoked by uid 99); 4 Dec 2009 20:45:41 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 04 Dec 2009 20:45:41 +0000 X-ASF-Spam-Status: No, hits=-0.0 required=10.0 tests=SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of rrabah@playdom.com designates 74.125.149.197 as permitted sender) Received: from [74.125.149.197] (HELO na3sys009aog107.obsmtp.com) (74.125.149.197) by apache.org (qpsmtpd/0.29) with SMTP; Fri, 04 Dec 2009 20:45:31 +0000 Received: from source ([209.85.160.46]) by na3sys009aob107.postini.com ([74.125.148.12]) with SMTP ID DSNKSxl01hrJMLF65mNroo7sVkmkWhUXHe81@postini.com; Fri, 04 Dec 2009 12:45:11 PST Received: by mail-pw0-f46.google.com with SMTP id 17so2525462pwj.25 for ; Fri, 04 Dec 2009 12:45:10 -0800 (PST) MIME-Version: 1.0 Received: by 10.141.41.18 with SMTP id t18mr232620rvj.264.1259959509927; Fri, 04 Dec 2009 12:45:09 -0800 (PST) In-Reply-To: References: Date: Fri, 4 Dec 2009 12:45:09 -0800 Message-ID: Subject: Re: Removes increasing disk space usage in Cassandra? From: Ramzi Rabah To: cassandra-user@incubator.apache.org Content-Type: text/plain; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable X-Virus-Checked: Checked by ClamAV on apache.org I have a two week old version of trunk. Probably need to update it to latest build. On Fri, Dec 4, 2009 at 12:34 PM, Jonathan Ellis wrote: > Are you testing trunk? =A0If not, you should check that first to see if > it's already fixed. > > On Fri, Dec 4, 2009 at 1:55 PM, Ramzi Rabah wrote: >> Just to be clear what I meant is that I ran the deletions and >> compaction with GCGraceSeconds set to 1 hour, so there was enough time >> for the tombstones to expire. >> Anyway I will try to make a simpler test case to hopefully reproduce >> this, and I will share the code if I can reproduce. >> >> Ray >> >> On Fri, Dec 4, 2009 at 11:04 AM, Ramzi Rabah wrote: >>> Hi Jonathan I have changed that to 3600(one hour) based on your >>> recommendation before. >>> >>> On Fri, Dec 4, 2009 at 11:01 AM, Jonathan Ellis wro= te: >>>> this is what I was referring to by "the period specified in your confi= g file": >>>> >>>> =A0 >>>> =A0864000 >>>> >>>> On Fri, Dec 4, 2009 at 12:51 PM, Ramzi Rabah wrot= e: >>>>> I think there might be a bug in the deletion logic. I removed all the >>>>> data on the cluster by running remove on every single key I entered, >>>>> and I run major compaction >>>>> nodeprobe -host hostname compact on a certain node, and after the >>>>> compaction is over, I am left with one data file/ one index file and >>>>> the bloom filter file, >>>>> and they are the same size of data as before I started doing the dele= tes. >>>>> >>>>> On Thu, Dec 3, 2009 at 6:09 PM, Jonathan Ellis wr= ote: >>>>>> cassandra never modifies data in-place. =A0so it writes tombstones t= o >>>>>> supress the older writes, and when compaction occurs the data and >>>>>> tombstones get GC'd (after the period specified in your config file)= . >>>>>> >>>>>> On Thu, Dec 3, 2009 at 8:07 PM, Ramzi Rabah wro= te: >>>>>>> Looking at jconsole I see a high number of writes when I do removes= , >>>>>>> so I am guessing these are tombstones being written? If that's the >>>>>>> case, is the data being removed and replaced by tombstones? and wil= l >>>>>>> they all be deleted eventually when compaction runs? >>>>>>> >>>>>>> >>>>>>> >>>>>>> On Thu, Dec 3, 2009 at 3:18 PM, Ramzi Rabah wr= ote: >>>>>>>> Hi all, >>>>>>>> >>>>>>>> I ran a test where I inserted about 1.2 Gigabytes worth of data in= to >>>>>>>> each node of a 4 node cluster. >>>>>>>> I ran a script that first calls a get on each column inserted foll= owed >>>>>>>> by a remove. Since I was basically removing every entry >>>>>>>> I inserted before, I expected that the disk space occupied by the >>>>>>>> nodes will go down and eventually become 0. The disk space >>>>>>>> actually goes up when I do the bulk removes to about 1.8 gigs per >>>>>>>> node. Am I missing something here? >>>>>>>> >>>>>>>> Thanks a lot for your help >>>>>>>> Ray >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >