Return-Path: X-Original-To: apmail-cassandra-user-archive@www.apache.org Delivered-To: apmail-cassandra-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id DC88EE90C for ; Sat, 29 Dec 2012 19:52:05 +0000 (UTC) Received: (qmail 24158 invoked by uid 500); 29 Dec 2012 19:52:02 -0000 Delivered-To: apmail-cassandra-user-archive@cassandra.apache.org Received: (qmail 24138 invoked by uid 500); 29 Dec 2012 19:52:02 -0000 Mailing-List: contact user-help@cassandra.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@cassandra.apache.org Delivered-To: mailing list user@cassandra.apache.org Received: (qmail 24129 invoked by uid 99); 29 Dec 2012 19:52:02 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 29 Dec 2012 19:52:02 +0000 X-ASF-Spam-Status: No, hits=0.2 required=5.0 tests=FREEMAIL_ENVFROM_END_DIGIT,RCVD_IN_DNSWL_NONE,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: local policy) Received: from [98.136.217.111] (HELO nm39-vm8.bullet.mail.gq1.yahoo.com) (98.136.217.111) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 29 Dec 2012 19:51:53 +0000 Received: from [98.137.12.191] by nm39.bullet.mail.gq1.yahoo.com with NNFMP; 29 Dec 2012 19:51:33 -0000 Received: from [98.136.185.40] by tm12.bullet.mail.gq1.yahoo.com with NNFMP; 29 Dec 2012 19:51:33 -0000 Received: from [127.0.0.1] by smtp101.mail.gq1.yahoo.com with NNFMP; 29 Dec 2012 19:51:33 -0000 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=yahoo.com; s=s1024; t=1356810693; bh=vmzRgt9Lu+pAnuCVRoJjV2E17xod2ZiFndX2/jJ2+ms=; h=X-Yahoo-Newman-Id:X-Yahoo-Newman-Property:X-YMail-OSG:X-Yahoo-SMTP:Received:Message-ID:Date:From:User-Agent:MIME-Version:To:Subject:Content-Type:Content-Transfer-Encoding; b=ctLQcCBEpynaW4conljHojlngPM9olw0eqhlPK+qlHMyv8H0uHtBjY+M3lDZkOpwxBn5mpa01/TvaiPsDEWdjXN8yUhpszk91vG4wyW1vuqnDuiSpTg4pjbJhhUiKrgx+LGd2Md+swxiPk4HsWz6fb9FV9f5NW2dAj3hTgx+7+Y= X-Yahoo-Newman-Id: 494904.62125.bm@smtp101.mail.gq1.yahoo.com X-Yahoo-Newman-Property: ymail-3 X-YMail-OSG: VRV_dp4VM1lKsZawD.08RzSsQu0YpFH5er0BmOcfG2xGams _lVjlSZ1SFkpnoaRxmKx283idg4jhgIqoIRjPtBdNiERJGssKDNyQV8ePV3u nrm8o8nhXRTEeoIwwODa_HW_BGphcrs_ZTAGFeyIL33eVJ8iUAk4W.vPm0lU GcwPo50_iDKUfYnN_D369FNZF99nDj7kPact52Qtx02DlF7klWMq5rhuPSY. vnabVrdXEr9bGUOejM5nZMSy5YUjXbFek0bQ8cqJ4sYTgqOFXfn5YHdH1tGa 14veBkzhS1EWY4aoPnsnFQAQgWQdhWvuQ87hzZ_2Qq_zbEPEQreQKNSs1Inu 1HNp03zQYRDqqd_YyBGY5S1QzYU65o1HCt8ZAwuaC2nsmZtWiVKAuboOzHKg FX0m0lvqaa3Wuc4nN7lpmnVp.GDPnCaN0O8tX._NtERDvVpB4gvnnUzmD09W px_ei3exUEGc- X-Yahoo-SMTP: t0UN_U2swBCFgwLIRu70LU92TrvpdQ-- Received: from [192.168.1.5] (mtheroux2@76.118.248.45 with plain) by smtp101.mail.gq1.yahoo.com with SMTP; 29 Dec 2012 11:51:33 -0800 PST Message-ID: <50DF49CF.2060800@yahoo.com> Date: Sat, 29 Dec 2012 14:51:43 -0500 From: Mike User-Agent: Mozilla/5.0 (Windows NT 6.1; WOW64; rv:17.0) Gecko/17.0 Thunderbird/17.0 MIME-Version: 1.0 To: user@cassandra.apache.org Subject: Column Family migration/tombstones Content-Type: text/plain; charset=ISO-8859-1; format=flowed Content-Transfer-Encoding: 7bit X-Virus-Checked: Checked by ClamAV on apache.org Hello, We are undergoing a change to our internal datamodel that will result in the eventual deletion of over a hundred million rows from a Cassandra column family. From what I understand, this will result in the generation of tombstones, which will be cleaned up during compaction, after gc_grace_period time (default: 10 days). A couple of questions: 1) As one can imagine, the index and bloom filter for this column family is large. Am I correct to assume that bloom filter and index space will not be reduced until after gc_grace_period? 2) If I would manually run repair across a cluster, is there a process I can use to safely remove these tombstones before gc_grace period to free this memory sooner? 3) Any words of warning when undergoing this? We are running Cassandra 1.1.2 on a 6 node cluster and a Replication Factor of 3. We use LOCAL_QUORM consistency for all operations. Thanks! -Mike