Return-Path: X-Original-To: apmail-cassandra-user-archive@www.apache.org Delivered-To: apmail-cassandra-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 5D2B6107EC for ; Tue, 4 Feb 2014 18:40:27 +0000 (UTC) Received: (qmail 23210 invoked by uid 500); 4 Feb 2014 18:40:23 -0000 Delivered-To: apmail-cassandra-user-archive@cassandra.apache.org Received: (qmail 23163 invoked by uid 500); 4 Feb 2014 18:40:23 -0000 Mailing-List: contact user-help@cassandra.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@cassandra.apache.org Delivered-To: mailing list user@cassandra.apache.org Received: (qmail 23155 invoked by uid 99); 4 Feb 2014 18:40:23 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 04 Feb 2014 18:40:23 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of rcoli@eventbrite.com designates 209.85.217.169 as permitted sender) Received: from [209.85.217.169] (HELO mail-lb0-f169.google.com) (209.85.217.169) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 04 Feb 2014 18:40:18 +0000 Received: by mail-lb0-f169.google.com with SMTP id q8so6905675lbi.28 for ; Tue, 04 Feb 2014 10:39:56 -0800 (PST) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:mime-version:in-reply-to:references:date :message-id:subject:from:to:content-type; bh=dNrxbtZy+67+oDpP7GtB2tG/HvDuuYraXksOwxAtU5c=; b=W5+vS4laoVkXyucS7Lbjlw57r8Xxv78mFS1dcnPEMnSejd14RoMK1VyLmft9ppf61T 8CchuqAZJan8Bu4arhriB+k1Z7DVZRGKNBfnSr/Q4YQDn4bAnYZuypg9btR55fSD504l ZoZoYOWAlbwcx/1+5XiVmPvkVhOFbmJJFa+hmJ29xJhPyWA6iOT4l6fGhrPRJpQe7lOk UdXfgfP4rEI1hMXnhu0IKjYd9W4KtX0W2ZF8f4hvn7+LeHxbqwd+YHIDbkqYuAtsj3XQ hQrE4wmrC2+PZ+9sk/+mHg16yGp1Mu+CcKc4A28xOnqBMC792r83fa2TXSbt4XD2OkDv WjZQ== X-Gm-Message-State: ALoCoQlOnaDZn2mrUJYsuHYYFE1Inn5VlLG3fvttNYO9R+ECKzSL38jQWo+Lk0LghDG/SCCV3l4R MIME-Version: 1.0 X-Received: by 10.112.41.100 with SMTP id e4mr120280lbl.84.1391539196741; Tue, 04 Feb 2014 10:39:56 -0800 (PST) Received: by 10.114.80.65 with HTTP; Tue, 4 Feb 2014 10:39:56 -0800 (PST) In-Reply-To: References: Date: Tue, 4 Feb 2014 10:39:56 -0800 Message-ID: Subject: Re: Data tombstoned during bulk loading 1.2.10 -> 2.0.3 From: Robert Coli To: "user@cassandra.apache.org" Content-Type: multipart/alternative; boundary=001a1135e87a835c7a04f198f73c X-Virus-Checked: Checked by ClamAV on apache.org --001a1135e87a835c7a04f198f73c Content-Type: text/plain; charset=ISO-8859-1 On Tue, Feb 4, 2014 at 12:21 AM, olek.stasiak@gmail.com < olek.stasiak@gmail.com> wrote: > I don't know what is the real cause of my problem. We are still guessing. > All operations I have done one cluster are described on timeline: > 1.1.7-> 1.2.10 -> upgradesstable -> 2.0.2 -> normal operations ->2.0.3 > -> normal operations -> now > normal operations means reads/writes/repairs. > Could you please, describe briefly how to recover data? I have a > problem with scenario described under link: > > http://thelastpickle.com/blog/2011/12/15/Anatomy-of-a-Cassandra-Partition.html, > I can't apply this solution to my case. > I think your only option is the following : 1) determine which SSTables contain rows have doomstones (tombstones from the far future) 2) determine whether these tombstones mask a live or dead version of the row, by looking at other row fragments 3) dump/filter/re-write all your data via some method, probably sstable2json/json2sstable 4) load the corrected sstables by starting a node with the sstables in the data directory I understand you have a lot of data, but I am pretty sure there is no way for you to fix it within Cassandra. Perhaps ask for advice on the JIRA ticket mentioned upthread if this answer is not sufficient? =Rob --001a1135e87a835c7a04f198f73c Content-Type: text/html; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable
On T= ue, Feb 4, 2014 at 12:21 AM, olek= .stasiak@gmail.com <olek.stasiak@gmail.com> wrote:<= br>
I don't know what is the real cause of m= y problem. We are still guessing.
All operations I have done one cluster are described on timeline:
1.1.7-> 1.2.10 -> upgradesstable -> 2.0.2 -> = normal operations ->2.0.3
-> normal operations -> now
normal operations means reads/writes/repairs.
Could you please, describe briefly how to recover data? I have a
problem with scenario described under link:
http://thelastpickle.com/blog/2011/12/15/= Anatomy-of-a-Cassandra-Partition.html ,
I can't apply this solution to my case.

=
I think your only option is the following :

1= ) determine which SSTables contain rows have doomstones (tombstones from th= e far future)
2) determine whether these tombstones mask a live or dead version of t= he row, by looking at other row fragments
3) dump/filter/re-write= all your data via some method, probably sstable2json/json2sstable
4) load the corrected sstables by starting a node with the sstables in= the data directory

I understand you have a lot of= data, but I am pretty sure there is no way for you to fix it within Cassan= dra. Perhaps ask for advice on the JIRA ticket mentioned upthread if this a= nswer is not sufficient?

=3DRob

--001a1135e87a835c7a04f198f73c--