Return-Path: X-Original-To: apmail-cassandra-user-archive@www.apache.org Delivered-To: apmail-cassandra-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 60E25DB02 for ; Tue, 16 Oct 2012 01:33:07 +0000 (UTC) Received: (qmail 39630 invoked by uid 500); 16 Oct 2012 01:33:05 -0000 Delivered-To: apmail-cassandra-user-archive@cassandra.apache.org Received: (qmail 39613 invoked by uid 500); 16 Oct 2012 01:33:05 -0000 Mailing-List: contact user-help@cassandra.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@cassandra.apache.org Delivered-To: mailing list user@cassandra.apache.org Received: (qmail 39597 invoked by uid 99); 16 Oct 2012 01:33:04 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 16 Oct 2012 01:33:04 +0000 X-ASF-Spam-Status: No, hits=-0.7 required=5.0 tests=RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of will@voodoolunchbox.com designates 209.85.210.172 as permitted sender) Received: from [209.85.210.172] (HELO mail-ia0-f172.google.com) (209.85.210.172) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 16 Oct 2012 01:32:55 +0000 Received: by mail-ia0-f172.google.com with SMTP id o25so4826130iad.31 for ; Mon, 15 Oct 2012 18:32:34 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20120113; h=content-type:mime-version:subject:from:in-reply-to:date :content-transfer-encoding:message-id:references:to:x-mailer :x-gm-message-state; bh=FySlGn3LVJkmrk5gMAsYtjVUqFi32CTwVRlo8Muhiyo=; b=HDsja93JKXkNcWAZlpsINlhFFckxwV+hdFkNSDRcthb6PaM9appZrDe2/BZ+EEn/sw K2zAp/yVxmcBe8jxRYsRAXL35VXFSE0bGUE96mCjZ9FmcbeKxfPyOWY4H1W/80Z3Au3c DMfOVtslIibESVnzSwwmyBvwAS7moWpx8hV5UJYbJmbzPYvjXUD4UGSfHs+MfRT1Nzzp 5+5n4rxtB39VOEsOE/2AgJ/rOA8XD+y126UzK3B2+FaetatVG45fbow1tL1U0VkAOPBn 9N3uB4Cd2++GvIZ/fR/VB+XF8M8lzCEfKCjS+zEx7gOxFmJJ9BCfOgnr3dKpqw1N+CRm yZNw== Received: by 10.50.36.169 with SMTP id r9mr10419836igj.67.1350351154001; Mon, 15 Oct 2012 18:32:34 -0700 (PDT) Received: from [192.168.2.2] ([64.79.58.29]) by mx.google.com with ESMTPS id pq3sm7360896igc.8.2012.10.15.18.32.32 (version=TLSv1/SSLv3 cipher=OTHER); Mon, 15 Oct 2012 18:32:33 -0700 (PDT) Content-Type: text/plain; charset=iso-8859-1 Mime-Version: 1.0 (Apple Message framework v1278) Subject: Re: RF update From: Will Martin In-Reply-To: Date: Mon, 15 Oct 2012 21:32:30 -0400 Content-Transfer-Encoding: quoted-printable Message-Id: <868AB5E4-E058-4A16-BDDB-A5849A0155E9@voodoolunchbox.com> References: To: user@cassandra.apache.org X-Mailer: Apple Mail (2.1278) X-Gm-Message-State: ALoCoQkAnSO+REwwOI0nyZjZ+R0iUe54Tudwe46DGmYXy63sDX6krsuxoDH09Klfau3CqJiZfjVx X-Virus-Checked: Checked by ClamAV on apache.org +1 It doesn't make sense that the xfr compactions are heavy unless = they are translating the file. This could be a protocol mismatch: = however the requirements for node level compaction and wire compaction I = would expect to be pretty different. On Oct 15, 2012, at 4:42 PM, Matthias Broecheler wrote: > Hey, >=20 > we are writing a lot of data into a cassandra cluster for a batch = loading use case. We cannot use the sstable batch loader, so in order to = speed up the loading process we are using RF=3D1 while the data is = loading. After the load is complete, we want to increase the RF. For = that, we are updating the RF in the schema and then run the node repair = tool on each cassandra instance to stream the data over. However, we are = noticing that this process is slowed down by a lot of compactions (the = actually streaming of data only takes a couple of minutes). >=20 > Cassandra is already running a major compaction after the data loading = process has completed. But then, there are to be two more compactions = (one on the sender and one on the receiver) happening and those take a = very long time even on the aws high i/o instance with no compaction = throttling.=20 >=20 > Question: These additional compactions seem redundant since there are = no reads or writes on the cluster after the first major compaction = (immediately after the data load), is that right? And if so, what can we = do to avoid them? We are currently waiting multiple days. >=20 > Thank you very much for your help, > Matthias >=20