Return-Path: X-Original-To: apmail-cassandra-user-archive@www.apache.org Delivered-To: apmail-cassandra-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 078524D67 for ; Thu, 7 Jul 2011 10:58:22 +0000 (UTC) Received: (qmail 59185 invoked by uid 500); 7 Jul 2011 10:58:18 -0000 Delivered-To: apmail-cassandra-user-archive@cassandra.apache.org Received: (qmail 58923 invoked by uid 500); 7 Jul 2011 10:58:14 -0000 Mailing-List: contact user-help@cassandra.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@cassandra.apache.org Delivered-To: mailing list user@cassandra.apache.org Received: (qmail 58915 invoked by uid 99); 7 Jul 2011 10:58:13 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 07 Jul 2011 10:58:13 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=FREEMAIL_FROM,HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_PASS,T_TO_NO_BRKTS_FREEMAIL X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of iecanfly@gmail.com designates 74.125.82.172 as permitted sender) Received: from [74.125.82.172] (HELO mail-wy0-f172.google.com) (74.125.82.172) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 07 Jul 2011 10:58:06 +0000 Received: by wyj26 with SMTP id 26so635385wyj.31 for ; Thu, 07 Jul 2011 03:57:45 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :content-type; bh=cDGCuRnjZFe/0VTAkMdCk7sex8qFJPks6QbQmoNnA5Q=; b=JLbrHyRJ0fasCC45Y2KOU4AtnScLn/LMJyYw8+bfSMYKc/khtYLqnkK2M8C9CkA9PX E4vuflDS07s6xHh2ih0LbW7YHgT9tiTkaXYR/NVsO0r7t0dU5tL3IaF0hDIV4R0F5WRr yODkS10WGEQR20+AeueSs+w8MiXuqb+z5+xhM= MIME-Version: 1.0 Received: by 10.216.63.68 with SMTP id z46mr6319268wec.82.1310036265617; Thu, 07 Jul 2011 03:57:45 -0700 (PDT) Received: by 10.216.60.136 with HTTP; Thu, 7 Jul 2011 03:57:45 -0700 (PDT) In-Reply-To: References: Date: Thu, 7 Jul 2011 19:57:45 +0900 Message-ID: Subject: Re: minimum number of machines for RF=3 From: david lee To: user@cassandra.apache.org Content-Type: multipart/alternative; boundary=000e0ce0b71a41499804a778965a --000e0ce0b71a41499804a778965a Content-Type: text/plain; charset=ISO-8859-1 cheers~ On 7 July 2011 19:41, Watanabe Maki wrote: > It is expected behaviour and not relate on number of node. > After the failed node bringing back, the ring will be busy by Hinted > Handoff rewriting and Read Repair. If you run repair, all your 3 nodes need > to build Merkel Tree, compare the hash values, then transfer latest data to > each other. > > You can tune the HH, read repair to reduce performance impact on self > healing activities. > > maki > > On 2011/07/07, at 18:10, david lee wrote: > > > hi guys, > > > > is there a minimum(recommended) number of machines for RF=3? > > > > i encountered a test result where > > # of nodes = 3 > > RF = 3 > > CL.READ=QUORUM > > and when 1 node was taken out and back in after which node repair was > run, > > the TPS dropped significantly. > > > > is this behaviour expected since the RF (for the duration of when 1 node > was taken out) > > is higher than the # of nodes? > > > > cheers, > > david > > --000e0ce0b71a41499804a778965a Content-Type: text/html; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable cheers~

On 7 July 2011 19:41, Watanabe Ma= ki <watanab= e.maki@gmail.com> wrote:
It is expected behaviour and not relate on number of node.
After the failed node bringing back, the ring will be busy by Hinted Handof= f rewriting and Read Repair. If you run repair, all your 3 nodes need to bu= ild Merkel Tree, compare the hash values, =A0then transfer latest data to e= ach other.

You can tune the HH, read repair to reduce performance impact on self heali= ng activities.

maki

On 2011/07/07, at 18:10, david lee <iecanfly@gmail.com> wrote:

> hi guys,
>
> is there a minimum(recommended) number of machines for RF=3D3?
>
> i encountered a test result where
> # of nodes =3D 3
> RF =A0=3D =A03
> CL.READ=3DQUORUM
> and when 1 node was taken out and back in after which node repair was = run,
> the TPS dropped significantly.
>
> is this behaviour expected since the RF (for the duration of when 1 no= de was taken out)
> is higher than the # of nodes?
>
> cheers,
> david


--000e0ce0b71a41499804a778965a--