Return-Path: X-Original-To: apmail-cassandra-user-archive@www.apache.org Delivered-To: apmail-cassandra-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 0B4D0454A for ; Fri, 27 May 2011 15:48:07 +0000 (UTC) Received: (qmail 46954 invoked by uid 500); 27 May 2011 15:48:04 -0000 Delivered-To: apmail-cassandra-user-archive@cassandra.apache.org Received: (qmail 46906 invoked by uid 500); 27 May 2011 15:48:04 -0000 Mailing-List: contact user-help@cassandra.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@cassandra.apache.org Delivered-To: mailing list user@cassandra.apache.org Received: (qmail 46898 invoked by uid 99); 27 May 2011 15:48:04 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 27 May 2011 15:48:04 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=FREEMAIL_FROM,HTML_MESSAGE,RCVD_IN_DNSWL_LOW,RFC_ABUSE_POST,SPF_PASS,T_TO_NO_BRKTS_FREEMAIL X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of keteracel@gmail.com designates 209.85.161.44 as permitted sender) Received: from [209.85.161.44] (HELO mail-fx0-f44.google.com) (209.85.161.44) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 27 May 2011 15:47:58 +0000 Received: by fxm15 with SMTP id 15so1596420fxm.31 for ; Fri, 27 May 2011 08:47:38 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:mime-version:in-reply-to:references:date :message-id:subject:from:to:content-type; bh=+2fbojGySa4hsC/Ef1i7dugd9NZyHkncFdxL7MJZisI=; b=Ul71NfuMuvrtEJnOLgsJTaByliVZVfiuKutf1ulMPQmzUJmjKm2uDTlsm0CmHLPK74 9QM7HBuotkc5PSnyykpiBi0WeAE+kTu1xL51f9lZYuZWowApm2KpM6fbPSH54NNQPleQ +ChX7v8dk52msu4HpZYZZjz6jPCQJxCyBkuPw= DomainKey-Signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :content-type; b=VZQeI7jT1mqAk32pwv8MDi2EImxov9RLOcpGL/A4lr5yuoCZhgowYXRkX5/QUH7lLy GFUpgWS6PEfroKwtpYuEMlDDaHcGGhW4XZPjUpYkwj6tQc9zsdWgrZVzgHX8a5sGxJA6 B/KqnAQyfKY/JcCQZiRAqpyhy9jxIAoB177bY= MIME-Version: 1.0 Received: by 10.223.75.15 with SMTP id w15mr2428987faj.134.1306511258384; Fri, 27 May 2011 08:47:38 -0700 (PDT) Received: by 10.223.113.134 with HTTP; Fri, 27 May 2011 08:47:38 -0700 (PDT) In-Reply-To: References: Date: Fri, 27 May 2011 16:47:38 +0100 Message-ID: Subject: Re: Cluster not recovering when a single node dies From: Paul Loy To: user@cassandra.apache.org Content-Type: multipart/alternative; boundary=0015174c19567378bf04a443db24 X-Virus-Checked: Checked by ClamAV on apache.org --0015174c19567378bf04a443db24 Content-Type: text/plain; charset=ISO-8859-1 I guess my next question is: the data should be complete somewhere in the ring with RF = 2. Does cassandra not redistribute the replication ring without a nodetool decommission call? On Fri, May 27, 2011 at 4:45 PM, Paul Loy wrote: > ahh, thanks. > > On Fri, May 27, 2011 at 4:43 PM, Jonathan Ellis wrote: > >> Quorum of 2 is 2. You need at least RF=3 for quorum to tolerate losing >> a node indefinitely. >> >> On Fri, May 27, 2011 at 10:37 AM, Paul Loy wrote: >> > We have a 4 node cluster with a replication factor of 2. When one node >> dies, >> > the other nodes throw UnavailableExceptions for quorum reads (as >> expected >> > initially). They never get out of that state. >> > >> > Is there something we can do in nodetool to make the remaining nodes >> > function? >> > >> > Thanks. >> > >> > -- >> > --------------------------------------------- >> > Paul Loy >> > paul@keteracel.com >> > http://uk.linkedin.com/in/paulloy >> > >> >> >> >> -- >> Jonathan Ellis >> Project Chair, Apache Cassandra >> co-founder of DataStax, the source for professional Cassandra support >> http://www.datastax.com >> > > > > -- > --------------------------------------------- > Paul Loy > paul@keteracel.com > http://uk.linkedin.com/in/paulloy > -- --------------------------------------------- Paul Loy paul@keteracel.com http://uk.linkedin.com/in/paulloy --0015174c19567378bf04a443db24 Content-Type: text/html; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable I guess my next question is: the data should be complete somewhere in the r= ing with RF =3D 2. Does cassandra not redistribute the replication ring wit= hout a nodetool decommission call?

On Fri= , May 27, 2011 at 4:45 PM, Paul Loy <keteracel@gmail.com> wrote:
ahh, thanks.

On Fri, May 27, 2011 at 4:43 PM, Jon= athan Ellis <jbellis@gmail.com> wrote:
Quorum of 2 is 2. You need at least RF=3D3 for quorum to tolerate losing a node indefinitely.

On Fri, May 27, 2011 at 10:37 AM, Paul Loy <keteracel@gmail.com> wrote:
> We have a 4 node cluster with a replication factor of 2. When one node= dies,
> the other nodes throw UnavailableExceptions for quorum reads (as expec= ted
> initially). They never get out of that state.
>
> Is there something we can do in nodetool to make the remaining nodes > function?
>
> Thanks.
>
> --
> ---------------------------------------------
> Paul Loy
> paul@keteracel= .com
> http:/= /uk.linkedin.com/in/paulloy
>



--
Jonathan Ellis
Project Chair, Apache Cassandra
co-founder of DataStax, the source for professional Cassandra support
http://www.datastax.c= om



--
---------------------------------------------=
Paul Loy
pau= l@keteracel.com
h= ttp://uk.linkedin.com/in/paulloy



--
-----------= ----------------------------------
Paul Loy
paul@keteracel.com
http://uk.linkedin.com/in/paulloy
--0015174c19567378bf04a443db24--