Return-Path: X-Original-To: apmail-cassandra-user-archive@www.apache.org Delivered-To: apmail-cassandra-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 2827910C43 for ; Mon, 21 Oct 2013 18:12:06 +0000 (UTC) Received: (qmail 3099 invoked by uid 500); 21 Oct 2013 18:12:03 -0000 Delivered-To: apmail-cassandra-user-archive@cassandra.apache.org Received: (qmail 3076 invoked by uid 500); 21 Oct 2013 18:12:03 -0000 Mailing-List: contact user-help@cassandra.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@cassandra.apache.org Delivered-To: mailing list user@cassandra.apache.org Received: (qmail 3068 invoked by uid 99); 21 Oct 2013 18:12:02 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 21 Oct 2013 18:12:02 +0000 X-ASF-Spam-Status: No, hits=-0.1 required=5.0 tests=HTML_MESSAGE,NORMAL_HTTP_TO_IP,RCVD_IN_DNSWL_MED,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of john.pyeatt@singlewire.com designates 74.125.149.242 as permitted sender) Received: from [74.125.149.242] (HELO na3sys009aog117.obsmtp.com) (74.125.149.242) by apache.org (qpsmtpd/0.29) with SMTP; Mon, 21 Oct 2013 18:11:55 +0000 Received: from mail-pb0-f51.google.com ([209.85.160.51]) (using TLSv1) by na3sys009aob117.postini.com ([74.125.148.12]) with SMTP ID DSNKUmVuVdYbJQFwmfucgEO/8NGqeziMNbhk@postini.com; Mon, 21 Oct 2013 11:11:34 PDT Received: by mail-pb0-f51.google.com with SMTP id wz7so2725475pbc.38 for ; Mon, 21 Oct 2013 11:11:33 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:mime-version:date:message-id:subject:from:to :content-type; bh=2HFg//gU9glxL3froMY/s8H4RxtX3wTko2PKb1WsOl4=; b=YaDs0BaBbUAHMnbYuu+kaE1/1DxKiFv3UDhcHn5F5cy3xLAgxgEJKQyZmj5QI4yNJL RXsEZT0eqm6Y3TNJ3tdyC9xgLq+Vvk7ju1aT0frEIJFRkS36ntqjHYDCx89N52QrUuaw S1/MYpCfe4/oxyNAxZhBDJoFl0bxr5WLGG15Pc0ZUJs/qjRFYOnxgy+M5LJxal3K/Df4 ySTu8V0fpo7uxekGzrZU+NBIDuXXskUIE2ccC5UXXZJPBi8tHd+hAqeUeEtksQvi4IyW WZWjTDU4ZdOoVZup8GTSV006tI13HdWfvVmLbyROZi2H6IiK/kxc08qC8KWdveQtSdM8 d3Ug== X-Gm-Message-State: ALoCoQmJtkq+13377iapd6+N2P3W2ZValZNsrrCtlAewv4GOaDnfPxQfKUh5+ojfd2wM1cVJFjWasRyeBD+bgmd/lQuEkQ8uK4ULLzQ8xiIzhSw3qQsbeLZ+xd4z30U2tBG1EeSoUIDCMyhtc5TfWcLMyN+fVz3kVdjKDd2rER2T6uvnmZDPhNI= X-Received: by 10.68.203.34 with SMTP id kn2mr18603075pbc.82.1382379093162; Mon, 21 Oct 2013 11:11:33 -0700 (PDT) MIME-Version: 1.0 X-Received: by 10.68.203.34 with SMTP id kn2mr18603049pbc.82.1382379092687; Mon, 21 Oct 2013 11:11:32 -0700 (PDT) Received: by 10.70.93.102 with HTTP; Mon, 21 Oct 2013 11:11:32 -0700 (PDT) Date: Mon, 21 Oct 2013 13:11:32 -0500 Message-ID: Subject: decommission of one EC2 node in cluster causes other nodes to go DOWN/UP and results in "May not be enough replicas..." From: John Pyeatt To: user@cassandra.apache.org Content-Type: multipart/alternative; boundary=047d7b10c851c44fec04e9443624 X-Virus-Checked: Checked by ClamAV on apache.org --047d7b10c851c44fec04e9443624 Content-Type: text/plain; charset=ISO-8859-1 We have a 6 node cassandra 1.2.10 cluster running on aws with NetworkTopologyStrategy, a replication factor of 3 and the EC2Snitch. Each AWS availability zone has 2 nodes in it. When we are reading or writing data with consistency of Quorum to the cluster while decommissioning a node we are getting 'May not be enough replicas present to handle consistency level". This doesn't make sense because we are only taking one node down, we have an RF of three so even if we take one node down with a quorum read/write there should still be enough nodes with the data (2). Looking at the cassandra log on a server that we are not decommissioning we are seeing this during the decommission of the other node. INFO [GossipTasks:1] 2013-10-21 15:18:10,695 Gossiper.java (line 803) InetAddress /10.0.22.142 *is now DOWN* INFO [GossipTasks:1] 2013-10-21 15:18:10,696 Gossiper.java (line 803) InetAddress /10.0.32.159 *is now DOWN* INFO [HANDSHAKE-/10.0.22.142] 2013-10-21 15:18:10,862 OutboundTcpConnection.java (line 399) Handshaking version with /10.0.22.142 INFO [GossipTasks:1] 2013-10-21 15:18:11,696 Gossiper.java (line 803) InetAddress /10.0.12.178* is now DOWN* INFO [GossipTasks:1] 2013-10-21 15:18:11,697 Gossiper.java (line 803) InetAddress /10.0.22.106* is now DOWN* INFO [GossipTasks:1] 2013-10-21 15:18:11,698 Gossiper.java (line 803) InetAddress /10.0.32.248 *is now DOWN* Eventually we are seeing a message that looks like this. INFO [GossipStage:3] 2013-10-21 15:18:19,429 Gossiper.java (line 789) InetAddress /10.0.32.248 is now UP for each of the nodes. So eventually the remaining nodes in the cluster come back to life. While these nodes are down I can see why we get the "May not be enough replicas..." message. Because everything is down. My question is *why does gossip shutdown for these nodes that we aren't decommissioning in the first place*? -- John Pyeatt Singlewire Software, LLC www.singlewire.com ------------------ 608.661.1184 john.pyeatt@singlewire.com --047d7b10c851c44fec04e9443624 Content-Type: text/html; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable
We have a 6 node cassandra 1.2.10 cluster running on = aws with NetworkTopologyStrategy, a replication factor of 3 and the EC2Snit= ch. Each AWS availability zone has 2 nodes in it.

Wh= en we are reading or writing data with consistency of Quorum to the cluster= while decommissioning a node we are getting 'May not be enough replica= s present to handle consistency level".

This doesn't make sense because we are only taking one n= ode down, we have an RF of three so even if we take one node down with a qu= orum read/write there should still be enough nodes with the data (2).

Looking at the cassandra log on a server that we are not dec= ommissioning we are seeing this during the decommission of the other node.<= br>
=A0INFO [GossipTasks:1] 2013-10-21 15:18:10,695 Gossiper.java (line = 803) InetAddress /10.0.22.= 142 is now DOWN
=A0INFO [GossipTasks:1] 2013-10-21 15:18:10,696 Gossiper.java (line 803) In= etAddress /10.0.32.159= is now DOWN
=A0INFO [HANDSHAKE-/10.0.2= 2.142] 2013-10-21 15:18:10,862 OutboundTcpConnection.java (line 399) Ha= ndshaking version with /10= .0.22.142
=A0INFO [GossipTasks:1] 2013-10-21 15:18:11,696 Gossiper.java (line 803) In= etAddress /10.0.12.178= is now DOWN
=A0INFO [GossipTasks:1] 2013-10-21 15:18:11,697 Gossiper.java (line 803) In= etAddress /10.0.22.106= is now DOWN
=A0INFO [GossipTasks:1] 2013-10-21 15:18:11,698 Gossiper.java (line 803) In= etAddress /10.0.32.248= is now DOWN

Eventually we are seeing a= message that looks like this.
=A0INFO [GossipStage:3] 2013-10-21 15:18:19,429 Gossiper.java (line 789) In= etAddress /10.0.32.248= is now UP

for each of the nodes. So eventuall= y the remaining nodes in the cluster come back to life.

While these nodes are down I can see why we get the "May not be en= ough replicas..." message. Because everything is down.

My quest= ion is why does gossip shutdown for these nodes that we aren't decom= missioning in the first place?

--
John Pyeatt
Singlewire Software, LLC
www.singlewire.com=
------------------
608.661.1184
john.pyeatt= @singlewire.com
--047d7b10c851c44fec04e9443624--