Return-Path: X-Original-To: apmail-cassandra-user-archive@www.apache.org Delivered-To: apmail-cassandra-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 057CD11DD5 for ; Mon, 8 Sep 2014 21:01:31 +0000 (UTC) Received: (qmail 8239 invoked by uid 500); 8 Sep 2014 21:01:28 -0000 Delivered-To: apmail-cassandra-user-archive@cassandra.apache.org Received: (qmail 8203 invoked by uid 500); 8 Sep 2014 21:01:28 -0000 Mailing-List: contact user-help@cassandra.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@cassandra.apache.org Delivered-To: mailing list user@cassandra.apache.org Received: (qmail 8192 invoked by uid 99); 8 Sep 2014 21:01:28 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 08 Sep 2014 21:01:28 +0000 X-ASF-Spam-Status: No, hits=-0.7 required=5.0 tests=RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of tim@pagerduty.com designates 209.85.218.42 as permitted sender) Received: from [209.85.218.42] (HELO mail-oi0-f42.google.com) (209.85.218.42) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 08 Sep 2014 21:01:24 +0000 Received: by mail-oi0-f42.google.com with SMTP id e131so2892172oig.15 for ; Mon, 08 Sep 2014 14:01:03 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:mime-version:in-reply-to:references:date :message-id:subject:from:to:content-type; bh=bwkXWgQjQx/gDwibXI6aYw/fLr7YycX2qTO3Gffx7U0=; b=DRmxLC1MWZ5hkb6BR8wpIlg/PtXW2q3ANazmf9KW4cQ3EjlMItJlhwxVLSbJAn59h+ RinIf0QsSCu3DQwu/9odpHHHEj8i/oMTqZuiqLHJXv04duSM0WjDkbIGgVKBzwc4K48+ 7IX9WLyGSBElXLmQEx/qa/oU6OvwHouv7GTiC+8upSN37Rjb5yNTH6dcOSkMGdmuxOg/ 1HVagE2TrL38IDw+JYCYB+SAkgYYT2YjGLd34bD3/qaUI+vNwujRsb5rvpfPAYuZi5BE 3oDY1YzyAd5lgTL2ZhBj+ZyKYBIQxtJkUeL9iHEyjOIelKQMKDz35KEaw0QLzwEJ+6ny oB0A== X-Gm-Message-State: ALoCoQk/Fke1W3iEdiZdpQRP2m0QTPuCIXg49rj+6WUvpPl1nohm9vDoOaoUeQRjKkbWmrW36q9L MIME-Version: 1.0 X-Received: by 10.60.94.65 with SMTP id da1mr15896056oeb.11.1410210063761; Mon, 08 Sep 2014 14:01:03 -0700 (PDT) Received: by 10.76.152.193 with HTTP; Mon, 8 Sep 2014 14:01:03 -0700 (PDT) In-Reply-To: References: Date: Mon, 8 Sep 2014 14:01:03 -0700 Message-ID: Subject: Re: "Failed to enable shuffling" error From: Tim Heckman To: user@cassandra.apache.org Content-Type: text/plain; charset=UTF-8 X-Virus-Checked: Checked by ClamAV on apache.org On Mon, Sep 8, 2014 at 1:45 PM, Jonathan Haddad wrote: > I believe shuffle has been removed recently. I do not recommend using > it for any reason. We're still using the 1.2.x branch of Cassandra, and will be for some time due to the thrift deprecation. Has it only been removed from the 2.x line? > If you really want to go vnodes, your only sane option is to add a new > DC that uses vnodes and switch to it. We use the NetworkTopologyStrategy across three geographically separated regions. Doing it this way feels a bit more risky based on our replication strategy. Also, I'm not sure where all we have our current datacenter names defined across our different internal repositories. So there could be quite a large number of changes going this route. > The downside in the 2.0.x branch to using vnodes is that repairs take > N times as long, where N is the number of tokens you put on each node. > I can't think of any other reasons why you wouldn't want to use vnodes > (but this may be significant enough for you by itself) > > 2.1 should address the repair issue for most use cases. > > Jon Thank you for the notes on the behaviors in the 2.x branch. If we do move to the 2.x version that's something we'll be keeping in mind. Cheers! -Tim > On Mon, Sep 8, 2014 at 1:28 PM, Robert Coli wrote: >> On Mon, Sep 8, 2014 at 1:21 PM, Tim Heckman wrote: >>> >>> We're still at the exploratory stage on systems that are not >>> production-facing but contain production-like data. Based on our >>> placement strategy we have some concerns that the new datacenter >>> approach may be riskier or more difficult. We're just trying to gauge >>> both paths and see what works best for us. >> >> >> Your case of RF=N is probably the best possible case for shuffle, but >> general statements about how much this code has been exercised remain. :) >> >>> >>> The cluster I'm testing this on is a 5 node cluster with a placement >>> strategy such that all nodes contain 100% of the data. In practice we >>> have six clusters of similar size that are used for different >>> services. These different clusters may need additional capacity at >>> different times, so it's hard to answer the maximum size question. For >>> now let's just assume that the clusters may never see an 11th >>> member... but no guarantees. >> >> >> With RF of 3, cluster sizes of under approximately 10 tend to net lose from >> vnodes. If these clusters are not very likely to ever have more than 10 >> nodes, consider not using Vnodes. >> >>> >>> We're looking to use vnodes to help with easing the administrative >>> work of scaling out the cluster. The improvements of streaming data >>> during repairs amongst others. >> >> >> Most of these wins don't occur until you have a lot of nodes, but the fixed >> costs of having many ranges are paid all the time. >> >>> >>> For shuffle, it looks like it may be easier than adding a new >>> datacenter and then have to adjust the schema for a new "datacenter" >>> to come to life. And we weren't sure whether the same pitfalls of >>> shuffle would effect us while having all data on all nodes. >> >> >> Let us know! Good luck! >> >> =Rob >> > > > > -- > Jon Haddad > http://www.rustyrazorblade.com > twitter: rustyrazorblade