Return-Path: Delivered-To: apmail-couchdb-user-archive@www.apache.org Received: (qmail 15393 invoked from network); 28 Feb 2011 02:21:11 -0000 Received: from hermes.apache.org (HELO mail.apache.org) (140.211.11.3) by minotaur.apache.org with SMTP; 28 Feb 2011 02:21:11 -0000 Received: (qmail 11642 invoked by uid 500); 28 Feb 2011 02:21:09 -0000 Delivered-To: apmail-couchdb-user-archive@couchdb.apache.org Received: (qmail 11600 invoked by uid 500); 28 Feb 2011 02:21:09 -0000 Mailing-List: contact user-help@couchdb.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@couchdb.apache.org Delivered-To: mailing list user@couchdb.apache.org Received: (qmail 11592 invoked by uid 99); 28 Feb 2011 02:21:09 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 28 Feb 2011 02:21:09 +0000 X-ASF-Spam-Status: No, hits=-0.7 required=5.0 tests=FREEMAIL_FROM,RCVD_IN_DNSWL_LOW,SPF_PASS,T_TO_NO_BRKTS_FREEMAIL X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of adam.kocoloski@gmail.com designates 209.85.216.52 as permitted sender) Received: from [209.85.216.52] (HELO mail-qw0-f52.google.com) (209.85.216.52) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 28 Feb 2011 02:20:59 +0000 Received: by qwf6 with SMTP id 6so2968499qwf.11 for ; Sun, 27 Feb 2011 18:20:39 -0800 (PST) Received: by 10.229.236.134 with SMTP id kk6mr3819270qcb.93.1298859639047; Sun, 27 Feb 2011 18:20:39 -0800 (PST) Received: from [10.0.1.6] (c-71-232-49-44.hsd1.ma.comcast.net [71.232.49.44]) by mx.google.com with ESMTPS id q12sm2812037qcu.6.2011.02.27.18.20.37 (version=TLSv1/SSLv3 cipher=OTHER); Sun, 27 Feb 2011 18:20:37 -0800 (PST) Content-Type: text/plain; charset=us-ascii Mime-Version: 1.0 (Apple Message framework v1082) Subject: Re: How many nodes can couchdb scale to? From: Adam Kocoloski In-Reply-To: Date: Sun, 27 Feb 2011 21:20:36 -0500 Content-Transfer-Encoding: quoted-printable Message-Id: <1286BC69-1352-4BB2-B647-5938E9ACF88C@apache.org> References: To: user@couchdb.apache.org X-Mailer: Apple Mail (2.1082) X-Virus-Checked: Checked by ClamAV on apache.org On Feb 27, 2011, at 8:56 PM, Isaac Force wrote: > On Sun, Feb 27, 2011 at 4:13 PM, niall el-assaad = wrote: >> I'm looking at developing an application that will have a couple of = central >> nodes (data centre) and around 2,000 remote nodes (in branch = offices). >>=20 >> I'd like to know if couchdb can scale to having this many nodes = working in a >> cluster. >=20 > For the list to write a more meaningful answer to address your needs, > it would be helpful if you gave more detail about the nature of the > problem you want to solve. >=20 > * Replication topology: is the plan to have replication from the = branch office > nodes to your centralized data center? (n:1) > * Replication type: continuous or triggered manually/programatically? > * Scope of data set: I would be more concerned with writes than reads. > You'll need to have an idea of what your current aggregate average = and > peak writes per second are, how much data is written for a given > period of time, and how far you think you will need this rate to = scale in > the future. > * Why Couch: is CouchDB going to be addressing a brand new need, or > is it going to replace existing systems for known reasons? If it's = the > latter, what is it about your current systems that aren't meeting = your > demands, and what do you hope Couch will provide that will fill the = gap? > (Specifically looking for performance data that you might have = already > collected, and if Couch is going to be living on your existing = hardware > or new hardware.) >=20 > I haven't dealt with large distributed Couch systems, but my instinct > would be that Couch wouldn't have any problem with a 2000:1 replicated > system. (See Ubuntu One as an example of a large CouchDB system with > many external replicators.) The ability to handle it would come down > to how well the aggregate data set matches the size of hardware and > replication layout in your data center, and of course available > ingress bandwidth. >=20 > -Isaac Well said Isaac.