Return-Path: X-Original-To: apmail-couchdb-user-archive@www.apache.org Delivered-To: apmail-couchdb-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 9FE741040E for ; Sun, 4 Aug 2013 15:33:04 +0000 (UTC) Received: (qmail 93002 invoked by uid 500); 4 Aug 2013 15:33:02 -0000 Delivered-To: apmail-couchdb-user-archive@couchdb.apache.org Received: (qmail 92931 invoked by uid 500); 4 Aug 2013 15:33:02 -0000 Mailing-List: contact user-help@couchdb.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@couchdb.apache.org Delivered-To: mailing list user@couchdb.apache.org Received: (qmail 92914 invoked by uid 99); 4 Aug 2013 15:33:00 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Sun, 04 Aug 2013 15:33:00 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_LOW X-Spam-Check-By: apache.org Received-SPF: error (nike.apache.org: local policy) Received: from [209.85.214.179] (HELO mail-ob0-f179.google.com) (209.85.214.179) by apache.org (qpsmtpd/0.29) with ESMTP; Sun, 04 Aug 2013 15:32:54 +0000 Received: by mail-ob0-f179.google.com with SMTP id fb19so3913845obc.24 for ; Sun, 04 Aug 2013 08:32:13 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20120113; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :content-type:x-gm-message-state; bh=3cdlg6O8bx0j9DTzQRM74Wfw1eAhD6DZPOo2mc/3Ig0=; b=Q9NJlXnaC1OhFuzVTdEBmdS3UWlRt3UrfRSLlqjpnxyH/wa3QiijLTLnME0Vc7cDwz vJDgupb4INH94IGreJCYdtx6+jucz1xbF9JOhdZ2Km9MPzoo/krR3Q3RrMibwsPM4+8P VDGQaSIptD58NxxXc8KXQ6oImoyhH/BVZvab0Cm8DxPQXisQ7tosXr45StTbhBORub50 E5dmELNaNTS5eR1Rr1rbw+cgByAcBI34NNZYqidljrmiqyvJTp3QwZV4BH+ioNBzourl MPhNS1uLQ69uMeOkMnFx2HZtemCQVweQ6KuqvPlbpncfdd9tCjPhfsADlGXoLtvwq3Li 0UtA== MIME-Version: 1.0 X-Received: by 10.182.22.5 with SMTP id z5mr11100853obe.42.1375630333091; Sun, 04 Aug 2013 08:32:13 -0700 (PDT) Received: by 10.182.246.33 with HTTP; Sun, 4 Aug 2013 08:32:12 -0700 (PDT) Received: by 10.182.246.33 with HTTP; Sun, 4 Aug 2013 08:32:12 -0700 (PDT) In-Reply-To: <20130804141807.GA1197@atypical.net> References: <20130804141807.GA1197@atypical.net> Date: Sun, 4 Aug 2013 08:32:12 -0700 Message-ID: Subject: Re: distributed map/reduce in BigCouch From: Stanley Iriele To: user@couchdb.apache.org Content-Type: multipart/alternative; boundary=001a1133232e58b1ae04e320e56f X-Gm-Message-State: ALoCoQmJkpk1+ACLF6aRI+PRNnjlhOwBlpbx4b7/LmxVWPplIBuZOQml5EBGV74yXo7GhsQSWyqW X-Virus-Checked: Checked by ClamAV on apache.org --001a1133232e58b1ae04e320e56f Content-Type: text/plain; charset=ISO-8859-1 Thanks joan, I worded part of that poorly...When I said hard drives I really meant physical machines.. But the bottleneck was disk space; so I said hard disks... But That completely answered my question actually.... I appreciate that...what is this 'q' value?..is that the number of nodes?.. Or is that the r + w > n thing... Either way that answers questions thank you! On Aug 4, 2013 7:18 AM, "Joan Touzet" wrote: > Hi Stanley, > > Let me provide a simplistic explanation, and others can help refine it > as necessary. > > On Sat, Aug 03, 2013 at 09:34:48AM -0700, Stanley Iriele wrote: > > How then does distributed map/reduce work? > > Each BigCouch node with a shard of the database also keeps that shard of > the view. When a request is made for a view, sufficient nodes are > queried to retrieve the view result, with the reduce operation occurring > as part of the return. > > > if not all nodes have replications of all things how does that > coordination > > happen during view building? > > This is not true, all nodes do not have replications of all things. If > you ask a node for a view on a database it does not have at all, it will > use information in the partition map to ask that question of a node that > has at least one shard of the database in question, which will in turn > complete the scatter/gather request to other nodes participating in that > database. > > > also its sharded right so certain nodes have a certain range of keys. > > Right, see above. > > > My problem is this. I need a solution that can incrementally scale across > > many hard disks...does big couch do this? with views and such?..if > > so..how?...and what are the drawbacks? > > I wouldn't necessarily recommend running 1 BigCouch process per HD you > have on a single machine, but there's no reason that it wouldn't work. > > The real challenge is that a database's partition map is determined > staticly at the time of database creation. If you choose to add more HDs > after this creation time, you will have to create a new database with > more shards, then replicate data to this new database to use the new > disks. The other option would be to use a very high number for 'q', then > rebalance the shard map onto the new disks and BigCouch processes. There > is a StackOverflow answer from Robert Newson that describes the process > for performing this operation. > > In short, neither is trivial nor automated. For a single-machine system, > you'd do far better to use some sort of Logical Volume Manager to deal > with expanding storage over time, such as Linux's lvm, some HW raid > cards, ZFS or similar features in OSX and Windows. > > > Thanks for any kind of response. > > > > Regards, > > > > Stanley > > -- > Joan Touzet | joant@atypical.net | wohali everywhere else > --001a1133232e58b1ae04e320e56f--