Return-Path: Delivered-To: apmail-hadoop-hbase-user-archive@minotaur.apache.org Received: (qmail 99312 invoked from network); 25 Nov 2009 09:55:54 -0000 Received: from hermes.apache.org (HELO mail.apache.org) (140.211.11.3) by minotaur.apache.org with SMTP; 25 Nov 2009 09:55:54 -0000 Received: (qmail 82109 invoked by uid 500); 25 Nov 2009 09:55:53 -0000 Delivered-To: apmail-hadoop-hbase-user-archive@hadoop.apache.org Received: (qmail 82051 invoked by uid 500); 25 Nov 2009 09:55:53 -0000 Mailing-List: contact hbase-user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: hbase-user@hadoop.apache.org Delivered-To: mailing list hbase-user@hadoop.apache.org Received: (qmail 82041 invoked by uid 99); 25 Nov 2009 09:55:53 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 25 Nov 2009 09:55:53 +0000 X-ASF-Spam-Status: No, hits=-0.0 required=10.0 tests=SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of ryanobjc@gmail.com designates 209.85.222.198 as permitted sender) Received: from [209.85.222.198] (HELO mail-pz0-f198.google.com) (209.85.222.198) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 25 Nov 2009 09:55:42 +0000 Received: by pzk36 with SMTP id 36so787927pzk.5 for ; Wed, 25 Nov 2009 01:55:20 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:mime-version:received:in-reply-to:references :date:message-id:subject:from:to:content-type :content-transfer-encoding; bh=zO4dKG9JE+6XY3NGTIw4IiGfA49gbCldi/aN68YXXXg=; b=hiOWuBrh0BP3H0qzBHX3WVzSR8pmtKyrygb2VTPae9k8awCK9+Vq3MPqH1VsgQseLZ FyehT4W2lBsYmmMI+HsukGy84kA/aF3+pt786wB7222nbMcFHfmtZ8l/wVtLepIK75o8 8qjnDYrQjJOss0WuxuYPQ0s0yiRJJFceXGNGk= DomainKey-Signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :content-type:content-transfer-encoding; b=O7MtvDmWnXt79lm7jC9U8HDDvYm20QNVkOVSw6HyNi1u8kOLwRmMfATWDfgnoLh5fw W1SbnrT598+tX3D3SpDH4H3OVgzCS7HmjWCYXmrwhO7dsa3bzVwR9SrB1KHZKzEVRbrt Jpml1Kx2+WnSzDfamolWGmzAsYWfx5l4AMquQ= MIME-Version: 1.0 Received: by 10.114.10.13 with SMTP id 13mr15045111waj.176.1259142920800; Wed, 25 Nov 2009 01:55:20 -0800 (PST) In-Reply-To: <659265.80398.qm@web7703.mail.in.yahoo.com> References: <659265.80398.qm@web7703.mail.in.yahoo.com> Date: Wed, 25 Nov 2009 01:55:20 -0800 Message-ID: <78568af10911250155kef67be1sa9c4dc904852ebf6@mail.gmail.com> Subject: Re: HBase High Availability From: Ryan Rawson To: hbase-user@hadoop.apache.org Content-Type: text/plain; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable X-Virus-Checked: Checked by ClamAV on apache.org With multiple masters, the election is mediated by zookeeper and the idle masters are awaiting the relection cycle. The problems with brining regions up after a failure isnt the actual speed of loading them, but bugs with the master. This is being fixed in 0.21. It will allow us to much more rapidly bring regions back online after a failure. As for loading a region across multiple servers, this would have to be thought about quite carefully to see if it is possible. Right now there is a substantial amount of state loaded that would be changed by other servers, and you would still have to reload that state anyways. We also need to ask ourselves, what does "availability" mean anyways? For example, if a regionserver failed, does that mean hbase is offline? The answer would have to be a "no", but certain sections of data might be offline temporarily. Thus HBase has 100% uptime by this definition, correct? In the annals of distributed computing, you are only protected with minimal downtime from limited hardware failures. Once you take out too many nodes, things start failing, that is a given. HBase solves the data scalability problem, it solves the limited machine failure problem. I highly suggest this presentation: http://www.cs.cornell.edu/projects/ladis2009/talks/dean-keynote-ladis2009.p= df BTW, what is your budget for "near 100% uptime" anyways? How many datacenters did you plan on using? On Wed, Nov 25, 2009 at 1:31 AM, Murali Krishna. P wrote: > Hi, > =A0 =A0This is regarding the region unavailability when a region server g= oes down. There will be cases where we have thousands of regions per RS and= it takes considerable amount of time to redistribute the regions when a no= de fails. The service will be unavailable during that period. I am evaluati= ng HBase for an application where we need to guarantee close to 100% availa= bility (namenode is still SPOF, leave that). > > =A0 =A0One simple idea would be to replicate the regions in memory. Can w= e load the same region in multiple region servers? I am not sure about the = feasibility yet, there will be issues like consistency across these in memo= ry replicas. Wanted to know whether there were any thoughts / work already = going on this area? I saw some related discussion here http://osdir.com/ml/= hbase-user-hadoop-apache/2009-09/msg00118.html, not sure what is the state. > > =A0Same needs to be done with the master as well or is it already done wi= th ZK? How fast is the master re-election and catalog load currently ? Do w= e always have multiple masters in ready to run state? > > > =A0Thanks, > Murali Krishna >