Return-Path: Delivered-To: apmail-hbase-user-archive@www.apache.org Received: (qmail 8673 invoked from network); 24 Nov 2010 19:53:07 -0000 Received: from unknown (HELO mail.apache.org) (140.211.11.3) by 140.211.11.9 with SMTP; 24 Nov 2010 19:53:07 -0000 Received: (qmail 96066 invoked by uid 500); 24 Nov 2010 19:53:38 -0000 Delivered-To: apmail-hbase-user-archive@hbase.apache.org Received: (qmail 96032 invoked by uid 500); 24 Nov 2010 19:53:38 -0000 Mailing-List: contact user-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hbase.apache.org Delivered-To: mailing list user@hbase.apache.org Received: (qmail 96008 invoked by uid 99); 24 Nov 2010 19:53:37 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 24 Nov 2010 19:53:37 +0000 X-ASF-Spam-Status: No, hits=0.0 required=10.0 tests=FREEMAIL_FROM,RCVD_IN_DNSWL_NONE,SPF_PASS,T_TO_NO_BRKTS_FREEMAIL X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of magnito@gmail.com designates 209.85.214.169 as permitted sender) Received: from [209.85.214.169] (HELO mail-iw0-f169.google.com) (209.85.214.169) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 24 Nov 2010 19:53:33 +0000 Received: by iwn36 with SMTP id 36so108298iwn.14 for ; Wed, 24 Nov 2010 11:53:13 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:mime-version:received:received:in-reply-to :references:date:message-id:subject:from:to:content-type :content-transfer-encoding; bh=y+4JayCXZs3D9YfQtsBpbY7YWcy7K+gTh66T0sB1X6Q=; b=DUEkjvQ6kn76gT+AP7H+OCtfEq4+S09pNIEm3F+vkdt1VvlFzgO2OSzWAmgb3thwfd Z2Yt7cBFiDWvA8UkwEUxtQwUncHGzbyRon81dwiS67g4LnMVOCJJLV9o1YtxEJZkB1sv 7NxZzUvmnfQfDRJ4pv7rzhhCVvpT3dtavU5no= DomainKey-Signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :content-type:content-transfer-encoding; b=whw5pFovaRhsOxSn0GVVC6Hd/9EOOMQ0ESLL++W8AUqRJsZMRnZ+dF74doHmCQNKha JKJfHXMzgMq1tHeCPjOwuKzQkVBe44qD9g7II7meKjsfZRD4pUm+p+PW9LTvddpaIav6 MQ1DP6lTTmcL7sGmMGHd0Rd1EaQkDX6njq0s0= MIME-Version: 1.0 Received: by 10.231.15.137 with SMTP id k9mr10007667iba.58.1290628392802; Wed, 24 Nov 2010 11:53:12 -0800 (PST) Received: by 10.231.48.134 with HTTP; Wed, 24 Nov 2010 11:53:12 -0800 (PST) In-Reply-To: References: <965525.19357.qm@web65510.mail.ac4.yahoo.com> Date: Wed, 24 Nov 2010 11:53:12 -0800 Message-ID: Subject: Re: question about meta data query intensity From: Jack Levin To: user@hbase.apache.org Content-Type: text/plain; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable Yes, I am game. -Jack On Wed, Nov 24, 2010 at 11:02 AM, Stack wrote: > If you are game for deploying an instrumented jar, we could log client > lookups in .META. and try and figure if it profligate. > St.Ack > > On Wed, Nov 24, 2010 at 9:25 AM, Jack Levin wrote: >> Yes, but that does not alleviate CPU contention should there be too >> many queries to a single region server. =A0 On a separate topic, is >> 'compression' in the works for REST gateway? =A0 Similar to >> mysql_client_compression? =A0We plan to drop in 500K or more queries at >> a time into the REST, and it would be interesting to see the >> performance gain against uncompressed data. >> >> Thanks. >> >> -Jack >> >> On Wed, Nov 24, 2010 at 9:04 AM, Andrew Purtell wr= ote: >>> The REST gateway (Stargate) is a long lived client. :-) >>> >>> It uses HTablePool internally so this will keep some warm table referen= ces around in addition to the region location caching that HConnectionManag= er does behind the scenes. (10 references, but this could be made configura= ble.) >>> >>> Best regards, >>> >>> =A0 =A0- Andy >>> >>> --- On Tue, 11/23/10, Jack Levin wrote: >>> >>>> From: Jack Levin >>>> Subject: Re: question about meta data query intensity >>>> To: user@hbase.apache.org >>>> Date: Tuesday, November 23, 2010, 11:06 AM >>>> its REST, and generally no long lived >>>> clients, yes, caching of regions >>>> helps however, we expect long tail hits that will be >>>> uncached, which >>>> may stress out meta region, that being said, is it possible >>>> create >>>> affinity and nail meta region into a beefy server or set of >>>> beefy >>>> servers? >>>> >>>> -Jack >>>> >>>> On Tue, Nov 23, 2010 at 10:58 AM, Jonathan Gray >>>> wrote: >>>> > Are you going to have long-lived clients? =A0How are >>>> you accessing HBase? =A0REST or Thrift gateways? =A0Caching of >>>> region locations should help significantly so that it's only >>>> a bottleneck right at the startup of the >>>> cluster/gateways/clients. >>>> > >>>> >> -----Original Message----- >>>> >> From: Jack Levin [mailto:magnito@gmail.com] >>>> >> Sent: Tuesday, November 23, 2010 10:53 AM >>>> >> To: user@hbase.apache.org >>>> >> Subject: Re: question about meta data query >>>> intensity >>>> >> >>>> >> my concern is that we plane to have 120 >>>> regionservers with 1000 >>>> >> Regions each, so the hits to meta could be quite >>>> intense. =A0(why so >>>> >> many regions? we are storing 1 Petabyte of data of >>>> images into hbase). >>>> >> >>>> >> -Jack >>>> >> >>>> >> On Tue, Nov 23, 2010 at 9:50 AM, Jonathan Gray >>>> >>>> wrote: >>>> >> > It is possible that it could be a bottleneck >>>> but usually is not. >>>> >> =A0Generally production HBase installations have >>>> long-lived clients, so >>>> >> the client-side caching is sufficient to reduce >>>> the amount of load to >>>> >> META (virtually 0 clean cluster is at steady-state >>>> / no region >>>> >> movement). >>>> >> > >>>> >> > For MapReduce, you do make new clients but >>>> generally only need to >>>> >> query for one region per task. >>>> >> > >>>> >> > It is not currently possible to split META. >>>> =A0We hard-coded some stuff >>>> >> a while back to make things easier and in the name >>>> of correctness. >>>> >> > >>>> >> > HBASE-3171 is about removing the ROOT region >>>> and putting the META >>>> >> region(s) locations into ZK directly. =A0When we >>>> make that change, we >>>> >> could probably also re-enable the splitting of >>>> META. >>>> >> > >>>> >> > JG >>>> >> > >>>> >> >> -----Original Message----- >>>> >> >> From: Jack Levin [mailto:magnito@gmail.com] >>>> >> >> Sent: Tuesday, November 23, 2010 9:31 AM >>>> >> >> To: user@hbase.apache.org >>>> >> >> Subject: question about meta data query >>>> intensity >>>> >> >> >>>> >> >> Hello, I am curious if there is a >>>> potential bottleneck in .META. >>>> >> >> ownership by a single region server. =A0Is >>>> it possible (safe) to split >>>> >> >> meta region into several? >>>> >> >> >>>> >> >> -Jack >>>> >> > >>>> > >>>> >>> >>> >>> >>> >> >