Return-Path: X-Original-To: apmail-hbase-user-archive@www.apache.org Delivered-To: apmail-hbase-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 774EF18342 for ; Thu, 14 May 2015 01:43:34 +0000 (UTC) Received: (qmail 49583 invoked by uid 500); 14 May 2015 01:43:32 -0000 Delivered-To: apmail-hbase-user-archive@hbase.apache.org Received: (qmail 49514 invoked by uid 500); 14 May 2015 01:43:32 -0000 Mailing-List: contact user-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hbase.apache.org Delivered-To: mailing list user@hbase.apache.org Received: (qmail 49499 invoked by uid 99); 14 May 2015 01:43:32 -0000 Received: from Unknown (HELO spamd4-us-west.apache.org) (209.188.14.142) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 14 May 2015 01:43:32 +0000 Received: from localhost (localhost [127.0.0.1]) by spamd4-us-west.apache.org (ASF Mail Server at spamd4-us-west.apache.org) with ESMTP id F2D7FC0B1C for ; Thu, 14 May 2015 01:43:31 +0000 (UTC) X-Virus-Scanned: Debian amavisd-new at spamd4-us-west.apache.org X-Spam-Flag: NO X-Spam-Score: 2.9 X-Spam-Level: ** X-Spam-Status: No, score=2.9 tagged_above=-999 required=6.31 tests=[DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, HTML_MESSAGE=3, SPF_PASS=-0.001, URIBL_BLOCKED=0.001] autolearn=disabled Authentication-Results: spamd4-us-west.apache.org (amavisd-new); dkim=pass (2048-bit key) header.d=gmail.com Received: from mx1-us-east.apache.org ([10.40.0.8]) by localhost (spamd4-us-west.apache.org [10.40.0.11]) (amavisd-new, port 10024) with ESMTP id 7KUfgBPJW8X7 for ; Thu, 14 May 2015 01:43:23 +0000 (UTC) Received: from mail-wg0-f44.google.com (mail-wg0-f44.google.com [74.125.82.44]) by mx1-us-east.apache.org (ASF Mail Server at mx1-us-east.apache.org) with ESMTPS id 660DB42AD1 for ; Thu, 14 May 2015 01:43:23 +0000 (UTC) Received: by wgin8 with SMTP id n8so60151597wgi.0 for ; Wed, 13 May 2015 18:43:16 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:in-reply-to:references:from:date:message-id:subject:to :content-type; bh=JNsDurw3BRQ1rwNozptxSDz8vsIgAk15u6CojxZFtIM=; b=KHVSBv774Vqz/nBh30PnAGjXHYEgDykdexweFOKFTP/r1+ltdX0Cbfpi6a2dJQzyOf UXP33GWViOxNczgaKpIKQ5nQSg5LUhr3GZDyrpqGb0r11be7uz+GJzKPP08+iF+bV+eG w/a7CdTDLNy2n8DUYzXoihiPWyGV9O619e87u3Ha71jAVjTqJkqip4FeytiShvruM7Qj QKsiUtyf+cZNEZglUQEPeVPwwvU4nQXW4/HvecQd0o3GhkSMhfh4lUlHKE4hTIGt/Rti QdfH3sGiCq5Qmt8fiw4aqCJD1B2ZzKWTg0XiastTWhA7Mf76yJZpnKH4jdL2c4w7eE/P COlw== X-Received: by 10.180.210.162 with SMTP id mv2mr19331847wic.59.1431567796344; Wed, 13 May 2015 18:43:16 -0700 (PDT) MIME-Version: 1.0 Received: by 10.194.136.242 with HTTP; Wed, 13 May 2015 18:42:55 -0700 (PDT) In-Reply-To: References: <21843.61664.956285.397408@bfym.cloudera.com> From: Matteo Bertozzi Date: Wed, 13 May 2015 18:42:55 -0700 Message-ID: Subject: Re: New post on hbase-1.1.0 throttling feature up on our Apache blog To: user@hbase.apache.org Content-Type: multipart/alternative; boundary=001a11c37daef91bb0051600d92a --001a11c37daef91bb0051600d92a Content-Type: text/plain; charset=UTF-8 @nick we have already something like that, which is HBASE-10993 and it is basically reordering requests based on how many scan.next you did. (see the picture) http://blog.cloudera.com/wp-content/uploads/2014/11/hbase-multi-f2.png the problem is that we can't eject requests in execution and we are not heavy enough on removing request from the queue and send a retry to the client in case someone with more priority is in. Matteo On Wed, May 13, 2015 at 6:38 PM, Nick Dimiduk wrote: > I guess what I'm thinking of is more about scheduling than > quota/throttling. I don't want my online requests to sit in a queue behind > MR requests while the MR work build up to it's quota amount. I want a > scheduler to do time-slicing of operations, with preferential treatment > given to online work over long scan ("analytical") work. For example, all > scan RPC's "known" to cover "lots" of Cells get de-prioritized vs gets and > short scans. Maybe this is synthesized with an RPC annotation marking it as > "long" vs "short" -- MR scans are marked "long". I'm not sure, and I need > to look more closely at recent scan improvements. IIRC, there's a heartbeat > now, which maybe is a general mechanism allowing for long operations to not > stomp over short operations. Heartbeat implies the long-running scan is > coming up for air from time to time, allowing itself to be interrupted and > defer to higher priority work. This isn't preemption, but does allow for an > upper bound on how long the next queued task waits. > > On Wed, May 13, 2015 at 6:11 PM, Matteo Bertozzi > wrote: > > > @nick what would you like to have? a match on a Job ID or something like > > that? > > currently only user/table/namespace are supported, > > but group support can be easily added. > > not sure about a job-id or job-name since we don't have that info on the > > scan. > > > > On Wed, May 13, 2015 at 6:04 PM, Nick Dimiduk > wrote: > > > > > Sorry. Yeah, sure, I can ask over there. > > > > > > The throttle was set by user in these tests. You cannot directly > > > > throttle a specific job, but do have the option to set the throttle > > > > for a table or a namespace. That might be sufficient for you to > > > > achieve your objective (unless those jobs are run by one user and > > > > access the same table.) > > > > > > > > > Maybe running as different users is the key, but this seems like a very > > > important use-case to support -- folks doing aggregate analysis > > > concurrently on an online table. > > > > > > On Wed, May 13, 2015 at 5:53 PM, Stack wrote: > > > > > > > Should we add in your comments on the blog Govind: i.e. the answers > to > > > > Nicks' questions? > > > > St.Ack > > > > > > > > On Wed, May 13, 2015 at 5:48 PM, Govind Kamat > > > wrote: > > > > > > > > > > This is a great demonstration of these new features, thanks for > > > > > pointing it > > > > > > out Stack. > > > > > > > > > > > > I'm curious: what percentile latencies are this reported? Does > the > > > > > > non-throttled user see significant latency improvements in the > 95, > > > > 99pct > > > > > > when the competing, scanning users are throttled? MB/s and req/s > > are > > > > > > managed at the region level? Region server level? Aggregate? > > > > > > > > > > The latencies reported in the post are average latencies. > > > > > > > > > > Yes, the non-throttled user sees an across-the-board improvement in > > > > > the 95th and 99th percentiles, in addition to the improvement in > > > > > average latency. The extent of improvement is significant as well > > but > > > > > varies with the throttle pressure, just as in the case of the > average > > > > > latencies. > > > > > > > > > > The total throughput numbers (req/s) are aggregate numbers reported > > by > > > > > the YCSB client. > > > > > > > > > > > These throttle points are by user? Is there a way for us to say > > "all > > > > MR > > > > > > jobs are lower priority than online queries"? > > > > > > > > > > > > > > > > The throttle was set by user in these tests. You cannot directly > > > > > throttle a specific job, but do have the option to set the throttle > > > > > for a table or a namespace. That might be sufficient for you to > > > > > achieve your objective (unless those jobs are run by one user and > > > > > access the same table.) > > > > > > > > > > Govind > > > > > > > > > > > > > > > > Thanks, > > > > > > Nick > > > > > > > > > > > > On Tue, May 12, 2015 at 1:58 PM, Stack > wrote: > > > > > > > > > > > > > .. by our Govind. > > > > > > > > > > > > > > See here: > > > > > > > > > > > > > > > > > > > > > > https://blogs.apache.org/hbase/entry/the_hbase_request_throttling_feature > > > > > > > > > > > > > > St.Ack > > > > > > > > > > > > > > > > > > > > > > --001a11c37daef91bb0051600d92a--