Return-Path: X-Original-To: apmail-hadoop-common-user-archive@www.apache.org Delivered-To: apmail-hadoop-common-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 0F621180BD for ; Tue, 23 Feb 2016 05:29:13 +0000 (UTC) Received: (qmail 87583 invoked by uid 500); 23 Feb 2016 05:29:08 -0000 Delivered-To: apmail-hadoop-common-user-archive@hadoop.apache.org Received: (qmail 86608 invoked by uid 500); 23 Feb 2016 05:29:08 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 86585 invoked by uid 99); 23 Feb 2016 05:29:08 -0000 Received: from pnap-us-west-generic-nat.apache.org (HELO spamd3-us-west.apache.org) (209.188.14.142) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 23 Feb 2016 05:29:08 +0000 Received: from localhost (localhost [127.0.0.1]) by spamd3-us-west.apache.org (ASF Mail Server at spamd3-us-west.apache.org) with ESMTP id B227E1806BB; Tue, 23 Feb 2016 05:29:07 +0000 (UTC) X-Virus-Scanned: Debian amavisd-new at spamd3-us-west.apache.org X-Spam-Flag: NO X-Spam-Score: 1.879 X-Spam-Level: * X-Spam-Status: No, score=1.879 tagged_above=-999 required=6.31 tests=[DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, HTML_MESSAGE=2, RCVD_IN_DNSWL_NONE=-0.0001, RCVD_IN_MSPIKE_H3=-0.01, RCVD_IN_MSPIKE_WL=-0.01, SPF_PASS=-0.001] autolearn=disabled Authentication-Results: spamd3-us-west.apache.org (amavisd-new); dkim=pass (2048-bit key) header.d=gmail.com Received: from mx2-lw-eu.apache.org ([10.40.0.8]) by localhost (spamd3-us-west.apache.org [10.40.0.10]) (amavisd-new, port 10024) with ESMTP id m13Voo4jK7fH; Tue, 23 Feb 2016 05:29:06 +0000 (UTC) Received: from mail-yw0-f175.google.com (mail-yw0-f175.google.com [209.85.161.175]) by mx2-lw-eu.apache.org (ASF Mail Server at mx2-lw-eu.apache.org) with ESMTPS id 99BE05FB10; Tue, 23 Feb 2016 05:29:05 +0000 (UTC) Received: by mail-yw0-f175.google.com with SMTP id e63so137918118ywc.3; Mon, 22 Feb 2016 21:29:05 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :cc:content-type; bh=NR0YsuF/kWrGKEtTlG44bQ9Tu6aslbSH10/knJ/A9Nk=; b=YR51xhW1ol2XsVWgKzBUvWWzvWc52l5OID02RBplRlLhzDbaan5kgOpZmAliRINXlk R2lCnbiavHS7NsO0ySX3ncZMyu/VmB9tuifuOGKO1ePtIIG/Gr8UAR814pJp9/4diAaS LXZP0NSTZzw+gRlHMedReeHwFHZL9uOKSHtaIwHLXfhms896ytjhWY4YMa4GY2S1miRj +25hkPuxgqo56ZUscLnq0Gs8ry77RC+ocR2wgiQ+EjHzFbkaVkVu6uyKscd6+B78Qy4C Fc89DSiuc4voD52XTyQRMsexqaNHkhwYrQ93T9vQCNKMxbzOWrGK3FE4Qk0MkVhkKaiQ rWEg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:mime-version:in-reply-to:references:date :message-id:subject:from:to:cc:content-type; bh=NR0YsuF/kWrGKEtTlG44bQ9Tu6aslbSH10/knJ/A9Nk=; b=Lu1oz5Q9By/QXxLCxnTekAtVsY/k7BXVfWNhJjMkQTOTd5pwqgDTxsb6yFZ1zqMCI8 2ICC2yOjsc80PokOIF407w0VmUzdXdGI0STI7xW+OZ9cuSjDLdv6o/+DLXiono/FYGwW nwL8WfHpQSjkcb9/ANW8OsdGh9YdJxJVzDd5QKFGCy8u9lb+a29NLHn4k1jiUc2Bo6oC GEr8o3byu/74KwVCxwS63XMui6GVM2RVgCyxfmBtAX2ff9ymeJeHxMpKRF1D9UfZFIUT kLivKlMEyUwXDF8LvpO/0d+LouxMfNb4/91yjgsvBd02stE1+geOUs3Hzxld0tu1az5G +0KQ== X-Gm-Message-State: AG10YOQmzq8qSrTy62JFLe9nvECDAMwNLGpOWqOnTYpSJg1Q7sGlC3sV1irC7TZkgQCM+/yZ38NoN2WKxnxtVw== MIME-Version: 1.0 X-Received: by 10.129.97.195 with SMTP id v186mr15547949ywb.269.1456205339453; Mon, 22 Feb 2016 21:28:59 -0800 (PST) Received: by 10.37.59.10 with HTTP; Mon, 22 Feb 2016 21:28:59 -0800 (PST) In-Reply-To: References: Date: Tue, 23 Feb 2016 10:58:59 +0530 Message-ID: Subject: Re: YARN Fair Scheduler From: Prabhu Joseph To: yarn-dev@hadoop.apache.org Cc: "user@hadoop.apache.org" Content-Type: multipart/alternative; boundary=001a11490122fa72af052c69390a --001a11490122fa72af052c69390a Content-Type: text/plain; charset=UTF-8 Hi Karthik, Yes all the queues are always active (atleast one job is running at a time) and thus the fair share of all queue is very less. How to design the fair scheduler for this kind of case. Do you have some Best Practices to design the fair-scheduler.xml. Weights - is the correct way to make critical queues get a bigger share. How Nesting of queues helps. And few more doubts 1. How to configure minResources of a queue, is the sum of minResources of all queue should be equal to Total YARN Cluster Resource. 2. What we need to consider when configuring YARN queue for Spark Jobs Thanks, Prabhu Joseph On Tue, Feb 23, 2016 at 10:35 AM, Karthik Kambatla wrote: > Hey Prabhu > > Are all the 250 queues always active? If not, the actual (instantaneous) > fairshare used by the scheduler only considers the active queues (i.e., > those that have running applications). Otherwise, you can tune your queues > (weights, nesting etc.) so the critical queues get a bigger share. > > Hope that helps. > > On Mon, Feb 22, 2016 at 5:07 PM, Prabhu Joseph > > wrote: > > > Hi All, > > > > When YARN Fair Scheduler is configured with a parent root and 250 > child > > queues for a big Cluster having total resource of 10TB and 3000 Cores. > The > > fair share of a child queue is very less. Fair Share is Total Cluster > > resource / total number of child queues. How to design a Fair Scheduler > > with many like 250 number of queues in such a way, each queue gets more > > fair share. > > > > Is having Nested Queues or configuring weight or any other way to design. > > > > Thanks, > > Prabhu Joseph > > > --001a11490122fa72af052c69390a Content-Type: text/html; charset=UTF-8 Content-Transfer-Encoding: quoted-printable
Hi Karthik,

=C2=A0=C2=A0 Yes all th= e queues are always active (atleast one job is running at a time) and thus = the fair share of all queue is very less. How to design the fair scheduler = for this kind of case. Do you have some Best Practices to design the fair-s= cheduler.xml.

Weights - is the correct way to make critic= al queues get a bigger share. How Nesting of queues helps. And few more dou= bts

1. How to configure minResources of a queue, is the s= um of minResources of all queue should be equal to Total YARN Cluster Resou= rce.
2. What we need to consider when configuring YARN queue = for Spark Jobs

Thanks,
Prabhu Jo= seph




<= div class=3D"gmail_quote">On Tue, Feb 23, 2016 at 10:35 AM, Karthik Kambatl= a <kasha@cloudera.com> wrote:
Hey Prabhu

Are all the 250 queues always active? If not, the actual (instantaneous) fairshare used by the scheduler only considers the active queues (i.e.,
those that have running applications). Otherwise, you can tune your queues<= br> (weights, nesting etc.) so the critical queues get a bigger share.

Hope that helps.

On Mon, Feb 22, 2016 at 5:07 PM, Prabhu Joseph <prabhujose.gates@gmail.com>
wrote:

> Hi All,
>
>=C2=A0 =C2=A0 When YARN Fair Scheduler is configured with a parent root= and 250 child
> queues for a big Cluster having total resource of 10TB and 3000 Cores.= The
> fair share of a child queue is very less. Fair Share is Total Cluster<= br> > resource / total number of child queues. How to design a Fair Schedule= r
> with many like 250 number of queues in such a way, each queue gets mor= e
> fair share.
>
> Is having Nested Queues or configuring weight or any other way to desi= gn.
>
> Thanks,
> Prabhu Joseph
>

--001a11490122fa72af052c69390a--