Return-Path: X-Original-To: apmail-hadoop-yarn-issues-archive@minotaur.apache.org Delivered-To: apmail-hadoop-yarn-issues-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id C13E617444 for ; Tue, 9 Jun 2015 21:29:01 +0000 (UTC) Received: (qmail 94384 invoked by uid 500); 9 Jun 2015 21:29:01 -0000 Delivered-To: apmail-hadoop-yarn-issues-archive@hadoop.apache.org Received: (qmail 94335 invoked by uid 500); 9 Jun 2015 21:29:01 -0000 Mailing-List: contact yarn-issues-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: yarn-issues@hadoop.apache.org Delivered-To: mailing list yarn-issues@hadoop.apache.org Received: (qmail 94322 invoked by uid 99); 9 Jun 2015 21:29:01 -0000 Received: from arcas.apache.org (HELO arcas.apache.org) (140.211.11.28) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 09 Jun 2015 21:29:01 +0000 Date: Tue, 9 Jun 2015 21:29:01 +0000 (UTC) From: "Ashwin Shankar (JIRA)" To: yarn-issues@hadoop.apache.org Message-ID: In-Reply-To: References: Subject: [jira] [Commented] (YARN-3453) Fair Scheduler : Parts of preemption logic uses DefaultResourceCalculator even in DRF mode causing thrashing MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: quoted-printable X-JIRA-FingerPrint: 30527f35849b9dde25b450d4833f0394 [ https://issues.apache.org/jira/browse/YARN-3453?page=3Dcom.atlassian.= jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=3D14579= 588#comment-14579588 ]=20 Ashwin Shankar commented on YARN-3453: -------------------------------------- hey Arun, Thanks for working on this ! Couple more comments in addition to Karthik's comments : 1. Why are we not using componentwisemin here ? Resource target =3D Resources.min(calc, clusterResource, sched.getMinShare(), sched.getDemand()); 2. FairScheduler.preemptResources() uses DefaultResourceCalculator and henc= e would look at only memory. This could lead to a problem in the following scenario : Preemption round0 : toPreempt =3D (100G,10 cores) ... Fair Scheduler : Parts of preemption logic uses DefaultResourceCalculator= even in DRF mode causing thrashing > -------------------------------------------------------------------------= ----------------------------------- > > Key: YARN-3453 > URL: https://issues.apache.org/jira/browse/YARN-3453 > Project: Hadoop YARN > Issue Type: Bug > Components: fairscheduler > Affects Versions: 2.6.0 > Reporter: Ashwin Shankar > Assignee: Arun Suresh > Attachments: YARN-3453.1.patch, YARN-3453.2.patch > > > There are two places in preemption code flow where DefaultResourceCalcula= tor is used, even in DRF mode. > Which basically results in more resources getting preempted than needed, = and those extra preempted containers aren=E2=80=99t even getting to the =E2= =80=9Cstarved=E2=80=9D queue since scheduling logic is based on DRF's Calcu= lator. > Following are the two places : > 1. {code:title=3DFSLeafQueue.java|borderStyle=3Dsolid} > private boolean isStarved(Resource share) > {code} > A queue shouldn=E2=80=99t be marked as =E2=80=9Cstarved=E2=80=9D if the d= ominant resource usage > is >=3D fair/minshare. > 2. {code:title=3DFairScheduler.java|borderStyle=3Dsolid} > protected Resource resToPreempt(FSLeafQueue sched, long curTime) > {code} > -------------------------------------------------------------- > One more thing that I believe needs to change in DRF mode is : during a p= reemption round,if preempting a few containers results in satisfying needs = of a resource type, then we should exit that preemption round, since the co= ntainers that we just preempted should bring the dominant resource usage to= min/fair share. -- This message was sent by Atlassian JIRA (v6.3.4#6332)