Return-Path: X-Original-To: apmail-hadoop-yarn-issues-archive@minotaur.apache.org Delivered-To: apmail-hadoop-yarn-issues-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 85561185F7 for ; Sat, 30 May 2015 18:18:18 +0000 (UTC) Received: (qmail 68941 invoked by uid 500); 30 May 2015 18:18:18 -0000 Delivered-To: apmail-hadoop-yarn-issues-archive@hadoop.apache.org Received: (qmail 68897 invoked by uid 500); 30 May 2015 18:18:18 -0000 Mailing-List: contact yarn-issues-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: yarn-issues@hadoop.apache.org Delivered-To: mailing list yarn-issues@hadoop.apache.org Received: (qmail 68886 invoked by uid 99); 30 May 2015 18:18:18 -0000 Received: from arcas.apache.org (HELO arcas.apache.org) (140.211.11.28) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 30 May 2015 18:18:18 +0000 Date: Sat, 30 May 2015 18:18:18 +0000 (UTC) From: "Naganarasimha G R (JIRA)" To: yarn-issues@hadoop.apache.org Message-ID: In-Reply-To: References: Subject: [jira] [Commented] (YARN-3730) scheduler reserve more resource than required MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 7bit X-JIRA-FingerPrint: 30527f35849b9dde25b450d4833f0394 [ https://issues.apache.org/jira/browse/YARN-3730?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14566130#comment-14566130 ] Naganarasimha G R commented on YARN-3730: ----------------------------------------- hi [~gu chi], Which version did you find this problem ? If its below 2.6.0, please test with the latest as there have been some improvements wrt to reservation in YARN-1769. If its with 2.6.0 and above version, share some RM logs with debug enabled, so that we can do further analysis. > scheduler reserve more resource than required > --------------------------------------------- > > Key: YARN-3730 > URL: https://issues.apache.org/jira/browse/YARN-3730 > Project: Hadoop YARN > Issue Type: Bug > Components: scheduler > Reporter: gu-chi > > Using capacity scheduler, environment is 3 NM and each has 9 vcores, I ran a spark task with 4 executors and each executor 5 cores, as suspected, only 1 executor not able to start and will be reserved, but actually more containers are reserved. This way, I can not run some other smaller tasks. As I checked the capacity scheduler, the 'needContainers' method in LeafQueue.java has a computation of 'starvation', this cause the scenario of more container reserved than required, any idea or suggestion on this? -- This message was sent by Atlassian JIRA (v6.3.4#6332)