Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id D3452200C01 for ; Thu, 19 Jan 2017 09:41:33 +0100 (CET) Received: by cust-asf.ponee.io (Postfix) id D1E8C160B54; Thu, 19 Jan 2017 08:41:33 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id D5F9C160B42 for ; Thu, 19 Jan 2017 09:41:32 +0100 (CET) Received: (qmail 61200 invoked by uid 500); 19 Jan 2017 08:41:32 -0000 Mailing-List: contact yarn-issues-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Delivered-To: mailing list yarn-issues@hadoop.apache.org Received: (qmail 61188 invoked by uid 99); 19 Jan 2017 08:41:32 -0000 Received: from pnap-us-west-generic-nat.apache.org (HELO spamd3-us-west.apache.org) (209.188.14.142) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 19 Jan 2017 08:41:32 +0000 Received: from localhost (localhost [127.0.0.1]) by spamd3-us-west.apache.org (ASF Mail Server at spamd3-us-west.apache.org) with ESMTP id 94A2A180661 for ; Thu, 19 Jan 2017 08:41:31 +0000 (UTC) X-Virus-Scanned: Debian amavisd-new at spamd3-us-west.apache.org X-Spam-Flag: NO X-Spam-Score: -1.199 X-Spam-Level: X-Spam-Status: No, score=-1.199 tagged_above=-999 required=6.31 tests=[KAM_ASCII_DIVIDERS=0.8, KAM_LAZY_DOMAIN_SECURITY=1, RP_MATCHES_RCVD=-2.999] autolearn=disabled Received: from mx1-lw-eu.apache.org ([10.40.0.8]) by localhost (spamd3-us-west.apache.org [10.40.0.10]) (amavisd-new, port 10024) with ESMTP id Vip5v2e01hOu for ; Thu, 19 Jan 2017 08:41:29 +0000 (UTC) Received: from mailrelay1-us-west.apache.org (mailrelay1-us-west.apache.org [209.188.14.139]) by mx1-lw-eu.apache.org (ASF Mail Server at mx1-lw-eu.apache.org) with ESMTP id D0B565FB80 for ; Thu, 19 Jan 2017 08:41:28 +0000 (UTC) Received: from jira-lw-us.apache.org (unknown [207.244.88.139]) by mailrelay1-us-west.apache.org (ASF Mail Server at mailrelay1-us-west.apache.org) with ESMTP id 80EBFE03A3 for ; Thu, 19 Jan 2017 08:41:27 +0000 (UTC) Received: from jira-lw-us.apache.org (localhost [127.0.0.1]) by jira-lw-us.apache.org (ASF Mail Server at jira-lw-us.apache.org) with ESMTP id A9B0C2528D for ; Thu, 19 Jan 2017 08:41:26 +0000 (UTC) Date: Thu, 19 Jan 2017 08:41:26 +0000 (UTC) From: "Bibin A Chundatt (JIRA)" To: yarn-issues@hadoop.apache.org Message-ID: In-Reply-To: References: Subject: [jira] [Commented] (YARN-3884) RMContainerImpl transition from RESERVED to KILL apphistory status not updated MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 7bit X-JIRA-FingerPrint: 30527f35849b9dde25b450d4833f0394 archived-at: Thu, 19 Jan 2017 08:41:34 -0000 [ https://issues.apache.org/jira/browse/YARN-3884?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15829524#comment-15829524 ] Bibin A Chundatt commented on YARN-3884: ---------------------------------------- YARN-5548 we had increased the scope of jira.. What about fixing only testcase failing randomly in YARN-5895. Failing randomly for long time now. Will upload patch after checkstlye fix > RMContainerImpl transition from RESERVED to KILL apphistory status not updated > ------------------------------------------------------------------------------ > > Key: YARN-3884 > URL: https://issues.apache.org/jira/browse/YARN-3884 > Project: Hadoop YARN > Issue Type: Bug > Components: resourcemanager > Environment: Suse11 Sp3 > Reporter: Bibin A Chundatt > Assignee: Bibin A Chundatt > Labels: oct16-easy > Attachments: 0001-YARN-3884.patch, Apphistory Container Status.jpg, Elapsed Time.jpg, Test Result-Container status.jpg, YARN-3884.0002.patch, YARN-3884.0003.patch, YARN-3884.0004.patch, YARN-3884.0005.patch, YARN-3884.0006.patch > > > Setup > =============== > 1 NM 3072 16 cores each > Steps to reproduce > =============== > 1.Submit apps to Queue 1 with 512 mb 1 core > 2.Submit apps to Queue 2 with 512 mb and 5 core > lots of containers get reserved and unreserved in this case > {code} > 2015-07-02 20:45:31,169 INFO org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerImpl: container_e24_1435849994778_0002_01_000013 Container Transitioned from NEW to RESERVED > 2015-07-02 20:45:31,170 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.LeafQueue: Reserved container application=application_1435849994778_0002 resource= queue=QueueA: capacity=0.4, absoluteCapacity=0.4, usedResources=, usedCapacity=1.6410257, absoluteUsedCapacity=0.65625, numApps=1, numContainers=5 usedCapacity=1.6410257 absoluteUsedCapacity=0.65625 used= cluster= > 2015-07-02 20:45:31,170 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.ParentQueue: Re-sorting assigned queue: root.QueueA stats: QueueA: capacity=0.4, absoluteCapacity=0.4, usedResources=, usedCapacity=2.0317461, absoluteUsedCapacity=0.8125, numApps=1, numContainers=6 > 2015-07-02 20:45:31,170 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.ParentQueue: assignedContainer queue=root usedCapacity=0.96875 absoluteUsedCapacity=0.96875 used= cluster= > 2015-07-02 20:45:31,191 INFO org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerImpl: container_e24_1435849994778_0001_01_000014 Container Transitioned from NEW to ALLOCATED > 2015-07-02 20:45:31,191 INFO org.apache.hadoop.yarn.server.resourcemanager.RMAuditLogger: USER=dsperf OPERATION=AM Allocated Container TARGET=SchedulerApp RESULT=SUCCESS APPID=application_1435849994778_0001 CONTAINERID=container_e24_1435849994778_0001_01_000014 > 2015-07-02 20:45:31,191 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerNode: Assigned container container_e24_1435849994778_0001_01_000014 of capacity on host host-10-19-92-117:64318, which has 6 containers, used and available after allocation > 2015-07-02 20:45:31,191 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.LeafQueue: assignedContainer application attempt=appattempt_1435849994778_0001_000001 container=Container: [ContainerId: container_e24_1435849994778_0001_01_000014, NodeId: host-10-19-92-117:64318, NodeHttpAddress: host-10-19-92-117:65321, Resource: , Priority: 20, Token: null, ] queue=default: capacity=0.2, absoluteCapacity=0.2, usedResources=, usedCapacity=2.0846906, absoluteUsedCapacity=0.41666666, numApps=1, numContainers=5 clusterResource= > 2015-07-02 20:45:31,191 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.ParentQueue: Re-sorting assigned queue: root.default stats: default: capacity=0.2, absoluteCapacity=0.2, usedResources=, usedCapacity=2.5016286, absoluteUsedCapacity=0.5, numApps=1, numContainers=6 > 2015-07-02 20:45:31,191 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.ParentQueue: assignedContainer queue=root usedCapacity=1.0 absoluteUsedCapacity=1.0 used= cluster= > 2015-07-02 20:45:32,143 INFO org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerImpl: container_e24_1435849994778_0001_01_000014 Container Transitioned from ALLOCATED to ACQUIRED > 2015-07-02 20:45:32,174 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler: Trying to fulfill reservation for application application_1435849994778_0002 on node: host-10-19-92-143:64318 > 2015-07-02 20:45:32,174 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.LeafQueue: Reserved container application=application_1435849994778_0002 resource= queue=QueueA: capacity=0.4, absoluteCapacity=0.4, usedResources=, usedCapacity=2.0317461, absoluteUsedCapacity=0.8125, numApps=1, numContainers=6 usedCapacity=2.0317461 absoluteUsedCapacity=0.8125 used= cluster= > 2015-07-02 20:45:32,174 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler: Skipping scheduling since node host-10-19-92-143:64318 is reserved by application appattempt_1435849994778_0002_000001 > 2015-07-02 20:45:32,213 INFO org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerImpl: container_e24_1435849994778_0001_01_000014 Container Transitioned from ACQUIRED to RUNNING > 2015-07-02 20:45:32,213 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler: Null container completed... > 2015-07-02 20:45:33,178 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler: Trying to fulfill reservation for application application_1435849994778_0002 on node: host-10-19-92-143:64318 > 2015-07-02 20:45:33,178 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.LeafQueue: Reserved container application=application_1435849994778_0002 resource= queue=QueueA: capacity=0.4, absoluteCapacity=0.4, usedResources=, usedCapacity=2.0317461, absoluteUsedCapacity=0.8125, numApps=1, numContainers=6 usedCapacity=2.0317461 absoluteUsedCapacity=0.8125 used= cluster= > 2015-07-02 20:45:33,178 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler: Skipping scheduling since node host-10-19-92-143:64318 is reserved by application appattempt_1435849994778_0002_000001 > 2015-07-02 20:45:33,704 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.common.fica.FiCaSchedulerApp: Application application_1435849994778_0002 unreserved on node host: host-10-19-92-143:64318 #containers=5 available= used=, currently has 0 at priority 20; currentReservation > 2015-07-02 20:45:33,704 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.LeafQueue: QueueA used= numContainers=5 user=dsperf user-resources= > 2015-07-02 20:45:33,710 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.LeafQueue: completedContainer container=Container: [ContainerId: container_e24_1435849994778_0002_01_000013, NodeId: host-10-19-92-143:64318, NodeHttpAddress: host-10-19-92-143:65321, Resource: , Priority: 20, Token: null, ] queue=QueueA: capacity=0.4, absoluteCapacity=0.4, usedResources=, usedCapacity=1.6410257, absoluteUsedCapacity=0.65625, numApps=1, numContainers=5 cluster= > 2015-07-02 20:45:33,710 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.ParentQueue: completedContainer queue=root usedCapacity=0.9166667 absoluteUsedCapacity=0.9166667 used= cluster= > 2015-07-02 20:45:33,711 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.ParentQueue: Re-sorting completed queue: root.QueueA stats: QueueA: capacity=0.4, absoluteCapacity=0.4, usedResources=, usedCapacity=1.6410257, absoluteUsedCapacity=0.65625, numApps=1, numContainers=5 > 2015-07-02 20:45:33,711 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler: Application attempt appattempt_1435849994778_0002_000001 released container container_e24_1435849994778_0002_01_000013 on node: host: host-10-19-92-143:64318 #containers=5 available= used= with event: KILL > {code} > *Impact:* > In application history server the status get updated to -1000 (INVALID) > but the end time not updated so Elapsed Time always changes. > Please check the snapshot attached -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: yarn-issues-unsubscribe@hadoop.apache.org For additional commands, e-mail: yarn-issues-help@hadoop.apache.org