Return-Path: X-Original-To: apmail-cloudstack-issues-archive@www.apache.org Delivered-To: apmail-cloudstack-issues-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 64A6FCE56 for ; Tue, 12 Aug 2014 11:07:12 +0000 (UTC) Received: (qmail 14858 invoked by uid 500); 12 Aug 2014 11:07:12 -0000 Delivered-To: apmail-cloudstack-issues-archive@cloudstack.apache.org Received: (qmail 14816 invoked by uid 500); 12 Aug 2014 11:07:12 -0000 Mailing-List: contact issues-help@cloudstack.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@cloudstack.apache.org Delivered-To: mailing list issues@cloudstack.apache.org Received: (qmail 14610 invoked by uid 500); 12 Aug 2014 11:07:12 -0000 Delivered-To: apmail-incubator-cloudstack-issues@incubator.apache.org Received: (qmail 14572 invoked by uid 99); 12 Aug 2014 11:07:12 -0000 Received: from arcas.apache.org (HELO arcas.apache.org) (140.211.11.28) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 12 Aug 2014 11:07:12 +0000 Date: Tue, 12 Aug 2014 11:07:12 +0000 (UTC) From: "Joris van Lieshout (JIRA)" To: cloudstack-issues@incubator.apache.org Message-ID: In-Reply-To: References: Subject: [jira] [Updated] (CLOUDSTACK-7319) Copy Snapshot command too heavy on XenServer Dom0 resources when using dd to copy incremental snapshots MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 7bit X-JIRA-FingerPrint: 30527f35849b9dde25b450d4833f0394 [ https://issues.apache.org/jira/browse/CLOUDSTACK-7319?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Joris van Lieshout updated CLOUDSTACK-7319: ------------------------------------------- Description: We noticed that the dd process was way to agressive on Dom0 causing all kinds of problems on a xenserver with medium workloads. ACS uses the dd command to copy incremental snapshots to secondary storage. This process is to heavy on Dom0 resources and even impacts DomU performance, and can even lead to domain freezes (including Dom0) of more then a minute. We've found that this is because the Dom0 kernel caches the read and write operations of dd. Some of the issues we have seen as a consequence of this are: - DomU performance/freezes - OVS freeze and not forwarding any traffic - Including LACPDUs resulting in the bond going down - keepalived heartbeat packets between RRVMs not being send/received resulting in flapping RRVM master state - Braking snapshot copy processes - the xenserver heartbeat script reaching it's timeout and fencing the server - poolmaster connection loss - ACS marking the host as down and fencing the instances even though they are still running on the origional host resulting in the same instance running on to hosts in one cluster - vhd corruption are a result of some of the issues mentioned above We've developed a patch on the xenserver scripts /etc/xapi.d/plugins/vmopsSnapshot that added the direct flag of both input and output files (iflag=direct oflag=direct). Our test have shown that Dom0 load during snapshot copy is way lower. was: We noticed that the dd process was way to agressive on Dom0 causing all kinds of problems on a xenserver with medium workloads. ACS uses the dd command to copy incremental snapshots to secondary storage. This process is to heavy on Dom0 resources and even impacts DomU performance, and can even lead to domain freezes (including Dom0) of more then a minute. We've found that this is because the Dom0 kernel caches the read and write operations of dd. We've developed a patch on the xenserver scripts /etc/xapi.d/plugins/vmopsSnapshot that added the direct flag of both input and output files. Our test have shown that Dom0 load during snapshot copy is way lower. I will upload the patch on review. > Copy Snapshot command too heavy on XenServer Dom0 resources when using dd to copy incremental snapshots > ------------------------------------------------------------------------------------------------------- > > Key: CLOUDSTACK-7319 > URL: https://issues.apache.org/jira/browse/CLOUDSTACK-7319 > Project: CloudStack > Issue Type: Bug > Security Level: Public(Anyone can view this level - this is the default.) > Components: Snapshot, XenServer > Affects Versions: 4.0.0, 4.0.1, 4.0.2, 4.1.0, 4.1.1, 4.2.0, Future, 4.2.1, 4.3.0, 4.4.0, 4.5.0, 4.3.1, 4.4.1 > Reporter: Joris van Lieshout > Priority: Critical > > We noticed that the dd process was way to agressive on Dom0 causing all kinds of problems on a xenserver with medium workloads. > ACS uses the dd command to copy incremental snapshots to secondary storage. This process is to heavy on Dom0 resources and even impacts DomU performance, and can even lead to domain freezes (including Dom0) of more then a minute. We've found that this is because the Dom0 kernel caches the read and write operations of dd. > Some of the issues we have seen as a consequence of this are: > - DomU performance/freezes > - OVS freeze and not forwarding any traffic > - Including LACPDUs resulting in the bond going down > - keepalived heartbeat packets between RRVMs not being send/received resulting in flapping RRVM master state > - Braking snapshot copy processes > - the xenserver heartbeat script reaching it's timeout and fencing the server > - poolmaster connection loss > - ACS marking the host as down and fencing the instances even though they are still running on the origional host resulting in the same instance running on to hosts in one cluster > - vhd corruption are a result of some of the issues mentioned above > We've developed a patch on the xenserver scripts /etc/xapi.d/plugins/vmopsSnapshot that added the direct flag of both input and output files (iflag=direct oflag=direct). > Our test have shown that Dom0 load during snapshot copy is way lower. -- This message was sent by Atlassian JIRA (v6.2#6252)