Return-Path: X-Original-To: apmail-hadoop-common-user-archive@www.apache.org Delivered-To: apmail-hadoop-common-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 4458E98A4 for ; Tue, 27 Nov 2012 09:51:18 +0000 (UTC) Received: (qmail 36986 invoked by uid 500); 27 Nov 2012 09:51:13 -0000 Delivered-To: apmail-hadoop-common-user-archive@hadoop.apache.org Received: (qmail 36688 invoked by uid 500); 27 Nov 2012 09:51:12 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hadoop.apache.org Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 36654 invoked by uid 99); 27 Nov 2012 09:51:11 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 27 Nov 2012 09:51:11 +0000 X-ASF-Spam-Status: No, hits=-0.1 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_MED,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of amits@infolinks.com designates 207.126.144.129 as permitted sender) Received: from [207.126.144.129] (HELO eu1sys200aog110.obsmtp.com) (207.126.144.129) by apache.org (qpsmtpd/0.29) with SMTP; Tue, 27 Nov 2012 09:51:01 +0000 Received: from mail-lb0-f198.google.com ([209.85.217.198]) (using TLSv1) by eu1sys200aob110.postini.com ([207.126.147.11]) with SMTP ID DSNKULSM8Ye+pBdm/xznqqu7oZwhAvW86nu5@postini.com; Tue, 27 Nov 2012 09:50:41 UTC Received: by mail-lb0-f198.google.com with SMTP id n1so7669683lba.5 for ; Tue, 27 Nov 2012 01:50:40 -0800 (PST) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20120113; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :content-type:x-gm-message-state; bh=vza0huFTY64UV8j7zYuUEPbJF1taML0xg0OOyvTB4Yc=; b=DqN8R3bRqNn8AZmxVX9VfBaX2JVVnn4ToZ+A/nVq0NEvNOBoq6fAmQvb/H4OPcGRf6 SOtikzRoIaipQgao8/BRtaGURQePfI4aSiroIgKz7SjA8hz+7jUfDKzpO03ExHCQF0n1 ay2XUB0wc1/ZV0Te84E9gCIMJXDiSME1KA5jN2TRzMq4n9s1SQ9i4TIv9wIYzelBVm2J xQ6iFQoI9AR4GlwPdSK/3S3/C4xQBUqy+jU4mlpPnMKw2+Io3G5Hs1Lk9IS7Z82NpPW9 8D14eEZzhRMKlSiDTWU3MEj20jhNpj3mI9yoxxxxcDVfeX6DBQPWXoJ66ebR+8ITH57W AZXA== Received: by 10.112.98.131 with SMTP id ei3mr6497630lbb.63.1354009840775; Tue, 27 Nov 2012 01:50:40 -0800 (PST) MIME-Version: 1.0 Received: by 10.112.98.131 with SMTP id ei3mr6497623lbb.63.1354009840636; Tue, 27 Nov 2012 01:50:40 -0800 (PST) Received: by 10.114.38.204 with HTTP; Tue, 27 Nov 2012 01:50:40 -0800 (PST) In-Reply-To: References: Date: Tue, 27 Nov 2012 11:50:40 +0200 Message-ID: Subject: Re: Hadoop 1.0.4 Performance Problem From: Amit Sela To: user@hadoop.apache.org Content-Type: multipart/alternative; boundary=f46d0401fb9592d2b404cf76fbb2 X-Gm-Message-State: ALoCoQkitpY//N4z9rJfCREQGuSBto18oUNLJvmzDCpvl/J2bChgXYODsZiOm2L4lxgNML7cwuGmca58MqWZLj4TsfyuQORqDmmdm+G+rPfiLCvuGK5QSq7y/IW3KcIigmcOZ3Z6IC/iNPbYDAH+L+tdbqCOPwtLbHwA16K22jK5HvdiCDdLsnA= X-Virus-Checked: Checked by ClamAV on apache.org --f46d0401fb9592d2b404cf76fbb2 Content-Type: text/plain; charset=ISO-8859-1 Hi Jon, I recently upgraded our cluster from Hadoop 0.20.3-append to Hadoop 1.0.4 and I haven't noticed any performance issues. By "multiple assignment feature" do you mean speculative execution (mapred.map.tasks.speculative.execution and mapred.reduce.tasks.speculative.execution) ? On Mon, Nov 26, 2012 at 11:49 PM, Jon Allen wrote: > Problem solved, but worth warning others about. > > Before the upgrade the reducers for the terasort process had been evenly > distributed around the cluster - one per task tracker in turn, looping > around the cluster until all tasks were allocated. After the upgrade all > reduce task had been submitted to small number of task trackers - submit > tasks until the task tracker slots were full and then move onto the next > task tracker. Skewing the reducers like this quite clearly hit the > benchmark performance. > > The reason for this turns out to be the fair scheduler rewrite > (MAPREDUCE-2981) that appears to have subtly modified the behaviour of the > assign multiple property. Previously this property caused a single map and > a single reduce task to be allocated in a task tracker heartbeat (rather > than the default of a map or a reduce). After the upgrade it allocates as > many tasks as there are available task slots. Turning off the multiple > assignment feature returned the terasort to its pre-upgrade performance. > > I can see potential benefits to this change and need to think through the > consequences to real world applications (though in practice we're likely to > move away from fair scheduler due to MAPREDUCE-4451). Investigating this > has been a pain so to warn other user is there anywhere central that can be > used to record upgrade gotchas like this? > > > On Fri, Nov 23, 2012 at 12:02 PM, Jon Allen wrote: > >> Hi, >> >> We've just upgraded our cluster from Hadoop 0.20.203 to 1.0.4 and have >> hit performance problems. Before the upgrade a 15TB terasort took about 45 >> minutes, afterwards it takes just over an hour. Looking in more detail it >> appears the shuffle phase has increased from 20 minutes to 40 minutes. >> Does anyone have any thoughts about what's changed between these releases >> that may have caused this? >> >> The only change to the system has been to Hadoop. We moved from a >> tarball install of 0.20.203 with all processes running as hadoop to an RPM >> deployment of 1.0.4 with processes running as hdfs and mapred. Nothing >> else has changed. >> >> As a related question, we're still running with a configuration that was >> tuned for version 0.20.1. Are there any recommendations for tuning >> properties that have been introduced in recent versions that are worth >> investigating? >> >> Thanks, >> Jon > > > --f46d0401fb9592d2b404cf76fbb2 Content-Type: text/html; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable
Hi Jon,=A0

I recently upgraded our clus= ter from Hadoop 0.20.3-append to Hadoop 1.0.4 and I haven't noticed any= =A0performance issues. By =A0"multiple assignment feature" do you= mean speculative execution (mapred.map.tasks.speculative.execution and=A0m= apred.reduce.tasks.speculative.execution) ?


On Mon, Nov 26, 2012 at = 11:49 PM, Jon Allen <jayayedev@gmail.com> wrote:
Problem solved, but worth warning others about.

Before t= he upgrade the reducers for the terasort process had been evenly distribute= d around the cluster - one per task tracker in turn, looping around the clu= ster until all tasks were allocated. =A0After the upgrade all reduce task h= ad been submitted to small number of task trackers - submit tasks until the= task tracker slots were full and then move onto the next task tracker. =A0= Skewing the reducers like this quite clearly hit the benchmark performance.=

The reason for this turns out to be the fair scheduler = rewrite (MAPREDUCE-2981) that appears to have subtly modified the behaviour= of the assign multiple property. Previously this property caused a single = map and a single reduce task to be allocated in a task tracker heartbeat (r= ather than the default of a map or a reduce). =A0After the upgrade it alloc= ates as many tasks as there are available task slots. =A0Turning off the mu= ltiple assignment feature returned the terasort to its pre-upgrade=A0perfor= mance.

I can see potential benefits to this change and need to= think through the consequences to real world applications (though in pract= ice we're likely to move away from fair scheduler due to=A0MAPREDUCE-4451). =A0Investigating this has been a pain so to warn other user is there anywhe= re central that can be used to record upgrade gotchas like this?


On Fri, Nov 23, 2012 at 12:02 PM, J= on Allen <jayayedev@gmail.com> wrote:
Hi,

We've just upgraded our cluster from Hadoop 0.20.203 to 1.0.4 and have = hit performance problems. =A0Before the upgrade a 15TB terasort took about = 45 minutes, afterwards it takes just over an hour. =A0Looking in more detai= l it appears the shuffle phase has increased from 20 minutes to 40 minutes.= =A0Does anyone have any thoughts about what's changed between these re= leases that may have caused this?

The only change to the system has been to Hadoop. =A0We moved from a tarbal= l install of 0.20.203 with all processes running as hadoop to an RPM deploy= ment of 1.0.4 with processes running as hdfs and mapred. =A0Nothing else ha= s changed.

As a related question, we're still running with a configuration that wa= s tuned for version 0.20.1. Are there any recommendations for tuning proper= ties that have been introduced in recent versions that are worth investigat= ing?

Thanks,
Jon


--f46d0401fb9592d2b404cf76fbb2--