Return-Path: Delivered-To: apmail-hadoop-core-user-archive@www.apache.org Received: (qmail 83166 invoked from network); 6 Dec 2008 09:22:42 -0000 Received: from hermes.apache.org (HELO mail.apache.org) (140.211.11.2) by minotaur.apache.org with SMTP; 6 Dec 2008 09:22:42 -0000 Received: (qmail 2304 invoked by uid 500); 6 Dec 2008 09:22:49 -0000 Delivered-To: apmail-hadoop-core-user-archive@hadoop.apache.org Received: (qmail 2257 invoked by uid 500); 6 Dec 2008 09:22:49 -0000 Mailing-List: contact core-user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: core-user@hadoop.apache.org Delivered-To: mailing list core-user@hadoop.apache.org Received: (qmail 2246 invoked by uid 99); 6 Dec 2008 09:22:49 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 06 Dec 2008 01:22:49 -0800 X-ASF-Spam-Status: No, hits=1.2 required=10.0 tests=SPF_NEUTRAL X-Spam-Check-By: apache.org Received-SPF: neutral (athena.apache.org: local policy) Received: from [69.147.107.20] (HELO mrout1-b.corp.re1.yahoo.com) (69.147.107.20) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 06 Dec 2008 09:21:17 +0000 Received: from [192.168.1.64] (snvvpn2-10-72-77-c13.hq.corp.yahoo.com [10.72.77.13]) by mrout1-b.corp.re1.yahoo.com (8.13.8/8.13.8/y.out) with ESMTP id mB69KIu7009349 for ; Sat, 6 Dec 2008 01:20:18 -0800 (PST) DomainKey-Signature: a=rsa-sha1; s=serpent; d=yahoo-inc.com; c=nofws; q=dns; h=message-id:from:to:in-reply-to:content-type: content-transfer-encoding:mime-version:subject:date:references:x-mailer; b=zKe7vNSwXLD5bLCV+LaAPPwroz12j1AYW9ubRRpeK0mhE/YebNCHAaK+vbnITRYH Message-Id: <615650B7-B996-43D3-BF41-5B2C52E99BD9@yahoo-inc.com> From: Arun C Murthy To: core-user@hadoop.apache.org In-Reply-To: <10927.75885.qm@web81802.mail.mud.yahoo.com> Content-Type: text/plain; charset=US-ASCII; format=flowed; delsp=yes Content-Transfer-Encoding: 7bit Mime-Version: 1.0 (Apple Message framework v929.2) Subject: Re: slow shuffle Date: Sat, 6 Dec 2008 01:20:17 -0800 References: <10927.75885.qm@web81802.mail.mud.yahoo.com> X-Mailer: Apple Mail (2.929.2) X-Virus-Checked: Checked by ClamAV on apache.org On Dec 5, 2008, at 2:43 PM, Songting Chen wrote: > To summarize the slow shuffle issue: > > 1. I think one problem is that the Reducer starts very > late in the process, slowing the entire job significantly. > > Is there a way to let reducer start earlier? > http://issues.apache.org/jira/browse/HADOOP-3136 should help you there, it's pretty close to getting in to 0.20. Arun > 2. Copying 300 files with 30K each took total 3 mins (after all map > finished). This really puzzles me what's behind the scene. (note > that sorting takes < 1 sec) > > Thanks, > -Songting > >> >> >> --- On Fri, 12/5/08, Songting Chen >> wrote: >> >>> From: Songting Chen >>> Subject: Re: slow shuffle >>> To: core-user@hadoop.apache.org >>> Date: Friday, December 5, 2008, 1:27 PM >>> We have 4 testing data nodes with 3 reduce tasks. The >>> parallel.copies parameter has been increased to 20,30, >> even >>> 50. But it doesn't really help... >>> >>> >>> --- On Fri, 12/5/08, Aaron Kimball >>> wrote: >>> >>>> From: Aaron Kimball >>>> Subject: Re: slow shuffle >>>> To: core-user@hadoop.apache.org >>>> Date: Friday, December 5, 2008, 12:28 PM >>>> How many reduce tasks do you have? Look into >>> increasing >>>> mapred.reduce.parallel.copies from the default of >> 5 to >>>> something more like >>>> 20 or 30. >>>> >>>> - Aaron >>>> >>>> On Fri, Dec 5, 2008 at 10:00 PM, Songting Chen >>>> wrote: >>>> >>>>> A little more information: >>>>> >>>>> We optimized our Map process quite a bit >> that now >>> the >>>> Shuffle becomes the >>>>> bottleneck. >>>>> >>>>> 1. There are 300 Map jobs (128M size block), >> each >>>> takes about 13 sec. >>>>> 2. The Reducer starts running at a very late >>> stage >>>> (80% maps are done) >>>>> 3. copy 300 map outputs (shuffle) takes as >> long >>> as the >>>> entire map process, >>>>> although each map output is just about >> 50Kbytes >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> --- On Fri, 12/5/08, Alex Loddengaard >>>> wrote: >>>>> >>>>>> From: Alex Loddengaard >>> >>>>>> Subject: Re: slow shuffle >>>>>> To: core-user@hadoop.apache.org >>>>>> Date: Friday, December 5, 2008, 11:43 >> AM >>>>>> These configuration options will be >> useful: >>>>>> >>>>>> >>>>>>> >>>>>> >>>> >>> >> mapred.job.shuffle.merge.percent >>>>>>> 0.66 >>>>>>> The usage >>> threshold at >>>> which an >>>>>> in-memory merge will be >>>>>>> initiated, expressed as a >> percentage >>> of >>>> the total >>>>>> memory allocated to >>>>>>> storing in-memory map outputs, >> as >>> defined >>>> by >>>>>>> >>> mapred.job.shuffle.input.buffer.percent. >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>> >>>> >>> >> mapred.job.shuffle.input.buffer.percent >>>>>>> 0.70 >>>>>>> The >> percentage of >>>> memory to be >>>>>> allocated from the maximum >>>>>>> heap >>>>>>> size to storing map outputs >> during >>> the >>>> shuffle. >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>> >>>> >>> >> mapred.job.reduce.input.buffer.percent >>>>>>> 0.0 >>>>>>> The >> percentage of >>>> memory- >>>>>> relative to the maximum heap size- >>>>>>> to >>>>>>> retain map outputs during the >> reduce. >>> When >>>> the >>>>>> shuffle is concluded, any >>>>>>> remaining map outputs in memory >> must >>>> consume less >>>>>> than this threshold >>>>>>> before >>>>>>> the reduce can begin. >>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>>> How long did the shuffle take relative >> to >>> the >>>> rest of the >>>>>> job? >>>>>> >>>>>> Alex >>>>>> >>>>>> On Fri, Dec 5, 2008 at 11:17 AM, >> Songting >>> Chen >>>>>> wrote: >>>>>> >>>>>>> We encountered a bottleneck during >> the >>>> shuffle phase. >>>>>> However, there is not >>>>>>> much data to be shuffled across >> the >>> network >>>> at all - >>>>>> total less than >>>>>>> 10MBytes (the combiner aggregated >> most >>> of >>>> the data). >>>>>>> >>>>>>> Are there any parameters or >> anything we >>> can >>>> tune to >>>>>> improve the shuffle >>>>>>> performance? >>>>>>> >>>>>>> Thanks, >>>>>>> -Songting >>>>>>> >>>>>