Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id 96EA7200CD6 for ; Mon, 31 Jul 2017 22:21:42 +0200 (CEST) Received: by cust-asf.ponee.io (Postfix) id 95936161A14; Mon, 31 Jul 2017 20:21:42 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id B353B161A13 for ; Mon, 31 Jul 2017 22:21:41 +0200 (CEST) Received: (qmail 9929 invoked by uid 500); 31 Jul 2017 20:21:39 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 9918 invoked by uid 99); 31 Jul 2017 20:21:39 -0000 Received: from pnap-us-west-generic-nat.apache.org (HELO spamd4-us-west.apache.org) (209.188.14.142) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 31 Jul 2017 20:21:39 +0000 Received: from localhost (localhost [127.0.0.1]) by spamd4-us-west.apache.org (ASF Mail Server at spamd4-us-west.apache.org) with ESMTP id 106F9C028C for ; Mon, 31 Jul 2017 20:21:39 +0000 (UTC) X-Virus-Scanned: Debian amavisd-new at spamd4-us-west.apache.org X-Spam-Flag: NO X-Spam-Score: 2.379 X-Spam-Level: ** X-Spam-Status: No, score=2.379 tagged_above=-999 required=6.31 tests=[DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, HTML_MESSAGE=2, RCVD_IN_DNSWL_NONE=-0.0001, RCVD_IN_MSPIKE_H3=-0.01, RCVD_IN_MSPIKE_WL=-0.01, RCVD_IN_SORBS_SPAM=0.5, SPF_PASS=-0.001] autolearn=disabled Authentication-Results: spamd4-us-west.apache.org (amavisd-new); dkim=pass (2048-bit key) header.d=gmail.com Received: from mx1-lw-us.apache.org ([10.40.0.8]) by localhost (spamd4-us-west.apache.org [10.40.0.11]) (amavisd-new, port 10024) with ESMTP id RUHkDKX_41KF for ; Mon, 31 Jul 2017 20:21:37 +0000 (UTC) Received: from mail-it0-f51.google.com (mail-it0-f51.google.com [209.85.214.51]) by mx1-lw-us.apache.org (ASF Mail Server at mx1-lw-us.apache.org) with ESMTPS id 8F44D5F522 for ; Mon, 31 Jul 2017 20:21:37 +0000 (UTC) Received: by mail-it0-f51.google.com with SMTP id 77so637385itj.1 for ; Mon, 31 Jul 2017 13:21:37 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=mime-version:in-reply-to:references:from:date:message-id:subject:to :cc; bh=PKLliqdBpYvRdQ+GO54AQvdFK59noa3QLvceJwD7UrE=; b=SydkSM+vpEVZc+Ab1k7XR530BGd+EUORNkolPzKrzZ4elfxg5hM+aoShA/3N+SdJYN e1LGY69Jd4jxG/tpa8KgKGGrswaImYNAi8tF8de+Uds8KsUmqmVaYvKvIowodIx6GcQo eisoDJ17vWD6GSigzZ5+CxjrUqWsanqvUOyz4DUN/03ZB1FRYUFqwd2eu+2ZxTBZ0vhX TY//sS3eFZV5WIJjU6mMBcl0hF+rlibVEdW79PI71p1Nicc1G00d4eG8s0T52uDbDyjm 4+WK/BrTtscb7cmgo9Ldv9kUXxbnjo1bgguR3iSADMNs/ce+Abe0d3tkNYtwJowsGUp0 n/oA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:in-reply-to:references:from:date :message-id:subject:to:cc; bh=PKLliqdBpYvRdQ+GO54AQvdFK59noa3QLvceJwD7UrE=; b=GqAYIfC3ek+lDfwSt8aSyozrPJBSecWcARO9QwuFzlJoB3uBjIM/7UyfUaZOf2fdB2 ogSbDM/LR/Gm9xzOAgwbOlCkOCm7fbyK3FVKPboBc7QtZj3fbG/XjbfBXcrLqodIpLmB sHU3Vo1G8Z8aF9jUZgFnNx45iOl78X0pBQGXsBNdoJ3HfIBCEjv5anz1NBPPSToeZpxF oyhj/n04Arpx0xRDuKNW57Gevh77UfeaBYuXkhQAldPMqp1TYd5133lsuEwMxWkEkcFE rSBjuPkXy0TmJ9M8cPtFNJIL3DwIROFC2mHZ+tQVwlrp6a5K+6R+KcR9o7Mrrana8n+/ sYqQ== X-Gm-Message-State: AIVw113DP59Q6xEsoF0519A+R9M+9w1YnFgPS0nJKeTC332IVEJh1Uhu q8b3NNOnR/pY0uMW2i3V4g5j+2lnVQ== X-Received: by 10.36.48.65 with SMTP id q62mr19348067itq.112.1501532496976; Mon, 31 Jul 2017 13:21:36 -0700 (PDT) MIME-Version: 1.0 Received: by 10.2.21.141 with HTTP; Mon, 31 Jul 2017 13:21:36 -0700 (PDT) In-Reply-To: References: From: Ravi Prakash Date: Mon, 31 Jul 2017 13:21:36 -0700 Message-ID: Subject: Re: Shuffle buffer size in presence of small partitions To: Robert Schmidtke Cc: "user@hadoop.apache.org user@hadoop.apache.org" Content-Type: multipart/alternative; boundary="001a11c0152419df760555a2c7cc" archived-at: Mon, 31 Jul 2017 20:21:42 -0000 --001a11c0152419df760555a2c7cc Content-Type: text/plain; charset="UTF-8" Hi Robert! I'm sorry I do not have a Windows box and probably don't understand the shuffle process well enough. Could you please create a JIRA in the mapreduce proect if you would like this fixed upstream? https://issues.apache.org/jira/secure/RapidBoard.jspa?rapidView=116&projectKey=MAPREDUCE Thanks Ravi On Mon, Jul 31, 2017 at 6:36 AM, Robert Schmidtke wrote: > Hi all, > > I just ran into an issue, which likely resulted from my not very > intelligent configuration, but nonetheless I'd like to share this with the > community. This is all on Hadoop 2.7.3. > > In my setup, each reducer roughly fetched 65K from each mapper's spill > file. I disabled transferTo during shuffle, because I wanted to have a look > at the file system statistics, which miss mmap calls, which is what > transferTo sometimes defaults to. I left the shuffle buffer size at 128K > (not knowing about the parameter at the time). This had the effect that I > observed roughly 100% more data being read during shuffle, since 128K were > read for each 65K needed. > > I added a quick fix to Hadoop which chooses the minimum of the partition > size and the shuffle buffer size: https://github.com/ > apache/hadoop/compare/branch-2.7.3...robert-schmidtke: > adaptive-shuffle-buffer > Benchmarking this version against transferTo.allowed=true yields the same > runtime and roughly 10% more reads in YARN during the shuffle phase > (compared to previous 100%). > Maybe this is something that should be added to Hadoop? Or do users have > to be more clever about their job configurations? I'd be happy to open a PR > if this is deemed useful. > > Anyway, thanks for the attention! > > Cheers > Robert > > -- > My GPG Key ID: 336E2680 > --001a11c0152419df760555a2c7cc Content-Type: text/html; charset="UTF-8" Content-Transfer-Encoding: quoted-printable
Hi Robert!

I'm sorry I do = not have a Windows box and probably don't understand the shuffle proces= s well enough. Could you please create a JIRA in the mapreduce proect if yo= u would like this fixed upstream?
htt= ps://issues.apache.org/jira/secure/RapidBoard.jspa?rapidView=3D116&proj= ectKey=3DMAPREDUCE

Thanks
Ravi

On Mon, Jul 31, 2017 at 6:= 36 AM, Robert Schmidtke <ro.schmidtke@gmail.com> wrote:=
Hi all,

<= div>I just ran into an issue, which likely resulted from my not very intell= igent configuration, but nonetheless I'd like to share this with the co= mmunity. This is all on Hadoop 2.7.3.

In my setup,= each reducer roughly fetched 65K from each mapper's spill file. I disa= bled transferTo during shuffle, because I wanted to have a look at the file= system statistics, which miss mmap calls, which is what transferTo sometim= es defaults to. I left the shuffle buffer size at 128K (not knowing about t= he parameter at the time). This had the effect that I observed roughly 100%= more data being read during shuffle, since 128K were read for each 65K nee= ded.

I added a quick fix to Hadoop which chooses t= he minimum of the partition size and the shuffle buffer size:=C2=A0https://github.com/apache= /hadoop/compare/branch-2.7.3...robert-schmidtke:adaptive-shuffle-= buffer
Benchmarking this version against transferTo.allowed= =3Dtrue yields the same runtime and roughly 10% more reads in YARN during t= he shuffle phase (compared to previous 100%).
Maybe this is somet= hing that should be added to Hadoop? Or do users have to be more clever abo= ut their job configurations? I'd be happy to open a PR if this is deeme= d useful.

Anyway, thanks for the attention!
<= div>
Cheers
Robert

--
My GPG Key ID: 336E2680
=

--001a11c0152419df760555a2c7cc--