Return-Path: X-Original-To: apmail-hadoop-mapreduce-user-archive@minotaur.apache.org Delivered-To: apmail-hadoop-mapreduce-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id B204D10F6B for ; Sat, 1 Feb 2014 00:47:15 +0000 (UTC) Received: (qmail 27056 invoked by uid 500); 1 Feb 2014 00:47:07 -0000 Delivered-To: apmail-hadoop-mapreduce-user-archive@hadoop.apache.org Received: (qmail 26739 invoked by uid 500); 1 Feb 2014 00:47:07 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hadoop.apache.org Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 26732 invoked by uid 99); 1 Feb 2014 00:47:06 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 01 Feb 2014 00:47:06 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of harsh@cloudera.com designates 209.85.223.172 as permitted sender) Received: from [209.85.223.172] (HELO mail-ie0-f172.google.com) (209.85.223.172) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 01 Feb 2014 00:47:00 +0000 Received: by mail-ie0-f172.google.com with SMTP id e14so5014577iej.17 for ; Fri, 31 Jan 2014 16:46:40 -0800 (PST) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:mime-version:in-reply-to:references:date :message-id:subject:from:to:content-type; bh=ppOht9QbfA+6uXy03gEQT8GGcy+UEOmzvk3KGM5Yh30=; b=AdO0uZVFUE1jnxTQU1YTnR0vkplRHXcajU+EVB+BpnlumI9GSCKcHNsaPamRl6behd 9MvDPGmRRi1Y7pFMJHa4f1Gf9ACxQA/y4BW+izBeH1NmpLQXqy6tSu9MgtpTFgjYVf51 crpy2ilo1PDZIDBYlU4OOGY+PsyuD2MCssi3Q6YWty8GvPRFZbzB6kbWvpcrkcHh9vnC ccY4FQZ1+ZUGFcxjysPVKOcjBjrmLti2djq6ml7HYjJK6KspPo2uq+C3D97SzIT2aGxq bx87rtsSjbQODOpK8+1os4HxwcxnRVD+ulOzRZZwwS7zbj2anReUGxB/zv3Rs0Z0oWlt Ec/Q== X-Gm-Message-State: ALoCoQmGwCPgETQZGLcl4m+nHzANh/A1qRBTZ35inDoJRSzgCBITS9lbSJeJxft7brHePMiMPtUm MIME-Version: 1.0 X-Received: by 10.42.47.201 with SMTP id p9mr17878949icf.4.1391215599914; Fri, 31 Jan 2014 16:46:39 -0800 (PST) Received: by 10.50.61.97 with HTTP; Fri, 31 Jan 2014 16:46:39 -0800 (PST) Received: by 10.50.61.97 with HTTP; Fri, 31 Jan 2014 16:46:39 -0800 (PST) In-Reply-To: <941970C0-69AA-4928-A8E6-B9F72FD8917D@keithwiley.com> References: <280A85F8-DBB1-483D-843D-0D99C3356689@keithwiley.com> <941970C0-69AA-4928-A8E6-B9F72FD8917D@keithwiley.com> Date: Sat, 1 Feb 2014 06:16:39 +0530 Message-ID: Subject: Re: Force one mapper per machine (not core)? From: Harsh J To: "" Content-Type: multipart/alternative; boundary=90e6ba6149b2a3a1a504f14d9fb8 X-Virus-Checked: Checked by ClamAV on apache.org --90e6ba6149b2a3a1a504f14d9fb8 Content-Type: text/plain; charset=ISO-8859-1 If it's job tracker you use, it's MR1. On Feb 1, 2014 12:23 AM, "Keith Wiley" wrote: > Hmmm, okay. I know it's running CDH4 4.4.0, as but for whether it was > specifically configured with MR1 or MR2 (is there a distinction between MR2 > and Yarn?) I'm not absolutely certain. I know that the cluster "behaves" > like the MR1 clusters I've worked with for years (I interact with the job > tracker in a classical way for example). Can I tell whether it's MR1 or > MR2 from the job tracker or namename web UIs? > > Thanks. > > On Jan 29, 2014, at 00:52 , Harsh J wrote: > > > Is your cluster running MR1 or MR2? On MR1, the CapacityScheduler > > would allow you to do this if you used appropriate memory based > > requests (see http://search-hadoop.com/m/gnFs91yIg1e), and on MR2 > > (depending on the YARN scheduler resource request limits config) you > > can request your job be run with the maximum-most requests that would > > soak up all provided resources (of CPU and Memory) of a node such that > > only one container runs on a host at any given time. > > > > On Wed, Jan 29, 2014 at 3:30 AM, Keith Wiley > wrote: > >> I'm running a program which in the streaming layer automatically > multithreads and does so by automatically detecting the number of cores on > the machine. I realize this model is somewhat in conflict with Hadoop, but > nonetheless, that's what I'm doing. Thus, for even resource utilization, > it would be nice to not only assign one mapper per core, but only one > mapper per machine. I realize that if I saturate the cluster none of this > really matters, but consider the following example for clarity: 4-core > nodes, 10-node cluster, thus 40 slots, fully configured across mappers and > reducers (40 slots of each). Say I run this program with just two mappers. > It would run much more efficiently (in essentially half the time) if I > could force the two mappers to go to slots on two separate machines instead > of running the risk that Hadoop may assign them both to the same machine. > >> > >> Can this be done? > >> > >> Thanks. > > > > ________________________________________________________________________________ > Keith Wiley kwiley@keithwiley.com keithwiley.com > music.keithwiley.com > > "I used to be with it, but then they changed what it was. Now, what I'm > with > isn't it, and what's it seems weird and scary to me." > -- Abe (Grandpa) Simpson > > ________________________________________________________________________________ > > --90e6ba6149b2a3a1a504f14d9fb8 Content-Type: text/html; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable

If it's job tracker you use, it's MR1.

On Feb 1, 2014 12:23 AM, "Keith Wiley"= <kwiley@keithwiley.com>= wrote:
Hmmm, okay. =A0I know it's running CDH4 4.4.0, as but for whether it wa= s specifically configured with MR1 or MR2 (is there a distinction between M= R2 and Yarn?) I'm not absolutely certain. =A0I know that the cluster &q= uot;behaves" like the MR1 clusters I've worked with for years (I i= nteract with the job tracker in a classical way for example). =A0Can I tell= whether it's MR1 or MR2 from the job tracker or namename web UIs?

Thanks.

On Jan 29, 2014, at 00:52 , Harsh J wrote:

> Is your cluster running MR1 or MR2? On MR1, the CapacityScheduler
> would allow you to do this if you used appropriate memory based
> requests (see http://search-hadoop.com/m/gnFs91yIg1e), and on MR2
> (depending on the YARN scheduler resource request limits config) you > can request your job be run with the maximum-most requests that would<= br> > soak up all provided resources (of CPU and Memory) of a node such that=
> only one container runs on a host at any given time.
>
> On Wed, Jan 29, 2014 at 3:30 AM, Keith Wiley <kwiley@keithwiley.com> wrote:
>> I'm running a program which in the streaming layer automatical= ly multithreads and does so by automatically detecting the number of cores = on the machine. =A0I realize this model is somewhat in conflict with Hadoop= , but nonetheless, that's what I'm doing. =A0Thus, for even resourc= e utilization, it would be nice to not only assign one mapper per core, but= only one mapper per machine. =A0I realize that if I saturate the cluster n= one of this really matters, but consider the following example for clarity:= 4-core nodes, 10-node cluster, thus 40 slots, fully configured across mapp= ers and reducers (40 slots of each). =A0Say I run this program with just tw= o mappers. =A0It would run much more efficiently (in essentially half the t= ime) if I could force the two mappers to go to slots on two separate machin= es instead of running the risk that Hadoop may assign them both to the same= machine.
>>
>> Can this be done?
>>
>> Thanks.


___________________________________________________________________________= _____
Keith Wiley =A0 =A0 kwiley@keithwi= ley.com =A0 =A0 kei= thwiley.com =A0 =A0music.keithwiley.com

"I used to be with it, but then they changed what it was. =A0Now, what= I'm with
isn't it, and what's it seems weird and scary to me."
=A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0= =A0 =A0 =A0-- =A0Abe (Grandpa) Simpson
___________________________________________________________________________= _____

--90e6ba6149b2a3a1a504f14d9fb8--