Return-Path: Delivered-To: apmail-hadoop-core-user-archive@www.apache.org Received: (qmail 5214 invoked from network); 3 Feb 2009 19:34:22 -0000 Received: from hermes.apache.org (HELO mail.apache.org) (140.211.11.2) by minotaur.apache.org with SMTP; 3 Feb 2009 19:34:22 -0000 Received: (qmail 21062 invoked by uid 500); 3 Feb 2009 19:34:16 -0000 Delivered-To: apmail-hadoop-core-user-archive@hadoop.apache.org Received: (qmail 21012 invoked by uid 500); 3 Feb 2009 19:34:15 -0000 Mailing-List: contact core-user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: core-user@hadoop.apache.org Delivered-To: mailing list core-user@hadoop.apache.org Received: (qmail 21001 invoked by uid 99); 3 Feb 2009 19:34:15 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 03 Feb 2009 11:34:15 -0800 X-ASF-Spam-Status: No, hits=-0.0 required=10.0 tests=SPF_HELO_PASS,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: local policy) Received: from [216.86.168.183] (HELO mxout-08.mxes.net) (216.86.168.183) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 03 Feb 2009 19:34:06 +0000 Received: from [192.168.10.105] (unknown [24.6.146.30]) (using TLSv1 with cipher AES128-SHA (128/128 bits)) (No client certificate requested) by smtp.mxes.net (Postfix) with ESMTPSA id A3BB1D05A7 for ; Tue, 3 Feb 2009 14:33:44 -0500 (EST) Message-Id: <3E9F6237-39EC-4026-8132-60439D4948B5@wensel.net> From: Chris K Wensel To: core-user@hadoop.apache.org In-Reply-To: <1f2901c98633$f6e8dd40$e4ba97c0$@com> Content-Type: text/plain; charset=US-ASCII; format=flowed; delsp=yes Content-Transfer-Encoding: 7bit Mime-Version: 1.0 (Apple Message framework v930.3) Subject: Re: Control over max map/reduce tasks per job Date: Tue, 3 Feb 2009 11:33:43 -0800 References: <1f2901c98633$f6e8dd40$e4ba97c0$@com> X-Mailer: Apple Mail (2.930.3) X-Virus-Checked: Checked by ClamAV on apache.org Hey Jonathan Are you looking to limit the total number of concurrent mapper/ reducers a single job can consume cluster wide, or limit the number per node? That is, you have X mappers/reducers, but only can allow N mappers/ reducers to run at a time globally, for a given job. Or, you are cool with all X running concurrently globally, but want to guarantee that no node can run more than N tasks from that job? Or both? just reconciling the conversation we had last week with this thread. ckw On Feb 3, 2009, at 11:16 AM, Jonathan Gray wrote: > All, > > > > I have a few relatively small clusters (5-20 nodes) and am having > trouble > keeping them loaded with my MR jobs. > > > > The primary issue is that I have different jobs that have drastically > different patterns. I have jobs that read/write to/from HBase or > Hadoop > with minimal logic (network throughput bound or io bound), others that > perform crawling (network latency bound), and one huge parsing > streaming job > (very CPU bound, each task eats a core). > > > > I'd like to launch very large numbers of tasks for network latency > bound > jobs, however the large CPU bound job means I have to keep the max > maps > allowed per node low enough as to not starve the Datanode and > Regionserver. > > > > I'm an HBase dev but not familiar enough with Hadoop MR code to even > know > what would be involved with implementing this. However, in talking > with > other users, it seems like this would be a well-received option. > > > > I wanted to ping the list before filing an issue because it seems like > someone may have thought about this in the past. > > > > Thanks. > > > > Jonathan Gray > -- Chris K Wensel chris@wensel.net http://www.cascading.org/ http://www.scaleunlimited.com/