Return-Path: Delivered-To: apmail-lucene-hadoop-user-archive@locus.apache.org Received: (qmail 21166 invoked from network); 26 Jul 2006 20:22:25 -0000 Received: from hermes.apache.org (HELO mail.apache.org) (209.237.227.199) by minotaur.apache.org with SMTP; 26 Jul 2006 20:22:25 -0000 Received: (qmail 11600 invoked by uid 500); 26 Jul 2006 20:22:25 -0000 Delivered-To: apmail-lucene-hadoop-user-archive@lucene.apache.org Received: (qmail 11409 invoked by uid 500); 26 Jul 2006 20:22:24 -0000 Mailing-List: contact hadoop-user-help@lucene.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: hadoop-user@lucene.apache.org Delivered-To: mailing list hadoop-user@lucene.apache.org Received: (qmail 11399 invoked by uid 99); 26 Jul 2006 20:22:24 -0000 Received: from asf.osuosl.org (HELO asf.osuosl.org) (140.211.166.49) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 26 Jul 2006 13:22:24 -0700 X-ASF-Spam-Status: No, hits=0.5 required=10.0 tests=DNS_FROM_RFC_ABUSE,HTML_MESSAGE,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (asf.osuosl.org: domain of sutter@gmail.com designates 64.233.162.207 as permitted sender) Received: from [64.233.162.207] (HELO nz-out-0102.google.com) (64.233.162.207) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 26 Jul 2006 13:22:21 -0700 Received: by nz-out-0102.google.com with SMTP id x3so648919nzd for ; Wed, 26 Jul 2006 13:22:00 -0700 (PDT) DomainKey-Signature: a=rsa-sha1; q=dns; c=nofws; s=beta; d=gmail.com; h=received:message-id:date:from:to:subject:in-reply-to:mime-version:content-type:references; b=jTIjRLz/wBC4k3qMXdVmA53JXqzn63jTa3eiRbnBZABxjtPEB/i2sfX5uk7daargz1MJ7liQ5PlDV7MuqlNoPxmzb9YXt1hVzsN5ifa/oACIjqikCJ4L/p3MbyL4r1v8IEQVTpQivmIwwzkbzyKcp8cXCIWThxwFo0rQ2X5kjjk= Received: by 10.65.59.10 with SMTP id m10mr7434097qbk; Wed, 26 Jul 2006 13:21:59 -0700 (PDT) Received: by 10.36.132.4 with HTTP; Wed, 26 Jul 2006 13:21:59 -0700 (PDT) Message-ID: Date: Wed, 26 Jul 2006 13:21:59 -0700 From: "Paul Sutter" To: hadoop-user@lucene.apache.org Subject: Re: Task type priorities during scheduling ? In-Reply-To: <44C708F5.8080602@apache.org> MIME-Version: 1.0 Content-Type: multipart/alternative; boundary="----=_Part_12373_6883118.1153945319699" References: <358D735BD7AB45429F2B1C14F38E10F70465C729@DEN-EXM-03.corp.ebay.com> <007501c6ac3c$8dc7bad0$a248480a@ds.corp.yahoo.com> <80CCE470-BF9D-4AC6-9B76-F55EE0E3E31B@yahoo-inc.com> <44C4801F.5010903@apache.org> <44C5D507.9080203@apache.org> <44C708F5.8080602@apache.org> X-Virus-Checked: Checked by ClamAV on apache.org X-Spam-Rating: minotaur.apache.org 1.6.2 0/1000/N ------=_Part_12373_6883118.1153945319699 Content-Type: text/plain; charset=ISO-8859-1; format=flowed Content-Transfer-Encoding: 7bit Content-Disposition: inline Doug, I agree that this isnt a high priority change, I'm just trying to start discussion towards what is needed to make multijob work well. I really like Yoram's suggestion of a single limit for map and reduce tasks. Not charging the copy(shuffle) phase to that limit could be part of making that work. Again, no urgency. We are already running two parallel clusters on the same boxes, we call them Blue (normal) and Yellow (nice'd), named after the colors on the Ganglia CPU display. We run long jobs on the nice'd cluster, and short jobs at normal priority. It works really well. Kevin should be submitting the two patches we needed to make it work. Paul On 7/25/06, Doug Cutting wrote: > > Paul Sutter wrote: > > First, It matters in the case of concurrent jobs. If you submit a 20 > > minute job while a 20 hour job is running, it would be nice if the > > reducers for the 20 minute job could get a chance to run before the 20 > > hour job's mappers have all finished. So even without a throughput > > improvement, you have an important capability (although it may require > > another minor tweak or two to make possible). > > I fear that more than a minor tweak or two are required to make > concurrent jobs work well. For example, you would also want to make > sure that the long-running job does not consume all of the reduce slots, > or the short job would again get stuck behind it. Pausing long-running > tasks might be required. > > The best way to do this at present is to run two job trackers, and two > tasktrackers per node, then submit long-runnning jobs to one "cluster" > and short-running jobs to the other. > > > Secondarily, we often have stragglers, where one mapper runs slower > > than the others. When this happens, we end up with a largely idle > > cluster for as long as an hour. In cases like these, good support for > > concurrent jobs _would_ improve throughput. > > Can you perhaps increase the number of map tasks, so that even a slow > task takes only a very small portion of the total execution time? > > Good support for concurrent jobs would be great to have, and I'd love to > see a patch that addresses this issue comprehensively. I am not > convinced that it is worth making minor tweaks that may-or-may-not > really help us to get there. > > Doug > ------=_Part_12373_6883118.1153945319699--