Return-Path: X-Original-To: apmail-hadoop-mapreduce-user-archive@minotaur.apache.org Delivered-To: apmail-hadoop-mapreduce-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 45820F97D for ; Tue, 9 Jul 2013 05:01:12 +0000 (UTC) Received: (qmail 52212 invoked by uid 500); 9 Jul 2013 05:01:06 -0000 Delivered-To: apmail-hadoop-mapreduce-user-archive@hadoop.apache.org Received: (qmail 51549 invoked by uid 500); 9 Jul 2013 05:00:58 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hadoop.apache.org Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 51535 invoked by uid 99); 9 Jul 2013 05:00:56 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 09 Jul 2013 05:00:56 +0000 X-ASF-Spam-Status: No, hits=2.2 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_NONE,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of sandeepvreddy@outlook.com designates 65.55.111.102 as permitted sender) Received: from [65.55.111.102] (HELO blu0-omc2-s27.blu0.hotmail.com) (65.55.111.102) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 09 Jul 2013 05:00:51 +0000 Received: from BLU171-W114 ([65.55.111.71]) by blu0-omc2-s27.blu0.hotmail.com with Microsoft SMTPSVC(6.0.3790.4675); Mon, 8 Jul 2013 22:00:31 -0700 X-TMN: [EBlTCl0FoaYql21ev7VymRWjsr3vrs3N] X-Originating-Email: [sandeepvreddy@outlook.com] Message-ID: Content-Type: multipart/alternative; boundary="_0ebe23b3-4506-428a-904e-83a992e062f7_" From: Sandeep L To: "user@hadoop.apache.org" Subject: RE: Using CapacityScheduler to divide resources between jobs (not users) Date: Tue, 9 Jul 2013 10:30:30 +0530 Importance: Normal In-Reply-To: References: MIME-Version: 1.0 X-OriginalArrivalTime: 09 Jul 2013 05:00:31.0125 (UTC) FILETIME=[3CA21450:01CE7C61] X-Virus-Checked: Checked by ClamAV on apache.org --_0ebe23b3-4506-428a-904e-83a992e062f7_ Content-Type: text/plain; charset="iso-8859-1" Content-Transfer-Encoding: quoted-printable One solution I can suggest is use multiple jobtrackers.Jobtracker1: 2 or 3 = machines as tasktrackersJobtracker2: around 7 machines as tasktrackersJobtr= acker3: around 10 machines as tasktrackers As per requirement you can change number of tasktracker machines and run jo= bs accordingly. Thanks=2CSandeep. Date: Sat=2C 6 Jul 2013 17:12:01 +0300 Subject: Using CapacityScheduler to divide resources between jobs (not user= s) From: amits@infolinks.com To: user@hadoop.apache.org Hi all=2C=20 I'm running Hadoop 1.0.4 on a modest cluster (~20 machines).The jobs runnin= g on the cluster can be divided (resource wise) as follows: =0A= 1. Very short jobs: less then 1 minute.2. Normal jobs: 2-3 minutes up to an= hour or two.3. Very long jobs: days of processing. (still not active and t= he reason for my inquiries here).=0A= I was thinking of using the CapacityScheduler and divide the cluster resour= ces so that the long jobs can run without disturbing the other jobs.I read = that such job queues should be upper bound as well since it may use the ent= ire cluster resources once it's free but since it takes a long time to fini= sh=2C it won't release them to other queues as it should. Is it so ?=0A= Any advise about using the CapacityScheduler in that use case ? Thanks=2C and sorry for re-sending this message. Amit. = --_0ebe23b3-4506-428a-904e-83a992e062f7_ Content-Type: text/html; charset="iso-8859-1" Content-Transfer-Encoding: quoted-printable
One solution I can suggest is us= e multiple jobtrackers.
  1. Jobtracker1: 2 or 3 machines as tasktra= ckers
  2. Jobtracker2: around 7 machines =3Bas tasktrackers
  3. Jobtracker3: around 10 machines =3Bas tasktrackers

<= /div>
As per requirement you can change number of tasktracker machines = and run jobs accordingly.

Thanks=2C
Sandeep.



Date: Sat=2C 6 Jul 2013 17:12:01 +0300
Subject:= Using CapacityScheduler to divide resources between jobs (not users)
Fr= om: amits@infolinks.com
To: user@hadoop.apache.org

Hi all=2C =3B

I'm running Hadoop 1.0.4= on a modest cluster (~20 machines).
The jobs running on the clus= ter can be divided (resource wise) as follows:

=0A=
1. Very short jobs: less then 1 minute.
2. Normal jobs: 2-3 = minutes up to an hour or two.
3. Very long jobs: days of processi= ng. (still not active and the reason for my inquiries here).
=0A=
I was thinking of using the CapacityScheduler and divide the= cluster resources so that the long jobs can run without disturbing the oth= er jobs.
I read that such job queues should be upper bound as wel= l since it may use the entire cluster resources once it's free but since it= takes a long time to finish=2C it won't release them to other queues as it= should. Is it so ?
=0A=
Any advise about using the =3BCapacityScheduler in that use case ?=

Thanks=2C and sorry for re-sending this message.<= /div>

Amit.
= --_0ebe23b3-4506-428a-904e-83a992e062f7_--