Return-Path: X-Original-To: apmail-spark-dev-archive@minotaur.apache.org Delivered-To: apmail-spark-dev-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 9B85A1083E for ; Sun, 11 Jan 2015 15:39:02 +0000 (UTC) Received: (qmail 11017 invoked by uid 500); 11 Jan 2015 15:39:00 -0000 Delivered-To: apmail-spark-dev-archive@spark.apache.org Received: (qmail 10932 invoked by uid 500); 11 Jan 2015 15:39:00 -0000 Mailing-List: contact dev-help@spark.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Delivered-To: mailing list dev@spark.apache.org Received: (qmail 10048 invoked by uid 99); 11 Jan 2015 15:38:59 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Sun, 11 Jan 2015 15:38:59 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of alexbaretta@gmail.com designates 209.85.214.171 as permitted sender) Received: from [209.85.214.171] (HELO mail-ob0-f171.google.com) (209.85.214.171) by apache.org (qpsmtpd/0.29) with ESMTP; Sun, 11 Jan 2015 15:38:55 +0000 Received: by mail-ob0-f171.google.com with SMTP id uz6so19269322obc.2; Sun, 11 Jan 2015 07:37:04 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:in-reply-to:references:from:date:message-id:subject:to :cc:content-type; bh=DUNtZ/wSvDBy1u7WeROOpjDeIhYskaoCsYaWe0vjbM4=; b=toBhdS5fruXKra77I4Nf627qNFRoZfMpDy3gdNwxSvLezuzkxoVXGfT5ecnH8ATmaM mevh051FLfLiK5NLR/EkZQBOBRyAvof6W7qv2Sz1t3I7XLPrhNHdzlm07/UkRJBOS0k0 6S336B0rRs0mDHfBbN6FwSzLcaYfYzvPVKbZzaj7LSdQybzIv/Rp8sXRlMby+JrWmIga 32cEq0y3dCWI6eHa+IsN9OWcRo47XqgFVF0FqZ0ePolp3DU5iTACNlk7CSLvfKVIK5HM ggaovcU6LioPMR1NzN2h/XM70VWRzPRVoio1lOhm2351wWupRToSPHOLHN3C9ig/64tm A+bA== X-Received: by 10.202.224.198 with SMTP id x189mr14222319oig.62.1420990624700; Sun, 11 Jan 2015 07:37:04 -0800 (PST) MIME-Version: 1.0 Received: by 10.76.111.148 with HTTP; Sun, 11 Jan 2015 07:36:44 -0800 (PST) In-Reply-To: References: From: Alessandro Baretta Date: Sun, 11 Jan 2015 07:36:44 -0800 Message-ID: Subject: Re: Job priority To: Cody Koeninger Cc: Mark Hamstra , "user@spark.apache.org" , "dev@spark.apache.org" Content-Type: multipart/alternative; boundary=001a113d34306a21d5050c6229c5 X-Virus-Checked: Checked by ClamAV on apache.org --001a113d34306a21d5050c6229c5 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: quoted-printable Cody, While I might be able to improve the scheduling of my jobs by using a few different pools with weights equal to, say, 1, 1e3 and 1e6, effectively getting a small handful of priority classes. Still, this is really not quite what I am describing. This is why my original post was on the dev list. Let me then ask if there is any interest in having priority queue job scheduling in Spark. This is something I might be able to pull off. Alex On Sun, Jan 11, 2015 at 6:21 AM, Cody Koeninger wrote: > If you set up a number of pools equal to the number of different priority > levels you want, make the relative weights of those pools very different, > and submit a job to the pool representing its priority, I think youll get > behavior equivalent to a priority queue. Try it and see. > > If I'm misunderstandng what youre trying to do, then I don't know. > > > On Sunday, January 11, 2015, Alessandro Baretta > wrote: > >> Cody, >> >> Maybe I'm not getting this, but it doesn't look like this page is >> describing a priority queue scheduling policy. What this section discuss= es >> is how resources are shared between queues. A weight-1000 pool will get >> 1000 times more resources allocated to it than a priority 1 queue. Great= , >> but not what I want. I want to be able to define an Ordering on make my >> tasks representing their priority, and have Spark allocate all resources= to >> the job that has the highest priority. >> >> Alex >> >> On Sat, Jan 10, 2015 at 10:11 PM, Cody Koeninger >> wrote: >> >>> >>> http://spark.apache.org/docs/latest/job-scheduling.html#configuring-poo= l-properties >>> >>> "Setting a high weight such as 1000 also makes it possible to implement >>> *priority* between pools=E2=80=94in essence, the weight-1000 pool will = always >>> get to launch tasks first whenever it has jobs active." >>> >>> On Sat, Jan 10, 2015 at 11:57 PM, Alessandro Baretta < >>> alexbaretta@gmail.com> wrote: >>> >>>> Mark, >>>> >>>> Thanks, but I don't see how this documentation solves my problem. You >>>> are referring me to documentation of fair scheduling; whereas, I am as= king >>>> about as unfair a scheduling policy as can be: a priority queue. >>>> >>>> Alex >>>> >>>> On Sat, Jan 10, 2015 at 5:00 PM, Mark Hamstra >>>> wrote: >>>> >>>>> -dev, +user >>>>> >>>>> http://spark.apache.org/docs/latest/job-scheduling.html >>>>> >>>>> >>>>> On Sat, Jan 10, 2015 at 4:40 PM, Alessandro Baretta < >>>>> alexbaretta@gmail.com> wrote: >>>>> >>>>>> Is it possible to specify a priority level for a job, such that the >>>>>> active >>>>>> jobs might be scheduled in order of priority? >>>>>> >>>>>> Alex >>>>>> >>>>> >>>>> >>>> >>> >> --001a113d34306a21d5050c6229c5--