Return-Path: X-Original-To: apmail-hadoop-common-user-archive@www.apache.org Delivered-To: apmail-hadoop-common-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 2ADDDD16A for ; Sun, 6 Jan 2013 14:47:49 +0000 (UTC) Received: (qmail 95248 invoked by uid 500); 6 Jan 2013 14:47:44 -0000 Delivered-To: apmail-hadoop-common-user-archive@hadoop.apache.org Received: (qmail 94913 invoked by uid 500); 6 Jan 2013 14:47:44 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hadoop.apache.org Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 94892 invoked by uid 99); 6 Jan 2013 14:47:43 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Sun, 06 Jan 2013 14:47:43 +0000 X-ASF-Spam-Status: No, hits=-0.7 required=5.0 tests=RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of harsh@cloudera.com designates 209.85.210.169 as permitted sender) Received: from [209.85.210.169] (HELO mail-ia0-f169.google.com) (209.85.210.169) by apache.org (qpsmtpd/0.29) with ESMTP; Sun, 06 Jan 2013 14:47:39 +0000 Received: by mail-ia0-f169.google.com with SMTP id u20so9004201iag.28 for ; Sun, 06 Jan 2013 06:47:18 -0800 (PST) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20120113; h=mime-version:in-reply-to:references:from:date:message-id:subject:to :content-type:content-transfer-encoding:x-gm-message-state; bh=ZEDOjShhMQw8LAelXzXAHzQgMmPmBi0NErMgV7uHpyA=; b=OqrmOUYmmljMM56aWaId+KiATUXn134bdKbfPFgDm8CSBiIhb8DYci2WtJ4axHgziS AAd2alyzchCkLIz2ZfyCxoVyz7VWZHBksm0d7QAnjjOD2ebwZO6cFAO8n6++uXxdMahA hQeTG1NI0EZ7+bxfQUY6fSbyQktX0WX4VeLx469WyyfJuFWPIGUJK6Kz8TFWkYQIhqwR D41aJW4Qs0pUibq0H/N9PmQ9oKFAbB5UvB0ZO5HXv1IPXVpOpxDqiseQ+4nv/ae3UZDs lgs96VGwf2LwYzjhurNw/DrIXqh+IJTCuedhuGon0Rsl+hqX77Kc8/feXtNHn8wfQ3As JOkw== Received: by 10.50.151.238 with SMTP id ut14mr3437047igb.58.1357483638493; Sun, 06 Jan 2013 06:47:18 -0800 (PST) MIME-Version: 1.0 Received: by 10.64.32.166 with HTTP; Sun, 6 Jan 2013 06:46:58 -0800 (PST) In-Reply-To: References: <022f01cdeb4e$ea7fc170$bf7f4450$@yahoo.com> From: Harsh J Date: Sun, 6 Jan 2013 20:16:58 +0530 Message-ID: Subject: Re: Skipping entire task To: "" Content-Type: text/plain; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable X-Gm-Message-State: ALoCoQm806FgXpWOI38n/DIpVbqVy4fIie2wlufsFWCkF5Cc4IAuMvwqgjPyJ2iPg8XWZb9l6D00 X-Virus-Checked: Checked by ClamAV on apache.org You can use the mapred.max.map.failures.percent and mapred.max.reduce.failures.percent features to control the percentage of allowed failures of tasks in a single job (despite which the job is marked successful). On Sun, Jan 6, 2013 at 8:04 PM, H=E5vard Wahl Kongsg=E5rd wrote: >> Are tasks being executed multiple times due to failures? Sorry, it was n= ot >> very clear from your question. > > yes, and I simply want to skip them if they fail more than x > times(after all this is big data :) ). > > -H=E5vard > > On Sun, Jan 6, 2013 at 3:01 PM, Hemanth Yamijala > wrote: >> Hi, >> >> Are tasks being executed multiple times due to failures? Sorry, it was n= ot >> very clear from your question. >> >> Thanks >> hemanth >> >> >> On Sat, Jan 5, 2013 at 7:44 PM, David Parks wro= te: >>> >>> Thinking here... if you submitted the task programmatically you should = be >>> able to capture the failure of the task and gracefully move past it to >>> your >>> next tasks. >>> >>> To say it in a long-winded way: Let's say you submit a job to Hadoop, = a >>> java jar, and your main class implements Tool. That code has the >>> responsibility to submit a series of jobs to hadoop, something like thi= s: >>> >>> try{ >>> Job myJob =3D new MyJob(getConf()); >>> myJob.submitAndWait(); >>> }catch(Exception uhhohh){ >>> //Deal with the issue and move on >>> } >>> Job myNextJob =3D new MyNextJob(getConf()); >>> myNextJob.submit(); >>> >>> Just pseudo code there to demonstrate my thought. >>> >>> David >>> >>> >>> >>> -----Original Message----- >>> From: H=E5vard Wahl Kongsg=E5rd [mailto:haavard.kongsgaard@gmail.com] >>> Sent: Saturday, January 05, 2013 4:54 PM >>> To: user >>> Subject: Skipping entire task >>> >>> Hi, hadoop can skip bad records >>> >>> http://devblog.factual.com/practical-hadoop-streaming-dealing-with-brit= tle-c >>> ode. >>> But it is also possible to skip entire tasks? >>> >>> -H=E5vard >>> >>> -- >>> H=E5vard Wahl Kongsg=E5rd >>> Faculty of Medicine & >>> Department of Mathematical Sciences >>> NTNU >>> >>> http://havard.security-review.net/ >>> >> > > > > -- > H=E5vard Wahl Kongsg=E5rd > Faculty of Medicine & > Department of Mathematical Sciences > NTNU > > http://havard.security-review.net/ --=20 Harsh J