Return-Path: X-Original-To: apmail-hadoop-common-user-archive@www.apache.org Delivered-To: apmail-hadoop-common-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 4B025FA33 for ; Sat, 27 Apr 2013 19:35:33 +0000 (UTC) Received: (qmail 14086 invoked by uid 500); 27 Apr 2013 19:35:28 -0000 Delivered-To: apmail-hadoop-common-user-archive@hadoop.apache.org Received: (qmail 13733 invoked by uid 500); 27 Apr 2013 19:35:28 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hadoop.apache.org Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 13725 invoked by uid 99); 27 Apr 2013 19:35:28 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 27 Apr 2013 19:35:28 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of sudhakara.st@gmail.com designates 209.85.214.174 as permitted sender) Received: from [209.85.214.174] (HELO mail-ob0-f174.google.com) (209.85.214.174) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 27 Apr 2013 19:35:22 +0000 Received: by mail-ob0-f174.google.com with SMTP id wc20so4475843obb.33 for ; Sat, 27 Apr 2013 12:35:01 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:x-received:in-reply-to:references:date:message-id :subject:from:to:content-type; bh=0TgI9OZqosUsXQpxA2RttozcIceLKwgKN3CInT1RQa0=; b=wq7+meCr91Vr7cG87T5QZzeGXYCT3CVyk1hmGitKO3811XwmJCFz+f0fvGDPhYvQxi 5hrCubJYP10+ZtG8vLbZrJYPmzRbgTg8xcJdU6l1pMg1wgjMcPFFm+lUT5CTXmkAu988 s8WgPFdDJWisdKbOPvlySrgZQ5K5mStRvThrt8QBKdUC47F4jE4Q+M5vTUhxEyArXc+y dlV5JFZo2rtDrwY151r3udXtoiLGK+NQwngorT03y4EdkhWa9Aaxp+++Ebl+OgQykvDf rx2FEjlfY8xsBC9FjirQepeLZg6Rfuua32lIcVQ8DyYn9zVpg/vhi5qXRKOq56BQhZoo q0SQ== MIME-Version: 1.0 X-Received: by 10.182.27.40 with SMTP id q8mr20560693obg.100.1367091301347; Sat, 27 Apr 2013 12:35:01 -0700 (PDT) Received: by 10.76.20.46 with HTTP; Sat, 27 Apr 2013 12:35:01 -0700 (PDT) In-Reply-To: References: Date: Sun, 28 Apr 2013 01:05:01 +0530 Message-ID: Subject: Re: Exception during reduce phase when running jobs remotely From: sudhakara st To: user@hadoop.apache.org, orenbum@gmail.com Content-Type: multipart/alternative; boundary=089e0122976264c54e04db5cbfd5 X-Virus-Checked: Checked by ClamAV on apache.org --089e0122976264c54e04db5cbfd5 Content-Type: text/plain; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable Hello, Check you hadoop.tmp.dir and mapred.local.dir configuration and permissions= . On Sat, Apr 27, 2013 at 12:20 AM, Oren Bumgarner wrote: > I have a small hadoop cluster running 1.0.4 and I'm trying to have it > setup so that I can run jobs remotely from a computer on the same network > that is not a part of the cluster. I've got a main java class that > implements org.apache.hadoop.util.Tool and I'm able to run this job from > the NameNode using ToolRunner.run(), setting up the JobConf, and submitti= ng > with JobClient.submitJob(). > > When I try to run the same class remotely from any machine that is not th= e > NameNode the job is submitted and it appears that the Map tasks > successfully complete, but I get the following exception for all of the > reduce tasks: > > org.apache.hadoop.util.DiskChecker$DiskErrorException: Could not find out= put/map_0.out in any of the configured local directories > at org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPa= thToRead(LocalDirAllocator.java:429) > at org.apache.hadoop.fs.LocalDirAllocator.getLocalPathToRead(LocalDirAll= ocator.java:160) > at org.apache.hadoop.mapred.MapOutputFile.getInputFile(MapOutputFile.jav= a:161) > at org.apache.hadoop.mapred.ReduceTask.getMapFiles(ReduceTask.java:220) > at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:398) > at org.apache.hadoop.mapred.Child$4.run(Child.java:255) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:396) > at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInforma= tion.java:1121) > at org.apache.hadoop.mapred.Child.main(Child.java:249) > > > I'm not sure how to interpret that error message. I think I'm missing som= e config files that are not present on the remote machine but I can't pin d= own exactly what I need. Does anyone have any guidance on what the above er= ror means or how to submit jobs remotely? > > Thanks, > > --Oren > > --=20 Regards, ..... Sudhakara.st --089e0122976264c54e04db5cbfd5 Content-Type: text/html; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable
Hello,

Check you hadoop.tmp.dir and mapr= ed.local.dir configuration and permissions.



On Sat, Apr 27, 2013= at 12:20 AM, Oren Bumgarner <orenbum@gmail.com> wrote:
I have a small hadoop cluster running 1.0.4 = and I'm trying to have it setup so that I can run jobs remotely from a = computer on the same network that is not a part of the cluster. I've go= t a main java class that implements=A0org.apache.hadoop.util.Tool and I'= ;m able to run this job from the NameNode using ToolRunner.run(), setting u= p the JobConf, and submitting with JobClient.submitJob().

When I try to run the same class remotely from any machine t= hat is not the NameNode the job is submitted and it appears that the Map ta= sks successfully complete, but I get the following exception for all of the= reduce tasks:

org.apache.hadoop.util.DiskChecker$DiskErrorExcept=
ion: Could not find output/map_0.out in any of the configured local directo=
ries
	at org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPath=
ToRead(LocalDirAllocator.java:429)
	at org.apache.hadoop.fs.LocalDirAllocator.getLocalPathToRead(LocalDirAlloc=
ator.java:160)
	at org.apache.hadoop.mapred.MapOutputFile.getInputFile(MapOutputFile.java:=
161)
	at org.apache.hadoop.mapred.ReduceTask.getMapFiles(ReduceTask.java:220)
	at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:398)
	at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:396)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformati=
on.java:1121)
	at org.apache.hadoop.mapred.Child.main(Child.java:249)

I'm not sure how to interpr=
et that error message. I think I'm missing some config files that are n=
ot present on the remote machine but I can't pin down exactly what I ne=
ed. Does anyone have any guidance on what the above error means or how to s=
ubmit jobs remotely?
Thanks,
--Oren<=
/pre>



--
=A0 =A0 =A0=A0
Rega= rds,
.....=A0 Sudhakara.st
=A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 = =A0=A0
--089e0122976264c54e04db5cbfd5--