Return-Path: Delivered-To: apmail-hadoop-general-archive@minotaur.apache.org Received: (qmail 31685 invoked from network); 11 Apr 2009 00:52:15 -0000 Received: from hermes.apache.org (HELO mail.apache.org) (140.211.11.3) by minotaur.apache.org with SMTP; 11 Apr 2009 00:52:15 -0000 Received: (qmail 75096 invoked by uid 500); 11 Apr 2009 00:52:14 -0000 Delivered-To: apmail-hadoop-general-archive@hadoop.apache.org Received: (qmail 75029 invoked by uid 500); 11 Apr 2009 00:52:14 -0000 Mailing-List: contact general-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: general@hadoop.apache.org Delivered-To: mailing list general@hadoop.apache.org Received: (qmail 75019 invoked by uid 99); 11 Apr 2009 00:52:14 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 11 Apr 2009 00:52:14 +0000 X-ASF-Spam-Status: No, hits=-0.0 required=10.0 tests=SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of sethladd@gmail.com designates 209.85.200.168 as permitted sender) Received: from [209.85.200.168] (HELO wf-out-1314.google.com) (209.85.200.168) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 11 Apr 2009 00:52:06 +0000 Received: by wf-out-1314.google.com with SMTP id 23so1126573wfg.2 for ; Fri, 10 Apr 2009 17:51:45 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:mime-version:received:in-reply-to:references :date:message-id:subject:from:to:content-type :content-transfer-encoding; bh=f8FXU3q+BGVRzSzBlkHc7/64F1zicAve2IqPFxXPEhM=; b=j32gOrOElq27UIb3QdIxEKSrgpI9jpRozolxwufl1RZIUvOwI0q7Q3i9Hcs4vVmOCE o1IkRYepkH/H5RxeS/LQKEDo20IYyoeuGQU8B0CwYH/xPTQHvXNr+66Y0EiXDs0Ms0VA s+jxQYjSOiXQmoJXVHxqhEF7cASG8c2CJaQBc= DomainKey-Signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :content-type:content-transfer-encoding; b=KlB7jd74j5ml6dpdGR9Ag4EO0BAfXnz5LJCXb5/Wf6Xn64Gg9EDj3VPYL7q8qMCCkU NXHnZaG5pdGzfWPn8eXmlmNd+eDWmGTP6FrfzDFi8HzQFAYnqgP0c1Ky51hCMv2e5tX0 No8icpwe6TCi9xKEBCmoZGOLsCDcIBswp4Jso= MIME-Version: 1.0 Received: by 10.143.29.17 with SMTP id g17mr1565061wfj.109.1239411105458; Fri, 10 Apr 2009 17:51:45 -0700 (PDT) In-Reply-To: <64f07e2c0904101728g7dfcfeeesc8fd7253b1f409c3@mail.gmail.com> References: <64f07e2c0904101728g7dfcfeeesc8fd7253b1f409c3@mail.gmail.com> Date: Fri, 10 Apr 2009 14:51:45 -1000 Message-ID: <64f07e2c0904101751x658e3abdke40274eda037015e@mail.gmail.com> Subject: Re: Error initializing attempt From: Seth Ladd To: general@hadoop.apache.org Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: quoted-printable X-Virus-Checked: Checked by ClamAV on apache.org To answer my own question, I added explicit entries for my hosts in the cluster to each /etc/hosts file. Apparently, our DNS was mucking up the name resolution (even though I had specified the IP address in the config files) Lession learned: ensure there's a solid entry for your master node in each slave node's /etc/hosts file. I hope this saves someone else next time it happens. On Fri, Apr 10, 2009 at 2:28 PM, Seth Ladd wrote: > Hello, > > I am using Hadoop 0.18.3 and Pig 0.2.0 across a small cluster of 4 > machines. =C2=A0I am using start-all.sh to boot up the cluster. =C2=A0The= HDFS > system appears to be working really well. =C2=A0I am able to copy files i= n > and out of the filesystem. > > However, when I try to submit a Pig job, I am greeted by this exception: > > 2009-04-10 14:17:27,145 INFO org.apache.hadoop.mapred.TaskTracker: > Starting tracker tracker_sdi-kenglish:localhost/127.0.0.1:37109 > 2009-04-10 14:17:27,249 INFO org.apache.hadoop.mapred.TaskTracker: > Starting thread: Map-events fetcher for all reduce tasks on > tracker_sdi-kenglish:localhost/127.0.0.1:37109 > 2009-04-10 14:18:07,361 INFO org.apache.hadoop.mapred.TaskTracker: > LaunchTaskAction: attempt_200904101417_0001_m_000002_0 > 2009-04-10 14:18:10,353 WARN org.apache.hadoop.mapred.TaskTracker: > Error initializing attempt_200904101417_0001_m_000002_0: > java.io.IOException: Call to sladd/208.67.216.132:8020 failed on local > exception: java.io.EOFException > =C2=A0 =C2=A0 =C2=A0 =C2=A0at org.apache.hadoop.ipc.Client.wrapException(= Client.java:751) > =C2=A0 =C2=A0 =C2=A0 =C2=A0at org.apache.hadoop.ipc.Client.call(Client.ja= va:719) > =C2=A0 =C2=A0 =C2=A0 =C2=A0at org.apache.hadoop.ipc.RPC$Invoker.invoke(RP= C.java:216) > =C2=A0 =C2=A0 =C2=A0 =C2=A0at org.apache.hadoop.dfs.$Proxy5.getProtocolVe= rsion(Unknown Source) > =C2=A0 =C2=A0 =C2=A0 =C2=A0at org.apache.hadoop.ipc.RPC.getProxy(RPC.java= :348) > =C2=A0 =C2=A0 =C2=A0 =C2=A0at org.apache.hadoop.dfs.DFSClient.createRPCNa= menode(DFSClient.java:103) > =C2=A0 =C2=A0 =C2=A0 =C2=A0at org.apache.hadoop.dfs.DFSClient.(DFSC= lient.java:172) > =C2=A0 =C2=A0 =C2=A0 =C2=A0at org.apache.hadoop.dfs.DistributedFileSystem= .initialize(DistributedFileSystem.java:67) > =C2=A0 =C2=A0 =C2=A0 =C2=A0at org.apache.hadoop.fs.FileSystem.createFileS= ystem(FileSystem.java:1339) > =C2=A0 =C2=A0 =C2=A0 =C2=A0at org.apache.hadoop.fs.FileSystem.access$300(= FileSystem.java:56) > =C2=A0 =C2=A0 =C2=A0 =C2=A0at org.apache.hadoop.fs.FileSystem$Cache.get(F= ileSystem.java:1351) > =C2=A0 =C2=A0 =C2=A0 =C2=A0at org.apache.hadoop.fs.FileSystem.get(FileSys= tem.java:213) > =C2=A0 =C2=A0 =C2=A0 =C2=A0at org.apache.hadoop.fs.Path.getFileSystem(Pat= h.java:175) > =C2=A0 =C2=A0 =C2=A0 =C2=A0at org.apache.hadoop.mapred.TaskTracker.locali= zeJob(TaskTracker.java:638) > =C2=A0 =C2=A0 =C2=A0 =C2=A0at org.apache.hadoop.mapred.TaskTracker.startN= ewTask(TaskTracker.java:1297) > =C2=A0 =C2=A0 =C2=A0 =C2=A0at org.apache.hadoop.mapred.TaskTracker.offerS= ervice(TaskTracker.java:937) > =C2=A0 =C2=A0 =C2=A0 =C2=A0at org.apache.hadoop.mapred.TaskTracker.run(Ta= skTracker.java:1334) > =C2=A0 =C2=A0 =C2=A0 =C2=A0at org.apache.hadoop.mapred.TaskTracker.main(T= askTracker.java:2343) > Caused by: java.io.EOFException > =C2=A0 =C2=A0 =C2=A0 =C2=A0at java.io.DataInputStream.readFully(DataInput= Stream.java:180) > =C2=A0 =C2=A0 =C2=A0 =C2=A0at java.io.DataInputStream.readFully(DataInput= Stream.java:152) > =C2=A0 =C2=A0 =C2=A0 =C2=A0at org.apache.hadoop.io.WritableUtils.readStri= ng(WritableUtils.java:115) > =C2=A0 =C2=A0 =C2=A0 =C2=A0at org.apache.hadoop.ipc.Client$Connection.rec= eiveResponse(Client.java:509) > =C2=A0 =C2=A0 =C2=A0 =C2=A0at org.apache.hadoop.ipc.Client$Connection.run= (Client.java:442) > > What's VERY odd about this, is I don't have a 208.67.216.132 node in > my system at all. =C2=A0Where is hadoop getting this IP from? > > My hadoop-site.xml is below: > > > =C2=A0 > =C2=A0 =C2=A0fs.default.name > =C2=A0 =C2=A0hdfs://10.0.6.110/ > =C2=A0 =C2=A0true > =C2=A0 > =C2=A0 > =C2=A0 =C2=A0mapred.job.tracker > =C2=A0 =C2=A010.0.6.110:8012 > =C2=A0 =C2=A0true > =C2=A0 > =C2=A0 > =C2=A0 =C2=A0hadoop.tmp.dir > =C2=A0 =C2=A0/opt/cluster/hadoop-tmp > =C2=A0 =C2=A0true > =C2=A0 > > > I've confirmed SSH works just fine, ping works, etc. =C2=A0HDFS works as = well. > > Does the above exception have any clues as to why I can't run a Pig > MapReduce job? > > Your help or tips are much appreciated, > Seth >