Return-Path: X-Original-To: apmail-hadoop-hdfs-user-archive@minotaur.apache.org Delivered-To: apmail-hadoop-hdfs-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 47267D9D5 for ; Sat, 10 Nov 2012 00:11:11 +0000 (UTC) Received: (qmail 13544 invoked by uid 500); 10 Nov 2012 00:11:05 -0000 Delivered-To: apmail-hadoop-hdfs-user-archive@hadoop.apache.org Received: (qmail 13452 invoked by uid 500); 10 Nov 2012 00:11:05 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hadoop.apache.org Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 13445 invoked by uid 99); 10 Nov 2012 00:11:05 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 10 Nov 2012 00:11:05 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=HTML_MESSAGE,NORMAL_HTTP_TO_IP,RCVD_IN_DNSWL_LOW,SPF_PASS,T_REMOTE_IMAGE,WEIRD_PORT X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of yinghua.hu@gmail.com designates 209.85.223.176 as permitted sender) Received: from [209.85.223.176] (HELO mail-ie0-f176.google.com) (209.85.223.176) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 10 Nov 2012 00:10:58 +0000 Received: by mail-ie0-f176.google.com with SMTP id k11so7693169iea.35 for ; Fri, 09 Nov 2012 16:10:37 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :content-type; bh=ysup8bagrs3DpM4yVmv/xMGkJk48Q2NhOqnLlPv3LnA=; b=ilvsyW9oTjUZzL/Tqod9hENxOy3DwuyeEw1LzGZ3wWIiTQhyVOI8azXV1oavPpYk5+ q2fXKB0u2/aXRFkbMbdiQRKm3IFCOnAhZNAMePG9o5SBzYkP6Xk4YxV/EGkRfimfEDld 6aovDZzPTqSvMuJihZq70bMomEvDkmC/pwLvbWfo+VVL0ZSK92DgBnU4nkZYBrjxb9CC CSnDhGPpyU3IgJpVES5h9ortSgtVzo+WJkHoN8ePZDyQCZtoqaZtDbMbuNvpkRqOGlDg ALWjrlUJ0eUiILwwiUvsaxLJnnIy7w335RQ+CLNv1bFr1pzhutfmm3+jdToK8Mm722fN e2Sg== MIME-Version: 1.0 Received: by 10.43.58.14 with SMTP id wi14mr12257505icb.9.1352506237604; Fri, 09 Nov 2012 16:10:37 -0800 (PST) Received: by 10.64.12.66 with HTTP; Fri, 9 Nov 2012 16:10:37 -0800 (PST) In-Reply-To: References: Date: Fri, 9 Nov 2012 16:10:37 -0800 Message-ID: Subject: Re: Erro running pi programm From: yinghua hu To: user@hadoop.apache.org Content-Type: multipart/alternative; boundary=bcaec517a9b2d9520a04ce18e5e0 X-Virus-Checked: Checked by ClamAV on apache.org --bcaec517a9b2d9520a04ce18e5e0 Content-Type: text/plain; charset=windows-1252 Content-Transfer-Encoding: quoted-printable Hi, Andy and Ted Thanks a lot for pointing out my silly error. The problem did go away after I change the hostname from 'master' to the result of =93hostname =96f=94. Since I shutdown the EC2 instance every night, I thought that using 'master','slave1','slave2' will save typing after the full host name change with reboot. But it did bring problems. Thanks! On Fri, Nov 9, 2012 at 12:43 PM, Ted Reynolds wrote: > yinghua, > > One other thing to note here is that on EC2 computer are given two > hostnames, one internal and one external, you need to use the external. > These host names are listed in the ec2 configuration pane on the AWS > management console. > > > Ted Reynolds > Technical Support Engineer > Hortonworks > Work Phone: 408-645-7079 > > > > > > > On Fri, Nov 9, 2012 at 11:31 AM, Kartashov, Andy = wrote: > >> Try running =93hostname =96f=94 on each node, take a note of the fully >> qualified host address and replace your =93master=94 with the your respe= ctive >> finding. >> >> >> >> Here are my configuration files >> >> >> >> core-site.xml >> >> >> >> >> >> >> >> fs.default.name >> >> hdfs://master:9000 >> >> >> >> >> >> >> hadoop.tmp.dir >> >> >> /usr/local/hadoop/tmp >> >> >> >> >> >> >> >> >> >> >> >> mapred-site.xml >> >> >> >> >> >> >> >> mapreduce.framework.name >> >> yarn >> >> >> >> >> >> >> >> >> >> hdfs-site.xml >> >> >> >> >> >> >> >> dfs.replication >> >> 2 >> >> >> >> >> >> >> dfs.permissions >> >> >> false >> >> >> >> >> >> >> >> >> dfs.namenode.name.dir >> >> >> file:/home/hduser/yarn_data/hdfs/namenode >> >> >> >> >> >> >> >> >> dfs.datanode.data.dir >> >> >> file:/home/hduser/yarn_data/hdfs/datanode >> >> >> >> >> >> >> >> >> >> yarn-site.xml >> >> >> >> >> >> >> >> >> >> yarn.nodemanager.aux-services >> >> >> >> mapreduce.shuffle >> >> >> >> >> >> yarn.nodemanager.aux-services.mapreduce.shuffle.class >> >> org.apache.hadoop.mapred.ShuffleHandler >> >> >> >> >> >> yarn.nodemanager.log-aggregation-enable >> >> true >> >> >> >> >> >> yarn.resourcemanager.resource-tracker.address >> >> master:8050 >> >> >> >> >> >> yarn.resourcemanager.scheduler.address >> >> master:8030 >> >> >> >> >> >> yarn.resourcemanager.scheduler.class >> >> >> >> org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.= CapacityScheduler >> >> >> >> >> >> >> >> >> >> yarn.resourcemanager.address >> >> >> master:60400 >> >> >> >> >> >> >> >> >> >> >> >> >> On Fri, Nov 9, 2012 at 9:51 AM, yinghua hu wrote: >> >> Hi, Andy >> >> >> >> Thanks for suggestions! >> >> >> >> I am running it on a four node cluster on EC2. All the services started >> fine, Namenode, Datanode, ResourceManager, NodeManager and >> JobHistoryServer. Each node can ssh to all the nodes without problem. >> >> >> >> But problem appears when trying to run any job. >> >> >> >> >> >> >> >> *From:* Kartashov, Andy >> *Sent:* Friday, November 09, 2012 12:37 PM >> *To:* user@hadoop.apache.org >> *Subject:* Erro running pi programm >> >> >> >> Yinghua, >> >> >> >> What mode are you running your hadoop in: Local/Pseud/Fully...? >> >> >> >> Your hostname is not recognised >> >> >> >> Your configuration setting seems to be wrong. >> >> >> >> >> >> >> >> >> >> >> >> Hi, all >> >> >> >> Could some help looking at this problem? I am setting up a four node >> cluster on EC2 and seems that the cluster is set up fine until I start >> testing. >> >> >> >> I have tried password-less ssh from each node to all the nodes and there >> is no problem connecting. Any advice will be greatly appreciated! >> >> >> >> [hduser@ip-XX-XX-XXX-XXX hadoop]$ bin/hadoop jar >> share/hadoop/mapreduce/hadoop-mapreduce-examples-0.23.4.jar pi - >> Dmapreduce.clientfactory.class.name=3Dorg.apache.hadoop.mapred.YarnClien= tFactory >> -libjars >> share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-0.23.4.jar 16 1= 0000 >> >> Number of Maps =3D 16 >> >> Samples per Map =3D 10000 >> >> Wrote input for Map #0 >> >> Wrote input for Map #1 >> >> Wrote input for Map #2 >> >> Wrote input for Map #3 >> >> Wrote input for Map #4 >> >> Wrote input for Map #5 >> >> Wrote input for Map #6 >> >> Wrote input for Map #7 >> >> Wrote input for Map #8 >> >> Wrote input for Map #9 >> >> Wrote input for Map #10 >> >> Wrote input for Map #11 >> >> Wrote input for Map #12 >> >> Wrote input for Map #13 >> >> Wrote input for Map #14 >> >> Wrote input for Map #15 >> >> Starting Job >> >> 12/11/09 12:02:59 INFO input.FileInputFormat: Total input paths to >> process : 16 >> >> 12/11/09 12:02:59 INFO mapreduce.JobSubmitter: number of splits:16 >> >> 12/11/09 12:02:59 WARN conf.Configuration: mapred.job.classpath.files is >> deprecated. Instead, use mapreduce.job.classpath.files >> >> 12/11/09 12:02:59 WARN conf.Configuration: mapred.jar is deprecated. >> Instead, use mapreduce.job.jar >> >> 12/11/09 12:02:59 WARN conf.Configuration: mapred.cache.files is >> deprecated. Instead, use mapreduce.job.cache.files >> >> 12/11/09 12:02:59 WARN conf.Configuration: >> mapred.map.tasks.speculative.execution is deprecated. Instead, use >> mapreduce.map.speculative >> >> 12/11/09 12:02:59 WARN conf.Configuration: mapred.reduce.tasks is >> deprecated. Instead, use mapreduce.job.reduces >> >> 12/11/09 12:02:59 WARN conf.Configuration: mapred.output.value.class is >> deprecated. Instead, use mapreduce.job.output.value.class >> >> 12/11/09 12:02:59 WARN conf.Configuration: >> mapred.reduce.tasks.speculative.execution is deprecated. Instead, use >> mapreduce.reduce.speculative >> >> 12/11/09 12:02:59 WARN conf.Configuration: >> mapred.used.genericoptionsparser is deprecated. Instead, use >> mapreduce.client.genericoptionsparser.used >> >> 12/11/09 12:02:59 WARN conf.Configuration: mapreduce.map.class is >> deprecated. Instead, use mapreduce.job.map.class >> >> 12/11/09 12:02:59 WARN conf.Configuration: mapred.job.name is >> deprecated. Instead, use mapreduce.job.name >> >> 12/11/09 12:02:59 WARN conf.Configuration: mapreduce.reduce.class is >> deprecated. Instead, use mapreduce.job.reduce.class >> >> 12/11/09 12:02:59 WARN conf.Configuration: mapreduce.inputformat.class i= s >> deprecated. Instead, use mapreduce.job.inputformat.class >> >> 12/11/09 12:02:59 WARN conf.Configuration: mapred.input.dir is >> deprecated. Instead, use mapreduce.input.fileinputformat.inputdir >> >> 12/11/09 12:02:59 WARN conf.Configuration: mapred.output.dir is >> deprecated. Instead, use mapreduce.output.fileoutputformat.outputdir >> >> 12/11/09 12:02:59 WARN conf.Configuration: mapreduce.outputformat.class >> is deprecated. Instead, use mapreduce.job.outputformat.class >> >> 12/11/09 12:02:59 WARN conf.Configuration: mapred.map.tasks is >> deprecated. Instead, use mapreduce.job.maps >> >> 12/11/09 12:02:59 WARN conf.Configuration: mapred.cache.files.timestamps >> is deprecated. Instead, use mapreduce.job.cache.files.timestamps >> >> 12/11/09 12:02:59 WARN conf.Configuration: mapred.output.key.class is >> deprecated. Instead, use mapreduce.job.output.key.class >> >> 12/11/09 12:02:59 WARN conf.Configuration: mapred.working.dir is >> deprecated. Instead, use mapreduce.job.working.dir >> >> 12/11/09 12:03:00 INFO mapred.ResourceMgrDelegate: Submitted application >> application_1352478937343_0002 to ResourceManager at master/ >> 10.12.181.233:60400 >> >> 12/11/09 12:03:00 INFO mapreduce.Job: The url to track the job: >> http://master:8088/proxy/application_1352478937343_0002/ >> >> 12/11/09 12:03:00 INFO mapreduce.Job: Running job: job_1352478937343_000= 2 >> >> 12/11/09 12:03:01 INFO mapreduce.Job: Job job_1352478937343_0002 running >> in uber mode : false >> >> 12/11/09 12:03:01 INFO mapreduce.Job: map 0% reduce 0% >> >> 12/11/09 12:03:01 INFO mapreduce.Job: Job job_1352478937343_0002 failed >> with state FAILED due to: Application application_1352478937343_0002 fai= led >> 1 times due to Error launching appattempt_1352478937343_0002_000001. Got >> exception: java.lang.reflect.UndeclaredThrowableException >> >> at >> org.apache.hadoop.yarn.api.impl.pb.client.ContainerManagerPBClientImpl.s= tartContainer(ContainerManagerPBClientImpl.java:111) >> >> at >> org.apache.hadoop.yarn.server.resourcemanager.amlauncher.AMLauncher.laun= ch(AMLauncher.java:115) >> >> at >> org.apache.hadoop.yarn.server.resourcemanager.amlauncher.AMLauncher.run(= AMLauncher.java:258) >> >> at >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.jav= a:1110) >> >> at >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.ja= va:603) >> >> at java.lang.Thread.run(Thread.java:722) >> >> Caused by: com.google.protobuf.ServiceException: >> java.net.UnknownHostException: Yinghua java.net.UnknownHostException; Fo= r >> more details see: http://wiki.apache.org/hadoop/UnknownHost >> >> at >> org.apache.hadoop.yarn.ipc.ProtoOverHadoopRpcEngine$Invoker.invoke(Proto= OverHadoopRpcEngine.java:144) >> >> at $Proxy24.startContainer(Unknown Source) >> >> at >> org.apache.hadoop.yarn.api.impl.pb.client.ContainerManagerPBClientImpl.s= tartContainer(ContainerManagerPBClientImpl.java:104) >> >> ... 5 more >> >> Caused by: java.net.UnknownHostException: Yinghua For more details see: >> http://wiki.apache.org/hadoop/UnknownHost >> >> at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:71= 3) >> >> at org.apache.hadoop.ipc.Client$Connection.(Client.java:23= 6) >> >> at org.apache.hadoop.ipc.Client.getConnection(Client.java:1188) >> >> at org.apache.hadoop.ipc.Client.call(Client.java:1068) >> >> at >> org.apache.hadoop.yarn.ipc.ProtoOverHadoopRpcEngine$Invoker.invoke(Proto= OverHadoopRpcEngine.java:141) >> >> ... 7 more >> >> Caused by: java.net.UnknownHostException >> >> ... 11 more >> >> . Failing the application. >> >> 12/11/09 12:03:01 INFO mapreduce.Job: Counters: 0 >> >> Job Finished in 2.672 seconds >> >> java.io.FileNotFoundException: File does not exist: >> hdfs://master:9000/user/hduser/QuasiMonteCarlo_TMP_3_141592654/out/reduc= e-out >> >> at >> org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFi= leSystem.java:738) >> >> at >> org.apache.hadoop.io.SequenceFile$Reader.(SequenceFile.java:1685) >> >> at >> org.apache.hadoop.io.SequenceFile$Reader.(SequenceFile.java:1709) >> >> at >> org.apache.hadoop.examples.QuasiMonteCarlo.estimatePi(QuasiMonteCarlo.ja= va:314) >> >> at >> org.apache.hadoop.examples.QuasiMonteCarlo.run(QuasiMonteCarlo.java:351) >> >> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:69) >> >> at >> org.apache.hadoop.examples.QuasiMonteCarlo.main(QuasiMonteCarlo.java:360= ) >> >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> >> at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.jav= a:57) >> >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessor= Impl.java:43) >> >> at java.lang.reflect.Method.invoke(Method.java:601) >> >> at >> org.apache.hadoop.util.ProgramDriver$ProgramDescription.invoke(ProgramDr= iver.java:72) >> >> at >> org.apache.hadoop.util.ProgramDriver.driver(ProgramDriver.java:144) >> >> at >> org.apache.hadoop.examples.ExampleDriver.main(ExampleDriver.java:68) >> >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> >> at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.jav= a:57) >> >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessor= Impl.java:43) >> >> at java.lang.reflect.Method.invoke(Method.java:601) >> >> at org.apache.hadoop.util.RunJar.main(RunJar.java:208) >> >> >> >> >> >> -- >> Regards, >> >> Yinghua >> >> >> NOTICE: This e-mail message and any attachments are confidential, >> subject to copyright and may be privileged. Any unauthorized use, copyin= g >> or disclosure is prohibited. If you are not the intended recipient, plea= se >> delete and contact the sender immediately. Please consider the environme= nt >> before printing this e-mail. AVIS : le pr=E9sent courriel et toute pi=E8= ce >> jointe qui l'accompagne sont confidentiels, prot=E9g=E9s par le droit d'= auteur >> et peuvent =EAtre couverts par le secret professionnel. Toute utilisatio= n, >> copie ou divulgation non autoris=E9e est interdite. Si vous n'=EAtes pas= le >> destinataire pr=E9vu de ce courriel, supprimez-le et contactez imm=E9dia= tement >> l'exp=E9diteur. Veuillez penser =E0 l'environnement avant d'imprimer le = pr=E9sent >> courriel >> > > --=20 Regards, Yinghua --bcaec517a9b2d9520a04ce18e5e0 Content-Type: text/html; charset=windows-1252 Content-Transfer-Encoding: quoted-printable Hi, Andy and Ted

Thanks a lot for pointing out my silly = error. The problem did go away after I change the hostname from 'master= ' to the result of=A0=93hostname =96f=94.=A0

= Since I shutdown the EC2 instance every night, I thought that using 'ma= ster','slave1','slave2' will save typing after the full= host name change with reboot. But it did bring problems.=A0

= Thanks!

On Fri, Nov 9, 2012 at 12:43 PM, Ted Re= ynolds <tedr@hortonworks.com> wrote:
yinghua,

One other thing to note here is that on EC2 com= puter are given two hostnames, one internal and one external, you need to u= se the external. =A0These host names are listed in the ec2 configuration pa= ne on the AWS management console.


Ted Reynolds
Technical= Support Engineer
Hortonworks
Work Phone: 408-645-7079





On Fri, Nov 9, 2012 at 11:31 AM, Kartash= ov, Andy <Andy.Kartashov@mpac.ca> wrote:

Try running =93hostname = =96f=94 on each node, take a note of the fully qualified host address and r= eplace your =93master=94 with the your respective finding.

=A0

Here are my configuration files

=A0

core-site.xml

=A0

<configuration>

=A0 =A0 =A0 =A0 <property>

=A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 <name>fs.default.name</name&= gt;

=A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 <value>hdfs://= master:9000</value>

=A0 =A0 =A0 =A0 </property>

=A0 =A0 =A0 =A0 <property> =A0 =A0 =A0 =A0 =A0= =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0=A0<= /p>

=A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 <name>hadoop.t= mp.dir</name> =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0= =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0

=A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 <value>/usr/lo= cal/hadoop/tmp</value> =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0=A0

=A0 =A0 =A0 =A0 </property> =A0 =A0 =A0 =A0 = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0= =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0

</configuration> =A0 =A0

=A0

=A0

mapred-site.xml

=A0

<configuration>

=A0 =A0 =A0 =A0 <property>

=A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 <name>mapreduce.framework.= name</name>

=A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 <value>yarn<= ;/value>

=A0 =A0 =A0 =A0 </property>

=A0

</configuration>

=A0

hdfs-site.xml

=A0

<configuration>

=A0 =A0 =A0 =A0 <property>

=A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 <name>dfs.repl= ication</name>

=A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 <value>2</v= alue>

=A0 =A0 =A0 =A0 </property>

=A0 =A0 =A0 =A0 =A0<property> =A0 =A0 =A0 =A0 = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0= =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0

=A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 <name>dfs.perm= issions</name> =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0=A0

=A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 <value>false&l= t;/value> =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0=A0

=A0 =A0 =A0 =A0 </property> =A0 =A0 =A0 =A0 = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0= =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0

=A0 =A0 =A0 =A0 <property> =A0 =A0 =A0 =A0 =A0= =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0=A0<= /p>

=A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 <name>dfs.name= node.name.dir</name> =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0=A0

=A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 <value>file:/h= ome/hduser/yarn_data/hdfs/namenode</value> =A0 =A0 =A0 =A0 =A0 =A0 = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0=A0

=A0 =A0 =A0 =A0 </property> =A0 =A0 =A0 =A0 = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0= =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0

=A0 =A0 =A0 =A0 <property> =A0 =A0 =A0 =A0 =A0= =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0=A0<= /p>

=A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 <name>dfs.data= node.data.dir</name> =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0=A0

=A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 <value>file:/h= ome/hduser/yarn_data/hdfs/datanode</value> =A0 =A0 =A0 =A0 =A0 =A0 = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0=A0

=A0 =A0 =A0 =A0 </property> =A0 =A0 =A0 =A0 = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0= =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0

</configuration> =A0

=A0

yarn-site.xml

=A0

<?xml version=3D"1.0"?>

<configuration>

=A0<property>

=A0<name>yarn.nodemanager.aux-services</nam= e>

=A0

=A0 <value>mapreduce.shuffle</value>

=A0</property>

=A0<property>

=A0 =A0 =A0 =A0 <name>yarn.nodemanager.aux-ser= vices.mapreduce.shuffle.class</name>

=A0 =A0 =A0 =A0 <value>org.apache.hadoop.mapre= d.ShuffleHandler</value>

=A0</property>

=A0<property>

=A0 =A0 =A0 =A0 <name>yarn.nodemanager.log-agg= regation-enable</name>

=A0 =A0 =A0 =A0 <value>true</value>

=A0</property>

<property>

=A0 =A0 =A0 =A0 <name>yarn.resourcemanager.res= ource-tracker.address</name>

=A0 =A0 =A0 =A0 <value>master:8050</value&g= t;

</property>

<property>

=A0 =A0 =A0 =A0 <name>yarn.resourcemanager.sch= eduler.address</name>

=A0 =A0 =A0 =A0 <value>master:8030</value&g= t;

</property>

<property>

=A0 =A0 =A0 =A0 <name>yarn.resourcemanager.sch= eduler.class</name> =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0

=A0 =A0 =A0 =A0 <value>org.apache.hadoop.yarn.= server.resourcemanager.scheduler.capacity.CapacityScheduler</value> = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0= =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0=A0

</property> =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0= =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0

<property> =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0= =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0=A0<= /p>

=A0 =A0 <name>yarn.resourcemanager.address<= /name> =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0

=A0 =A0<value>master:60400</value> =A0 = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0= =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0=A0

=A0</property> =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0= =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0=A0<= /p>

=A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 = =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0= =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 =A0 = =A0

</configuration>

=A0

On Fri, Nov 9, 2012 at 9:51 AM, yinghua hu <yinghua.hu@gmail.com> wrote:

Hi, Andy

=A0

Thanks for suggestions!

=A0

I am running it on a four node cluster on EC2. All t= he services started fine, Namenode, Datanode, ResourceManager, NodeManager = and JobHistoryServer.=A0Each node can ssh to all the nodes without problem.= =A0

=A0

But problem appears when trying to run any job.=A0

=A0

=A0

=A0

=A0

Yinghua,

=A0

What mode are you running your hadoop in: Local/Pseu= d/Fully...?

=A0

Your hostname is not recognised

=A0

Your configuration setting seems to be wrong.

=A0

=A0

=A0

=A0

=A0

Hi, all

=A0

Could some help looking at this problem? I am settin= g up a four node cluster on EC2 and seems that the cluster is set up fine u= ntil I start testing.

=A0

I have tried password-less ssh from each node to all= the nodes and there is no problem connecting. Any advice will be greatly= =A0appreciated!

=A0

[hduser@ip-XX-XX-XXX-XXX hadoop]$ bin/hadoop jar sha= re/hadoop/mapreduce/hadoop-mapreduce-examples-0.23.4.jar pi -Dmapreduce.clien= tfactory.class.name=3Dorg.apache.hadoop.mapred.YarnClientFactory -libjars share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-0.23.4.j= ar 16 10000

Number of Maps =A0=3D 16

Samples per Map =3D 10000

Wrote input for Map #0

Wrote input for Map #1

Wrote input for Map #2

Wrote input for Map #3

Wrote input for Map #4

Wrote input for Map #5

Wrote input for Map #6

Wrote input for Map #7

Wrote input for Map #8

Wrote input for Map #9

Wrote input for Map #10

Wrote input for Map #11

Wrote input for Map #12

Wrote input for Map #13

Wrote input for Map #14

Wrote input for Map #15

Starting Job

12/11/09 12:02:59 INFO input.FileInputFormat: Total = input paths to process : 16

12/11/09 12:02:59 INFO mapreduce.JobSubmitter: numbe= r of splits:16

12/11/09 12:02:59 WARN conf.Configuration: mapred.jo= b.classpath.files is deprecated. Instead, use mapreduce.job.classpath.files=

12/11/09 12:02:59 WARN conf.Configuration: mapred.ja= r is deprecated. Instead, use mapreduce.job.jar

12/11/09 12:02:59 WARN conf.Configuration: mapred.ca= che.files is deprecated. Instead, use mapreduce.job.cache.files

12/11/09 12:02:59 WARN conf.Configuration: mapred.ma= p.tasks.speculative.execution is deprecated. Instead, use mapreduce.map.spe= culative

12/11/09 12:02:59 WARN conf.Configuration: mapred.re= duce.tasks is deprecated. Instead, use mapreduce.job.reduces

12/11/09 12:02:59 WARN conf.Configuration: mapred.ou= tput.value.class is deprecated. Instead, use mapreduce.job.output.value.cla= ss

12/11/09 12:02:59 WARN conf.Configuration: mapred.re= duce.tasks.speculative.execution is deprecated. Instead, use mapreduce.redu= ce.speculative

12/11/09 12:02:59 WARN conf.Configuration: mapred.us= ed.genericoptionsparser is deprecated. Instead, use mapreduce.client.generi= coptionsparser.used

12/11/09 12:02:59 WARN conf.Configuration: mapreduce= .map.class is deprecated. Instead, use mapreduce.job.map.class

12/11/09 12:02:59 WARN conf.Configuration: mapred.job.name is deprecated. Instead, use mapreduce.job.name

12/11/09 12:02:59 WARN conf.Configuration: mapreduce= .reduce.class is deprecated. Instead, use mapreduce.job.reduce.class

12/11/09 12:02:59 WARN conf.Configuration: mapreduce= .inputformat.class is deprecated. Instead, use mapreduce.job.inputformat.cl= ass

12/11/09 12:02:59 WARN conf.Configuration: mapred.in= put.dir is deprecated. Instead, use mapreduce.input.fileinputformat.inputdi= r

12/11/09 12:02:59 WARN conf.Configuration: mapred.ou= tput.dir is deprecated. Instead, use mapreduce.output.fileoutputformat.outp= utdir

12/11/09 12:02:59 WARN conf.Configuration: mapreduce= .outputformat.class is deprecated. Instead, use mapreduce.job.outputformat.= class

12/11/09 12:02:59 WARN conf.Configuration: mapred.ma= p.tasks is deprecated. Instead, use mapreduce.job.maps

12/11/09 12:02:59 WARN conf.Configuration: mapred.ca= che.files.timestamps is deprecated. Instead, use mapreduce.job.cache.files.= timestamps

12/11/09 12:02:59 WARN conf.Configuration: mapred.ou= tput.key.class is deprecated. Instead, use mapreduce.job.output.key.class

12/11/09 12:02:59 WARN conf.Configuration: mapred.wo= rking.dir is deprecated. Instead, use mapreduce.job.working.dir

12/11/09 12:03:00 INFO mapred.ResourceMgrDelegate: S= ubmitted application application_1352478937343_0002 to ResourceManager at m= aster/10.12.181.23= 3:60400

12/11/09 12:03:00 INFO mapreduce.Job: The url to tra= ck the job: http://master:8088/proxy/application_1352478937343_0002/

12/11/09 12:03:00 INFO mapreduce.Job: Running job: j= ob_1352478937343_0002

12/11/09 12:03:01 INFO mapreduce.Job: Job job_135247= 8937343_0002 running in uber mode : false

12/11/09 12:03:01 INFO mapreduce.Job: =A0map 0% redu= ce 0%

12/11/09 12:03:01 INFO mapreduce.Job: Job job_135247= 8937343_0002 failed with state FAILED due to: Application application_13524= 78937343_0002 failed 1 times due to Error launching appattempt_135247893734= 3_0002_000001. Got exception: java.lang.reflect.UndeclaredThrowableExceptio= n

=A0 =A0 =A0 =A0 at org.apache.hadoop.yarn.api.impl.p= b.client.ContainerManagerPBClientImpl.startContainer(ContainerManagerPBClie= ntImpl.java:111)

=A0 =A0 =A0 =A0 at org.apache.hadoop.yarn.server.res= ourcemanager.amlauncher.AMLauncher.launch(AMLauncher.java:115)

=A0 =A0 =A0 =A0 at org.apache.hadoop.yarn.server.res= ourcemanager.amlauncher.AMLauncher.run(AMLauncher.java:258)

=A0 =A0 =A0 =A0 at java.util.concurrent.ThreadPoolEx= ecutor.runWorker(ThreadPoolExecutor.java:1110)

=A0 =A0 =A0 =A0 at java.util.concurrent.ThreadPoolEx= ecutor$Worker.run(ThreadPoolExecutor.java:603)

=A0 =A0 =A0 =A0 at java.lang.Thread.run(Thread.java:= 722)

Caused by: com.google.protobuf.ServiceException: jav= a.net.UnknownHostException: Yinghua java.net.UnknownHostException; For more= details see: =A0http://wiki.apache.org/hadoop/UnknownHost

=A0 =A0 =A0 =A0 at org.apache.hadoop.yarn.ipc.ProtoO= verHadoopRpcEngine$Invoker.invoke(ProtoOverHadoopRpcEngine.java:144)

=A0 =A0 =A0 =A0 at $Proxy24.startContainer(Unknown S= ource)

=A0 =A0 =A0 =A0 at org.apache.hadoop.yarn.api.impl.p= b.client.ContainerManagerPBClientImpl.startContainer(ContainerManagerPBClie= ntImpl.java:104)

=A0 =A0 =A0 =A0 ... 5 more

Caused by: java.net.UnknownHostException: Yinghua Fo= r more details see: =A0http://wiki.apache.org/hadoop/UnknownHost

=A0 =A0 =A0 =A0 at org.apache.hadoop.net.NetUtils.wr= apException(NetUtils.java:713)

=A0 =A0 =A0 =A0 at org.apache.hadoop.ipc.Client$Conn= ection.<init>(Client.java:236)

=A0 =A0 =A0 =A0 at org.apache.hadoop.ipc.Client.getC= onnection(Client.java:1188)

=A0 =A0 =A0 =A0 at org.apache.hadoop.ipc.Client.call= (Client.java:1068)

=A0 =A0 =A0 =A0 at org.apache.hadoop.yarn.ipc.ProtoO= verHadoopRpcEngine$Invoker.invoke(ProtoOverHadoopRpcEngine.java:141)

=A0 =A0 =A0 =A0 ... 7 more

Caused by: java.net.UnknownHostException

=A0 =A0 =A0 =A0 ... 11 more

. Failing the application.

12/11/09 12:03:01 INFO mapreduce.Job: Counters: 0

Job Finished in 2.672 seconds

java.io.FileNotFoundException: File does not exist: = hdfs://master:9000/user/hduser/QuasiMonteCarlo_TMP_3_141592654/out/reduce-o= ut

=A0 =A0 =A0 =A0 at org.apache.hadoop.hdfs.Distribute= dFileSystem.getFileStatus(DistributedFileSystem.java:738)

=A0 =A0 =A0 =A0 at org.apache.hadoop.io.SequenceFile= $Reader.<init>(SequenceFile.java:1685)

=A0 =A0 =A0 =A0 at org.apache.hadoop.io.SequenceFile= $Reader.<init>(SequenceFile.java:1709)

=A0 =A0 =A0 =A0 at org.apache.hadoop.examples.QuasiM= onteCarlo.estimatePi(QuasiMonteCarlo.java:314)

=A0 =A0 =A0 =A0 at org.apache.hadoop.examples.QuasiM= onteCarlo.run(QuasiMonteCarlo.java:351)

=A0 =A0 =A0 =A0 at org.apache.hadoop.util.ToolRunner= .run(ToolRunner.java:69)

=A0 =A0 =A0 =A0 at org.apache.hadoop.examples.QuasiM= onteCarlo.main(QuasiMonteCarlo.java:360)

=A0 =A0 =A0 =A0 at sun.reflect.NativeMethodAccessorI= mpl.invoke0(Native Method)

=A0 =A0 =A0 =A0 at sun.reflect.NativeMethodAccessorI= mpl.invoke(NativeMethodAccessorImpl.java:57)

=A0 =A0 =A0 =A0 at sun.reflect.DelegatingMethodAcces= sorImpl.invoke(DelegatingMethodAccessorImpl.java:43)

=A0 =A0 =A0 =A0 at java.lang.reflect.Method.invoke(M= ethod.java:601)

=A0 =A0 =A0 =A0 at org.apache.hadoop.util.ProgramDri= ver$ProgramDescription.invoke(ProgramDriver.java:72)

=A0 =A0 =A0 =A0 at org.apache.hadoop.util.ProgramDri= ver.driver(ProgramDriver.java:144)

=A0 =A0 =A0 =A0 at org.apache.hadoop.examples.Exampl= eDriver.main(ExampleDriver.java:68)

=A0 =A0 =A0 =A0 at sun.reflect.NativeMethodAccessorI= mpl.invoke0(Native Method)

=A0 =A0 =A0 =A0 at sun.reflect.NativeMethodAccessorI= mpl.invoke(NativeMethodAccessorImpl.java:57)

=A0 =A0 =A0 =A0 at sun.reflect.DelegatingMethodAcces= sorImpl.invoke(DelegatingMethodAccessorImpl.java:43)

=A0 =A0 =A0 =A0 at java.lang.reflect.Method.invoke(M= ethod.java:601)

=A0 =A0 =A0 =A0 at org.apache.hadoop.util.RunJar.mai= n(RunJar.java:208)

=A0

=A0

--
Regards,

Yinghua

=A0

NOTICE: This e-mail message and any attachments are confidential, subject t= o copyright and may be privileged. Any unauthorized use, copying or disclos= ure is prohibited. If you are not the intended recipient, please delete and= contact the sender immediately. Please consider the environment before printing this e-mail. AVIS : le pr= =E9sent courriel et toute pi=E8ce jointe qui l'accompagne sont confiden= tiels, prot=E9g=E9s par le droit d'auteur et peuvent =EAtre couverts pa= r le secret professionnel. Toute utilisation, copie ou divulgation non autoris=E9e est interdite. Si vous n'=EAtes pas le = destinataire pr=E9vu de ce courriel, supprimez-le et contactez imm=E9diatem= ent l'exp=E9diteur. Veuillez penser =E0 l'environnement avant d'= ;imprimer le pr=E9sent courriel




--
Regards,
=
Yinghua
--bcaec517a9b2d9520a04ce18e5e0--