Return-Path: X-Original-To: apmail-spark-user-archive@minotaur.apache.org Delivered-To: apmail-spark-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 6D2AB10FB1 for ; Tue, 19 Nov 2013 15:56:10 +0000 (UTC) Received: (qmail 80951 invoked by uid 500); 19 Nov 2013 15:56:10 -0000 Delivered-To: apmail-spark-user-archive@spark.apache.org Received: (qmail 80762 invoked by uid 500); 19 Nov 2013 15:56:09 -0000 Mailing-List: contact user-help@spark.incubator.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@spark.incubator.apache.org Delivered-To: mailing list user@spark.incubator.apache.org Received: (qmail 80747 invoked by uid 99); 19 Nov 2013 15:56:09 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 19 Nov 2013 15:56:09 +0000 X-ASF-Spam-Status: No, hits=2.5 required=5.0 tests=HTML_FONT_FACE_BAD,HTML_MESSAGE,RCVD_IN_DNSWL_NONE,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: local policy) Received: from [98.138.90.159] (HELO nm11-vm2.bullet.mail.ne1.yahoo.com) (98.138.90.159) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 19 Nov 2013 15:56:03 +0000 Received: from [98.138.100.112] by nm11.bullet.mail.ne1.yahoo.com with NNFMP; 19 Nov 2013 15:55:42 -0000 Received: from [98.138.87.10] by tm103.bullet.mail.ne1.yahoo.com with NNFMP; 19 Nov 2013 15:55:41 -0000 Received: from [127.0.0.1] by omp1010.mail.ne1.yahoo.com with NNFMP; 19 Nov 2013 15:55:41 -0000 X-Yahoo-Newman-Property: ymail-3 X-Yahoo-Newman-Id: 968693.83071.bm@omp1010.mail.ne1.yahoo.com Received: (qmail 48168 invoked by uid 60001); 19 Nov 2013 15:55:41 -0000 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=yahoo.com; s=s1024; t=1384876541; bh=452lZzMVLGTXCF83hcsg7kdML9OG4oQyAYd3tg0QkDo=; h=X-YMail-OSG:Received:X-Rocket-MIMEInfo:X-Mailer:References:Message-ID:Date:From:Reply-To:Subject:To:In-Reply-To:MIME-Version:Content-Type; b=lXkuG+YZ2sIs/UUHdd1xw5lUoDXfSePgKaGwMU+Z45/jMIuevIc5bKlxIQSWOOa56+w4HUpynlvROYaYLDZmQivvqlB82S0HLWpmsFQbhK8+5QF6+/vniJigKmIUVMVS3I2vepSIB9GSjrkkfouAUyZQ6dtj7GvjMH8UgElJ10E= DomainKey-Signature: a=rsa-sha1; q=dns; c=nofws; s=s1024; d=yahoo.com; h=X-YMail-OSG:Received:X-Rocket-MIMEInfo:X-Mailer:References:Message-ID:Date:From:Reply-To:Subject:To:In-Reply-To:MIME-Version:Content-Type; b=l/r5n8Q2eWoxDvJWhLaB7YayYdNjRBn5mUmgKA5XIALXEmRmRNs7t1AWHIjdJx5sA1Q2DFVI/rT0hmlhAvVQ/fTV9T0lY4Rew9TzPADs3L8HBs/W3gugeX1/TZD9TBlyyUMWgiyLlQMbNQv6R1wA5HSyXM8z3a+IFtnArATtUVI=; X-YMail-OSG: I50wlZIVM1mOKZXeYdIiIlWQhMPYPYEDBsUrVUOaMXsa.iT rXA_cye6QTNT_175gtIB30qu46syzxWFj8X0moHdzno4npqchhtHwYXZbNzK tvd18K3suckSJFCBg_6_NnW.Je6kiHQttZKjG7iXIm.FgaShSOgYP.8e9XKp 0tj2lSfS4OwPI2smqkGdrXsh3EzlVIMzsXClt1H6L1IIb9SiQ0aXsb21Oa0n _vqlLO8V.q12RxVy6K.Z7A3gTqAkagAhQbqgZMgGj_XNhyl_0wELuaW8tw9A tOA7pi8Pxbz1L1jH85snJdQJofSwuRkHOAT0n8OhMCUlWqz6QjkQfw2r9OBb gu6C56O8FjxS1_XZX2UHrosfVU921C7i6o.vDVlTejCdChnETvKNX0LvWWbs K2swQ_ravNCNs6NDDzwwFHY_ZUrM0PjsbC24aj4A94Rzss0kIlB82BZYiWGV nFac1XUVlyeXPqXNimZAeGjIwuRoJeVelQRyxv1g_Z.YkIrTfmmvQpWymgTn Pcj0PYDz47nyccwVRlDdpFAy7I.WCLBQMJ1O0pVR.U8vbgO0yjPpRnginqlH LyiYoZusV8bMuxD.Y7b_oTHBpuMuWlgr78CGo0ZxartGKCp5utSVSCSQJ18E AEdi7THPamrVPRrH1E7XQ2kAxNR6GTMOuHZrdHGaALsT6xf3y8vCzRi1VlwW GehazZUovjMZgMZXia1Hvd2eN1sf4wOZ9gDB_mWNUc_cFhMv7.pY.Mk_Dh93 xG1gDd.IZbiuSa742sAZ.2EYGAZAayE7PMu1BCoQGqxb8BQcg_CnrJeiDuCc BfjWtq9HefmhCM3IeWoCY6l0DkJJf5txciQUGaJN5JZOEXzbUcK5EDIXdNwp rGKQzKksnOQEZXh0TTgv4WhOM7aHIrpD1pYFy0R_u6LiRV9LFnKVnxr6jOfF vHx5CSE8tnv.7iB2WuEUbN0S6_tKgzDDAb5EkmwDhoHd3nd3sUFm1FMfIWfR OkGB3HPThAo9ixkN.jaiwIrU4x1iwR2O6zT1vv4SOWfgoJC9ItjfoNUEYIrZ _SOuXST1HaEM0FvXzsfxGaGoZWs_DveGjiuOvCg-- Received: from [204.11.79.50] by web161205.mail.bf1.yahoo.com via HTTP; Tue, 19 Nov 2013 07:55:41 PST X-Rocket-MIMEInfo: 002.001,VGhlIHByb3BlcnR5IGlzIGRlcHJlY2F0ZWQgYnV0IHdpbGwgc3RpbGwgd29yay4gRWl0aGVyIG9uZSBpcyBmaW5lLgoKTGF1bmNoaW5nIHRoZSBqb2IgZnJvbSB0aGUgbmFtZW5vZGUgaXMgZmluZSAuwqAKCkkgYnJvdWdodCB1cCBhIGNsdXN0ZXIgd2l0aCAyLjAuNS1hbHBoYSBhbmQgYnVpbHQgdGhlIGxhdGVzdCBzcGFyayBtYXN0ZXIgYnJhbmNoIGFuZCBpdCBydW5zIGZpbmUgZm9yIG1lLiBJdCBsb29rcyBsaWtlIG5hbWVub2RlIDIuMC41LWFscGhhIHdvbid0IGV2ZW4gc3RhcnQgd2l0aCB0aGUgZGVmYXUBMAEBAQE- X-Mailer: YahooMailWebService/0.8.166.601 References: <1384785248.10520.YahooMailNeo@web140306.mail.bf1.yahoo.com> <1384833057.82865.YahooMailNeo@web140304.mail.bf1.yahoo.com> Message-ID: <1384876541.48001.YahooMailNeo@web161205.mail.bf1.yahoo.com> Date: Tue, 19 Nov 2013 07:55:41 -0800 (PST) From: Tom Graves Reply-To: Tom Graves Subject: Re: App master failed to find application jar in the master branch on YARN To: "user@spark.incubator.apache.org" In-Reply-To: MIME-Version: 1.0 Content-Type: multipart/alternative; boundary="537826103-1223718939-1384876541=:48001" X-Virus-Checked: Checked by ClamAV on apache.org --537826103-1223718939-1384876541=:48001 Content-Type: text/plain; charset=iso-8859-1 Content-Transfer-Encoding: quoted-printable The property is deprecated but will still work. Either one is fine.=0A=0ALa= unching the job from the namenode is fine .=A0=0A=0AI brought up a cluster = with 2.0.5-alpha and built the latest spark master branch and it runs fine = for me. It looks like namenode 2.0.5-alpha won't even start with the defaul= Fs of file:///. =A0Please make sure your namenode is actually up and runnin= g and you are pointing to it because you can run some jobs successfully wit= hout it (on a single node cluster), but when you have a multinode cluster = =A0here is the error I get when I run without a namenode up and it looks ve= ry similar to your error message:=0A=0A=A0 =A0 =A0 =A0 appDiagnostics: Appl= ication application_1384876319080_0001 failed 1 times due to AM Container f= or appattempt_1384876319080_0001_000001 exited with =A0exitCode: -1000 due = to: java.io.FileNotFoundException: File file:/home/tgravescs/spark-master/a= ssembly/target/scala-2.9.3/spark-assembly-0.9.0-incubating-SNAPSHOT-hadoop2= .0.5-alpha.jar does not exist=0A=0A=0AWhen you changed the default fs confi= g did you restart the cluster?=0A=0A=0ACan you try just running the example= s jar:=0A=0ASPARK_JAR=3Dassembly/target/scala-2.9.3/spark-assembly-0.9.0-in= cubating-SNAPSHOT-hadoop2.0.5-alpha.jar=0A=0A./spark-class =A0org.apache.sp= ark.deploy.yarn.Client --jar examples/target/scala-2.9.3/spark-examples-ass= embly-0.9.0-incubating-SNAPSHOT.jar =A0--class org.apache.spark.examples.Sp= arkPi =A0--args yarn-standalone =A0--num-workers 2 =A0--master-memory 2g --= worker-memory 2g --worker-cores 1=0A=0AOn the client side you should see me= ssages like this:=0A13/11/19 15:41:30 INFO yarn.Client: Uploading file:/hom= e/tgravescs/spark-master/examples/target/scala-2.9.3/spark-examples-assembl= y-0.9.0-incubating-SNAPSHOT.jar to hdfs://namenode.host.com:9000/user/tgrav= escs/.sparkStaging/application_1384874528558_0003/spark-examples-assembly-0= .9.0-incubating-SNAPSHOT.jar=0A13/11/19 15:41:31 INFO yarn.Client: Uploadin= g file:/home/tgravescs/spark-master/assembly/target/scala-2.9.3/spark-assem= bly-0.9.0-incubating-SNAPSHOT-hadoop2.0.5-alpha.jar to hdfs://namenode.host= .com:9000/user/tgravescs/.sparkStaging/application_1384874528558_0003/spark= -assembly-0.9.0-incubating-SNAPSHOT-hadoop2.0.5-alpha.jar=0A=0ATom=0A=0A=0A= =0AOn Tuesday, November 19, 2013 5:35 AM, guojc wrote:= =0A =0AHi Tom,=0A=A0 =A0Thank you for your response. I =A0have double check= ed that I had upload both jar in the same folder on hdfs. I think the=A0fs.default.name you pointed out is the old deprecated name for=A0= fs.defaultFS=A0config accordiing =A0http://hadoop.apache.org/docs/r2.0.2-al= pha/hadoop-project-dist/hadoop-common/DeprecatedProperties.html . =A0Anyway= , we have tried both =A0fs.default.name and =A0fs.defaultFS=A0set to hdfs n= amenode, and the situation remained same. And we have removed SPARK_HOME en= v variable on worker node. =A0An additional information might be related is= that job=A0submission=A0is done on the same machine of hdfs namenode. =A0B= ut I'm not sure this will cause the problem.=0A=0AThanks,=0AJiacheng Guo=0A= =0A=0A=0AOn Tue, Nov 19, 2013 at 11:50 AM, Tom Graves wrote:=0A=0ASorry for the delay. What is the default filesystem on your H= DFS setup? =A0It looks like its set to file: rather then hdfs://. =A0That i= s the only reason I can think its listing the directory as=A0=A0file:/home/= work/.sparkStaging/application_1384588058297_0056. =A0Its basically just co= pying it local rather then uploading to hdfs and its just trying to use the= local=A0=A0file:/home/work/guojiacheng/spark/assembly/target/scala-2.9.3/s= park-assembly-0.9.0-incubating-SNAPSHOT-hadoop2.0.5-alpha.jar.=A0 It genera= lly would create that in hdfs so it accessible on all the nodes. =A0Is your= /home/work nfs mounted on all the nodes? =A0 =A0=0A>=0A>=0A>You can find t= he default fs by looking at the Hadoop config files. =A0Generally in core-s= ite.xml. =A0its specified by:=A0=A0 =A0 =A0 =A0 fs.default.name=0A>=0A>=0A>Its pretty odd if those are its erroring with file:// when you= specified hdfs://.=0A>when you tried the hdfs:// did you upload both the s= park jar and your client jar=A0(SparkAUC-assembly-0.1.jar)? =A0If not try t= hat and make sure to put hdfs:// on them when you export SPARK_JAR and spec= ify the --jar option. =A0=0A>=0A>=0A>=0A>I'll try to reproduce the error to= morrow to see if a bug was introduced when I added the feature to run spark= from HDFS.=0A>=0A>=0A>Tom=0A>=0A>=0A>=0A>On Monday, November 18, 2013 11:1= 3 AM, guojc wrote:=0A> =0A>Hi Tom,=0A>=A0 =A0I'm on Had= oop 2.05. =A0I can launch application spark 0.8 release normally. However I= switch to git master branch version with application built with it, I got = the jar not found exception and same happens to the example application. I = have tried both file:// protocol and hdfs:// protocol with jar in local fil= e system and hdfs respectively, and even tried jar list parameter when new = spark context. =A0The exception is slightly different for hdfs protocol and= local file path. My application launch command is =A0=A0=0A>=0A>=0A>=A0SPA= RK_JAR=3D/home/work/guojiacheng/spark/assembly/target/scala-2.9.3/spark-ass= embly-0.9.0-incubating-SNAPSHOT-hadoop2.0.5-alpha.jar /home/work/guojiachen= g/spark/spark-class =A0org.apache.spark.deploy.yarn.Client --jar /home/work= /guojiacheng/spark-auc/target/scala-2.9.3/SparkAUC-assembly-0.1.jar --class= =A0myClass.SparkAUC --args -c --args yarn-standalone =A0--args -i --args h= dfs://{hdfs_host}:9000/user/work/guojiacheng/data --args -m --args hdfs://{= hdfs_host}:9000/user/work/guojiacheng/model_large --args -o --args hdfs://{= hdfs_host}:9000/user/work/guojiacheng/score --num-workers 60 =A0--master-me= mory 6g --worker-memory 7g --worker-cores 1=0A>=0A>=0A>And my build command= is=A0SPARK_HADOOP_VERSION=3D2.0.5-alpha SPARK_YARN=3Dtrue sbt/sbt assembly= =0A>=0A>=0A>Only thing I can think of might be related is on each cluster n= ode, it has a env SPARK_HOME point to a copy of 0.8 version's position, and= its bin fold is in Path environment variable. And 0.9 version is not there= . =A0It was something left over, when cluster was setup. =A0But I don't kno= w whether it is related, as my understand is the yarn version try to distri= bute spark through yarn.=0A>=0A>=0A>hdfs version error message:=0A>=0A>=0A>= =A0 =A0 =A0 =A0 =A0appDiagnostics: Application application_1384588058297_00= 56 failed 1 times due to AM Container for appattempt_1384588058297_0056_000= 001 exited with =A0exitCode: -1000 due to: RemoteTrace:=A0=0A>java.io.FileN= otFoundException: File file:/home/work/.sparkStaging/application_1384588058= 297_0056/SparkAUC-assembly-0.1.jar does not exist=0A>=A0 =A0=0A>local versi= on error message.=0A>appDiagnostics: Application application_1384588058297_= 0066 failed 1 times due to AM Container for appattempt_1384588058297_0066_0= 00001 exited with =A0exitCode: -1000 due to: java.io.FileNotFoundException:= File file:/home/work/guojiacheng/spark/assembly/target/scala-2.9.3/spark-a= ssembly-0.9.0-incubating-SNAPSHOT-hadoop2.0.5-alpha.jar does not exist=0A>= =0A>=0A>=0A>Best Regards,=0A>Jiacheng GUo=0A>=0A>=0A>=0A>=0A>=0A>On Mon, No= v 18, 2013 at 10:34 PM, Tom Graves wrote:=0A>=0A>Hey= =A0Jiacheng Guo,=0A>>=0A>>=0A>>do you have SPARK_EXAMPLES_JAR env variable = set? =A0If you do, you have to add the --addJars parameter to the yarn clie= nt and point to the spark examples jar. =A0Or just unset SPARK_EXAMPLES_JAR= env variable.=0A>>=0A>>=0A>>You should only have to set SPARK_JAR env vari= able. =A0=0A>>=0A>>=0A>>If that isn't the issue let me know the build comma= nd you used and hadoop version, and your defaultFs or hadoop.=0A>>=0A>>=0A>= >Tom=0A>>=0A>>=0A>>=0A>>On Saturday, November 16, 2013 2:32 AM, guojc wrote:=0A>> =0A>>hi,=0A>>=A0 =A0After reading about the exit= ing progress in consolidating shuffle, I'm eager to trying out the last mas= ter branch. However up to launch the example application, the job failed wi= th prompt the app master failed to find the target jar. appDiagnostics: App= lication application_1384588058297_0017 failed 1 times due to AM Container = for appattempt_1384588058297_0017_000001 exited with =A0exitCode: -1000 due= to: java.io.FileNotFoundException: File file:/${my_work_dir}/spark/example= s/target/scala-2.9.3/spark-examples-assembly-0.9.0-incubating-SNAPSHOT.jar = does not exist.=0A>>=0A>>=0A>>=A0 Is there any change on how to launch a ya= rn job now?=0A>>=0A>>=0A>>Best Regards,=0A>>Jiacheng Guo=0A>>=0A>>=0A>>=0A>= >=0A>=0A>=0A> --537826103-1223718939-1384876541=:48001 Content-Type: text/html; charset=iso-8859-1 Content-Transfer-Encoding: quoted-printable
The property is deprecated but will still work. Ei= ther one is fine.
Launching the job from the namenode is fine . 

I brought up a cluster wit= h 2.0.5-alpha and built the latest spark master branch and it runs fine for= me. It looks like namenode 2.0.5-alpha won't even start with the defaulFs = of file:///.  Please make sure your namenode is actually up and runnin= g and you are pointing to it because you can run some jobs successfully wit= hout it (on a single node cluster), but when you have a multinode cluster &= nbsp;here is the error I get when I= run without a namenode up and it looks very similar to your error message:=

        appDiagnostics: Application application_1= 384876319080_0001 failed 1 times due to AM Container for appattempt_1384876= 319080_0001_000001 exited with  exitCode: -1000 due to: java.io.FileNo= tFoundException: File file:/home/tgravescs/spark-master/assembly/target/sca= la-2.9.3/spark-assembly-0.9.0-incubating-SNAPSHOT-hadoop2.0.5-alpha.jar doe= s not exist
=
When you changed the default fs config did you restart the cluster?

=
Can you try just= running the examples jar:

SPARK_JAR=3Dassembly/target/scala-2.9.3/spark-assembly-0.9.0-incubating-= SNAPSHOT-hadoop2.0.5-alpha.jar

./spark-class  org.apache.spark.deploy.yar= n.Client --jar examples/target/scala-2.9.3/spark-examples-assembly-0.9.0-in= cubating-SNAPSHOT.jar  --class org.apache.spark.examples.SparkPi  = ;--args yarn-standalone  --num-workers 2  --master-memory 2g --wo= rker-memory 2g --worker-cores 1

On the client side you should see messages like this:=
13/11/19 15:41:3= 0 INFO yarn.Client: Uploading file:/home/tgravescs/spark-master/examples/ta= rget/scala-2.9.3/spark-examples-assembly-0.9.0-incubating-SNAPSHOT.jar to h= dfs://namenode.host.com:9000/user/tgravescs/.sparkStaging/application_13848= 74528558_0003/spark-examples-assembly-0.9.0-incubating-SNAPSHOT.jar
13/11/19 15:41:31 INFO yarn.Client: Upl= oading file:/home/tgravescs/spark-master/assembly/target/scala-2.9.3/spark-= assembly-0.9.0-incubating-SNAPSHOT-hadoop2.0.5-alpha.jar to hdfs://namenode.host.com:9000/user/tgravescs/.sparkStaging/application_138= 4874528558_0003/spark-assembly-0.9.0-incubating-SNAPSHOT-hadoop2.0.5-alpha.= jar

Tom


On Tues= day, November 19, 2013 5:35 AM, guojc <guojc03@gmail.com> wrote:
Hi Tom,
   Thank you for your response. I  have= double checked that I had upload both jar in the same folder on hdfs. I th= ink the <name>f= s.default.name</name> you pointed = out is the old deprecated name for fs.defaultFS config accordiing  fs.default.name and  fs.defaultFS = ;set to hdfs namenode, and the situation remained same. And we have removed= SPARK_HOME env variable on worker node.  An additional information mi= ght be related is that job submission is done on the same machine= of hdfs namenode.  But I'm not sure this will cause the problem.
=0A

= Thanks,
Jiacheng Guo
=0A


On Tue, Nov 19, 2013 at 11:50 AM, Tom Grav= es <tgraves_cs@yahoo.com> wrote:
=0A
=0ASorry for the delay. What is the default filesys= tem on your HDFS setup?  It looks like its set to file: rather then hd= fs://.  That is the only reason I can think its listing the directory = as  file:/home/work/.sparkStaging/app= lication_1384588058297_0056.  Its basically just copying it local rath= er then uploading to hdfs and its just trying to use the local =  file:/home/work/guojiacheng/spark/assembly/target/scala-2.9.3/spark= -assembly-0.9.0-incubating-SNAPSHOT-hadoop2.0.5-alpha.jar.  It generally would create that= in hdfs so it accessible on all the nodes.  Is your /home/work nfs mo= unted on all the nodes?    
=0A

=0A
You can fi= nd the default fs by looking at the Hadoop config files.  Generally in= core-site.xml.  its specified by:         <name>fs.default.name</name>
=0A

=0A
Its pretty odd if those a= re its erroring with file:// when you specified hdfs://.
=0Awhen you = tried the hdfs:// did you upload both the spark jar and your client jar (SparkAUC-assembly-0= .1.jar)?  If not try that and make sure to put hdfs:// on them when yo= u export SPARK_JAR and specify the --jar option.  
=0A

=0A
I'll try to reproduce the error tomor= row to see if a bug was introduced when I added the feature to run spark fr= om HDFS.
=0A
=0A
<= div style=3D"font-style: normal; font-size: 13px; background-color: transpa= rent; font-family: 'Helvetica Neue', 'Segoe UI', Helvetica, Arial, 'Lucida = Grande', sans-serif;">Tom
=0A


=0A=
On Monday, November 18, 2013 11:13= AM, guojc <guojc03@gma= il.com> wrote:
Hi Tom,
=0A   I'm on Hadoop 2.05.  I can l= aunch application spark 0.8 release normally. However I switch to git maste= r branch version with application built with it, I got the jar not found ex= ception and same happens to the example application. I have tried both file= :// protocol and hdfs:// protocol with jar in local file system and hdfs re= spectively, and even tried jar list parameter when new spark context.  = ;The exception is slightly different for hdfs protocol and local file path.= =0A My application launch command is   
=0A

 SPARK_JAR=3D/home/work/guojiacheng/spark/assembly/t= arget/scala-2.9.3/spark-assembly-0.9.0-incubating-SNAPSHOT-hadoop2.0.5-alph= a.jar /home/work/guojiacheng/spark/spark-class  org.apache.spark.deplo= y.yarn.Client --jar /home/work/guojiacheng/spark-auc/target/scala-2.9.3/Spa= rkAUC-assembly-0.1.jar --class  myClass.SparkAUC --args -c --args yarn= -standalone  --args -i --args hdfs://{hdfs_host}:9000/user/work/guojia= cheng/data --args -m --args hdfs://{hdfs_host}:9000/user/work/guojiacheng/m= odel_large --args -o --args hdfs://{hdfs_host}:9000/user/work/guojiacheng/s= core --num-workers 60  --master-memory 6g --worker-memory 7g --worker-= cores 1
=0A=0A

And my build command = is SPARK_HADOOP_VERSION=3D2.0.5-alpha SPARK_YARN=3Dtrue sbt/sbt assemb= ly

Only thing I can think of might = be related is on each cluster node, it has a env SPARK_HOME point to a copy= of 0.8 version's position, and its bin fold is in Path environment variabl= e. And 0.9 version is not there.  It was something left over, when clu= ster was setup.  But I don't know whether it is related, as my underst= and is the yarn version try to distribute spark through yarn.
=0A=0A
hdfs version error message:

         appDiagnos= tics: Application application_1384588058297_0056 failed 1 times due to AM C= ontainer for appattempt_1384588058297_0056_000001 exited with  exitCod= e: -1000 due to: RemoteTrace: 
=0A=0A
java.io.FileNotFoundExc= eption: File file:/home/work/.sparkStaging/application_1384588058297_0056/S= parkAUC-assembly-0.1.jar does not exist
   
=
local version error message.
appDiagnostics: Application app= lication_1384588058297_0066 failed 1 times due to AM Container for appattem= pt_1384588058297_0066_000001 exited with  exitCode: -1000 due to: java= .io.FileNotFoundException: File file:/home/work/guojiacheng/spark/assembly/= target/scala-2.9.3/spark-assembly-0.9.0-incubating-SNAPSHOT-hadoop2.0.5-alp= ha.jar does not exist
=0A=0A

Best Regards,
Jiacheng GUo



On Mon,= Nov 18, 2013 at 10:34 PM, Tom Graves <tgraves_cs@yahoo.com>= wrote:
=0A=0A
=0A=0AHey Jiacheng Guo,

=
=0A=0Ado you have SPARK_EXAMPLES_JAR en= v variable set?  If you do, you have to add the --addJars parameter to= =0A the yarn client and point to the spark examples jar.  Or just unse= t SPARK_EXAMPLES_JAR env variable.
=0A=0A
You should only have to set SPARK= _JAR env variable.  
=0A=0A

=0A=0AIf=0A that isn't the issue let me kn= ow the build command you used and hadoop version, and your defaultFs or had= oop.
Tom
= =0A=0A


=0A=0A
= On Saturday, November 16, 2013 2:32 AM, guojc <guojc03@gmail.com> wrote:
=
=0A
hi,
   After r= eading about the exiting progress in consolidating shuffle, I'm eager to tr= ying out the last master branch. However up to launch the example applicati= on, the job failed with prompt the app master failed to find the target jar= . appDiagnostics: Application application_1384588058297_0017 failed 1 times= due to AM Container for appattempt_1384588058297_0017_000001 exited with &= nbsp;exitCode: -1000 due to: java.io.FileNotFoundException: File file:/${my= _work_dir}/spark/examples/target/scala-2.9.3/spark-examples-assembly-0.9.0-= incubating-SNAPSHOT.jar does not exist.
=0A=0A=0A

  Is there any change on how to launch a yarn job now?

Best Regards,
Jiacheng Guo=


=0A
=0A



=



--537826103-1223718939-1384876541=:48001--