flink-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Robert Metzger <rmetz...@apache.org>
Subject Re: flink k-means on hadoop cluster
Date Thu, 04 Jun 2015 15:04:25 GMT
It looks like the user "yarn" which is running Flink doesn't have
permission to access the files.

Can you do "sudo su yarn" to become the "yarn" user. Then, you can do
"hadoop fs -chmod 777" to make the files accessible for everyone.


On Thu, Jun 4, 2015 at 4:59 PM, Pa Rö <paul.roewer1990@googlemail.com>
wrote:

> okay, it's work, i get a exception:
>
> [cloudera@quickstart Desktop]$ cd flink-0.9-SNAPSHOT/bin/
> [cloudera@quickstart bin]$ flink run /home/cloudera/Desktop/ma-flink.jar
> bash: flink: command not found
> [cloudera@quickstart bin]$ ./flink run /home/cloudera/Desktop/ma-flink.jar
> log4j:WARN No appenders could be found for logger
> (org.apache.hadoop.metrics2.lib.MutableMetricsFactory).
> log4j:WARN Please initialize the log4j system properly.
> log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for
> more info.
> Found YARN properties file
> /home/cloudera/Desktop/flink-0.9-SNAPSHOT/bin/../conf/.yarn-properties
> Using JobManager address from YARN properties quickstart.cloudera/
> 127.0.0.1:53874
> java.io.IOException: Mkdirs failed to create /user/cloudera/outputs
>     at
> org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:438)
>     at
> org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:424)
>     at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:905)
>     at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:886)
>     at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:783)
>     at
> mgm.tp.bigdata.ma_commons.commons.Seeding.randomSeeding(Seeding.java:21)
>     at
> mgm.tp.bigdata.ma_flink.FlinkMain.getCentroidDataSet(FlinkMain.java:178)
>     at mgm.tp.bigdata.ma_flink.FlinkMain.main(FlinkMain.java:47)
>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>     at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>     at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>     at java.lang.reflect.Method.invoke(Method.java:606)
>     at
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:437)
>     at
> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:353)
>     at org.apache.flink.client.program.Client.run(Client.java:315)
>     at
> org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:584)
>     at org.apache.flink.client.CliFrontend.run(CliFrontend.java:290)
>     at
> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:880)
>     at org.apache.flink.client.CliFrontend.main(CliFrontend.java:922)
> org.apache.flink.client.program.ProgramInvocationException: The program
> execution failed: Failed to submit job 934743a5c49c6d5e31c9e8201452e36d
> (KMeans Flink)
>     at org.apache.flink.client.program.Client.run(Client.java:412)
>     at org.apache.flink.client.program.Client.run(Client.java:355)
>     at org.apache.flink.client.program.Client.run(Client.java:348)
>     at
> org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:63)
>     at mgm.tp.bigdata.ma_flink.FlinkMain.main(FlinkMain.java:70)
>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>     at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>     at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>     at java.lang.reflect.Method.invoke(Method.java:606)
>     at
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:437)
>     at
> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:353)
>     at org.apache.flink.client.program.Client.run(Client.java:315)
>     at
> org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:584)
>     at org.apache.flink.client.CliFrontend.run(CliFrontend.java:290)
>     at
> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:880)
>     at org.apache.flink.client.CliFrontend.main(CliFrontend.java:922)
> Caused by: org.apache.flink.runtime.client.JobExecutionException: Failed
> to submit job 934743a5c49c6d5e31c9e8201452e36d (KMeans Flink)
>     at org.apache.flink.runtime.jobmanager.JobManager.org
> $apache$flink$runtime$jobmanager$JobManager$$submitJob(JobManager.scala:595)
>     at
> org.apache.flink.runtime.jobmanager.JobManager$$anonfun$receiveWithLogMessages$1.applyOrElse(JobManager.scala:192)
>     at
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
>     at
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
>     at
> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
>     at
> org.apache.flink.yarn.ApplicationMasterActor$$anonfun$receiveYarnMessages$1.applyOrElse(ApplicationMasterActor.scala:99)
>     at scala.PartialFunction$OrElse.apply(PartialFunction.scala:162)
>     at
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:36)
>     at
> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:29)
>     at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
>     at
> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:29)
>     at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
>     at
> org.apache.flink.runtime.jobmanager.JobManager.aroundReceive(JobManager.scala:94)
>     at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
>     at akka.actor.ActorCell.invoke(ActorCell.scala:487)
>     at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
>     at akka.dispatch.Mailbox.run(Mailbox.scala:221)
>     at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
>     at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>     at
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
>     at
> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
>     at
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> Caused by: org.apache.flink.runtime.JobException: Creating the input
> splits caused an error: File /user/cloudera/outputs/seed-1 does not exist
> or the user running Flink ('yarn') has insufficient permissions to access
> it.
>     at
> org.apache.flink.runtime.executiongraph.ExecutionJobVertex.<init>(ExecutionJobVertex.java:162)
>     at
> org.apache.flink.runtime.executiongraph.ExecutionGraph.attachJobGraph(ExecutionGraph.java:471)
>     at org.apache.flink.runtime.jobmanager.JobManager.org
> $apache$flink$runtime$jobmanager$JobManager$$submitJob(JobManager.scala:535)
>     ... 21 more
> Caused by: java.io.FileNotFoundException: File
> /user/cloudera/outputs/seed-1 does not exist or the user running Flink
> ('yarn') has insufficient permissions to access it.
>     at
> org.apache.flink.core.fs.local.LocalFileSystem.getFileStatus(LocalFileSystem.java:106)
>     at
> org.apache.flink.api.common.io.FileInputFormat.createInputSplits(FileInputFormat.java:390)
>     at
> org.apache.flink.api.common.io.FileInputFormat.createInputSplits(FileInputFormat.java:51)
>     at
> org.apache.flink.runtime.executiongraph.ExecutionJobVertex.<init>(ExecutionJobVertex.java:146)
>     ... 23 more
>
>
> how i must set the files in the hdfs?
> quickstart.cloudera:50075/home/cloudera/output?
>
> 2015-06-04 16:51 GMT+02:00 Robert Metzger <rmetzger@apache.org>:
>
>> Once you've started the YARN session, you can submit a Flink job with
>> "./bin/flink run <pathToYourJar>".
>>
>> The jar file of your job doesn't need to be in HDFS. It has to be in the
>> local file system and flink will send it to all machines.
>>
>> On Thu, Jun 4, 2015 at 4:48 PM, Pa Rö <paul.roewer1990@googlemail.com>
>> wrote:
>>
>>> okay, now it run on my hadoop.
>>> how i can start my flink job? and where must the jar file save, at hdfs
>>> or as local file?
>>>
>>> 2015-06-04 16:31 GMT+02:00 Robert Metzger <rmetzger@apache.org>:
>>>
>>>> Yes, you have to run these commands in the command line of the Cloudera
>>>> VM.
>>>>
>>>> On Thu, Jun 4, 2015 at 4:28 PM, Pa Rö <paul.roewer1990@googlemail.com>
>>>> wrote:
>>>>
>>>>> you mean run this command on terminal/shell and not define a hue job?
>>>>>
>>>>> 2015-06-04 16:25 GMT+02:00 Robert Metzger <rmetzger@apache.org>:
>>>>>
>>>>>> It should be certainly possible to run Flink on a cloudera live VM
>>>>>>
>>>>>> I think these are the commands you need to execute:
>>>>>>
>>>>>> wget
>>>>>> http://stratosphere-bin.s3-website-us-east-1.amazonaws.com/flink-0.9-SNAPSHOT-bin-hadoop2.tgz
>>>>>> tar xvzf flink-0.9-SNAPSHOT-bin-hadoop2.tgz
>>>>>> cd flink-0.9-SNAPSHOT/
>>>>>> *export HADOOP_CONF_DIR=/usr/lib/hadoop/etc/hadoop/*
>>>>>> ./bin/yarn-session.sh -n 1 -jm 1024 -tm 1024
>>>>>>
>>>>>> If that is not working for you, please post the exact error message
>>>>>> you are getting and I can help you to get it to run.
>>>>>>
>>>>>>
>>>>>> On Thu, Jun 4, 2015 at 4:18 PM, Pa Rö <paul.roewer1990@googlemail.com
>>>>>> > wrote:
>>>>>>
>>>>>>> hi robert,
>>>>>>>
>>>>>>> i think the problem is the hue api,
>>>>>>> i had the same problem with spark submit script,
>>>>>>> but on the new hue release, they have a spark submit api.
>>>>>>>
>>>>>>> i asked the group for the same problem with spark, no reply.
>>>>>>>
>>>>>>> i want test my app on local cluster, before i run it on the big
>>>>>>> cluster,
>>>>>>> for that i use cloudera live. maybe it give an other way to test
>>>>>>> flink on a local cluster vm?
>>>>>>>
>>>>>>> 2015-06-04 16:12 GMT+02:00 Robert Metzger <rmetzger@apache.org>:
>>>>>>>
>>>>>>>> Hi Paul,
>>>>>>>>
>>>>>>>> why did running Flink from the regular scripts not work for
you?
>>>>>>>>
>>>>>>>> I'm not an expert on Hue, I would recommend asking in the
Hue user
>>>>>>>> forum / mailing list:
>>>>>>>> https://groups.google.com/a/cloudera.org/forum/#!forum/hue-user.
>>>>>>>>
>>>>>>>> On Thu, Jun 4, 2015 at 4:09 PM, Pa Rö <
>>>>>>>> paul.roewer1990@googlemail.com> wrote:
>>>>>>>>
>>>>>>>>> thanks,
>>>>>>>>> now i want run my app on cloudera live vm single node,
>>>>>>>>> how i can define my flink job with hue?
>>>>>>>>> i try to run the flink script in the hdfs, it's not work.
>>>>>>>>>
>>>>>>>>> best regards,
>>>>>>>>> paul
>>>>>>>>>
>>>>>>>>> 2015-06-02 14:50 GMT+02:00 Robert Metzger <rmetzger@apache.org>:
>>>>>>>>>
>>>>>>>>>> I would recommend using HDFS.
>>>>>>>>>> For that, you need to specify the paths like this:
>>>>>>>>>> hdfs:///path/to/data.
>>>>>>>>>>
>>>>>>>>>> On Tue, Jun 2, 2015 at 2:48 PM, Pa Rö <
>>>>>>>>>> paul.roewer1990@googlemail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> nice,
>>>>>>>>>>>
>>>>>>>>>>> which file system i must use for the cluster?
java.io or
>>>>>>>>>>> hadoop.fs or flink?
>>>>>>>>>>>
>>>>>>>>>>> 2015-06-02 14:29 GMT+02:00 Robert Metzger <rmetzger@apache.org>:
>>>>>>>>>>>
>>>>>>>>>>>> Hi,
>>>>>>>>>>>> you can start Flink on YARN on the Cloudera
distribution.
>>>>>>>>>>>>
>>>>>>>>>>>> See here for more:
>>>>>>>>>>>> http://ci.apache.org/projects/flink/flink-docs-master/setup/yarn_setup.html
>>>>>>>>>>>>
>>>>>>>>>>>> These are the commands you need to execute
>>>>>>>>>>>>
>>>>>>>>>>>> wget http://stratosphere-bin.s3-website-us-east-1.amazonaws.com/flink-0.9-SNAPSHOT-bin-hadoop2.tgz
>>>>>>>>>>>> tar xvzf flink-0.9-SNAPSHOT-bin-hadoop2.tgzcd
flink-0.9-SNAPSHOT/
>>>>>>>>>>>> ./bin/yarn-session.sh -n 4 -jm 1024 -tm 4096
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On Tue, Jun 2, 2015 at 2:03 PM, Pa Rö <
>>>>>>>>>>>> paul.roewer1990@googlemail.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> hi community,
>>>>>>>>>>>>>
>>>>>>>>>>>>> i want test my flink k-means on a hadoop
cluster. i use the
>>>>>>>>>>>>> cloudera live distribution. how i can
run flink on this cluster? maybe only
>>>>>>>>>>>>> the java dependencies are engouth?
>>>>>>>>>>>>>
>>>>>>>>>>>>> best regards,
>>>>>>>>>>>>> paul
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Mime
View raw message