[cloudera@quickstart bin]$ sudo su yarn bash-4.1$ hadoop fs -chmod 777 -chmod: Not enough arguments: expected 2 but got 1 Usage: hadoop fs [generic options] -chmod [-R] <MODE[,MODE]... | OCTALMODE> PATH... bash-4.1$
you understand? 2015-06-04 17:04 GMT+02:00 Robert Metzger <rmetz...@apache.org>: > It looks like the user "yarn" which is running Flink doesn't have > permission to access the files. > > Can you do "sudo su yarn" to become the "yarn" user. Then, you can do > "hadoop fs -chmod 777" to make the files accessible for everyone. > > > On Thu, Jun 4, 2015 at 4:59 PM, Pa Rö <paul.roewer1...@googlemail.com> > wrote: > >> okay, it's work, i get a exception: >> >> [cloudera@quickstart Desktop]$ cd flink-0.9-SNAPSHOT/bin/ >> [cloudera@quickstart bin]$ flink run /home/cloudera/Desktop/ma-flink.jar >> bash: flink: command not found >> [cloudera@quickstart bin]$ ./flink run >> /home/cloudera/Desktop/ma-flink.jar >> log4j:WARN No appenders could be found for logger >> (org.apache.hadoop.metrics2.lib.MutableMetricsFactory). >> log4j:WARN Please initialize the log4j system properly. >> log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for >> more info. >> Found YARN properties file >> /home/cloudera/Desktop/flink-0.9-SNAPSHOT/bin/../conf/.yarn-properties >> Using JobManager address from YARN properties quickstart.cloudera/ >> 127.0.0.1:53874 >> java.io.IOException: Mkdirs failed to create /user/cloudera/outputs >> at >> org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:438) >> at >> org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:424) >> at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:905) >> at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:886) >> at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:783) >> at >> mgm.tp.bigdata.ma_commons.commons.Seeding.randomSeeding(Seeding.java:21) >> at >> mgm.tp.bigdata.ma_flink.FlinkMain.getCentroidDataSet(FlinkMain.java:178) >> at mgm.tp.bigdata.ma_flink.FlinkMain.main(FlinkMain.java:47) >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:606) >> at >> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:437) >> at >> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:353) >> at org.apache.flink.client.program.Client.run(Client.java:315) >> at >> org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:584) >> at org.apache.flink.client.CliFrontend.run(CliFrontend.java:290) >> at >> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:880) >> at org.apache.flink.client.CliFrontend.main(CliFrontend.java:922) >> org.apache.flink.client.program.ProgramInvocationException: The program >> execution failed: Failed to submit job 934743a5c49c6d5e31c9e8201452e36d >> (KMeans Flink) >> at org.apache.flink.client.program.Client.run(Client.java:412) >> at org.apache.flink.client.program.Client.run(Client.java:355) >> at org.apache.flink.client.program.Client.run(Client.java:348) >> at >> org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:63) >> at mgm.tp.bigdata.ma_flink.FlinkMain.main(FlinkMain.java:70) >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:606) >> at >> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:437) >> at >> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:353) >> at org.apache.flink.client.program.Client.run(Client.java:315) >> at >> org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:584) >> at org.apache.flink.client.CliFrontend.run(CliFrontend.java:290) >> at >> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:880) >> at org.apache.flink.client.CliFrontend.main(CliFrontend.java:922) >> Caused by: org.apache.flink.runtime.client.JobExecutionException: Failed >> to submit job 934743a5c49c6d5e31c9e8201452e36d (KMeans Flink) >> at org.apache.flink.runtime.jobmanager.JobManager.org >> $apache$flink$runtime$jobmanager$JobManager$$submitJob(JobManager.scala:595) >> at >> org.apache.flink.runtime.jobmanager.JobManager$$anonfun$receiveWithLogMessages$1.applyOrElse(JobManager.scala:192) >> at >> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33) >> at >> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33) >> at >> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25) >> at >> org.apache.flink.yarn.ApplicationMasterActor$$anonfun$receiveYarnMessages$1.applyOrElse(ApplicationMasterActor.scala:99) >> at scala.PartialFunction$OrElse.apply(PartialFunction.scala:162) >> at >> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:36) >> at >> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:29) >> at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118) >> at >> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:29) >> at akka.actor.Actor$class.aroundReceive(Actor.scala:465) >> at >> org.apache.flink.runtime.jobmanager.JobManager.aroundReceive(JobManager.scala:94) >> at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516) >> at akka.actor.ActorCell.invoke(ActorCell.scala:487) >> at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254) >> at akka.dispatch.Mailbox.run(Mailbox.scala:221) >> at akka.dispatch.Mailbox.exec(Mailbox.scala:231) >> at >> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) >> at >> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) >> at >> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) >> at >> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) >> Caused by: org.apache.flink.runtime.JobException: Creating the input >> splits caused an error: File /user/cloudera/outputs/seed-1 does not exist >> or the user running Flink ('yarn') has insufficient permissions to access >> it. >> at >> org.apache.flink.runtime.executiongraph.ExecutionJobVertex.<init>(ExecutionJobVertex.java:162) >> at >> org.apache.flink.runtime.executiongraph.ExecutionGraph.attachJobGraph(ExecutionGraph.java:471) >> at org.apache.flink.runtime.jobmanager.JobManager.org >> $apache$flink$runtime$jobmanager$JobManager$$submitJob(JobManager.scala:535) >> ... 21 more >> Caused by: java.io.FileNotFoundException: File >> /user/cloudera/outputs/seed-1 does not exist or the user running Flink >> ('yarn') has insufficient permissions to access it. >> at >> org.apache.flink.core.fs.local.LocalFileSystem.getFileStatus(LocalFileSystem.java:106) >> at >> org.apache.flink.api.common.io.FileInputFormat.createInputSplits(FileInputFormat.java:390) >> at >> org.apache.flink.api.common.io.FileInputFormat.createInputSplits(FileInputFormat.java:51) >> at >> org.apache.flink.runtime.executiongraph.ExecutionJobVertex.<init>(ExecutionJobVertex.java:146) >> ... 23 more >> >> >> how i must set the files in the hdfs? >> quickstart.cloudera:50075/home/cloudera/output? >> >> 2015-06-04 16:51 GMT+02:00 Robert Metzger <rmetz...@apache.org>: >> >>> Once you've started the YARN session, you can submit a Flink job with >>> "./bin/flink run <pathToYourJar>". >>> >>> The jar file of your job doesn't need to be in HDFS. It has to be in the >>> local file system and flink will send it to all machines. >>> >>> On Thu, Jun 4, 2015 at 4:48 PM, Pa Rö <paul.roewer1...@googlemail.com> >>> wrote: >>> >>>> okay, now it run on my hadoop. >>>> how i can start my flink job? and where must the jar file save, at hdfs >>>> or as local file? >>>> >>>> 2015-06-04 16:31 GMT+02:00 Robert Metzger <rmetz...@apache.org>: >>>> >>>>> Yes, you have to run these commands in the command line of the >>>>> Cloudera VM. >>>>> >>>>> On Thu, Jun 4, 2015 at 4:28 PM, Pa Rö <paul.roewer1...@googlemail.com> >>>>> wrote: >>>>> >>>>>> you mean run this command on terminal/shell and not define a hue job? >>>>>> >>>>>> 2015-06-04 16:25 GMT+02:00 Robert Metzger <rmetz...@apache.org>: >>>>>> >>>>>>> It should be certainly possible to run Flink on a cloudera live VM >>>>>>> >>>>>>> I think these are the commands you need to execute: >>>>>>> >>>>>>> wget >>>>>>> http://stratosphere-bin.s3-website-us-east-1.amazonaws.com/flink-0.9-SNAPSHOT-bin-hadoop2.tgz >>>>>>> tar xvzf flink-0.9-SNAPSHOT-bin-hadoop2.tgz >>>>>>> cd flink-0.9-SNAPSHOT/ >>>>>>> *export HADOOP_CONF_DIR=/usr/lib/hadoop/etc/hadoop/* >>>>>>> ./bin/yarn-session.sh -n 1 -jm 1024 -tm 1024 >>>>>>> >>>>>>> If that is not working for you, please post the exact error message >>>>>>> you are getting and I can help you to get it to run. >>>>>>> >>>>>>> >>>>>>> On Thu, Jun 4, 2015 at 4:18 PM, Pa Rö < >>>>>>> paul.roewer1...@googlemail.com> wrote: >>>>>>> >>>>>>>> hi robert, >>>>>>>> >>>>>>>> i think the problem is the hue api, >>>>>>>> i had the same problem with spark submit script, >>>>>>>> but on the new hue release, they have a spark submit api. >>>>>>>> >>>>>>>> i asked the group for the same problem with spark, no reply. >>>>>>>> >>>>>>>> i want test my app on local cluster, before i run it on the big >>>>>>>> cluster, >>>>>>>> for that i use cloudera live. maybe it give an other way to test >>>>>>>> flink on a local cluster vm? >>>>>>>> >>>>>>>> 2015-06-04 16:12 GMT+02:00 Robert Metzger <rmetz...@apache.org>: >>>>>>>> >>>>>>>>> Hi Paul, >>>>>>>>> >>>>>>>>> why did running Flink from the regular scripts not work for you? >>>>>>>>> >>>>>>>>> I'm not an expert on Hue, I would recommend asking in the Hue user >>>>>>>>> forum / mailing list: >>>>>>>>> https://groups.google.com/a/cloudera.org/forum/#!forum/hue-user. >>>>>>>>> >>>>>>>>> On Thu, Jun 4, 2015 at 4:09 PM, Pa Rö < >>>>>>>>> paul.roewer1...@googlemail.com> wrote: >>>>>>>>> >>>>>>>>>> thanks, >>>>>>>>>> now i want run my app on cloudera live vm single node, >>>>>>>>>> how i can define my flink job with hue? >>>>>>>>>> i try to run the flink script in the hdfs, it's not work. >>>>>>>>>> >>>>>>>>>> best regards, >>>>>>>>>> paul >>>>>>>>>> >>>>>>>>>> 2015-06-02 14:50 GMT+02:00 Robert Metzger <rmetz...@apache.org>: >>>>>>>>>> >>>>>>>>>>> I would recommend using HDFS. >>>>>>>>>>> For that, you need to specify the paths like this: >>>>>>>>>>> hdfs:///path/to/data. >>>>>>>>>>> >>>>>>>>>>> On Tue, Jun 2, 2015 at 2:48 PM, Pa Rö < >>>>>>>>>>> paul.roewer1...@googlemail.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> nice, >>>>>>>>>>>> >>>>>>>>>>>> which file system i must use for the cluster? java.io or >>>>>>>>>>>> hadoop.fs or flink? >>>>>>>>>>>> >>>>>>>>>>>> 2015-06-02 14:29 GMT+02:00 Robert Metzger <rmetz...@apache.org> >>>>>>>>>>>> : >>>>>>>>>>>> >>>>>>>>>>>>> Hi, >>>>>>>>>>>>> you can start Flink on YARN on the Cloudera distribution. >>>>>>>>>>>>> >>>>>>>>>>>>> See here for more: >>>>>>>>>>>>> http://ci.apache.org/projects/flink/flink-docs-master/setup/yarn_setup.html >>>>>>>>>>>>> >>>>>>>>>>>>> These are the commands you need to execute >>>>>>>>>>>>> >>>>>>>>>>>>> wget >>>>>>>>>>>>> http://stratosphere-bin.s3-website-us-east-1.amazonaws.com/flink-0.9-SNAPSHOT-bin-hadoop2.tgz >>>>>>>>>>>>> tar xvzf flink-0.9-SNAPSHOT-bin-hadoop2.tgzcd flink-0.9-SNAPSHOT/ >>>>>>>>>>>>> ./bin/yarn-session.sh -n 4 -jm 1024 -tm 4096 >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On Tue, Jun 2, 2015 at 2:03 PM, Pa Rö < >>>>>>>>>>>>> paul.roewer1...@googlemail.com> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> hi community, >>>>>>>>>>>>>> >>>>>>>>>>>>>> i want test my flink k-means on a hadoop cluster. i use the >>>>>>>>>>>>>> cloudera live distribution. how i can run flink on this cluster? >>>>>>>>>>>>>> maybe only >>>>>>>>>>>>>> the java dependencies are engouth? >>>>>>>>>>>>>> >>>>>>>>>>>>>> best regards, >>>>>>>>>>>>>> paul >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >