i have change the permissions from the cloudera user and try the following command. and the files exist on hdfs ;) i set the files in my properties file like "flink.output=/user/cloudera/outputs/output_flink" i get the same exception again, maybe the problem have an other reason?
[cloudera@quickstart bin]$ sudo su yarn bash-4.1$ hadoop fs -chmod 777 /user/cloudera/inputs bash-4.1$ hadoop fs -chmod 777 /user/cloudera/outputs bash-4.1$ exit exit [cloudera@quickstart bin]$ sudo ./flink run /home/cloudera/Desktop/ma-flink.jar log4j:WARN No appenders could be found for logger (org.apache.hadoop.metrics2.lib.MutableMetricsFactory). log4j:WARN Please initialize the log4j system properly. log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info. Found YARN properties file /home/cloudera/Desktop/flink-0.9-SNAPSHOT/bin/../conf/.yarn-properties Using JobManager address from YARN properties quickstart.cloudera/ 127.0.0.1:52601 org.apache.flink.client.program.ProgramInvocationException: The program execution failed: Failed to submit job fe78e9ee50cf76ac8b487919e1c951fa (KMeans Flink) at org.apache.flink.client.program.Client.run(Client.java:412) at org.apache.flink.client.program.Client.run(Client.java:355) at org.apache.flink.client.program.Client.run(Client.java:348) at org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:63) at mgm.tp.bigdata.ma_flink.FlinkMain.main(FlinkMain.java:70) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:437) at org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:353) at org.apache.flink.client.program.Client.run(Client.java:315) at org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:584) at org.apache.flink.client.CliFrontend.run(CliFrontend.java:290) at org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:880) at org.apache.flink.client.CliFrontend.main(CliFrontend.java:922) Caused by: org.apache.flink.runtime.client.JobExecutionException: Failed to submit job fe78e9ee50cf76ac8b487919e1c951fa (KMeans Flink) at org.apache.flink.runtime.jobmanager.JobManager.org $apache$flink$runtime$jobmanager$JobManager$$submitJob(JobManager.scala:595) at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$receiveWithLogMessages$1.applyOrElse(JobManager.scala:192) at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33) at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33) at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25) at org.apache.flink.yarn.ApplicationMasterActor$$anonfun$receiveYarnMessages$1.applyOrElse(ApplicationMasterActor.scala:99) at scala.PartialFunction$OrElse.apply(PartialFunction.scala:162) at org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:36) at org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:29) at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118) at org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:29) at akka.actor.Actor$class.aroundReceive(Actor.scala:465) at org.apache.flink.runtime.jobmanager.JobManager.aroundReceive(JobManager.scala:94) at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516) at akka.actor.ActorCell.invoke(ActorCell.scala:487) at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254) at akka.dispatch.Mailbox.run(Mailbox.scala:221) at akka.dispatch.Mailbox.exec(Mailbox.scala:231) at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) Caused by: org.apache.flink.runtime.JobException: Creating the input splits caused an error: File /user/cloudera/inputs does not exist or the user running Flink ('yarn') has insufficient permissions to access it. at org.apache.flink.runtime.executiongraph.ExecutionJobVertex.<init>(ExecutionJobVertex.java:162) at org.apache.flink.runtime.executiongraph.ExecutionGraph.attachJobGraph(ExecutionGraph.java:471) at org.apache.flink.runtime.jobmanager.JobManager.org $apache$flink$runtime$jobmanager$JobManager$$submitJob(JobManager.scala:535) ... 21 more Caused by: java.io.FileNotFoundException: File /user/cloudera/inputs does not exist or the user running Flink ('yarn') has insufficient permissions to access it. at org.apache.flink.core.fs.local.LocalFileSystem.getFileStatus(LocalFileSystem.java:106) at org.apache.flink.api.common.io.FileInputFormat.createInputSplits(FileInputFormat.java:390) at org.apache.flink.api.common.io.FileInputFormat.createInputSplits(FileInputFormat.java:51) at org.apache.flink.runtime.executiongraph.ExecutionJobVertex.<init>(ExecutionJobVertex.java:146) ... 23 more 2015-06-04 17:38 GMT+02:00 Pa Rö <paul.roewer1...@googlemail.com>: > sorry, i see my yarn end before i can run my app, i must set the write > access for yarn, maybe this solve my problem. > > 2015-06-04 17:33 GMT+02:00 Pa Rö <paul.roewer1...@googlemail.com>: > >> i start the yarn-session.sh with sudo >> and than the flink run command with sudo, >> i get the following exception: >> >> cloudera@quickstart bin]$ sudo ./flink run >> /home/cloudera/Desktop/ma-flink.jar >> log4j:WARN No appenders could be found for logger >> (org.apache.hadoop.metrics2.lib.MutableMetricsFactory). >> log4j:WARN Please initialize the log4j system properly. >> log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for >> more info. >> org.apache.flink.client.program.ProgramInvocationException: Failed to >> resolve JobManager >> at org.apache.flink.client.program.Client.run(Client.java:378) >> at org.apache.flink.client.program.Client.run(Client.java:355) >> at org.apache.flink.client.program.Client.run(Client.java:348) >> at >> org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:63) >> at mgm.tp.bigdata.ma_flink.FlinkMain.main(FlinkMain.java:70) >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:606) >> at >> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:437) >> at >> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:353) >> at org.apache.flink.client.program.Client.run(Client.java:315) >> at >> org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:584) >> at org.apache.flink.client.CliFrontend.run(CliFrontend.java:290) >> at >> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:880) >> at org.apache.flink.client.CliFrontend.main(CliFrontend.java:922) >> Caused by: java.io.IOException: JobManager at akka.tcp:// >> flink@127.0.0.1:6123/user/jobmanager not reachable. Please make sure >> that the JobManager is running and its port is reachable. >> at >> org.apache.flink.runtime.jobmanager.JobManager$.getJobManagerRemoteReference(JobManager.scala:1198) >> at >> org.apache.flink.runtime.jobmanager.JobManager$.getJobManagerRemoteReference(JobManager.scala:1222) >> at >> org.apache.flink.runtime.jobmanager.JobManager$.getJobManagerRemoteReference(JobManager.scala:1240) >> at >> org.apache.flink.runtime.jobmanager.JobManager.getJobManagerRemoteReference(JobManager.scala) >> at org.apache.flink.client.program.Client.run(Client.java:375) >> ... 15 more >> Caused by: akka.actor.ActorNotFound: Actor not found for: >> ActorSelection[Anchor(akka.tcp://flink@127.0.0.1:6123/), >> Path(/user/jobmanager)] >> at >> akka.actor.ActorSelection$$anonfun$resolveOne$1.apply(ActorSelection.scala:65) >> at >> akka.actor.ActorSelection$$anonfun$resolveOne$1.apply(ActorSelection.scala:63) >> at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32) >> at >> akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.processBatch$1(BatchingExecutor.scala:67) >> at >> akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.apply$mcV$sp(BatchingExecutor.scala:82) >> at >> akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.apply(BatchingExecutor.scala:59) >> at >> akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.apply(BatchingExecutor.scala:59) >> at >> scala.concurrent.BlockContext$.withBlockContext(BlockContext.scala:72) >> at akka.dispatch.BatchingExecutor$Batch.run(BatchingExecutor.scala:58) >> at >> akka.dispatch.ExecutionContexts$sameThreadExecutionContext$.unbatchedExecute(Future.scala:74) >> at >> akka.dispatch.BatchingExecutor$class.execute(BatchingExecutor.scala:110) >> at >> akka.dispatch.ExecutionContexts$sameThreadExecutionContext$.execute(Future.scala:73) >> at >> scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40) >> at >> scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248) >> at akka.pattern.PromiseActorRef.$bang(AskSupport.scala:267) >> at akka.actor.EmptyLocalActorRef.specialHandle(ActorRef.scala:508) >> at akka.actor.DeadLetterActorRef.specialHandle(ActorRef.scala:541) >> at akka.actor.DeadLetterActorRef.$bang(ActorRef.scala:531) >> at >> akka.remote.RemoteActorRefProvider$RemoteDeadLetterActorRef.$bang(RemoteActorRefProvider.scala:87) >> at akka.remote.EndpointWriter.postStop(Endpoint.scala:561) >> at akka.actor.Actor$class.aroundPostStop(Actor.scala:475) >> at akka.remote.EndpointActor.aroundPostStop(Endpoint.scala:415) >> at >> akka.actor.dungeon.FaultHandling$class.akka$actor$dungeon$FaultHandling$$finishTerminate(FaultHandling.scala:210) >> at >> akka.actor.dungeon.FaultHandling$class.terminate(FaultHandling.scala:172) >> at akka.actor.ActorCell.terminate(ActorCell.scala:369) >> at akka.actor.ActorCell.invokeAll$1(ActorCell.scala:462) >> at akka.actor.ActorCell.systemInvoke(ActorCell.scala:478) >> at akka.dispatch.Mailbox.processAllSystemMessages(Mailbox.scala:279) >> at akka.dispatch.Mailbox.run(Mailbox.scala:220) >> at akka.dispatch.Mailbox.exec(Mailbox.scala:231) >> at >> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) >> at >> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) >> at >> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) >> at >> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) >> >> the FlinkMain.java: 70 is: >> >> env.execute("KMeans Flink"); >> >> >> 2015-06-04 17:17 GMT+02:00 Pa Rö <paul.roewer1...@googlemail.com>: >> >>> i try this: >>> >>> [cloudera@quickstart bin]$ sudo su yarn >>> bash-4.1$ hadoop fs -chmod 777 /user/cloudera/outputs >>> chmod: changing permissions of '/user/cloudera/outputs': Permission >>> denied. user=yarn is not the owner of inode=outputs >>> bash-4.1$ hadoop fs -chmod 777 /user/cloudera/inputs >>> chmod: changing permissions of '/user/cloudera/inputs': Permission >>> denied. user=yarn is not the owner of inode=inputs >>> bash-4.1$ exit >>> exit >>> [cloudera@quickstart bin]$ sudo ./flink run >>> /home/cloudera/Desktop/ma-flink.jar >>> log4j:WARN No appenders could be found for logger >>> (org.apache.hadoop.metrics2.lib.MutableMetricsFactory). >>> log4j:WARN Please initialize the log4j system properly. >>> log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig >>> for more info. >>> Found YARN properties file >>> /home/cloudera/Desktop/flink-0.9-SNAPSHOT/bin/../conf/.yarn-properties >>> Using JobManager address from YARN properties quickstart.cloudera/ >>> 127.0.0.1:53874 >>> org.apache.flink.client.program.ProgramInvocationException: The program >>> execution failed: Failed to submit job 2f46ef5dff4ecf5552b3477ed1c6f4b9 >>> (KMeans Flink) >>> at org.apache.flink.client.program.Client.run(Client.java:412) >>> at org.apache.flink.client.program.Client.run(Client.java:355) >>> at org.apache.flink.client.program.Client.run(Client.java:348) >>> at >>> org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:63) >>> at mgm.tp.bigdata.ma_flink.FlinkMain.main(FlinkMain.java:70) >>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>> at >>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>> at >>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>> at java.lang.reflect.Method.invoke(Method.java:606) >>> at >>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:437) >>> at >>> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:353) >>> at org.apache.flink.client.program.Client.run(Client.java:315) >>> at >>> org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:584) >>> at org.apache.flink.client.CliFrontend.run(CliFrontend.java:290) >>> at >>> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:880) >>> at org.apache.flink.client.CliFrontend.main(CliFrontend.java:922) >>> Caused by: org.apache.flink.runtime.client.JobExecutionException: Failed >>> to submit job 2f46ef5dff4ecf5552b3477ed1c6f4b9 (KMeans Flink) >>> at org.apache.flink.runtime.jobmanager.JobManager.org >>> $apache$flink$runtime$jobmanager$JobManager$$submitJob(JobManager.scala:595) >>> at >>> org.apache.flink.runtime.jobmanager.JobManager$$anonfun$receiveWithLogMessages$1.applyOrElse(JobManager.scala:192) >>> at >>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33) >>> at >>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33) >>> at >>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25) >>> at >>> org.apache.flink.yarn.ApplicationMasterActor$$anonfun$receiveYarnMessages$1.applyOrElse(ApplicationMasterActor.scala:99) >>> at scala.PartialFunction$OrElse.apply(PartialFunction.scala:162) >>> at >>> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:36) >>> at >>> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:29) >>> at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118) >>> at >>> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:29) >>> at akka.actor.Actor$class.aroundReceive(Actor.scala:465) >>> at >>> org.apache.flink.runtime.jobmanager.JobManager.aroundReceive(JobManager.scala:94) >>> at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516) >>> at akka.actor.ActorCell.invoke(ActorCell.scala:487) >>> at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254) >>> at akka.dispatch.Mailbox.run(Mailbox.scala:221) >>> at akka.dispatch.Mailbox.exec(Mailbox.scala:231) >>> at >>> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) >>> at >>> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) >>> at >>> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) >>> at >>> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) >>> Caused by: org.apache.flink.runtime.JobException: Creating the input >>> splits caused an error: File /user/cloudera/inputs does not exist or the >>> user running Flink ('yarn') has insufficient permissions to access it. >>> at >>> org.apache.flink.runtime.executiongraph.ExecutionJobVertex.<init>(ExecutionJobVertex.java:162) >>> at >>> org.apache.flink.runtime.executiongraph.ExecutionGraph.attachJobGraph(ExecutionGraph.java:471) >>> at org.apache.flink.runtime.jobmanager.JobManager.org >>> $apache$flink$runtime$jobmanager$JobManager$$submitJob(JobManager.scala:535) >>> ... 21 more >>> Caused by: java.io.FileNotFoundException: File /user/cloudera/inputs >>> does not exist or the user running Flink ('yarn') has insufficient >>> permissions to access it. >>> at >>> org.apache.flink.core.fs.local.LocalFileSystem.getFileStatus(LocalFileSystem.java:106) >>> at >>> org.apache.flink.api.common.io.FileInputFormat.createInputSplits(FileInputFormat.java:390) >>> at >>> org.apache.flink.api.common.io.FileInputFormat.createInputSplits(FileInputFormat.java:51) >>> at >>> org.apache.flink.runtime.executiongraph.ExecutionJobVertex.<init>(ExecutionJobVertex.java:146) >>> ... 23 more >>> >>> >>> 2015-06-04 17:15 GMT+02:00 Robert Metzger <rmetz...@apache.org>: >>> >>>> As the output of the "hadoop" tool indicates, it expects two arguments, >>>> you only passed one (777). >>>> The second argument it is expecting is the path to the file you want to >>>> change. >>>> >>>> In your case, it is: >>>> hadoop fs -chmod 777 /user/cloudera/outputs >>>> >>>> >>>> The reason why >>>> hadoop fs -chmod 777 * >>>> does not work is the following: the * is evaluated by your local bash >>>> and expanded to the files which are present in your current, local >>>> directory. The bash expansion is not able to expand to the files in HDFS. >>>> >>>> >>>> On Thu, Jun 4, 2015 at 5:08 PM, Pa Rö <paul.roewer1...@googlemail.com> >>>> wrote: >>>> >>>>> [cloudera@quickstart bin]$ sudo su yarn >>>>> bash-4.1$ hadoop fs -chmod 777 >>>>> -chmod: Not enough arguments: expected 2 but got 1 >>>>> Usage: hadoop fs [generic options] -chmod [-R] <MODE[,MODE]... | >>>>> OCTALMODE> PATH... >>>>> bash-4.1$ >>>>> >>>>> you understand? >>>>> >>>>> 2015-06-04 17:04 GMT+02:00 Robert Metzger <rmetz...@apache.org>: >>>>> >>>>>> It looks like the user "yarn" which is running Flink doesn't have >>>>>> permission to access the files. >>>>>> >>>>>> Can you do "sudo su yarn" to become the "yarn" user. Then, you can do >>>>>> "hadoop fs -chmod 777" to make the files accessible for everyone. >>>>>> >>>>>> >>>>>> On Thu, Jun 4, 2015 at 4:59 PM, Pa Rö <paul.roewer1...@googlemail.com >>>>>> > wrote: >>>>>> >>>>>>> okay, it's work, i get a exception: >>>>>>> >>>>>>> [cloudera@quickstart Desktop]$ cd flink-0.9-SNAPSHOT/bin/ >>>>>>> [cloudera@quickstart bin]$ flink run >>>>>>> /home/cloudera/Desktop/ma-flink.jar >>>>>>> bash: flink: command not found >>>>>>> [cloudera@quickstart bin]$ ./flink run >>>>>>> /home/cloudera/Desktop/ma-flink.jar >>>>>>> log4j:WARN No appenders could be found for logger >>>>>>> (org.apache.hadoop.metrics2.lib.MutableMetricsFactory). >>>>>>> log4j:WARN Please initialize the log4j system properly. >>>>>>> log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig >>>>>>> for more info. >>>>>>> Found YARN properties file >>>>>>> /home/cloudera/Desktop/flink-0.9-SNAPSHOT/bin/../conf/.yarn-properties >>>>>>> Using JobManager address from YARN properties quickstart.cloudera/ >>>>>>> 127.0.0.1:53874 >>>>>>> java.io.IOException: Mkdirs failed to create /user/cloudera/outputs >>>>>>> at >>>>>>> org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:438) >>>>>>> at >>>>>>> org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:424) >>>>>>> at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:905) >>>>>>> at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:886) >>>>>>> at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:783) >>>>>>> at >>>>>>> mgm.tp.bigdata.ma_commons.commons.Seeding.randomSeeding(Seeding.java:21) >>>>>>> at >>>>>>> mgm.tp.bigdata.ma_flink.FlinkMain.getCentroidDataSet(FlinkMain.java:178) >>>>>>> at mgm.tp.bigdata.ma_flink.FlinkMain.main(FlinkMain.java:47) >>>>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>>>>> at >>>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>>>>>> at >>>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>>>>> at java.lang.reflect.Method.invoke(Method.java:606) >>>>>>> at >>>>>>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:437) >>>>>>> at >>>>>>> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:353) >>>>>>> at org.apache.flink.client.program.Client.run(Client.java:315) >>>>>>> at >>>>>>> org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:584) >>>>>>> at org.apache.flink.client.CliFrontend.run(CliFrontend.java:290) >>>>>>> at >>>>>>> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:880) >>>>>>> at org.apache.flink.client.CliFrontend.main(CliFrontend.java:922) >>>>>>> org.apache.flink.client.program.ProgramInvocationException: The >>>>>>> program execution failed: Failed to submit job >>>>>>> 934743a5c49c6d5e31c9e8201452e36d (KMeans Flink) >>>>>>> at org.apache.flink.client.program.Client.run(Client.java:412) >>>>>>> at org.apache.flink.client.program.Client.run(Client.java:355) >>>>>>> at org.apache.flink.client.program.Client.run(Client.java:348) >>>>>>> at >>>>>>> org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:63) >>>>>>> at mgm.tp.bigdata.ma_flink.FlinkMain.main(FlinkMain.java:70) >>>>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>>>>> at >>>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>>>>>> at >>>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>>>>> at java.lang.reflect.Method.invoke(Method.java:606) >>>>>>> at >>>>>>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:437) >>>>>>> at >>>>>>> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:353) >>>>>>> at org.apache.flink.client.program.Client.run(Client.java:315) >>>>>>> at >>>>>>> org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:584) >>>>>>> at org.apache.flink.client.CliFrontend.run(CliFrontend.java:290) >>>>>>> at >>>>>>> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:880) >>>>>>> at org.apache.flink.client.CliFrontend.main(CliFrontend.java:922) >>>>>>> Caused by: org.apache.flink.runtime.client.JobExecutionException: >>>>>>> Failed to submit job 934743a5c49c6d5e31c9e8201452e36d (KMeans Flink) >>>>>>> at org.apache.flink.runtime.jobmanager.JobManager.org >>>>>>> $apache$flink$runtime$jobmanager$JobManager$$submitJob(JobManager.scala:595) >>>>>>> at >>>>>>> org.apache.flink.runtime.jobmanager.JobManager$$anonfun$receiveWithLogMessages$1.applyOrElse(JobManager.scala:192) >>>>>>> at >>>>>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33) >>>>>>> at >>>>>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33) >>>>>>> at >>>>>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25) >>>>>>> at >>>>>>> org.apache.flink.yarn.ApplicationMasterActor$$anonfun$receiveYarnMessages$1.applyOrElse(ApplicationMasterActor.scala:99) >>>>>>> at scala.PartialFunction$OrElse.apply(PartialFunction.scala:162) >>>>>>> at >>>>>>> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:36) >>>>>>> at >>>>>>> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:29) >>>>>>> at >>>>>>> scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118) >>>>>>> at >>>>>>> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:29) >>>>>>> at akka.actor.Actor$class.aroundReceive(Actor.scala:465) >>>>>>> at >>>>>>> org.apache.flink.runtime.jobmanager.JobManager.aroundReceive(JobManager.scala:94) >>>>>>> at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516) >>>>>>> at akka.actor.ActorCell.invoke(ActorCell.scala:487) >>>>>>> at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254) >>>>>>> at akka.dispatch.Mailbox.run(Mailbox.scala:221) >>>>>>> at akka.dispatch.Mailbox.exec(Mailbox.scala:231) >>>>>>> at >>>>>>> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) >>>>>>> at >>>>>>> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) >>>>>>> at >>>>>>> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) >>>>>>> at >>>>>>> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) >>>>>>> Caused by: org.apache.flink.runtime.JobException: Creating the input >>>>>>> splits caused an error: File /user/cloudera/outputs/seed-1 does not >>>>>>> exist >>>>>>> or the user running Flink ('yarn') has insufficient permissions to >>>>>>> access >>>>>>> it. >>>>>>> at >>>>>>> org.apache.flink.runtime.executiongraph.ExecutionJobVertex.<init>(ExecutionJobVertex.java:162) >>>>>>> at >>>>>>> org.apache.flink.runtime.executiongraph.ExecutionGraph.attachJobGraph(ExecutionGraph.java:471) >>>>>>> at org.apache.flink.runtime.jobmanager.JobManager.org >>>>>>> $apache$flink$runtime$jobmanager$JobManager$$submitJob(JobManager.scala:535) >>>>>>> ... 21 more >>>>>>> Caused by: java.io.FileNotFoundException: File >>>>>>> /user/cloudera/outputs/seed-1 does not exist or the user running Flink >>>>>>> ('yarn') has insufficient permissions to access it. >>>>>>> at >>>>>>> org.apache.flink.core.fs.local.LocalFileSystem.getFileStatus(LocalFileSystem.java:106) >>>>>>> at >>>>>>> org.apache.flink.api.common.io.FileInputFormat.createInputSplits(FileInputFormat.java:390) >>>>>>> at >>>>>>> org.apache.flink.api.common.io.FileInputFormat.createInputSplits(FileInputFormat.java:51) >>>>>>> at >>>>>>> org.apache.flink.runtime.executiongraph.ExecutionJobVertex.<init>(ExecutionJobVertex.java:146) >>>>>>> ... 23 more >>>>>>> >>>>>>> >>>>>>> how i must set the files in the hdfs? >>>>>>> quickstart.cloudera:50075/home/cloudera/output? >>>>>>> >>>>>>> 2015-06-04 16:51 GMT+02:00 Robert Metzger <rmetz...@apache.org>: >>>>>>> >>>>>>>> Once you've started the YARN session, you can submit a Flink job >>>>>>>> with "./bin/flink run <pathToYourJar>". >>>>>>>> >>>>>>>> The jar file of your job doesn't need to be in HDFS. It has to be >>>>>>>> in the local file system and flink will send it to all machines. >>>>>>>> >>>>>>>> On Thu, Jun 4, 2015 at 4:48 PM, Pa Rö < >>>>>>>> paul.roewer1...@googlemail.com> wrote: >>>>>>>> >>>>>>>>> okay, now it run on my hadoop. >>>>>>>>> how i can start my flink job? and where must the jar file save, at >>>>>>>>> hdfs or as local file? >>>>>>>>> >>>>>>>>> 2015-06-04 16:31 GMT+02:00 Robert Metzger <rmetz...@apache.org>: >>>>>>>>> >>>>>>>>>> Yes, you have to run these commands in the command line of the >>>>>>>>>> Cloudera VM. >>>>>>>>>> >>>>>>>>>> On Thu, Jun 4, 2015 at 4:28 PM, Pa Rö < >>>>>>>>>> paul.roewer1...@googlemail.com> wrote: >>>>>>>>>> >>>>>>>>>>> you mean run this command on terminal/shell and not define a hue >>>>>>>>>>> job? >>>>>>>>>>> >>>>>>>>>>> 2015-06-04 16:25 GMT+02:00 Robert Metzger <rmetz...@apache.org>: >>>>>>>>>>> >>>>>>>>>>>> It should be certainly possible to run Flink on a cloudera live >>>>>>>>>>>> VM >>>>>>>>>>>> >>>>>>>>>>>> I think these are the commands you need to execute: >>>>>>>>>>>> >>>>>>>>>>>> wget >>>>>>>>>>>> http://stratosphere-bin.s3-website-us-east-1.amazonaws.com/flink-0.9-SNAPSHOT-bin-hadoop2.tgz >>>>>>>>>>>> tar xvzf flink-0.9-SNAPSHOT-bin-hadoop2.tgz >>>>>>>>>>>> cd flink-0.9-SNAPSHOT/ >>>>>>>>>>>> *export HADOOP_CONF_DIR=/usr/lib/hadoop/etc/hadoop/* >>>>>>>>>>>> ./bin/yarn-session.sh -n 1 -jm 1024 -tm 1024 >>>>>>>>>>>> >>>>>>>>>>>> If that is not working for you, please post the exact error >>>>>>>>>>>> message you are getting and I can help you to get it to run. >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Thu, Jun 4, 2015 at 4:18 PM, Pa Rö < >>>>>>>>>>>> paul.roewer1...@googlemail.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> hi robert, >>>>>>>>>>>>> >>>>>>>>>>>>> i think the problem is the hue api, >>>>>>>>>>>>> i had the same problem with spark submit script, >>>>>>>>>>>>> but on the new hue release, they have a spark submit api. >>>>>>>>>>>>> >>>>>>>>>>>>> i asked the group for the same problem with spark, no reply. >>>>>>>>>>>>> >>>>>>>>>>>>> i want test my app on local cluster, before i run it on the >>>>>>>>>>>>> big cluster, >>>>>>>>>>>>> for that i use cloudera live. maybe it give an other way to >>>>>>>>>>>>> test flink on a local cluster vm? >>>>>>>>>>>>> >>>>>>>>>>>>> 2015-06-04 16:12 GMT+02:00 Robert Metzger <rmetz...@apache.org >>>>>>>>>>>>> >: >>>>>>>>>>>>> >>>>>>>>>>>>>> Hi Paul, >>>>>>>>>>>>>> >>>>>>>>>>>>>> why did running Flink from the regular scripts not work for >>>>>>>>>>>>>> you? >>>>>>>>>>>>>> >>>>>>>>>>>>>> I'm not an expert on Hue, I would recommend asking in the Hue >>>>>>>>>>>>>> user forum / mailing list: >>>>>>>>>>>>>> https://groups.google.com/a/cloudera.org/forum/#!forum/hue-user >>>>>>>>>>>>>> . >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Thu, Jun 4, 2015 at 4:09 PM, Pa Rö < >>>>>>>>>>>>>> paul.roewer1...@googlemail.com> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> thanks, >>>>>>>>>>>>>>> now i want run my app on cloudera live vm single node, >>>>>>>>>>>>>>> how i can define my flink job with hue? >>>>>>>>>>>>>>> i try to run the flink script in the hdfs, it's not work. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> best regards, >>>>>>>>>>>>>>> paul >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> 2015-06-02 14:50 GMT+02:00 Robert Metzger < >>>>>>>>>>>>>>> rmetz...@apache.org>: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> I would recommend using HDFS. >>>>>>>>>>>>>>>> For that, you need to specify the paths like this: >>>>>>>>>>>>>>>> hdfs:///path/to/data. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On Tue, Jun 2, 2015 at 2:48 PM, Pa Rö < >>>>>>>>>>>>>>>> paul.roewer1...@googlemail.com> wrote: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> nice, >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> which file system i must use for the cluster? java.io or >>>>>>>>>>>>>>>>> hadoop.fs or flink? >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> 2015-06-02 14:29 GMT+02:00 Robert Metzger < >>>>>>>>>>>>>>>>> rmetz...@apache.org>: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>>>>> you can start Flink on YARN on the Cloudera distribution. >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> See here for more: >>>>>>>>>>>>>>>>>> http://ci.apache.org/projects/flink/flink-docs-master/setup/yarn_setup.html >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> These are the commands you need to execute >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> wget >>>>>>>>>>>>>>>>>> http://stratosphere-bin.s3-website-us-east-1.amazonaws.com/flink-0.9-SNAPSHOT-bin-hadoop2.tgz >>>>>>>>>>>>>>>>>> tar xvzf flink-0.9-SNAPSHOT-bin-hadoop2.tgzcd >>>>>>>>>>>>>>>>>> flink-0.9-SNAPSHOT/ >>>>>>>>>>>>>>>>>> ./bin/yarn-session.sh -n 4 -jm 1024 -tm 4096 >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> On Tue, Jun 2, 2015 at 2:03 PM, Pa Rö < >>>>>>>>>>>>>>>>>> paul.roewer1...@googlemail.com> wrote: >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> hi community, >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> i want test my flink k-means on a hadoop cluster. i use >>>>>>>>>>>>>>>>>>> the cloudera live distribution. how i can run flink on this >>>>>>>>>>>>>>>>>>> cluster? maybe >>>>>>>>>>>>>>>>>>> only the java dependencies are engouth? >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> best regards, >>>>>>>>>>>>>>>>>>> paul >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >