i have change the permissions from the cloudera user and try the following
command.
and the files exist on hdfs ;) i set the files in my properties file like
"flink.output=/user/cloudera/outputs/output_flink"
i get the same exception again, maybe the problem have an other reason?

[cloudera@quickstart bin]$ sudo su yarn
bash-4.1$ hadoop fs -chmod 777 /user/cloudera/inputs
bash-4.1$ hadoop fs -chmod 777 /user/cloudera/outputs
bash-4.1$ exit
exit
[cloudera@quickstart bin]$ sudo ./flink run
/home/cloudera/Desktop/ma-flink.jar
log4j:WARN No appenders could be found for logger
(org.apache.hadoop.metrics2.lib.MutableMetricsFactory).
log4j:WARN Please initialize the log4j system properly.
log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for
more info.
Found YARN properties file
/home/cloudera/Desktop/flink-0.9-SNAPSHOT/bin/../conf/.yarn-properties
Using JobManager address from YARN properties quickstart.cloudera/
127.0.0.1:52601
org.apache.flink.client.program.ProgramInvocationException: The program
execution failed: Failed to submit job fe78e9ee50cf76ac8b487919e1c951fa
(KMeans Flink)
    at org.apache.flink.client.program.Client.run(Client.java:412)
    at org.apache.flink.client.program.Client.run(Client.java:355)
    at org.apache.flink.client.program.Client.run(Client.java:348)
    at
org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:63)
    at mgm.tp.bigdata.ma_flink.FlinkMain.main(FlinkMain.java:70)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
    at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
    at java.lang.reflect.Method.invoke(Method.java:606)
    at
org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:437)
    at
org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:353)
    at org.apache.flink.client.program.Client.run(Client.java:315)
    at
org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:584)
    at org.apache.flink.client.CliFrontend.run(CliFrontend.java:290)
    at
org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:880)
    at org.apache.flink.client.CliFrontend.main(CliFrontend.java:922)
Caused by: org.apache.flink.runtime.client.JobExecutionException: Failed to
submit job fe78e9ee50cf76ac8b487919e1c951fa (KMeans Flink)
    at org.apache.flink.runtime.jobmanager.JobManager.org
$apache$flink$runtime$jobmanager$JobManager$$submitJob(JobManager.scala:595)
    at
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$receiveWithLogMessages$1.applyOrElse(JobManager.scala:192)
    at
scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
    at
scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
    at
scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
    at
org.apache.flink.yarn.ApplicationMasterActor$$anonfun$receiveYarnMessages$1.applyOrElse(ApplicationMasterActor.scala:99)
    at scala.PartialFunction$OrElse.apply(PartialFunction.scala:162)
    at
org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:36)
    at
org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:29)
    at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
    at
org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:29)
    at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
    at
org.apache.flink.runtime.jobmanager.JobManager.aroundReceive(JobManager.scala:94)
    at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
    at akka.actor.ActorCell.invoke(ActorCell.scala:487)
    at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
    at akka.dispatch.Mailbox.run(Mailbox.scala:221)
    at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
    at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
    at
scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
    at
scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
    at
scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
Caused by: org.apache.flink.runtime.JobException: Creating the input splits
caused an error: File /user/cloudera/inputs does not exist or the user
running Flink ('yarn') has insufficient permissions to access it.
    at
org.apache.flink.runtime.executiongraph.ExecutionJobVertex.<init>(ExecutionJobVertex.java:162)
    at
org.apache.flink.runtime.executiongraph.ExecutionGraph.attachJobGraph(ExecutionGraph.java:471)
    at org.apache.flink.runtime.jobmanager.JobManager.org
$apache$flink$runtime$jobmanager$JobManager$$submitJob(JobManager.scala:535)
    ... 21 more
Caused by: java.io.FileNotFoundException: File /user/cloudera/inputs does
not exist or the user running Flink ('yarn') has insufficient permissions
to access it.
    at
org.apache.flink.core.fs.local.LocalFileSystem.getFileStatus(LocalFileSystem.java:106)
    at
org.apache.flink.api.common.io.FileInputFormat.createInputSplits(FileInputFormat.java:390)
    at
org.apache.flink.api.common.io.FileInputFormat.createInputSplits(FileInputFormat.java:51)
    at
org.apache.flink.runtime.executiongraph.ExecutionJobVertex.<init>(ExecutionJobVertex.java:146)
    ... 23 more


2015-06-04 17:38 GMT+02:00 Pa Rö <paul.roewer1...@googlemail.com>:

> sorry, i see my yarn end before i can run my app, i must set the write
> access for yarn, maybe this solve my problem.
>
> 2015-06-04 17:33 GMT+02:00 Pa Rö <paul.roewer1...@googlemail.com>:
>
>> i start the yarn-session.sh with sudo
>> and than the flink run command with sudo,
>> i get the following exception:
>>
>> cloudera@quickstart bin]$ sudo ./flink run
>> /home/cloudera/Desktop/ma-flink.jar
>> log4j:WARN No appenders could be found for logger
>> (org.apache.hadoop.metrics2.lib.MutableMetricsFactory).
>> log4j:WARN Please initialize the log4j system properly.
>> log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for
>> more info.
>> org.apache.flink.client.program.ProgramInvocationException: Failed to
>> resolve JobManager
>>     at org.apache.flink.client.program.Client.run(Client.java:378)
>>     at org.apache.flink.client.program.Client.run(Client.java:355)
>>     at org.apache.flink.client.program.Client.run(Client.java:348)
>>     at
>> org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:63)
>>     at mgm.tp.bigdata.ma_flink.FlinkMain.main(FlinkMain.java:70)
>>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>     at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>     at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>     at java.lang.reflect.Method.invoke(Method.java:606)
>>     at
>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:437)
>>     at
>> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:353)
>>     at org.apache.flink.client.program.Client.run(Client.java:315)
>>     at
>> org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:584)
>>     at org.apache.flink.client.CliFrontend.run(CliFrontend.java:290)
>>     at
>> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:880)
>>     at org.apache.flink.client.CliFrontend.main(CliFrontend.java:922)
>> Caused by: java.io.IOException: JobManager at akka.tcp://
>> flink@127.0.0.1:6123/user/jobmanager not reachable. Please make sure
>> that the JobManager is running and its port is reachable.
>>     at
>> org.apache.flink.runtime.jobmanager.JobManager$.getJobManagerRemoteReference(JobManager.scala:1198)
>>     at
>> org.apache.flink.runtime.jobmanager.JobManager$.getJobManagerRemoteReference(JobManager.scala:1222)
>>     at
>> org.apache.flink.runtime.jobmanager.JobManager$.getJobManagerRemoteReference(JobManager.scala:1240)
>>     at
>> org.apache.flink.runtime.jobmanager.JobManager.getJobManagerRemoteReference(JobManager.scala)
>>     at org.apache.flink.client.program.Client.run(Client.java:375)
>>     ... 15 more
>> Caused by: akka.actor.ActorNotFound: Actor not found for:
>> ActorSelection[Anchor(akka.tcp://flink@127.0.0.1:6123/),
>> Path(/user/jobmanager)]
>>     at
>> akka.actor.ActorSelection$$anonfun$resolveOne$1.apply(ActorSelection.scala:65)
>>     at
>> akka.actor.ActorSelection$$anonfun$resolveOne$1.apply(ActorSelection.scala:63)
>>     at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32)
>>     at
>> akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.processBatch$1(BatchingExecutor.scala:67)
>>     at
>> akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.apply$mcV$sp(BatchingExecutor.scala:82)
>>     at
>> akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.apply(BatchingExecutor.scala:59)
>>     at
>> akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.apply(BatchingExecutor.scala:59)
>>     at
>> scala.concurrent.BlockContext$.withBlockContext(BlockContext.scala:72)
>>     at akka.dispatch.BatchingExecutor$Batch.run(BatchingExecutor.scala:58)
>>     at
>> akka.dispatch.ExecutionContexts$sameThreadExecutionContext$.unbatchedExecute(Future.scala:74)
>>     at
>> akka.dispatch.BatchingExecutor$class.execute(BatchingExecutor.scala:110)
>>     at
>> akka.dispatch.ExecutionContexts$sameThreadExecutionContext$.execute(Future.scala:73)
>>     at
>> scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
>>     at
>> scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
>>     at akka.pattern.PromiseActorRef.$bang(AskSupport.scala:267)
>>     at akka.actor.EmptyLocalActorRef.specialHandle(ActorRef.scala:508)
>>     at akka.actor.DeadLetterActorRef.specialHandle(ActorRef.scala:541)
>>     at akka.actor.DeadLetterActorRef.$bang(ActorRef.scala:531)
>>     at
>> akka.remote.RemoteActorRefProvider$RemoteDeadLetterActorRef.$bang(RemoteActorRefProvider.scala:87)
>>     at akka.remote.EndpointWriter.postStop(Endpoint.scala:561)
>>     at akka.actor.Actor$class.aroundPostStop(Actor.scala:475)
>>     at akka.remote.EndpointActor.aroundPostStop(Endpoint.scala:415)
>>     at
>> akka.actor.dungeon.FaultHandling$class.akka$actor$dungeon$FaultHandling$$finishTerminate(FaultHandling.scala:210)
>>     at
>> akka.actor.dungeon.FaultHandling$class.terminate(FaultHandling.scala:172)
>>     at akka.actor.ActorCell.terminate(ActorCell.scala:369)
>>     at akka.actor.ActorCell.invokeAll$1(ActorCell.scala:462)
>>     at akka.actor.ActorCell.systemInvoke(ActorCell.scala:478)
>>     at akka.dispatch.Mailbox.processAllSystemMessages(Mailbox.scala:279)
>>     at akka.dispatch.Mailbox.run(Mailbox.scala:220)
>>     at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
>>     at
>> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>>     at
>> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
>>     at
>> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
>>     at
>> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
>>
>> the FlinkMain.java: 70 is:
>>
>> env.execute("KMeans Flink");
>>
>>
>> 2015-06-04 17:17 GMT+02:00 Pa Rö <paul.roewer1...@googlemail.com>:
>>
>>> i try this:
>>>
>>> [cloudera@quickstart bin]$ sudo su yarn
>>> bash-4.1$ hadoop fs -chmod 777 /user/cloudera/outputs
>>> chmod: changing permissions of '/user/cloudera/outputs': Permission
>>> denied. user=yarn is not the owner of inode=outputs
>>> bash-4.1$ hadoop fs -chmod 777 /user/cloudera/inputs
>>> chmod: changing permissions of '/user/cloudera/inputs': Permission
>>> denied. user=yarn is not the owner of inode=inputs
>>> bash-4.1$ exit
>>> exit
>>> [cloudera@quickstart bin]$ sudo ./flink run
>>> /home/cloudera/Desktop/ma-flink.jar
>>> log4j:WARN No appenders could be found for logger
>>> (org.apache.hadoop.metrics2.lib.MutableMetricsFactory).
>>> log4j:WARN Please initialize the log4j system properly.
>>> log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig
>>> for more info.
>>> Found YARN properties file
>>> /home/cloudera/Desktop/flink-0.9-SNAPSHOT/bin/../conf/.yarn-properties
>>> Using JobManager address from YARN properties quickstart.cloudera/
>>> 127.0.0.1:53874
>>> org.apache.flink.client.program.ProgramInvocationException: The program
>>> execution failed: Failed to submit job 2f46ef5dff4ecf5552b3477ed1c6f4b9
>>> (KMeans Flink)
>>>     at org.apache.flink.client.program.Client.run(Client.java:412)
>>>     at org.apache.flink.client.program.Client.run(Client.java:355)
>>>     at org.apache.flink.client.program.Client.run(Client.java:348)
>>>     at
>>> org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:63)
>>>     at mgm.tp.bigdata.ma_flink.FlinkMain.main(FlinkMain.java:70)
>>>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>     at
>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>     at
>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>     at java.lang.reflect.Method.invoke(Method.java:606)
>>>     at
>>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:437)
>>>     at
>>> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:353)
>>>     at org.apache.flink.client.program.Client.run(Client.java:315)
>>>     at
>>> org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:584)
>>>     at org.apache.flink.client.CliFrontend.run(CliFrontend.java:290)
>>>     at
>>> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:880)
>>>     at org.apache.flink.client.CliFrontend.main(CliFrontend.java:922)
>>> Caused by: org.apache.flink.runtime.client.JobExecutionException: Failed
>>> to submit job 2f46ef5dff4ecf5552b3477ed1c6f4b9 (KMeans Flink)
>>>     at org.apache.flink.runtime.jobmanager.JobManager.org
>>> $apache$flink$runtime$jobmanager$JobManager$$submitJob(JobManager.scala:595)
>>>     at
>>> org.apache.flink.runtime.jobmanager.JobManager$$anonfun$receiveWithLogMessages$1.applyOrElse(JobManager.scala:192)
>>>     at
>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
>>>     at
>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
>>>     at
>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
>>>     at
>>> org.apache.flink.yarn.ApplicationMasterActor$$anonfun$receiveYarnMessages$1.applyOrElse(ApplicationMasterActor.scala:99)
>>>     at scala.PartialFunction$OrElse.apply(PartialFunction.scala:162)
>>>     at
>>> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:36)
>>>     at
>>> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:29)
>>>     at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
>>>     at
>>> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:29)
>>>     at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
>>>     at
>>> org.apache.flink.runtime.jobmanager.JobManager.aroundReceive(JobManager.scala:94)
>>>     at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
>>>     at akka.actor.ActorCell.invoke(ActorCell.scala:487)
>>>     at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
>>>     at akka.dispatch.Mailbox.run(Mailbox.scala:221)
>>>     at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
>>>     at
>>> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>>>     at
>>> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
>>>     at
>>> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
>>>     at
>>> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
>>> Caused by: org.apache.flink.runtime.JobException: Creating the input
>>> splits caused an error: File /user/cloudera/inputs does not exist or the
>>> user running Flink ('yarn') has insufficient permissions to access it.
>>>     at
>>> org.apache.flink.runtime.executiongraph.ExecutionJobVertex.<init>(ExecutionJobVertex.java:162)
>>>     at
>>> org.apache.flink.runtime.executiongraph.ExecutionGraph.attachJobGraph(ExecutionGraph.java:471)
>>>     at org.apache.flink.runtime.jobmanager.JobManager.org
>>> $apache$flink$runtime$jobmanager$JobManager$$submitJob(JobManager.scala:535)
>>>     ... 21 more
>>> Caused by: java.io.FileNotFoundException: File /user/cloudera/inputs
>>> does not exist or the user running Flink ('yarn') has insufficient
>>> permissions to access it.
>>>     at
>>> org.apache.flink.core.fs.local.LocalFileSystem.getFileStatus(LocalFileSystem.java:106)
>>>     at
>>> org.apache.flink.api.common.io.FileInputFormat.createInputSplits(FileInputFormat.java:390)
>>>     at
>>> org.apache.flink.api.common.io.FileInputFormat.createInputSplits(FileInputFormat.java:51)
>>>     at
>>> org.apache.flink.runtime.executiongraph.ExecutionJobVertex.<init>(ExecutionJobVertex.java:146)
>>>     ... 23 more
>>>
>>>
>>> 2015-06-04 17:15 GMT+02:00 Robert Metzger <rmetz...@apache.org>:
>>>
>>>> As the output of the "hadoop" tool indicates, it expects two arguments,
>>>> you only passed one (777).
>>>> The second argument it is expecting is the path to the file you want to
>>>> change.
>>>>
>>>> In your case, it is:
>>>> hadoop fs -chmod 777 /user/cloudera/outputs
>>>>
>>>>
>>>> The reason why
>>>> hadoop fs -chmod 777 *
>>>> does not work is the following: the * is evaluated by your local bash
>>>> and expanded to the files which are present in your current, local
>>>> directory. The bash expansion is not able to expand to the files in HDFS.
>>>>
>>>>
>>>> On Thu, Jun 4, 2015 at 5:08 PM, Pa Rö <paul.roewer1...@googlemail.com>
>>>> wrote:
>>>>
>>>>> [cloudera@quickstart bin]$ sudo su yarn
>>>>> bash-4.1$ hadoop fs -chmod 777
>>>>> -chmod: Not enough arguments: expected 2 but got 1
>>>>> Usage: hadoop fs [generic options] -chmod [-R] <MODE[,MODE]... |
>>>>> OCTALMODE> PATH...
>>>>> bash-4.1$
>>>>>
>>>>> you understand?
>>>>>
>>>>> 2015-06-04 17:04 GMT+02:00 Robert Metzger <rmetz...@apache.org>:
>>>>>
>>>>>> It looks like the user "yarn" which is running Flink doesn't have
>>>>>> permission to access the files.
>>>>>>
>>>>>> Can you do "sudo su yarn" to become the "yarn" user. Then, you can do
>>>>>> "hadoop fs -chmod 777" to make the files accessible for everyone.
>>>>>>
>>>>>>
>>>>>> On Thu, Jun 4, 2015 at 4:59 PM, Pa Rö <paul.roewer1...@googlemail.com
>>>>>> > wrote:
>>>>>>
>>>>>>> okay, it's work, i get a exception:
>>>>>>>
>>>>>>> [cloudera@quickstart Desktop]$ cd flink-0.9-SNAPSHOT/bin/
>>>>>>> [cloudera@quickstart bin]$ flink run
>>>>>>> /home/cloudera/Desktop/ma-flink.jar
>>>>>>> bash: flink: command not found
>>>>>>> [cloudera@quickstart bin]$ ./flink run
>>>>>>> /home/cloudera/Desktop/ma-flink.jar
>>>>>>> log4j:WARN No appenders could be found for logger
>>>>>>> (org.apache.hadoop.metrics2.lib.MutableMetricsFactory).
>>>>>>> log4j:WARN Please initialize the log4j system properly.
>>>>>>> log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig
>>>>>>> for more info.
>>>>>>> Found YARN properties file
>>>>>>> /home/cloudera/Desktop/flink-0.9-SNAPSHOT/bin/../conf/.yarn-properties
>>>>>>> Using JobManager address from YARN properties quickstart.cloudera/
>>>>>>> 127.0.0.1:53874
>>>>>>> java.io.IOException: Mkdirs failed to create /user/cloudera/outputs
>>>>>>>     at
>>>>>>> org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:438)
>>>>>>>     at
>>>>>>> org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:424)
>>>>>>>     at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:905)
>>>>>>>     at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:886)
>>>>>>>     at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:783)
>>>>>>>     at
>>>>>>> mgm.tp.bigdata.ma_commons.commons.Seeding.randomSeeding(Seeding.java:21)
>>>>>>>     at
>>>>>>> mgm.tp.bigdata.ma_flink.FlinkMain.getCentroidDataSet(FlinkMain.java:178)
>>>>>>>     at mgm.tp.bigdata.ma_flink.FlinkMain.main(FlinkMain.java:47)
>>>>>>>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>>>>     at
>>>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>>>>>     at
>>>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>>>>>     at java.lang.reflect.Method.invoke(Method.java:606)
>>>>>>>     at
>>>>>>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:437)
>>>>>>>     at
>>>>>>> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:353)
>>>>>>>     at org.apache.flink.client.program.Client.run(Client.java:315)
>>>>>>>     at
>>>>>>> org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:584)
>>>>>>>     at org.apache.flink.client.CliFrontend.run(CliFrontend.java:290)
>>>>>>>     at
>>>>>>> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:880)
>>>>>>>     at org.apache.flink.client.CliFrontend.main(CliFrontend.java:922)
>>>>>>> org.apache.flink.client.program.ProgramInvocationException: The
>>>>>>> program execution failed: Failed to submit job
>>>>>>> 934743a5c49c6d5e31c9e8201452e36d (KMeans Flink)
>>>>>>>     at org.apache.flink.client.program.Client.run(Client.java:412)
>>>>>>>     at org.apache.flink.client.program.Client.run(Client.java:355)
>>>>>>>     at org.apache.flink.client.program.Client.run(Client.java:348)
>>>>>>>     at
>>>>>>> org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:63)
>>>>>>>     at mgm.tp.bigdata.ma_flink.FlinkMain.main(FlinkMain.java:70)
>>>>>>>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>>>>     at
>>>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>>>>>     at
>>>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>>>>>     at java.lang.reflect.Method.invoke(Method.java:606)
>>>>>>>     at
>>>>>>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:437)
>>>>>>>     at
>>>>>>> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:353)
>>>>>>>     at org.apache.flink.client.program.Client.run(Client.java:315)
>>>>>>>     at
>>>>>>> org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:584)
>>>>>>>     at org.apache.flink.client.CliFrontend.run(CliFrontend.java:290)
>>>>>>>     at
>>>>>>> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:880)
>>>>>>>     at org.apache.flink.client.CliFrontend.main(CliFrontend.java:922)
>>>>>>> Caused by: org.apache.flink.runtime.client.JobExecutionException:
>>>>>>> Failed to submit job 934743a5c49c6d5e31c9e8201452e36d (KMeans Flink)
>>>>>>>     at org.apache.flink.runtime.jobmanager.JobManager.org
>>>>>>> $apache$flink$runtime$jobmanager$JobManager$$submitJob(JobManager.scala:595)
>>>>>>>     at
>>>>>>> org.apache.flink.runtime.jobmanager.JobManager$$anonfun$receiveWithLogMessages$1.applyOrElse(JobManager.scala:192)
>>>>>>>     at
>>>>>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
>>>>>>>     at
>>>>>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
>>>>>>>     at
>>>>>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
>>>>>>>     at
>>>>>>> org.apache.flink.yarn.ApplicationMasterActor$$anonfun$receiveYarnMessages$1.applyOrElse(ApplicationMasterActor.scala:99)
>>>>>>>     at scala.PartialFunction$OrElse.apply(PartialFunction.scala:162)
>>>>>>>     at
>>>>>>> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:36)
>>>>>>>     at
>>>>>>> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:29)
>>>>>>>     at
>>>>>>> scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
>>>>>>>     at
>>>>>>> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:29)
>>>>>>>     at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
>>>>>>>     at
>>>>>>> org.apache.flink.runtime.jobmanager.JobManager.aroundReceive(JobManager.scala:94)
>>>>>>>     at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
>>>>>>>     at akka.actor.ActorCell.invoke(ActorCell.scala:487)
>>>>>>>     at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
>>>>>>>     at akka.dispatch.Mailbox.run(Mailbox.scala:221)
>>>>>>>     at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
>>>>>>>     at
>>>>>>> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>>>>>>>     at
>>>>>>> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
>>>>>>>     at
>>>>>>> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
>>>>>>>     at
>>>>>>> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
>>>>>>> Caused by: org.apache.flink.runtime.JobException: Creating the input
>>>>>>> splits caused an error: File /user/cloudera/outputs/seed-1 does not 
>>>>>>> exist
>>>>>>> or the user running Flink ('yarn') has insufficient permissions to 
>>>>>>> access
>>>>>>> it.
>>>>>>>     at
>>>>>>> org.apache.flink.runtime.executiongraph.ExecutionJobVertex.<init>(ExecutionJobVertex.java:162)
>>>>>>>     at
>>>>>>> org.apache.flink.runtime.executiongraph.ExecutionGraph.attachJobGraph(ExecutionGraph.java:471)
>>>>>>>     at org.apache.flink.runtime.jobmanager.JobManager.org
>>>>>>> $apache$flink$runtime$jobmanager$JobManager$$submitJob(JobManager.scala:535)
>>>>>>>     ... 21 more
>>>>>>> Caused by: java.io.FileNotFoundException: File
>>>>>>> /user/cloudera/outputs/seed-1 does not exist or the user running Flink
>>>>>>> ('yarn') has insufficient permissions to access it.
>>>>>>>     at
>>>>>>> org.apache.flink.core.fs.local.LocalFileSystem.getFileStatus(LocalFileSystem.java:106)
>>>>>>>     at
>>>>>>> org.apache.flink.api.common.io.FileInputFormat.createInputSplits(FileInputFormat.java:390)
>>>>>>>     at
>>>>>>> org.apache.flink.api.common.io.FileInputFormat.createInputSplits(FileInputFormat.java:51)
>>>>>>>     at
>>>>>>> org.apache.flink.runtime.executiongraph.ExecutionJobVertex.<init>(ExecutionJobVertex.java:146)
>>>>>>>     ... 23 more
>>>>>>>
>>>>>>>
>>>>>>> how i must set the files in the hdfs?
>>>>>>> quickstart.cloudera:50075/home/cloudera/output?
>>>>>>>
>>>>>>> 2015-06-04 16:51 GMT+02:00 Robert Metzger <rmetz...@apache.org>:
>>>>>>>
>>>>>>>> Once you've started the YARN session, you can submit a Flink job
>>>>>>>> with "./bin/flink run <pathToYourJar>".
>>>>>>>>
>>>>>>>> The jar file of your job doesn't need to be in HDFS. It has to be
>>>>>>>> in the local file system and flink will send it to all machines.
>>>>>>>>
>>>>>>>> On Thu, Jun 4, 2015 at 4:48 PM, Pa Rö <
>>>>>>>> paul.roewer1...@googlemail.com> wrote:
>>>>>>>>
>>>>>>>>> okay, now it run on my hadoop.
>>>>>>>>> how i can start my flink job? and where must the jar file save, at
>>>>>>>>> hdfs or as local file?
>>>>>>>>>
>>>>>>>>> 2015-06-04 16:31 GMT+02:00 Robert Metzger <rmetz...@apache.org>:
>>>>>>>>>
>>>>>>>>>> Yes, you have to run these commands in the command line of the
>>>>>>>>>> Cloudera VM.
>>>>>>>>>>
>>>>>>>>>> On Thu, Jun 4, 2015 at 4:28 PM, Pa Rö <
>>>>>>>>>> paul.roewer1...@googlemail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> you mean run this command on terminal/shell and not define a hue
>>>>>>>>>>> job?
>>>>>>>>>>>
>>>>>>>>>>> 2015-06-04 16:25 GMT+02:00 Robert Metzger <rmetz...@apache.org>:
>>>>>>>>>>>
>>>>>>>>>>>> It should be certainly possible to run Flink on a cloudera live
>>>>>>>>>>>> VM
>>>>>>>>>>>>
>>>>>>>>>>>> I think these are the commands you need to execute:
>>>>>>>>>>>>
>>>>>>>>>>>> wget
>>>>>>>>>>>> http://stratosphere-bin.s3-website-us-east-1.amazonaws.com/flink-0.9-SNAPSHOT-bin-hadoop2.tgz
>>>>>>>>>>>> tar xvzf flink-0.9-SNAPSHOT-bin-hadoop2.tgz
>>>>>>>>>>>> cd flink-0.9-SNAPSHOT/
>>>>>>>>>>>> *export HADOOP_CONF_DIR=/usr/lib/hadoop/etc/hadoop/*
>>>>>>>>>>>> ./bin/yarn-session.sh -n 1 -jm 1024 -tm 1024
>>>>>>>>>>>>
>>>>>>>>>>>> If that is not working for you, please post the exact error
>>>>>>>>>>>> message you are getting and I can help you to get it to run.
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On Thu, Jun 4, 2015 at 4:18 PM, Pa Rö <
>>>>>>>>>>>> paul.roewer1...@googlemail.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> hi robert,
>>>>>>>>>>>>>
>>>>>>>>>>>>> i think the problem is the hue api,
>>>>>>>>>>>>> i had the same problem with spark submit script,
>>>>>>>>>>>>> but on the new hue release, they have a spark submit api.
>>>>>>>>>>>>>
>>>>>>>>>>>>> i asked the group for the same problem with spark, no reply.
>>>>>>>>>>>>>
>>>>>>>>>>>>> i want test my app on local cluster, before i run it on the
>>>>>>>>>>>>> big cluster,
>>>>>>>>>>>>> for that i use cloudera live. maybe it give an other way to
>>>>>>>>>>>>> test flink on a local cluster vm?
>>>>>>>>>>>>>
>>>>>>>>>>>>> 2015-06-04 16:12 GMT+02:00 Robert Metzger <rmetz...@apache.org
>>>>>>>>>>>>> >:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> Hi Paul,
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> why did running Flink from the regular scripts not work for
>>>>>>>>>>>>>> you?
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> I'm not an expert on Hue, I would recommend asking in the Hue
>>>>>>>>>>>>>> user forum / mailing list:
>>>>>>>>>>>>>> https://groups.google.com/a/cloudera.org/forum/#!forum/hue-user
>>>>>>>>>>>>>> .
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Thu, Jun 4, 2015 at 4:09 PM, Pa Rö <
>>>>>>>>>>>>>> paul.roewer1...@googlemail.com> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> thanks,
>>>>>>>>>>>>>>> now i want run my app on cloudera live vm single node,
>>>>>>>>>>>>>>> how i can define my flink job with hue?
>>>>>>>>>>>>>>> i try to run the flink script in the hdfs, it's not work.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> best regards,
>>>>>>>>>>>>>>> paul
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> 2015-06-02 14:50 GMT+02:00 Robert Metzger <
>>>>>>>>>>>>>>> rmetz...@apache.org>:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> I would recommend using HDFS.
>>>>>>>>>>>>>>>> For that, you need to specify the paths like this:
>>>>>>>>>>>>>>>> hdfs:///path/to/data.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On Tue, Jun 2, 2015 at 2:48 PM, Pa Rö <
>>>>>>>>>>>>>>>> paul.roewer1...@googlemail.com> wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> nice,
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> which file system i must use for the cluster? java.io or
>>>>>>>>>>>>>>>>> hadoop.fs or flink?
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> 2015-06-02 14:29 GMT+02:00 Robert Metzger <
>>>>>>>>>>>>>>>>> rmetz...@apache.org>:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Hi,
>>>>>>>>>>>>>>>>>> you can start Flink on YARN on the Cloudera distribution.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> See here for more:
>>>>>>>>>>>>>>>>>> http://ci.apache.org/projects/flink/flink-docs-master/setup/yarn_setup.html
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> These are the commands you need to execute
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> wget 
>>>>>>>>>>>>>>>>>> http://stratosphere-bin.s3-website-us-east-1.amazonaws.com/flink-0.9-SNAPSHOT-bin-hadoop2.tgz
>>>>>>>>>>>>>>>>>> tar xvzf flink-0.9-SNAPSHOT-bin-hadoop2.tgzcd 
>>>>>>>>>>>>>>>>>> flink-0.9-SNAPSHOT/
>>>>>>>>>>>>>>>>>> ./bin/yarn-session.sh -n 4 -jm 1024 -tm 4096
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> On Tue, Jun 2, 2015 at 2:03 PM, Pa Rö <
>>>>>>>>>>>>>>>>>> paul.roewer1...@googlemail.com> wrote:
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> hi community,
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> i want test my flink k-means on a hadoop cluster. i use
>>>>>>>>>>>>>>>>>>> the cloudera live distribution. how i can run flink on this 
>>>>>>>>>>>>>>>>>>> cluster? maybe
>>>>>>>>>>>>>>>>>>> only the java dependencies are engouth?
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> best regards,
>>>>>>>>>>>>>>>>>>> paul
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to