Hi I'm currently using flink 0.9.0 which by maven support Hadoop 1 . By using flink-clients-0.7.0-hadoop2-incubating.jar with executePlan(Plan p) method instead, I'm getting the same exception
Hanan On Fri, Aug 28, 2015 at 8:35 AM, Hanan Meyer <ha...@scalabill.it> wrote: > > Hi > > 1. I have restarted Flink service via stop/start-loval.sh - it have been > restarted successfully ,no errors in log folder > 2. default flink port is -6123 > > Getting this via Eclips IDE: > > Thanks > > > org.apache.flink.client.program.ProgramInvocationException: Failed to > resolve JobManager > at org.apache.flink.client.program.Client.run(Client.java:379) > at org.apache.flink.client.program.Client.run(Client.java:356) > at org.apache.flink.client.program.Client.run(Client.java:349) > at > org.apache.flink.client.RemoteExecutor.executePlanWithJars(RemoteExecutor.java:89) > at > org.apache.flink.client.RemoteExecutor.executePlan(RemoteExecutor.java:82) > at > org.apache.flink.api.java.RemoteEnvironment.execute(RemoteEnvironment.java:71) > at > org.apache.flink.api.java.ExecutionEnvironment.execute(ExecutionEnvironment.java:789) > at Test.main(Test.java:39) > Caused by: java.io.IOException: JobManager at > akka.tcp://flink@FLINK_SERVER_URL:6123/user/jobmanager not reachable. > Please make sure that the JobManager is running and its port is reachable. > at > org.apache.flink.runtime.jobmanager.JobManager$.getJobManagerRemoteReference(JobManager.scala:1197) > at > org.apache.flink.runtime.jobmanager.JobManager$.getJobManagerRemoteReference(JobManager.scala:1221) > at > org.apache.flink.runtime.jobmanager.JobManager$.getJobManagerRemoteReference(JobManager.scala:1239) > at > org.apache.flink.runtime.jobmanager.JobManager.getJobManagerRemoteReference(JobManager.scala) > at org.apache.flink.client.program.Client.run(Client.java:376) > ... 7 more > Caused by: akka.actor.ActorNotFound: Actor not found for: > ActorSelection[Anchor(akka.tcp://flink@FLINK_SERVER_URL:6123/), > Path(/user/jobmanager)] > at > akka.actor.ActorSelection$$anonfun$resolveOne$1.apply(ActorSelection.scala:65) > at > akka.actor.ActorSelection$$anonfun$resolveOne$1.apply(ActorSelection.scala:63) > at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32) > at akka.dispatch.BatchingExecutor$ > Batch$$anonfun$run$1.processBatch$1(BatchingExecutor.scala:67) > at > akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.apply$mcV$sp(BatchingExecutor.scala:82) > at > akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.apply(BatchingExecutor.scala:59) > at > akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.apply(BatchingExecutor.scala:59) > at scala.concurrent.BlockContext$.withBlockContext(BlockContext.scala:72) > at akka.dispatch.BatchingExecutor$Batch.run(BatchingExecutor.scala:58) > at > akka.dispatch.ExecutionContexts$sameThreadExecutionContext$.unbatchedExecute(Future.scala:74) > at akka.dispatch.BatchingExecutor$class.execute(BatchingExecutor.scala:110) > at > akka.dispatch.ExecutionContexts$sameThreadExecutionContext$.execute(Future.scala:73) > at > scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40) > at > scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248) > at akka.pattern.PromiseActorRef.$bang(AskSupport.scala:267) > at akka.actor.EmptyLocalActorRef.specialHandle(ActorRef.scala:508) > at akka.actor.DeadLetterActorRef.specialHandle(ActorRef.scala:541) > at akka.actor.DeadLetterActorRef.$bang(ActorRef.scala:531) > at > akka.remote.RemoteActorRefProvider$RemoteDeadLetterActorRef.$bang(RemoteActorRefProvider.scala:87) > at akka.remote.EndpointWriter.postStop(Endpoint.scala:561) > at akka.actor.Actor$class.aroundPostStop(Actor.scala:475) > at akka.remote.EndpointActor.aroundPostStop(Endpoint.scala:415) > at > akka.actor.dungeon.FaultHandling$class.akka$actor$dungeon$FaultHandling$$finishTerminate(FaultHandling.scala:210) > at > akka.actor.dungeon.FaultHandling$class.terminate(FaultHandling.scala:172) > at akka.actor.ActorCell.terminate(ActorCell.scala:369) > at akka.actor.ActorCell.invokeAll$1(ActorCell.scala:462) > at akka.actor.ActorCell.systemInvoke(ActorCell.scala:478) > at akka.dispatch.Mailbox.processAllSystemMessages(Mailbox.scala:279) > at akka.dispatch.Mailbox.run(Mailbox.scala:220) > at akka.dispatch.Mailbox.exec(Mailbox.scala:231) > at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) > at > scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253) > at > scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346) > at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) > at > scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) > > > On Thu, Aug 27, 2015 at 10:47 PM, Robert Metzger <rmetz...@apache.org> > wrote: > >> I guess you are getting an entire exception after the "org.apache.flink >> .client.program.ProgramInvocationException: Failed to >> resolve JobManager". >> Can you post it here to help us understanding the issue? >> >> On Thu, Aug 27, 2015 at 6:55 PM, Alexey Sapozhnikov <ale...@scalabill.it> >> wrote: >> >> > Hello all. >> > >> > Some clarification: locally everything works great. >> > However once we run our Flink on remote linux machine and try to run the >> > client program from our machine, using create remote environment- Flink >> > JobManager is raising this exception >> > >> > On Thu, Aug 27, 2015 at 7:41 PM, Stephan Ewen <se...@apache.org> wrote: >> > >> > > If you start the job via the "bin/flink" script, then simply use >> > > "ExecutionEnvironment.getExecutionEnvironment()" rather then creating >> a >> > > remote environment manually. >> > > >> > > That way, hosts and ports are configured automatically. >> > > >> > > On Thu, Aug 27, 2015 at 6:39 PM, Robert Metzger <rmetz...@apache.org> >> > > wrote: >> > > >> > >> Hi, >> > >> >> > >> Which values did you use for FLINK_SERVER_URL and FLINK_PORT? >> > >> Every time you deploy Flink on YARN, the host and port change, >> because >> > the >> > >> JobManager is started on a different YARN container. >> > >> >> > >> >> > >> On Thu, Aug 27, 2015 at 6:32 PM, Hanan Meyer <ha...@scalabill.it> >> > wrote: >> > >> >> > >> > Hello All >> > >> > >> > >> > When using Eclipse IDE to submit Flink to Yarn single node cluster >> I'm >> > >> > getting : >> > >> > "org.apache.flink.client.program.ProgramInvocationException: >> Failed to >> > >> > resolve JobManager" >> > >> > >> > >> > Using Flink 0.9.0 >> > >> > >> > >> > The Jar copy a file from one location in Hdfs to another and works >> > fine >> > >> > while executed locally on the single node Yarn cluster - >> > >> > bin/flink run -c Test ./examples/MyJar.jar >> > >> > hdfs://localhost:9000/flink/in.txt >> hdfs://localhost:9000/flink/out.txt >> > >> > >> > >> > The code skeleton: >> > >> > >> > >> > ExecutionEnvironment envRemote = >> > >> > ExecutionEnvironment.createRemoteEnvironment >> > >> > (FLINK_SERVER_URL,FLINK PORT,JAR_PATH_ON_CLIENT); >> > >> > DataSet<String> data = >> > >> > envRemote.readTextFile("hdfs://localhost:9000/flink/in.txt"); >> > >> > data.writeAsText("hdfs://localhost:9000/flink/out.txt"); >> > >> > envRemote.execute(); >> > >> > >> > >> > >> > >> > Please advise, >> > >> > >> > >> > Hanan Meyer >> > >> > >> > >> >> > > >> > > >> > >> > >