Can you try to not manually create a "RemoteExecutionEnvironment", but to
simply use the recommended way of doing this:

Please use "ExecutionEnvironment.getExecutionEnvironment()" if you run the
program through the command line anyways.

On Fri, Aug 28, 2015 at 1:04 PM, Hanan Meyer <ha...@scalabill.it> wrote:

> Hi
>
> I'm running with a formal server ip but for securuty reasons I can't share
> with you the real ip .
> I put "FLINK_SERVER_URL" in order to replace the actual ip only in my post
> .
>
> Hanan Meyer
>
> On Fri, Aug 28, 2015 at 10:27 AM, Robert Metzger <rmetz...@apache.org>
> wrote:
>
> > Hi,
> >
> > in the exception you've posted earlier, you can see the following root
> > cause:
> >
> > Caused by: akka.actor.ActorNotFound: Actor not found for:
> > ActorSelection[Anchor(akka.tcp://flink@FLINK_SERVER_URL:6123/),
> > Path(/user/jobmanager)]
> >
> > This string "akka.tcp://flink@FLINK_SERVER_URL:6123/" usually looks like
> > this: "akka.tcp://flink@1.2.3.4:6123/". So it seems that you are
> > passing FLINK_SERVER_URL
> > as the server hostname (or ip).
> > Can you pass the correct hostname when you call ExecutionEnvironment.
> > createRemoteEnvironment().
> >
> > On Fri, Aug 28, 2015 at 7:52 AM, Hanan Meyer <ha...@scalabill.it> wrote:
> >
> > > Hi
> > >  I'm currently using flink 0.9.0 which by maven support Hadoop 1 .
> > > By using flink-clients-0.7.0-hadoop2-incubating.jar with
> executePlan(Plan
> > > p) method  instead, I'm getting the same exception
> > >
> > > Hanan
> > >
> > > On Fri, Aug 28, 2015 at 8:35 AM, Hanan Meyer <ha...@scalabill.it>
> wrote:
> > >
> > > >
> > > > Hi
> > > >
> > > > 1. I have restarted Flink service via stop/start-loval.sh - it have
> > been
> > > > restarted successfully ,no errors in log folder
> > > > 2. default flink port is -6123
> > > >
> > > > Getting this via Eclips IDE:
> > > >
> > > > Thanks
> > > >
> > > >
> > > > org.apache.flink.client.program.ProgramInvocationException: Failed to
> > > > resolve JobManager
> > > > at org.apache.flink.client.program.Client.run(Client.java:379)
> > > > at org.apache.flink.client.program.Client.run(Client.java:356)
> > > > at org.apache.flink.client.program.Client.run(Client.java:349)
> > > > at
> > > >
> > >
> >
> org.apache.flink.client.RemoteExecutor.executePlanWithJars(RemoteExecutor.java:89)
> > > > at
> > > >
> > >
> >
> org.apache.flink.client.RemoteExecutor.executePlan(RemoteExecutor.java:82)
> > > > at
> > > >
> > >
> >
> org.apache.flink.api.java.RemoteEnvironment.execute(RemoteEnvironment.java:71)
> > > > at
> > > >
> > >
> >
> org.apache.flink.api.java.ExecutionEnvironment.execute(ExecutionEnvironment.java:789)
> > > > at Test.main(Test.java:39)
> > > > Caused by: java.io.IOException: JobManager at
> > > > akka.tcp://flink@FLINK_SERVER_URL:6123/user/jobmanager not
> reachable.
> > > > Please make sure that the JobManager is running and its port is
> > > reachable.
> > > > at
> > > >
> > >
> >
> org.apache.flink.runtime.jobmanager.JobManager$.getJobManagerRemoteReference(JobManager.scala:1197)
> > > > at
> > > >
> > >
> >
> org.apache.flink.runtime.jobmanager.JobManager$.getJobManagerRemoteReference(JobManager.scala:1221)
> > > > at
> > > >
> > >
> >
> org.apache.flink.runtime.jobmanager.JobManager$.getJobManagerRemoteReference(JobManager.scala:1239)
> > > > at
> > > >
> > >
> >
> org.apache.flink.runtime.jobmanager.JobManager.getJobManagerRemoteReference(JobManager.scala)
> > > > at org.apache.flink.client.program.Client.run(Client.java:376)
> > > > ... 7 more
> > > > Caused by: akka.actor.ActorNotFound: Actor not found for:
> > > > ActorSelection[Anchor(akka.tcp://flink@FLINK_SERVER_URL:6123/),
> > > > Path(/user/jobmanager)]
> > > > at
> > > >
> > >
> >
> akka.actor.ActorSelection$$anonfun$resolveOne$1.apply(ActorSelection.scala:65)
> > > > at
> > > >
> > >
> >
> akka.actor.ActorSelection$$anonfun$resolveOne$1.apply(ActorSelection.scala:63)
> > > > at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32)
> > > > at akka.dispatch.BatchingExecutor$
> > > > Batch$$anonfun$run$1.processBatch$1(BatchingExecutor.scala:67)
> > > > at
> > > >
> > >
> >
> akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.apply$mcV$sp(BatchingExecutor.scala:82)
> > > > at
> > > >
> > >
> >
> akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.apply(BatchingExecutor.scala:59)
> > > > at
> > > >
> > >
> >
> akka.dispatch.BatchingExecutor$Batch$$anonfun$run$1.apply(BatchingExecutor.scala:59)
> > > > at
> > scala.concurrent.BlockContext$.withBlockContext(BlockContext.scala:72)
> > > > at
> akka.dispatch.BatchingExecutor$Batch.run(BatchingExecutor.scala:58)
> > > > at
> > > >
> > >
> >
> akka.dispatch.ExecutionContexts$sameThreadExecutionContext$.unbatchedExecute(Future.scala:74)
> > > > at
> > >
> akka.dispatch.BatchingExecutor$class.execute(BatchingExecutor.scala:110)
> > > > at
> > > >
> > >
> >
> akka.dispatch.ExecutionContexts$sameThreadExecutionContext$.execute(Future.scala:73)
> > > > at
> > > >
> > scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
> > > > at
> > > >
> > >
> >
> scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248)
> > > > at akka.pattern.PromiseActorRef.$bang(AskSupport.scala:267)
> > > > at akka.actor.EmptyLocalActorRef.specialHandle(ActorRef.scala:508)
> > > > at akka.actor.DeadLetterActorRef.specialHandle(ActorRef.scala:541)
> > > > at akka.actor.DeadLetterActorRef.$bang(ActorRef.scala:531)
> > > > at
> > > >
> > >
> >
> akka.remote.RemoteActorRefProvider$RemoteDeadLetterActorRef.$bang(RemoteActorRefProvider.scala:87)
> > > > at akka.remote.EndpointWriter.postStop(Endpoint.scala:561)
> > > > at akka.actor.Actor$class.aroundPostStop(Actor.scala:475)
> > > > at akka.remote.EndpointActor.aroundPostStop(Endpoint.scala:415)
> > > > at
> > > >
> > >
> >
> akka.actor.dungeon.FaultHandling$class.akka$actor$dungeon$FaultHandling$$finishTerminate(FaultHandling.scala:210)
> > > > at
> > > >
> > akka.actor.dungeon.FaultHandling$class.terminate(FaultHandling.scala:172)
> > > > at akka.actor.ActorCell.terminate(ActorCell.scala:369)
> > > > at akka.actor.ActorCell.invokeAll$1(ActorCell.scala:462)
> > > > at akka.actor.ActorCell.systemInvoke(ActorCell.scala:478)
> > > > at akka.dispatch.Mailbox.processAllSystemMessages(Mailbox.scala:279)
> > > > at akka.dispatch.Mailbox.run(Mailbox.scala:220)
> > > > at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
> > > > at
> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> > > > at
> > > >
> > >
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
> > > > at
> > > >
> > >
> >
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
> > > > at
> > >
> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> > > > at
> > > >
> > >
> >
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> > > >
> > > >
> > > > On Thu, Aug 27, 2015 at 10:47 PM, Robert Metzger <
> rmetz...@apache.org>
> > > > wrote:
> > > >
> > > >> I guess you are getting an entire exception after the
> > "org.apache.flink
> > > >> .client.program.ProgramInvocationException: Failed to
> > > >> resolve JobManager".
> > > >> Can you post it here to help us understanding the issue?
> > > >>
> > > >> On Thu, Aug 27, 2015 at 6:55 PM, Alexey Sapozhnikov <
> > > ale...@scalabill.it>
> > > >> wrote:
> > > >>
> > > >> > Hello all.
> > > >> >
> > > >> > Some clarification: locally everything works great.
> > > >> > However once we run our Flink on remote linux machine and try to
> run
> > > the
> > > >> > client program from our machine, using create remote environment-
> > > Flink
> > > >> > JobManager is raising this exception
> > > >> >
> > > >> > On Thu, Aug 27, 2015 at 7:41 PM, Stephan Ewen <se...@apache.org>
> > > wrote:
> > > >> >
> > > >> > > If you start the job via the "bin/flink" script, then simply use
> > > >> > > "ExecutionEnvironment.getExecutionEnvironment()" rather then
> > > creating
> > > >> a
> > > >> > > remote environment manually.
> > > >> > >
> > > >> > > That way, hosts and ports are configured automatically.
> > > >> > >
> > > >> > > On Thu, Aug 27, 2015 at 6:39 PM, Robert Metzger <
> > > rmetz...@apache.org>
> > > >> > > wrote:
> > > >> > >
> > > >> > >> Hi,
> > > >> > >>
> > > >> > >> Which values did you use for FLINK_SERVER_URL and FLINK_PORT?
> > > >> > >> Every time you deploy Flink on YARN, the host and port change,
> > > >> because
> > > >> > the
> > > >> > >> JobManager is started on a different YARN container.
> > > >> > >>
> > > >> > >>
> > > >> > >> On Thu, Aug 27, 2015 at 6:32 PM, Hanan Meyer <
> ha...@scalabill.it
> > >
> > > >> > wrote:
> > > >> > >>
> > > >> > >> > Hello All
> > > >> > >> >
> > > >> > >> > When using Eclipse IDE to submit Flink to Yarn single node
> > > cluster
> > > >> I'm
> > > >> > >> > getting :
> > > >> > >> > "org.apache.flink.client.program.ProgramInvocationException:
> > > >> Failed to
> > > >> > >> > resolve JobManager"
> > > >> > >> >
> > > >> > >> > Using Flink 0.9.0
> > > >> > >> >
> > > >> > >> > The Jar copy a file from one location in Hdfs to another and
> > > works
> > > >> > fine
> > > >> > >> > while executed locally on the single node Yarn cluster -
> > > >> > >> > bin/flink run -c Test ./examples/MyJar.jar
> > > >> > >> > hdfs://localhost:9000/flink/in.txt
> > > >> hdfs://localhost:9000/flink/out.txt
> > > >> > >> >
> > > >> > >> > The code skeleton:
> > > >> > >> >
> > > >> > >> >     ExecutionEnvironment envRemote =
> > > >> > >> > ExecutionEnvironment.createRemoteEnvironment
> > > >> > >> > (FLINK_SERVER_URL,FLINK PORT,JAR_PATH_ON_CLIENT);
> > > >> > >> > DataSet<String> data =
> > > >> > >> > envRemote.readTextFile("hdfs://localhost:9000/flink/in.txt");
> > > >> > >> > data.writeAsText("hdfs://localhost:9000/flink/out.txt");
> > > >> > >> > envRemote.execute();
> > > >> > >> >
> > > >> > >> >
> > > >> > >> > Please advise,
> > > >> > >> >
> > > >> > >> > Hanan Meyer
> > > >> > >> >
> > > >> > >>
> > > >> > >
> > > >> > >
> > > >> >
> > > >>
> > > >
> > > >
> > >
> >
>

Reply via email to