If you start the job via the "bin/flink" script, then simply use "ExecutionEnvironment.getExecutionEnvironment()" rather then creating a remote environment manually.
That way, hosts and ports are configured automatically. On Thu, Aug 27, 2015 at 6:39 PM, Robert Metzger <rmetz...@apache.org> wrote: > Hi, > > Which values did you use for FLINK_SERVER_URL and FLINK_PORT? > Every time you deploy Flink on YARN, the host and port change, because the > JobManager is started on a different YARN container. > > > On Thu, Aug 27, 2015 at 6:32 PM, Hanan Meyer <ha...@scalabill.it> wrote: > > > Hello All > > > > When using Eclipse IDE to submit Flink to Yarn single node cluster I'm > > getting : > > "org.apache.flink.client.program.ProgramInvocationException: Failed to > > resolve JobManager" > > > > Using Flink 0.9.0 > > > > The Jar copy a file from one location in Hdfs to another and works fine > > while executed locally on the single node Yarn cluster - > > bin/flink run -c Test ./examples/MyJar.jar > > hdfs://localhost:9000/flink/in.txt hdfs://localhost:9000/flink/out.txt > > > > The code skeleton: > > > > ExecutionEnvironment envRemote = > > ExecutionEnvironment.createRemoteEnvironment > > (FLINK_SERVER_URL,FLINK PORT,JAR_PATH_ON_CLIENT); > > DataSet<String> data = > > envRemote.readTextFile("hdfs://localhost:9000/flink/in.txt"); > > data.writeAsText("hdfs://localhost:9000/flink/out.txt"); > > envRemote.execute(); > > > > > > Please advise, > > > > Hanan Meyer > > >