Oh, I thought the flink job could not be submitted. I don't know why the storm's example could not be submitted. Because I have never used it.
Maybe Till, Chesnay or Gary can help you. Ping them for you. Thanks, vino. jing <hanjingz...@163.com> 于2018年9月11日周二 上午10:26写道: > Hi vino, > My job mangaer log is as below. I can submit regular flink job to this > jobmanger, it worked. But the flink-storm example doesn's work. > Thanks. > Hanjing > > 2018-09-11 18:22:48,937 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - > -------------------------------------------------------------------------------- > 2018-09-11 18:22:48,938 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Starting > StandaloneSessionClusterEntrypoint (Version: 1.6.0, Rev:ff472b4, > Date:07.08.2018 @ 13:31:13 UTC) > 2018-09-11 18:22:48,938 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - OS current > user: hadoop3 > 2018-09-11 18:22:49,143 WARN org.apache.hadoop.util.NativeCodeLoader > - Unable to load native-hadoop library for your platform... > using builtin-java classes where applicable > 2018-09-11 18:22:49,186 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Current > Hadoop/Kerberos user: hadoop3 > 2018-09-11 18:22:49,186 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - JVM: Java > HotSpot(TM) 64-Bit Server VM - Oracle Corporation - 1.8/25.172-b11 > 2018-09-11 18:22:49,186 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Maximum heap > size: 981 MiBytes > 2018-09-11 18:22:49,186 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - JAVA_HOME: > /usr/java/jdk1.8.0_172-amd64 > 2018-09-11 18:22:49,188 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Hadoop > version: 2.7.5 > 2018-09-11 18:22:49,188 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - JVM Options: > 2018-09-11 18:22:49,188 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - -Xms1024m > 2018-09-11 18:22:49,188 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - -Xmx1024m > 2018-09-11 18:22:49,188 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - > -Dlog.file=/home/hadoop3/zh/flink-1.6.0/log/flink-hadoop3-standalonesession-0-p-a36-72.log > 2018-09-11 18:22:49,188 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - > -Dlog4j.configuration=file:/home/hadoop3/zh/flink-1.6.0/conf/log4j.properties > 2018-09-11 18:22:49,188 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - > -Dlogback.configurationFile=file:/home/hadoop3/zh/flink-1.6.0/conf/logback.xml > 2018-09-11 18:22:49,188 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Program > Arguments: > 2018-09-11 18:22:49,188 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - > --configDir > 2018-09-11 18:22:49,188 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - > /home/hadoop3/zh/flink-1.6.0/conf > 2018-09-11 18:22:49,188 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - > --executionMode > 2018-09-11 18:22:49,189 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - cluster > 2018-09-11 18:22:49,189 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Classpath: > /home/hadoop3/zh/flink-1.6.0/lib/flink-python_2.11-1.6.0.jar:/home/hadoop3/zh/flink-1.6.0/lib/flink-shaded-hadoop2-uber-1.6.0.jar:/home/hadoop3/zh/flink-1.6.0/lib/log4j-1.2.17.jar:/home/hadoop3/zh/flink-1.6.0/lib/slf4j-log4j12-1.7.7.jar:/home/hadoop3/zh/flink-1.6.0/lib/flink-dist_2.11-1.6.0.jar::: > 2018-09-11 18:22:49,189 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - > -------------------------------------------------------------------------------- > 2018-09-11 18:22:49,189 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Registered > UNIX signal handlers for [TERM, HUP, INT] > 2018-09-11 18:22:49,197 INFO > org.apache.flink.configuration.GlobalConfiguration - Loading > configuration property: jobmanager.rpc.address, p-a36-72 > 2018-09-11 18:22:49,197 INFO > org.apache.flink.configuration.GlobalConfiguration - Loading > configuration property: jobmanager.rpc.port, 6123 > 2018-09-11 18:22:49,197 INFO > org.apache.flink.configuration.GlobalConfiguration - Loading > configuration property: jobmanager.heap.size, 1024m > 2018-09-11 18:22:49,197 INFO > org.apache.flink.configuration.GlobalConfiguration - Loading > configuration property: taskmanager.heap.size, 10240m > 2018-09-11 18:22:49,197 INFO > org.apache.flink.configuration.GlobalConfiguration - Loading > configuration property: taskmanager.numberOfTaskSlots, 16 > 2018-09-11 18:22:49,197 INFO > org.apache.flink.configuration.GlobalConfiguration - Loading > configuration property: parallelism.default, 2 > 2018-09-11 18:22:49,198 INFO > org.apache.flink.configuration.GlobalConfiguration - Loading > configuration property: rest.port, 8081 > 2018-09-11 18:22:49,207 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Starting > StandaloneSessionClusterEntrypoint. > 2018-09-11 18:22:49,207 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Install > default filesystem. > 2018-09-11 18:22:49,214 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Install > security context. > 2018-09-11 18:22:49,237 INFO > org.apache.flink.runtime.security.modules.HadoopModule - Hadoop user > set to hadoop3 (auth:SIMPLE) > 2018-09-11 18:22:49,247 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Initializing > cluster services. > 2018-09-11 18:22:49,253 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Trying to > start actor system at p-a36-72:6123 > 2018-09-11 18:22:49,576 INFO akka.event.slf4j.Slf4jLogger > - Slf4jLogger started > 2018-09-11 18:22:49,611 INFO akka.remote.Remoting > - Starting remoting > 2018-09-11 18:22:49,718 INFO akka.remote.Remoting > - Remoting started; listening on addresses > :[akka.tcp://flink@p-a36-72:6123] > 2018-09-11 18:22:49,722 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Actor system > started at akka.tcp://flink@p-a36-72:6123 > 2018-09-11 18:22:49,732 WARN org.apache.flink.configuration.Configuration > - Config uses deprecated configuration key > 'jobmanager.rpc.address' instead of proper key 'rest.address' > 2018-09-11 18:22:49,737 INFO org.apache.flink.runtime.blob.BlobServer > - Created BLOB server storage directory > /tmp/blobStore-62c16996-0f38-43ae-9e40-ac4206329d93 > 2018-09-11 18:22:49,739 INFO org.apache.flink.runtime.blob.BlobServer > - Started BLOB server at 0.0.0.0:3706 - max concurrent > requests: 50 - max backlog: 1000 > 2018-09-11 18:22:49,749 INFO > org.apache.flink.runtime.metrics.MetricRegistryImpl - No metrics > reporter configured, no metrics will be exposed/reported. > 2018-09-11 18:22:49,751 INFO > org.apache.flink.runtime.dispatcher.FileArchivedExecutionGraphStore - > Initializing FileArchivedExecutionGraphStore: Storage directory > /tmp/executionGraphStore-fecb7e34-9d33-4af2-a623-ee96d8572800, expiration > time 3600000, maximum cache size 52428800 bytes. > 2018-09-11 18:22:49,766 INFO > org.apache.flink.runtime.blob.TransientBlobCache - Created BLOB > cache storage directory /tmp/blobStore-c1d1946d-9e19-40b1-800d-42598900e253 > 2018-09-11 18:22:49,771 WARN org.apache.flink.configuration.Configuration > - Config uses deprecated configuration key > 'jobmanager.rpc.address' instead of proper key 'rest.address' > 2018-09-11 18:22:49,772 WARN > org.apache.flink.runtime.dispatcher.DispatcherRestEndpoint - Upload > directory > /tmp/flink-web-c1cc0dde-0f4b-458e-84ba-841f405f3c78/flink-web-upload does not > exist, or has been deleted externally. Previously uploaded files are no > longer available. > 2018-09-11 18:22:49,772 INFO > org.apache.flink.runtime.dispatcher.DispatcherRestEndpoint - Created > directory > /tmp/flink-web-c1cc0dde-0f4b-458e-84ba-841f405f3c78/flink-web-upload for file > uploads. > 2018-09-11 18:22:49,774 INFO > org.apache.flink.runtime.dispatcher.DispatcherRestEndpoint - Starting rest > endpoint. > 2018-09-11 18:22:49,905 INFO > org.apache.flink.runtime.webmonitor.WebMonitorUtils - Determined > location of main cluster component log file: > /home/hadoop3/zh/flink-1.6.0/log/flink-hadoop3-standalonesession-0-p-a36-72.log > 2018-09-11 18:22:49,905 INFO > org.apache.flink.runtime.webmonitor.WebMonitorUtils - Determined > location of main cluster component stdout file: > /home/hadoop3/zh/flink-1.6.0/log/flink-hadoop3-standalonesession-0-p-a36-72.out > 2018-09-11 18:22:49,997 INFO > org.apache.flink.runtime.dispatcher.DispatcherRestEndpoint - Rest endpoint > listening at p-a36-72:8081 > 2018-09-11 18:22:49,997 INFO > org.apache.flink.runtime.dispatcher.DispatcherRestEndpoint - > http://p-a36-72:8081 was granted leadership with > leaderSessionID=00000000-0000-0000-0000-000000000000 > 2018-09-11 18:22:49,997 INFO > org.apache.flink.runtime.dispatcher.DispatcherRestEndpoint - Web frontend > listening at http://p-a36-72:8081. > 2018-09-11 18:22:50,004 INFO > org.apache.flink.runtime.rpc.akka.AkkaRpcService - Starting RPC > endpoint for > org.apache.flink.runtime.resourcemanager.StandaloneResourceManager at > akka://flink/user/resourcemanager . > 2018-09-11 18:22:50,045 INFO > org.apache.flink.runtime.rpc.akka.AkkaRpcService - Starting RPC > endpoint for org.apache.flink.runtime.dispatcher.StandaloneDispatcher at > akka://flink/user/dispatcher . > 2018-09-11 18:22:50,055 INFO > org.apache.flink.runtime.resourcemanager.StandaloneResourceManager - > ResourceManager akka.tcp://flink@p-a36-72:6123/user/resourcemanager was > granted leadership with fencing token 00000000000000000000000000000000 > 2018-09-11 18:22:50,055 INFO > org.apache.flink.runtime.resourcemanager.slotmanager.SlotManager - Starting > the SlotManager. > 2018-09-11 18:22:50,064 INFO > org.apache.flink.runtime.dispatcher.StandaloneDispatcher - Dispatcher > akka.tcp://flink@p-a36-72:6123/user/dispatcher was granted leadership with > fencing token 00000000-0000-0000-0000-000000000000 > 2018-09-11 18:22:50,064 INFO > org.apache.flink.runtime.dispatcher.StandaloneDispatcher - Recovering > all persisted jobs. > 2018-09-11 18:22:55,316 INFO > org.apache.flink.runtime.resourcemanager.slotmanager.SlotManager - > Registering TaskManager c82e0b779fe62e3b3e6efda75c97cd18 under > c19ad7a1b58afdfc79b0fbbf08d43653 at the SlotManager. > > > Hanjing > > <https://maas.mail.163.com/dashi-web-extend/html/proSignature.html?ftlId=1&name=Hanjing&uid=hanjingzuzu%40163.com&iconUrl=http%3A%2F%2Fmail-online.nosdn.127.net%2Fsmda6015df3a52ec22402a83bdad785acb.jpg&items=%5B%22%22%2C%22%22%2C%22%22%2C%22%22%2C%22%22%5D> > 签名由 网易邮箱大师 <https://mail.163.com/dashi/dlpro.html?from=mail81> 定制 > On 9/11/2018 10:14,vino yang<yanghua1...@gmail.com> > <yanghua1...@gmail.com> wrote: > > Hi Hanjing, > > Is your JobManager working properly? Can you share your JobManager log? > > Thanks, vino. > > jing <hanjingz...@163.com> 于2018年9月11日周二 上午10:06写道: > >> Hi vino, >> >> I tried change "localhost" to the real IP. But still throw >> exception as below. JobManager configuration is as below. >> >> >> >> Thanks. >> >> Hanjing >> >> ------------------------------------ >> >> flink-conf.yaml: >> >> jobmanager.rpc.address: 170.0.0.46 >> >> >> >> # The RPC port where the JobManager is reachable. >> >> >> >> jobmanager.rpc.port: 6123 >> >> >> >> >> >> # The heap size for the JobManager JVM >> >> >> >> jobmanager.heap.size: 1024m >> >> >> >> >> >> # The heap size for the TaskManager JVM >> >> >> >> taskmanager.heap.size: 10240m >> >> >> >> >> >> # The number of task slots that each TaskManager offers. Each slot runs >> one parallel pipeline. >> >> >> >> taskmanager.numberOfTaskSlots: 16 >> >> >> >> # The parallelism used for programs that did not specify and other >> parallelism. >> >> >> >> parallelism.default: 2 >> >> >> >> >> >> Excepiton log: >> >> >> >> Starting execution of program >> >> >> >> ------------------------------------------------------------ >> >> The program finished with the following exception: >> >> >> >> org.apache.flink.client.program.ProgramInvocationException: The main >> method caused an error. >> >> at >> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:546) >> >> at >> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:421) >> >> at >> org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:426) >> >> at >> org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:804) >> >> at >> org.apache.flink.client.cli.CliFrontend.runProgram(CliFrontend.java:280) >> >> at >> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:215) >> >> at >> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:1044) >> >> at >> org.apache.flink.client.cli.CliFrontend.lambda$main$11(CliFrontend.java:1120) >> >> at java.security.AccessController.doPrivileged(Native Method) >> >> at javax.security.auth.Subject.doAs(Subject.java:422) >> >> at >> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1836) >> >> at >> org.apache.flink.runtime.security.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:41) >> >> at >> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1120) >> >> Caused by: java.lang.RuntimeException: Could not connect to Flink >> JobManager with address 170.0.0.46:6123 >> >> at >> org.apache.flink.storm.api.FlinkClient.getTopologyJobId(FlinkClient.java:304) >> >> at >> org.apache.flink.storm.api.FlinkSubmitter.submitTopology(FlinkSubmitter.java:107) >> >> at >> org.apache.flink.storm.wordcount.WordCountRemoteBySubmitter.main(WordCountRemoteBySubmitter.java:75) >> >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> >> at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >> >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >> >> at java.lang.reflect.Method.invoke(Method.java:498) >> >> at >> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:529) >> >> ... 12 more >> >> Caused by: java.io.IOException: Actor at akka.tcp:// >> flink@170.0.0.46:6123/user/jobmanager not reachable. Please make sure >> that the actor is running and its port is reachable. >> >> at >> org.apache.flink.runtime.akka.AkkaUtils$.getActorRef(AkkaUtils.scala:547) >> >> at >> org.apache.flink.runtime.akka.AkkaUtils.getActorRef(AkkaUtils.scala) >> >> at >> org.apache.flink.storm.api.FlinkClient.getJobManager(FlinkClient.java:339) >> >> at >> org.apache.flink.storm.api.FlinkClient.getTopologyJobId(FlinkClient.java:278) >> >> ... 19 more >> >> Caused by: akka.actor.ActorNotFound: Actor not found for: >> ActorSelection[Anchor(akka.tcp://flink@170.0.0.46:6123/), >> Path(/user/jobmanager)] >> >> at >> akka.actor.ActorSelection$$anonfun$resolveOne$1.apply(ActorSelection.scala:68) >> >> at >> akka.actor.ActorSelection$$anonfun$resolveOne$1.apply(ActorSelection.scala:66) >> >> at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:36) >> >> at >> akka.dispatch.BatchingExecutor$AbstractBatch.processBatch(BatchingExecutor.scala:55) >> >> at >> akka.dispatch.BatchingExecutor$Batch.run(BatchingExecutor.scala:73) >> >> at >> akka.dispatch.ExecutionContexts$sameThreadExecutionContext$.unbatchedExecute(Future.scala:76) >> >> at >> akka.dispatch.BatchingExecutor$class.execute(BatchingExecutor.scala:120) >> >> at >> akka.dispatch.ExecutionContexts$sameThreadExecutionContext$.execute(Future.scala:75) >> >> at >> scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:44) >> >> at >> scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:252) >> >> at akka.pattern.PromiseActorRef.$bang(AskSupport.scala:534) >> >> at >> akka.remote.DefaultMessageDispatcher.dispatch(Endpoint.scala:97) >> >> at >> akka.remote.EndpointReader$$anonfun$receive$2.applyOrElse(Endpoint.scala:982) >> >> at akka.actor.Actor$class.aroundReceive(Actor.scala:502) >> >> at akka.remote.EndpointActor.aroundReceive(Endpoint.scala:446) >> >> at akka.actor.ActorCell.receiveMessage(ActorCell.scala:526) >> >> at akka.actor.ActorCell.invoke(ActorCell.scala:495) >> >> at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:257) >> >> at akka.dispatch.Mailbox.run(Mailbox.scala:224) >> >> at akka.dispatch.Mailbox.exec(Mailbox.scala:234) >> >> at >> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) >> >> at >> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) >> >> at >> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) >> >> at >> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) >> >> >> On 9/10/2018 20:17,vino yang<yanghua1...@gmail.com> >> <yanghua1...@gmail.com> wrote: >> >> Hi Hanjing, >> >> OK, I mean you change the "localhost" to the real IP. >> >> Try it. >> >> Thanks, vino. >> >> jing <hanjingz...@163.com> 于2018年9月10日周一 下午8:07写道: >> >>> Hi vino, >>> jonmanager rpc address value is setted by localhost. >>> hadoop3@p-a36-72 is the node host the jobmanager jvm. >>> >>> Thanks. >>> Hanjing >>> >>> >>> >>> jing >>> 邮箱hanjingz...@163.com >>> >>> <https://maas.mail.163.com/dashi-web-extend/html/proSignature.html?ftlId=1&name=jing&uid=hanjingzuzu%40163.com&iconUrl=https%3A%2F%2Fmail-online.nosdn.127.net%2Fqiyelogo%2FdefaultAvatar.png&items=%5B%22%E9%82%AE%E7%AE%B1hanjingzuzu%40163.com%22%5D> >>> >>> 签名由 网易邮箱大师 <https://mail.163.com/dashi/dlpro.html?from=mail88> 定制 >>> >>> On 09/10/2018 19:25, vino yang <yanghua1...@gmail.com> wrote: >>> Hi Hanjing, >>> >>> I mean this configuration key.[1] >>> >>> What's more, Is the "hadoop3@p-a36-72" also the node which host >>> JobManager's jvm process? >>> >>> Thanks, vino. >>> >>> [1]: >>> https://ci.apache.org/projects/flink/flink-docs-release-1.5/ops/config.html#jobmanager-rpc-address >>> >>> jing <hanjingz...@163.com> 于2018年9月10日周一 下午6:57写道: >>> >>>> Hi vino, >>>> I commit the job on the jvm code with the command below. >>>> hadoop3@p-a36-72 flink-1.6.0]$ ./bin/flink run >>>> WordCount-StormTopology.jar input output >>>> >>>> And I'm a new user, which configuation name should be set. All the >>>> configuations are the default setting now. >>>> >>>> Thanks. >>>> Hanjing >>>> >>>> jing >>>> 邮箱hanjingz...@163.com >>>> >>>> <https://maas.mail.163.com/dashi-web-extend/html/proSignature.html?ftlId=1&name=jing&uid=hanjingzuzu%40163.com&iconUrl=https%3A%2F%2Fmail-online.nosdn.127.net%2Fqiyelogo%2FdefaultAvatar.png&items=%5B%22%E9%82%AE%E7%AE%B1hanjingzuzu%40163.com%22%5D> >>>> >>>> 签名由 网易邮箱大师 <https://mail.163.com/dashi/dlpro.html?from=mail88> 定制 >>>> >>>> On 09/10/2018 15:49, vino yang <yanghua1...@gmail.com> wrote: >>>> Hi Hanjing, >>>> >>>> Did you perform a CLI commit on the JM node? Is the address bound to >>>> "localhost" in the Flink JM configuration? >>>> >>>> Thanks, vino. >>>> >>>> jing <hanjingz...@163.com> 于2018年9月10日周一 上午11:00写道: >>>> >>>>> Hello, >>>>> >>>>> I’m trying to run flink-storm-example on standalone clusters. >>>>> But there’s some exception I can’t sovle. Could anyone please help me >>>>> with trouble. >>>>> >>>>> flink-storm-example version: 1.60 >>>>> >>>>> flink version: 1.60 >>>>> >>>>> The log below is the Exception. My job manager status is as the >>>>> picture. >>>>> >>>>> I’v tried to changed the IP address and port, but it doesn’t’ >>>>> work. >>>>> >>>>> >>>>> >>>>> Thanks a lot. >>>>> >>>>> ------------------------------------------- >>>>> >>>>> [hadoop3@p-a36-72 flink-1.6.0]$ ./bin/flink run >>>>> WordCount-StormTopology.jar input output >>>>> >>>>> Starting execution of program >>>>> >>>>> >>>>> ------------------------------------------------------------ >>>>> >>>>> The program finished with the following exception: >>>>> >>>>> >>>>> org.apache.flink.client.program.ProgramInvocationException: The main >>>>> method caused an error. >>>>> >>>>> at >>>>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:546) >>>>> >>>>> at >>>>> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:421) >>>>> >>>>> at >>>>> org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:426) >>>>> >>>>> at >>>>> org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:804) >>>>> >>>>> at >>>>> org.apache.flink.client.cli.CliFrontend.runProgram(CliFrontend.java:280) >>>>> >>>>> at >>>>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:215) >>>>> >>>>> at >>>>> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:1044) >>>>> >>>>> at >>>>> org.apache.flink.client.cli.CliFrontend.lambda$main$11(CliFrontend.java:1120) >>>>> >>>>> at java.security.AccessController.doPrivileged(Native Method) >>>>> >>>>> at javax.security.auth.Subject.doAs(Subject.java:422) >>>>> >>>>> at >>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1754) >>>>> >>>>> at >>>>> org.apache.flink.runtime.security.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:41) >>>>> >>>>> at >>>>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1120) >>>>> >>>>> Caused by: java.lang.RuntimeException: Could not connect to Flink >>>>> JobManager with address localhost:6123 >>>>> >>>>> at >>>>> org.apache.flink.storm.api.FlinkClient.getTopologyJobId(FlinkClient.java:304) >>>>> >>>>> at >>>>> org.apache.flink.storm.api.FlinkSubmitter.submitTopology(FlinkSubmitter.java:107) >>>>> >>>>> at >>>>> org.apache.flink.storm.wordcount.WordCountRemoteBySubmitter.main(WordCountRemoteBySubmitter.java:75) >>>>> >>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>>> >>>>> at >>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >>>>> >>>>> at >>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>>> >>>>> at java.lang.reflect.Method.invoke(Method.java:498) >>>>> >>>>> at >>>>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:529) >>>>> >>>>> ... 12 more >>>>> >>>>> Caused by: java.io.IOException: Actor at >>>>> akka.tcp://flink@localhost:6123/user/jobmanager >>>>> not reachable. Please make sure that the actor is running and its port is >>>>> reachable. >>>>> >>>>> at >>>>> org.apache.flink.runtime.akka.AkkaUtils$.getActorRef(AkkaUtils.scala:547) >>>>> >>>>> at >>>>> org.apache.flink.runtime.akka.AkkaUtils.getActorRef(AkkaUtils.scala) >>>>> >>>>> at >>>>> org.apache.flink.storm.api.FlinkClient.getJobManager(FlinkClient.java:339) >>>>> >>>>> at >>>>> org.apache.flink.storm.api.FlinkClient.getTopologyJobId(FlinkClient.java:278) >>>>> >>>>> ... 19 more >>>>> >>>>> Caused by: akka.actor.ActorNotFound: Actor not found for: >>>>> ActorSelection[Anchor(akka.tcp://flink@localhost:6123/), >>>>> Path(/user/jobmanager)] >>>>> >>>>> at >>>>> akka.actor.ActorSelection$$anonfun$resolveOne$1.apply(ActorSelection.scala:68) >>>>> >>>>> at >>>>> akka.actor.ActorSelection$$anonfun$resolveOne$1.apply(ActorSelection.scala:66) >>>>> >>>>> at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:36) >>>>> >>>>> at >>>>> akka.dispatch.BatchingExecutor$AbstractBatch.processBatch(BatchingExecutor.scala:55) >>>>> >>>>> at >>>>> akka.dispatch.BatchingExecutor$Batch.run(BatchingExecutor.scala:73) >>>>> >>>>> at >>>>> akka.dispatch.ExecutionContexts$sameThreadExecutionContext$.unbatchedExecute(Future.scala:76) >>>>> >>>>> at >>>>> akka.dispatch.BatchingExecutor$class.execute(BatchingExecutor.scala:120) >>>>> >>>>> at >>>>> akka.dispatch.ExecutionContexts$sameThreadExecutionContext$.execute(Future.scala:75) >>>>> >>>>> at >>>>> scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:44) >>>>> >>>>> at >>>>> scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:252) >>>>> >>>>> at akka.pattern.PromiseActorRef.$bang(AskSupport.scala:534) >>>>> >>>>> at >>>>> akka.remote.DefaultMessageDispatcher.dispatch(Endpoint.scala:97) >>>>> >>>>> at >>>>> akka.remote.EndpointReader$$anonfun$receive$2.applyOrElse(Endpoint.scala:982) >>>>> >>>>> at akka.actor.Actor$class.aroundReceive(Actor.scala:502) >>>>> >>>>> at akka.remote.EndpointActor.aroundReceive(Endpoint.scala:446) >>>>> >>>>> at akka.actor.ActorCell.receiveMessage(ActorCell.scala:526) >>>>> >>>>> at akka.actor.ActorCell.invoke(ActorCell.scala:495) >>>>> >>>>> at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:257) >>>>> >>>>> at akka.dispatch.Mailbox.run(Mailbox.scala:224) >>>>> >>>>> at akka.dispatch.Mailbox.exec(Mailbox.scala:234) >>>>> >>>>> at >>>>> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) >>>>> >>>>> at >>>>> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) >>>>> >>>>> at >>>>> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) >>>>> >>>>> at >>>>> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> Hanjing >>>>> >>>>> >>>>> >>>>> >>>>