I will thanks. Once I had it set up and working. I switched my computers around from client to server to server to client. With your excellent instructions I was able to do it in 5 .minutes
On Mon, 20 Apr 2020, 00:05 Jeff Zhang, <zjf...@gmail.com> wrote: > Som, Let us know when you have any problems > > Som Lima <somplastic...@gmail.com> 于2020年4月20日周一 上午2:31写道: > >> Thanks for the info and links. >> >> I had a lot of problems I am not sure what I was doing wrong. >> >> May be conflicts with setup from apache spark. I think I may need to >> setup users for each development. >> >> >> Anyway I kept doing fresh installs about four altogether I think. >> >> Everything works fine now >> Including remote access of zeppelin on machines across the local area >> network. >> >> Next step setup remote clusters >> Wish me luck ! >> >> >> >> >> >> >> >> On Sun, 19 Apr 2020, 14:58 Jeff Zhang, <zjf...@gmail.com> wrote: >> >>> Hi Som, >>> >>> You can take a look at flink on zeppelin, in zeppelin you can connect to >>> a remote flink cluster via a few configuration, and you don't need to worry >>> about the jars. Flink interpreter will ship necessary jars for you. Here's >>> a list of tutorials. >>> >>> 1) Get started https://link.medium.com/oppqD6dIg5 >>> <https://t.co/PTouUYYTrv?amp=1> 2) Batch https:// >>> link.medium.com/3qumbwRIg5 <https://t.co/Yo9QAY0Joj?amp=1> 3) Streaming >>> https://link.medium.com/RBHa2lTIg5 <https://t.co/sUapN40tvI?amp=1> 4) >>> Advanced usage https://link.medium.com/CAekyoXIg5 >>> <https://t.co/MXolULmafZ?amp=1> >>> >>> >>> Zahid Rahman <zahidr1...@gmail.com> 于2020年4月19日周日 下午7:27写道: >>> >>>> Hi Tison, >>>> >>>> I think I may have found what I want in example 22. >>>> >>>> https://www.programcreek.com/java-api-examples/?api=org.apache.flink.configuration.Configuration >>>> >>>> I need to create Configuration object first as shown . >>>> >>>> Also I think flink-conf.yaml file may contain configuration for client >>>> rather than server. So before starting is irrelevant. >>>> I am going to play around and see but if the Configuration class allows >>>> me to set configuration programmatically and overrides the yaml file then >>>> that would be great. >>>> >>>> >>>> >>>> On Sun, 19 Apr 2020, 11:35 Som Lima, <somplastic...@gmail.com> wrote: >>>> >>>>> Thanks. >>>>> flink-conf.yaml does allow me to do what I need to do without making >>>>> any changes to client source code. >>>>> >>>>> But >>>>> RemoteStreamEnvironment constructor expects a jar file as the third >>>>> parameter also. >>>>> >>>>> RemoteStreamEnvironment >>>>> <https://ci.apache.org/projects/flink/flink-docs-release-1.7/api/java/org/apache/flink/streaming/api/environment/RemoteStreamEnvironment.html#RemoteStreamEnvironment-java.lang.String-int-java.lang.String...-> >>>>> (String >>>>> <http://docs.oracle.com/javase/7/docs/api/java/lang/String.html?is-external=true> >>>>> host, >>>>> int port, String >>>>> <http://docs.oracle.com/javase/7/docs/api/java/lang/String.html?is-external=true> >>>>> ... jarFiles) >>>>> Creates a new RemoteStreamEnvironment that points to the master >>>>> (JobManager) described by the given host name and port. >>>>> >>>>> On Sun, 19 Apr 2020, 11:02 tison, <wander4...@gmail.com> wrote: >>>>> >>>>>> You can change flink-conf.yaml "jobmanager.address" or >>>>>> "jobmanager.port" options before run the program or take a look at >>>>>> RemoteStreamEnvironment which enables configuring host and port. >>>>>> >>>>>> Best, >>>>>> tison. >>>>>> >>>>>> >>>>>> Som Lima <somplastic...@gmail.com> 于2020年4月19日周日 下午5:58写道: >>>>>> >>>>>>> Hi, >>>>>>> >>>>>>> After running >>>>>>> >>>>>>> $ ./bin/start-cluster.sh >>>>>>> >>>>>>> The following line of code defaults jobmanager to localhost:6123 >>>>>>> >>>>>>> final ExecutionEnvironment env = >>>>>>> Environment.getExecutionEnvironment(); >>>>>>> >>>>>>> which is same on spark. >>>>>>> >>>>>>> val spark = >>>>>>> SparkSession.builder.master(local[*]).appname("anapp").getOrCreate >>>>>>> >>>>>>> However if I wish to run the servers on a different physical >>>>>>> computer. >>>>>>> Then in Spark I can do it this way using the spark URI in my IDE. >>>>>>> >>>>>>> Conf = >>>>>>> SparkConf().setMaster("spark://<hostip>:<port>").setAppName("anapp") >>>>>>> >>>>>>> Can you please tell me the equivalent change to make so I can run my >>>>>>> servers and my IDE from different physical computers. >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>> >>> -- >>> Best Regards >>> >>> Jeff Zhang >>> >> > > -- > Best Regards > > Jeff Zhang >