How to write stream data to other Hadoop Cluster by StreamingFileSink

2019-10-04 Thread Jun Zhang
Hi,all: I have 2 hadoop cluster (hdfs://mycluster1 and hdfs://mycluster2),both of them configured the HA, I have a job ,read from streaming data from kafka, and write it to hdfs by StreamingFileSink,now I deployed my job on mycluster1 (flink on yarn),and I want to write the data to mycluster2

Re: How to write stream data to other Hadoop Cluster by StreamingFileSink

2019-10-08 Thread Yang Wang
Hi Jun Zhang, I think you could add two hdfs cluster configurations in your hdfs-site.xml. The following config keys need to be added. Then you could use both two hdfs clusters in your flink job. dfs.nameservices: mycluster1,mycluster2 dfs.ha.namenodes.mycluster1: nn1,nn2 dfs.client.failover.prox

Re: How to write stream data to other Hadoop Cluster by StreamingFileSink

2019-10-08 Thread Jun Zhang
Hi,Yang : thank you very much for your reply. I had add the configurations on my hadoop cluster client , both hdfs-site.xml and core-site.xml are configured, the client can read mycluster1 and mycluter2, but when I submit the flink job to yarn cluster , the hadoop client configurations is inva