Hi Gyula, Big +1 for this, it would be very helpful for flink jobs and cluster operations. Do you call flink rest api to gather the job info ? I hope this history server could work with multiple versions of flink as long as the flink rest api is compatible.
Gyula Fóra <gyula.f...@gmail.com> 于2020年5月13日周三 下午4:13写道: > Hi All! > > With the growing number of Flink streaming applications the current HS > implementation is starting to lose its value. Users running streaming > applications mostly care about what is running right now on the cluster and > a centralised view on history is not very useful. > > We have been experimenting with reworking the current HS into a Global > Flink Dashboard that would show all running and completed/failed jobs on > all the running Flink clusters the users have. > > In essence we would get a view similar to the current HS but it would also > show the running jobs with a link redirecting to the actual cluster > specific dashboard. > > This is how it looks now: > > > In this version we took a very simple approach of introducing a cluster > discovery abstraction to collect all the running Flink clusters (by listing > yarn apps for instance). > > The main pages aggregating jobs from different clusters would then simply > make calls to all clusters and aggregate the response. Job specific > endpoints would be simply routed to the correct target cluster. This way > the changes required are localised to the current HS implementation and > cluster rest endpoints don't need to be changed. > > In addition to getting a fully working global dashboard this also gets us a > fully functioning rest endpoint for accessing all jobs in all clusters > without having to provide the clusterId (yarn app id for instance) that we > can use to enhance CLI experience in multi cluster (lot of per-job > clusters) environments. Please let us know what you think! Gyula > -- Best Regards Jeff Zhang