It seems that not everyone can see the screenshot in the email, so here is
a link:

https://drive.google.com/open?id=1abrlpI976NFqOZSX20k2FoiAfVhBbER9

On Wed, May 13, 2020 at 11:29 AM Gyula Fóra <gyula.f...@gmail.com> wrote:

> Oops I forgot the screenshot, thanks Ufuk :D
>
>
> @Jeff Zhang <zjf...@gmail.com> : Yes we simply call to the individual
> cluster's rest endpoints so it would work with multiple flink versions yes.
> Gyula
>
>
> On Wed, May 13, 2020 at 10:56 AM Jeff Zhang <zjf...@gmail.com> wrote:
>
>> Hi Gyula,
>>
>> Big +1 for this, it would be very helpful for flink jobs and cluster
>> operations. Do you call flink rest api to gather the job info ? I hope
>> this
>> history server could work with multiple versions of flink as long as the
>> flink rest api is compatible.
>>
>> Gyula Fóra <gyula.f...@gmail.com> 于2020年5月13日周三 下午4:13写道:
>>
>> > Hi All!
>> >
>> > With the growing number of Flink streaming applications the current HS
>> > implementation is starting to lose its value. Users running streaming
>> > applications mostly care about what is running right now on the cluster
>> and
>> > a centralised view on history is not very useful.
>> >
>> > We have been experimenting with reworking the current HS into a Global
>> > Flink Dashboard that would show all running and completed/failed jobs on
>> > all the running Flink clusters the users have.
>> >
>> > In essence we would get a view similar to the current HS but it would
>> also
>> > show the running jobs with a link redirecting to the actual cluster
>> > specific dashboard.
>> >
>> > This is how it looks now:
>> >
>> >
>> > In this version we took a very simple approach of introducing a cluster
>> > discovery abstraction to collect all the running Flink clusters (by
>> listing
>> > yarn apps for instance).
>> >
>> > The main pages aggregating jobs from different clusters would then
>> simply
>> > make calls to all clusters and aggregate the response. Job specific
>> > endpoints would be simply routed to the correct target cluster. This way
>> > the changes required are localised to the current HS implementation and
>> > cluster rest endpoints don't need to be changed.
>> >
>> > In addition to getting a fully working global dashboard this also gets
>> us a
>> > fully functioning rest endpoint for accessing all jobs in all clusters
>> > without having to provide the clusterId (yarn app id for instance) that
>> we
>> > can use to enhance CLI experience in multi cluster (lot of per-job
>> > clusters) environments. Please let us know what you think! Gyula
>> >
>>
>>
>> --
>> Best Regards
>>
>> Jeff Zhang
>>
>

Reply via email to