Thanks Fabian and Tony for the info. It's very helpful.

Looks like the general approach is to implement a job topology containing
parameterized (CoXXXMapFunction) operators. The user defined parameters
will be ingested using the extra input the CoXXXMapFunction take.

Ken

On Wed, Nov 29, 2017 at 5:27 AM, Fabian Hueske <fhue...@gmail.com> wrote:

> Another example is King's RBEA platform [1] which was built on Flink.
> In a nutshell, RBEA runs a single large Flink job, to which users can add
> queries that should be computed.
> Of course, the query language is restricted because they queries must
> match on the structure of the running job.
>
> Hope this helps,
> Fabian
>
> [1] http://2016.flink-forward.org/kb_sessions/rbea-scalable-
> real-time-analytics-at-king/
>
> 2017-11-29 3:32 GMT+01:00 Tony Wei <tony19920...@gmail.com>:
>
>> Hi KZ,
>>
>> https://data-artisans.com/blog/real-time-fraud-detection-
>> ing-bank-apache-flink
>> This article seems to be a good example to trigger a new calculation on a
>> running job. Maybe you can get some help from it.
>>
>> Best Regards,
>> Tony Wei
>>
>> 2017-11-29 4:53 GMT+08:00 zanqing zhang <zanqingzh...@gmail.com>:
>>
>>> Hi All,
>>>
>>> Has anyone done any stream processing driven by a user request? What's
>>> the recommended way of doing this? Or is this completely wrong direction to
>>> go for applications running on top of Flink?
>>>
>>> Basically we need to tweak the stream processing based on parameters
>>> provided by a user, e.g. show me the total # of application failures due to
>>> "ABC", which is provided by the user. We are thinking of starting a flink
>>> job with "ABC" as a parameter but this would result in a huge number of
>>> flink jobs, is there a better way for this? Can we trigger the calculation
>>> on a running job?
>>>
>>> Thanks in advance.
>>>
>>> KZ
>>>
>>>
>>
>


-- 
Ken Zhang
-----------------------------------------------------------
*Smart && Creative && Open == Innovative*

Reply via email to