Hi Xingcan, Ken and Till,
OK, thank you. It is clear.
I have various option then:
* the one suggested by Ken where I can find a way to build a key that
will be well distributed (1 key per task)
o it relies on the way Flink partitions the key, but it will do
the job
* or I can also go with another way to build my key where I will have
more keys than the parallelism, so the distribution will be better
o I will still have few number of requests (much less than the
number of resource ids as 1 key will be for multiple resource ids)
o I will potentially do multiple requests on the same task, but it
may be acceptable, especially if I go with AsyncIO
* or I can go with the OperatorState and implements my own firing logic
o I am in a case where the memory-based mechanism should be fine
Thanks again,
Regards.
Julien.
On 20/02/2018 02:48, Xingcan Cui wrote:
Hi Julien,
you could use the OperatorState
<https://ci.apache.org/projects/flink/flink-docs-master/dev/stream/state/state.html#using-managed-operator-state> to
cache the data in a window and the last time your window fired. Then
you check the ctx.timerService().currentProcessingTime() in
processElement() and once it exceeds the next window boundary, all the
cached data should be processed as if the window is fired.
Note that currently, there are only memory-based operator states provided.
Hope this helps,
Xingcan
On 19 Feb 2018, at 4:34 PM, Julien <jmassio...@gmail.com
<mailto:jmassio...@gmail.com>> wrote:
Hello,
I've already tried to key my stream with
"resourceId.hashCode%parallelism" (with parallelism of 4 in my example).
So all my keys will be either 0,1, 2 or 3. I can then benefit from a
time window on this keyed stream and do only 4 queries to my external
system.
But it is not well distributed with the default partitioner on keyed
stream. (keys 0, 1, 2 and 3 only goes to operator idx 2, 3).
I think I should explore the customer partitioner, as you suggested
Xingcan.
Maybe my last question on this will be: "can you give me more details
on this point "and simulate a window operation by yourself in a
ProcessFunction" ?
When I look at the documentation about the custom partitioner, I can
see that the result of partitionCustom is a DataStream.
It is not a KeyedStream.
So the only window I have will be windowAll (which will bring me back
to a parallelism of 1, no ?).
And if I do something like "myStream.partitionCustom(<my new
partitioner>,<my key>).keyBy(<myKey>).window(...)", will it preserve
my custom partitioner ?
When looking at the "KeyedStream" class, it seems that it will go
back to the "KeyGroupStreamPartitioner" and forget my custom
partitioner ?
Thanks again for your feedback,
Julien.
On 19/02/2018 03:45, 周思华 wrote:
Hi Julien,
If I am not misunderstand, I think you can key your stream on a
`Random.nextInt() % parallesm`, this way you can "group" together
alerts from different and benefit from multi parallems.
发自网易邮箱大师
On 02/19/2018 09:08,Xingcan Cui<xingc...@gmail.com
<mailto:xingc...@gmail.com>> wrote:
Hi Julien,
sorry for my misunderstanding before. For now, the window can only
be defined on a KeyedStream or an ordinary DataStream but with
parallelism = 1. I’d like to provide three options for your scenario.
1. If your external data is static and can be fit into the memory,
you can use ManagedStates to cache them without considering the
querying problem.
2. Or you can use a CustomPartitioner to manually distribute your
alert data and simulate an window operation by yourself in a
ProcessFuncton.
3. You may also choose to use some external systems such as
in-memory store, which can work as a cache for your queries.
Best,
Xingcan
On 19 Feb 2018, at 5:55 AM, Julien <jmassio...@gmail.com
<mailto:jmassio...@gmail.com>> wrote:
Hi Xingcan,
Thanks for your answer.
Yes, I understand that point:
• if I have 100 resource IDs with parallelism of 4, then each
operator instance will handle about 25 keys
The issue I have is that I want, on a given operator instance, to
group those 25 keys together in order to do only 1 query to an
external system per operator instance:
• on a given operator instance, I will do 1 query for my 25 keys
• so with the 4 operator instances, I will do 4 query in parallel
(with about 25 keys per query)
I do not know how I can do that.
If I define a window on my keyed stream (with for
example stream.key(_.resourceId).window(TumblingProcessingTimeWindows.of(Time.milliseconds(500))), then
my understanding is that the window is "associated" to the key. So
in this case, on a given operator instance, I will have 25 of those
windows (one per key), and I will do 25 queries (instead of 1).
Do you understand my point ?
Or maybe am I missing something ?
I'd like to find a way on operator instance 1 to group all the
alerts received on those 25 resource ids and do 1 query for those
25 resource ids.
Same thing for operator instance 2, 3 and 4.
Thank you,
Regards.
On 18/02/2018 14:43, Xingcan Cui wrote:
Hi Julien,
the cardinality of your keys (e.g., resource ID) will not be
restricted to the parallelism. For instance, if you have 100
resource IDs processed by KeyedStream with parallelism 4, each
operator instance will handle about 25 keys.
Hope that helps.
Best,
Xingcan
On 18 Feb 2018, at 8:49 PM, Julien <jmassio...@gmail.com> wrote:
Hi,
I am pretty new to flink and I don't know what will be the best
way to deal with the following use case:
• as an input, I recieve some alerts from a kafka topic
• an alert is linked to a network resource (like router-1,
router-2, switch-1, switch-2, ...)
• so an alert has two main information (the alert id and the
resource id of the resource on which this alert has been raised)
• then I need to do a query to an external system in order to
enrich the alert with additional information on the resource
(A "natural" candidate for the key on this stream will be the
resource id)
The issue I have is that regarding the query to the external system:
• I do not want to do 1 query per resource id
• I want to do a small number of queries in parallel (for example
4 queries in parallel every 500ms), each query requesting the
external system for several alerts linked to several resource id
Currently, I don't know what will be the best way to deal with that:
• I can key my stream on the resource id and then define a
processing time window of 500ms and when the trigger is ok, then
I do my query
• by doing so, I will "group" several alerts in a single query,
but they will all be linked to the same resource.
• so I will do 1 query per resource id (which will be too much in
my use case)
• I can also do a windowAll on a non keyed stream
• by doing so, I will "group" together alerts from different
resource ids, but from what I've read in such a case the
parallelism will always be one.
• so in this case, I will only do 1 query whereas I'd like to
have some parallelism
I am thinking that a way to deal with that will be:
• define the resource id as the key of stream and put a
parallelism of 4
• and then having a way to do a windowAll on this keyed stream
• which is that, on a given operator instance, I will "group" on
the same window all the keys (ie all the resource ids) managed by
this operator instance
• with a parallelism of 4, I will do 4 queries in parallel (1 per
operator instance, and each query will be for several alerts
linked to several resource ids)
But after looking at the documentation, I cannot see this ability
(having a windowAll on a keyed stream).
Am I missing something?
What will be the best way to deal with such a use case?
I've tried for example to review my key and to do something
like "resourceId.hahsCode%<max nb of queries in parallel>" and
then to use a time window.
In my example above, the <max nb of queries in parallel> will be
4. And all my keys will be 0, 1, 2 or 3.
The issue with this approach is that due to the way the
operatorIdx is computed based on the key, it does not distribute
well my processing:
• when this partitioning logic from the "KeyGroupRangeAssignment"
class is applied
• /**
* Assigns the given key to a parallel operator index.
*
* @param key the key to assign
* @param maxParallelism the maximum supported parallelism,
aka the number of key-groups.
* @param parallelism the current parallelism of the operator
* @return the index of the parallel operator to which the
given key should be routed.
*/
public static int assignKeyToParallelOperator(Object key, int
maxParallelism, int parallelism) {
return computeOperatorIndexForKeyGroup(maxParallelism,
parallelism, assignToKeyGroup(key, maxParallelism));
}
/**
* Assigns the given key to a key-group index.
*
* @param key the key to assign
* @param maxParallelism the maximum supported parallelism,
aka the number of key-groups.
* @return the key-group to which the given key is assigned
*/
public static int assignToKeyGroup(Object key, int
maxParallelism) {
return computeKeyGroupForKeyHash(key.hashCode(),
maxParallelism);
}
• key 0, 1, 2 and 3 are only assigned to operator 2 and 3 (so 2
over my 4 operators will not have anything to do)
So, what will be the best way to deal with that?
Thank you in advance for your support.
Regards.
Julien.