Re: Implement Joins with Lookup Data

2018-08-28 Thread Till Rohrmann
Ankit that this problem could actually be solved quite >>>>> elegantly with Flink's state. If you can ingest the product/account >>>>> information changes as a stream, you can keep the latest version of it in >>>>> Flink state by using a co-map function [1,

Re: Implement Joins with Lookup Data

2018-08-23 Thread Hequn Cheng
incoming event. >>>> >>>> [1] >>>> https://ci.apache.org/projects/flink/flink-docs-master/dev/stream/operators/ >>>> [2] >>>> https://ci.apache.org/projects/flink/flink-docs-master/api/java/org/apache/flink/streami

Re: Implement Joins with Lookup Data

2018-08-22 Thread Harshvardhan Agrawal
t; >>> On Tue, Jul 24, 2018 at 2:15 PM Harshvardhan Agrawal < >>> harshvardhan.ag...@gmail.com> wrote: >>> >>> Hi, >>> >>> Thanks for your responses. >>> >>> There is no fixed interval for the data being updated. It’s more like >

Re: Implement Joins with Lookup Data

2018-08-19 Thread Hequn Cheng
mandates that change >> will trigger the reference data to change. >> >> It’s not just the enrichment we are doing here. Once we have enriched the >> data we will be performing a bunch of aggregations using the enriched data. >> >> Which approach wo

Re: Implement Joins with Lookup Data

2018-08-19 Thread Harshvardhan Agrawal
o, just based on this feature, flink doesn’t seem to add a lot of value > on top of Kafka. As Jorn said below, you can very well store all the events > in an external store and then periodically run a cron to enrich later since > your processing doesn’t seem to require absolute real time.

Re: Implement Joins with Lookup Data

2018-07-25 Thread ashish pok
Ankit   From: Jörn Franke Date: Monday, July 23, 2018 at 10:10 PM To: Harshvardhan Agrawal Cc: Subject: Re: Implement Joins with Lookup Data   For the first one (lookup of single entries) you could use a NoSQL db (eg key value store) - a relational database will not scale.   Depending

Re: Implement Joins with Lookup Data

2018-07-25 Thread Michael Gendelman
ell store all the events > in an external store and then periodically run a cron to enrich later since > your processing doesn’t seem to require absolute real time. > > > > Thanks > > Ankit > > > > *From: *Jörn Franke > *Date: *Monday, July 23, 2018 at 10:10 PM

Re: Implement Joins with Lookup Data

2018-07-24 Thread ashish pok
: Subject: Re: Implement Joins with Lookup Data   For the first one (lookup of single entries) you could use a NoSQL db (eg key value store) - a relational database will not scale.   Depending on when you need to do the enrichment you could also first store the data and enrich it later as part

Re: Implement Joins with Lookup Data

2018-07-24 Thread Harshvardhan Agrawal
n top of Kafka. As Jorn said below, you can very well store all the events > in an external store and then periodically run a cron to enrich later since > your processing doesn’t seem to require absolute real time. > > > > Thanks > > Ankit > > > > *From: *Jörn Franke > *Da

Re: Implement Joins with Lookup Data

2018-07-24 Thread ashish pok
.   Thanks Ankit   From: Jörn Franke Date: Monday, July 23, 2018 at 10:10 PM To: Harshvardhan Agrawal Cc: Subject: Re: Implement Joins with Lookup Data   For the first one (lookup of single entries) you could use a NoSQL db (eg key value store) - a relational database will not scale

Re: Implement Joins with Lookup Data

2018-07-24 Thread Elias Levy
the product db updated? Based on that you can store >>>>> product metadata as state in Flink, maybe setup the state on cluster >>>>> startup and then update daily etc. >>>>> >>>>> >>>>> >>>>> Also, just based on this featur

Re: Implement Joins with Lookup Data

2018-07-24 Thread Till Rohrmann
;> Also, just based on this feature, flink doesn’t seem to add a lot of >>>> value on top of Kafka. As Jorn said below, you can very well store all the >>>> events in an external store and then periodically run a cron to enrich >>>> later since your processing doesn’t

Re: Implement Joins with Lookup Data

2018-07-24 Thread Harshvardhan Agrawal
hen periodically run a cron to enrich >>> later since your processing doesn’t seem to require absolute real time. >>> >>> >>> >>> Thanks >>> >>> Ankit >>> >>> >>> >>> *From: *Jörn Franke >>>

Re: Implement Joins with Lookup Data

2018-07-24 Thread Till Rohrmann
;> >> Ankit >> >> >> >> *From: *Jörn Franke >> *Date: *Monday, July 23, 2018 at 10:10 PM >> *To: *Harshvardhan Agrawal >> *Cc: * >> *Subject: *Re: Implement Joins with Lookup Data >> >> >> >> For the first one (looku

Re: Implement Joins with Lookup Data

2018-07-24 Thread Harshvardhan Agrawal
cron to enrich later since > your processing doesn’t seem to require absolute real time. > > > > Thanks > > Ankit > > > > *From: *Jörn Franke > *Date: *Monday, July 23, 2018 at 10:10 PM > *To: *Harshvardhan Agrawal > *Cc: * > *Subject: *Re: Implement Jo

Re: Implement Joins with Lookup Data

2018-07-24 Thread Jain, Ankit
well store all the events in an external store and then periodically run a cron to enrich later since your processing doesn’t seem to require absolute real time. Thanks Ankit From: Jörn Franke Date: Monday, July 23, 2018 at 10:10 PM To: Harshvardhan Agrawal Cc: Subject: Re: Implement Joins

Re: Implement Joins with Lookup Data

2018-07-23 Thread Jörn Franke
For the first one (lookup of single entries) you could use a NoSQL db (eg key value store) - a relational database will not scale. Depending on when you need to do the enrichment you could also first store the data and enrich it later as part of a batch process. > On 24. Jul 2018, at 05:25,