Re: KafkaStream groupBy + count on KTable behaviour

2020-05-14 Thread Raffaele Esposito
Thanks a lot Bruno, much clearer now. It's only my opinion but since the Topology is a concept of the API as well as the repartitioning logic, for me also this mechanism should be a bit more transparent, but it aslo maybe that I'm plain wrong here :) Thanks ! On Thu, May 14, 2020 at 9:24 PM

Re: data structures used by GlobalKTable, KTable

2020-05-14 Thread Matthias J. Sax
Yeah, the current API doesn't make it very clear how to do it. You can set an in-memory like this: > builder.globalTable("topic", > Materialized.as(Stores.inMemoryKeyValueStore("store-name"))); We are already working on an improved API via KIP-591:

Re: KafkaStream groupBy + count on KTable behaviour

2020-05-14 Thread Bruno Cadonna
Hi Raffaele, Change is an internal class in Streams and also its SerDes are internal. To consume the repartition topic you mention outside of Streams you would need to use those internal classes (note: I've never tried this). Those classes can change at any time. So consuming from repartition

Re: KafkaStream groupBy + count on KTable behaviour

2020-05-14 Thread Raffaele Esposito
Hi Bruno, Also when you mention: The record structure key, (oldValue, newValue) is called Change in Kafka Streams and it is used where updates are emitted downstream Does it also mean the same happen when we convert a KTable to a KStream ? Do you know any docs or article about this topics?

Re: KafkaStream groupBy + count on KTable behaviour

2020-05-14 Thread Raffaele Esposito
Hi Bruno, Thanks, One more thing, As I told you I was consuming the repartitioning topic created by group by and I just saw the old and new value, as you are telling me now they are indeed marked as old and new, is this mark visible somehow consuming the repartitioning topic ? Raffaele On Thu,

JDBC source connector

2020-05-14 Thread vishnu murali
Hi Guys, I am using the mode *bulk *and poll.interval.ms *10* in the Source connector configuration. But I don't need to load data another time.? I need to load the data only once ?? How can I able to do this ?

Re: What to use to copy data from one kafka to another ?

2020-05-14 Thread Liam Clarke-Hutchinson
You could use Mirror Maker 1, it's a basic "consume from topic A in cluster X, produce to topic A in cluster Y" app. On Fri, 15 May 2020, 12:13 am JP MB, wrote: > Hi guys, > My use case is simply copying data from one Kafka to another. When > searching on google, the immediate answer seems to

KafkaStream groupBy + count on KTable behaviour

2020-05-14 Thread Raffaele Esposito
I m trying to better understand KTable and I have encountered a behaviour I cannot wrap my mind around it. So* groupByKey()* can only be applied to KStream and not to KTable, that's because of the nature of KTable that and its UPSERT logic. What I don't understand correctly and therefore ask your

Re: [External] kafka connector distributed mode balance stratety

2020-05-14 Thread Tauzell, Dave
When clients connect the Kafka broker assigns each client a subset of the topic partitions. When a client becomes unresponsive then the broker assigns those partitions to another client. The state of your connector task isn't going to be transferred but another connector task will eventually

Re: JDBC source connector

2020-05-14 Thread Liam Clarke-Hutchinson
Why not use autoincrement? It'll only emit new records on subsequent polls then. On Thu, 14 May 2020, 11:15 pm vishnu murali, wrote: > Hi Guys, > > I am using the mode *bulk *and poll.interval.ms *10* in the Source > connector configuration. > > But I don't need to load data another time.?

Need help regarding missing kafka producer metrics while upgrade from 1.1.0 to 2.0.0

2020-05-14 Thread Rajkumar Natarajan
Hi Kafka community, we are currently using kafka client version 1.1.0 in our production. I'm working on upgrade it to kafka version 2.0.0. I see there are some metrics which are present in version 1.1.1 are removed in kafka version 2.0.0. S# version 1.1.1 2.0.0 1

Re: JDBC source connector

2020-05-14 Thread vishnu murali
Thanks Liam But I am asking like assume I am having 10. Using JDBC source I need to push that once.. No more additional data will be added in future in that table. In that case i need to push that only once not more than one... For this scenario I am asking!! On Thu, May 14, 2020, 19:20

Re: What to use to copy data from one kafka to another ?

2020-05-14 Thread JP MB
Yes, I saw that it would fit my use case but that brings another question. Is MM1 deprecated or it will be discontinued ? Regards Em qui., 14 de mai. de 2020 às 14:46, Liam Clarke-Hutchinson < liam.cla...@adscale.co.nz> escreveu: > You could use Mirror Maker 1, it's a basic "consume from topic

What to use to copy data from one kafka to another ?

2020-05-14 Thread JP MB
Hi guys, My use case is simply copying data from one Kafka to another. When searching on google, the immediate answer seems to be Mirror Maker, so we jumped to the most recent version MM2. The thing is I don't want active/active replication, the consumers from cluster A will be different from the

Re: JDBC source connector

2020-05-14 Thread Robin Moffatt
If you just want it once then delete the connector once it's processed all the data -- Robin Moffatt | Senior Developer Advocate | ro...@confluent.io | @rmoff On Thu, 14 May 2020 at 16:14, vishnu murali wrote: > Thanks Liam > > But I am asking like assume I am having 10. > > Using

ACL/SSL Issue with Configuration = ssl.principal.mapping.rules

2020-05-14 Thread Jonathan Goings
I'm working through a kafka implementation. I'm having issues with the ssl.principal.mapping.rules configuration. I've successfully started up Kafka with the SSL encrypt/authenticate in place and I've successfully set up the super user using the full principal name. The issue arises when I

InvalidRecordException: Inner record LegacyRecordBatch

2020-05-14 Thread Gérald Quintana
Hello, Since we upgraded from Kafka 2.3 to 2.4 we have big amount logs telling: [2020-05-14 19:34:27,623] ERROR [ReplicaManager broker=1] Error processing append operation on partition ourtopic-1 (kafka.server.ReplicaManager) org.apache.kafka.common.InvalidRecordException: Inner record

Re: KafkaStream groupBy + count on KTable behaviour

2020-05-14 Thread Bruno Cadonna
Hi Raffaele, In your example, Kafka Streams would send the new and the old value downstream. More specifically, the groupBy() would send (as you also observed) London, (old value: London, new value: null) Berlin, (old value: null, new value: Berlin) At the count() record London, (old value: