Hi Chris, Take it easy, I think the compile failure may not affect a lot. I am 
glad to hear that higher Kafka client can be backward compatible well. ---- 
Vino yang Thanks. On 2018-04-04 23:21 , Christofer Dutz Wrote: Hi Vino, Yeah 
... but I did it without an ASF header ... that's why the build was failing for 
23 days ( (I am really ashamed about that) I tried updating the two Kafka 
dependencies to the 1.1.0 verison (and to Scala 2.12) and that worked without 
any noticeable problems. Chris Am 04.04.18, 13:38 schrieb "vino yang" 
<yanghua1...@gmail.com>: Hi Chris, I rechecked the old mails between you and 
me. I misunderstand your message. I thought you will create the annotation. In 
fact, you have created the annotation. I will do this work soon, hold on. Vino 
yang. Thanks. 2018-04-04 19:32 GMT+08:00 vino yang <yanghua1...@gmail.com>: > 
Hi Chris, > > I have not done this. And I would upgrade it soon. > > Vino yang 
> Thanks! > > 2018-04-04 19:23 GMT+08:00 Christofer Dutz 
<christofer.d...@c-ware.de>: > >> Hi, >> >> so I updated the libs locally, 
built and re-ran the example with this >> version and it now worked without any 
problems. >> >> Chris >> >> >> >> Am 04.04.18, 12:58 schrieb "Christofer Dutz" 
<christofer.d...@c-ware.de >> >: >> >>     Hi all, >> >>     reporting back 
from my easter holidays :-) >> >>     Today I had to help a customer with 
getting a POC working that uses >> PLC4X and Edgent. Unfortunately it seems 
that in order to use the kafka >> connector I can only use 0.x versions of 
Kafka. When connecting to 1.x >> versions I get stack-overflows and OutOfMemory 
errors. I did a quick test >> with updating the kafaka libs from the ancient 
0.8.2.2 to 1.1.0 seemed to >> not break anything ... I'll do some local tests 
with an updated Kafka >> client. >> >>     @vino yang ... have you been working 
on adding the Annotations to the >> client? >> >>     @all others ... does 
anyone have objections to updating the kafka >> client libs to 1.1.0? It 
shouldn't break anything as it should be backward >> compatible. As we are 
currently not using anything above the API level of >> 0.8.2 there should also 
not be any Exceptions (I don't know of any removed >> things, which could be a 
problem). >> >>     Chris >> >> >> >>     Am 20.03.18, 10:33 schrieb 
"Christofer Dutz" < >> christofer.d...@c-ware.de>: >> >>         Ok, >> >>      
   So I just added a new Annotation type to the Kafka module. >> >>         
org.apache.edgent.connectors.kafka.annotations.KafkaVersion >> >>         It 
has a fromVersion and a toVersion attribute. Both should be >> optional so just 
adding the annotation would have no effect (besides a few >> additional CPU 
operations). The annotation can be applied to methods or >> classes (every 
method then inherits this). I hope that's ok, because >> implementing this on a 
parameter Level would make things extremely >> difficult. >> >>         @vino 
yang With this you should be able to provide Kafka version >> constraints to 
your code changes. Just tell me if something's missing or >> needs to be done 
differently >> >>         For now this annotation will have no effect as I 
haven't >> implemented the Aspect for doing the checks, but I'll start working 
on that >> as soon as you have annotated something. >> >>         Chris >> >>   
      Am 20.03.18, 10:11 schrieb "Christofer Dutz" < >> 
christofer.d...@c-ware.de>: >> >>             Ok ... maybe I should add the 
Annotation prior to continuing >> my work on the AWS connector ... >> >> >>     
        Chris >> >>             Am 04.03.18, 08:10 schrieb "vino yang" 
<yanghua1...@gmail.com >> >: >> >>                 The reason is that Kafka 
0.9+ provided a new consumer API >> which has more >>                 features 
and better performance. >> >>                 Just like Flink's implementation 
: >>                 https://github.com/apache/flin >> 
k/tree/master/flink-connectors. >> >>                 vinoyang >>               
  Thanks. >> >> >> >> >> >> >> >> >> >

Reply via email to