This problem is kind of the other way around, as our 2.10 build has 2.11 dependencies pulled in by kafka. But let's see what happens. :)
Gyula Till Rohrmann <trohrm...@apache.org> ezt írta (időpont: 2016. márc. 3., Cs, 9:46): > Hi Gyula, > > we discovered yesterday that our build process for Scala 2.11 is broken > for the Kafka connector. The reason is that a property value is not > properly resolved and thus pulls in the 2.10 Kafka dependencies. Max > already opened a PR to fix this problem. I hope this will also solve your > problem. > > Cheers, > Till > > On Mar 3, 2016 9:36 AM, "Gyula Fóra" <gyula.f...@gmail.com> wrote: > > > > Hey, > > > > Do we have any idea why this is happening in the snapshot repo? We have > run into the same issue again... > > > > Cheers, > > Gyula > > > > Gyula Fóra <gyula.f...@gmail.com> ezt írta (időpont: 2016. febr. 26., > P, 11:17): > >> > >> Thanks Robert, so apparently the snapshot version was screwed up > somehow and included the 2.11 dependencies. > >> > >> Now it works. > >> > >> Cheers, > >> Gyula > >> > >> Gyula Fóra <gyula.f...@gmail.com> ezt írta (időpont: 2016. febr. 26., > P, 11:09): > >>> > >>> That actually seemed to be the issue, not that I compiled my own > version it doesnt have these wrond jars in the dependency tree... > >>> > >>> Gyula Fóra <gyula.f...@gmail.com> ezt írta (időpont: 2016. febr. 26., > P, 11:01): > >>>> > >>>> I was using the snapshot repo in this case, let me try building my > own version... > >>>> > >>>> Maybe this is interesting: > >>>> mvn dependency:tree | grep 2.11 > >>>> [INFO] | \- org.apache.kafka:kafka_2.11:jar:0.8.2.2:compile > >>>> [INFO] | +- > org.scala-lang.modules:scala-xml_2.11:jar:1.0.2:compile > >>>> [INFO] | +- > org.scala-lang.modules:scala-parser-combinators_2.11:jar:1.0.2:compile > >>>> > >>>> > >>>> Robert Metzger <rmetz...@apache.org> ezt írta (időpont: 2016. febr. > 26., P, 10:56): > >>>>> > >>>>> Are you building 1.0-SNAPSHOT yourself or are you relying on the > snapshot repository? > >>>>> > >>>>> We had issues in the past that jars in the snapshot repo were > incorrect > >>>>> > >>>>> On Fri, Feb 26, 2016 at 10:45 AM, Gyula Fóra <gyula.f...@gmail.com> > wrote: > >>>>>> > >>>>>> I am not sure what is happening. I tried running against a Flink > cluster that is definitely running the correct Scala version (2.10) and I > still got the error. So it might be something with the pom.xml but we just > don't see how it is different from the correct one. > >>>>>> > >>>>>> Gyula > >>>>>> > >>>>>> Till Rohrmann <trohrm...@apache.org> ezt írta (időpont: 2016. > febr. 26., P, 10:42): > >>>>>>> > >>>>>>> Hi Gyula, > >>>>>>> > >>>>>>> could it be that you compiled against a different Scala version > than the one you're using for running the job? This usually happens when > you compile against 2.10 and let it run with version 2.11. > >>>>>>> > >>>>>>> Cheers, > >>>>>>> Till > >>>>>>> > >>>>>>> On Fri, Feb 26, 2016 at 10:09 AM, Gyula Fóra <gyula.f...@gmail.com> > wrote: > >>>>>>>> > >>>>>>>> Hey, > >>>>>>>> > >>>>>>>> For one of our jobs we ran into this issue. It's probably some > dependency issue but we cant figure it out as a very similar setup works > without issues for a different program. > >>>>>>>> > >>>>>>>> java.lang.NoSuchMethodError: > scala.Predef$.ArrowAssoc(Ljava/lang/Object;)Ljava/lang/Object; > >>>>>>>> at > kafka.consumer.FetchRequestAndResponseMetrics.<init>(FetchRequestAndResponseStats.scala:32) > >>>>>>>> at > kafka.consumer.FetchRequestAndResponseStats.<init>(FetchRequestAndResponseStats.scala:46) > >>>>>>>> at > kafka.consumer.FetchRequestAndResponseStatsRegistry$$anonfun$2.apply(FetchRequestAndResponseStats.scala:59) > >>>>>>>> at > kafka.consumer.FetchRequestAndResponseStatsRegistry$$anonfun$2.apply(FetchRequestAndResponseStats.scala:59) > >>>>>>>> at kafka.utils.Pool.getAndMaybePut(Pool.scala:61) > >>>>>>>> at > kafka.consumer.FetchRequestAndResponseStatsRegistry$.getFetchRequestAndResponseStats(FetchRequestAndResponseStats.scala:63) > >>>>>>>> at kafka.consumer.SimpleConsumer.<init>(SimpleConsumer.scala:39) > >>>>>>>> at > kafka.javaapi.consumer.SimpleConsumer.<init>(SimpleConsumer.scala:34) > >>>>>>>> at > org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer08.getPartitionsForTopic(FlinkKafkaConsumer08.java:518) > >>>>>>>> at > org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer08.<init>(FlinkKafkaConsumer08.java:218) > >>>>>>>> at > org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer08.<init>(FlinkKafkaConsumer08.java:193) > >>>>>>>> at > com.king.deduplo.source.EventSource.readRawInput(EventSource.java:46) > >>>>>>>> at com.king.deduplo.DeduploProgram.main(DeduploProgram.java:33) > >>>>>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > >>>>>>>> at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > >>>>>>>> at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > >>>>>>>> at java.lang.reflect.Method.invoke(Method.java:497) > >>>>>>>> at > org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:505) > >>>>>>>> at > org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:403) > >>>>>>>> at > org.apache.flink.client.program.Client.runBlocking(Client.java:248) > >>>>>>>> at > org.apache.flink.client.CliFrontend.executeProgramBlocking(CliFrontend.java:866) > >>>>>>>> at org.apache.flink.client.CliFrontend.run(CliFrontend.java:333) > >>>>>>>> at > org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:1189) > >>>>>>>> at org.apache.flink.client.CliFrontend.main(CliFrontend.java:1239) > >>>>>>>> > >>>>>>>> Any insights? > >>>>>>>> > >>>>>>>> Cheers, > >>>>>>>> Gyula > >>>>>>> > >>>>>>> > >>>>> >