I was using the snapshot repo in this case, let me try building my own version...
Maybe this is interesting: mvn dependency:tree | grep 2.11 [INFO] | \- org.apache.kafka:kafka_2.11:jar:0.8.2.2:compile [INFO] | +- org.scala-lang.modules:scala-xml_2.11:jar:1.0.2:compile [INFO] | +- org.scala-lang.modules:scala-parser-combinators_2.11:jar:1.0.2:compile Robert Metzger <rmetz...@apache.org> ezt írta (időpont: 2016. febr. 26., P, 10:56): > Are you building 1.0-SNAPSHOT yourself or are you relying on the snapshot > repository? > > We had issues in the past that jars in the snapshot repo were incorrect > > On Fri, Feb 26, 2016 at 10:45 AM, Gyula Fóra <gyula.f...@gmail.com> wrote: > >> I am not sure what is happening. I tried running against a Flink cluster >> that is definitely running the correct Scala version (2.10) and I still got >> the error. So it might be something with the pom.xml but we just don't see >> how it is different from the correct one. >> >> Gyula >> >> Till Rohrmann <trohrm...@apache.org> ezt írta (időpont: 2016. febr. 26., >> P, 10:42): >> >>> Hi Gyula, >>> >>> could it be that you compiled against a different Scala version than the >>> one you're using for running the job? This usually happens when you compile >>> against 2.10 and let it run with version 2.11. >>> >>> Cheers, >>> Till >>> >>> On Fri, Feb 26, 2016 at 10:09 AM, Gyula Fóra <gyula.f...@gmail.com> >>> wrote: >>> >>>> Hey, >>>> >>>> For one of our jobs we ran into this issue. It's probably some >>>> dependency issue but we cant figure it out as a very similar setup works >>>> without issues for a different program. >>>> >>>> java.lang.NoSuchMethodError: >>>> scala.Predef$.ArrowAssoc(Ljava/lang/Object;)Ljava/lang/Object; >>>> at >>>> kafka.consumer.FetchRequestAndResponseMetrics.<init>(FetchRequestAndResponseStats.scala:32) >>>> at >>>> kafka.consumer.FetchRequestAndResponseStats.<init>(FetchRequestAndResponseStats.scala:46) >>>> at >>>> kafka.consumer.FetchRequestAndResponseStatsRegistry$$anonfun$2.apply(FetchRequestAndResponseStats.scala:59) >>>> at >>>> kafka.consumer.FetchRequestAndResponseStatsRegistry$$anonfun$2.apply(FetchRequestAndResponseStats.scala:59) >>>> at kafka.utils.Pool.getAndMaybePut(Pool.scala:61) >>>> at >>>> kafka.consumer.FetchRequestAndResponseStatsRegistry$.getFetchRequestAndResponseStats(FetchRequestAndResponseStats.scala:63) >>>> at kafka.consumer.SimpleConsumer.<init>(SimpleConsumer.scala:39) >>>> at kafka.javaapi.consumer.SimpleConsumer.<init>(SimpleConsumer.scala:34) >>>> at >>>> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer08.getPartitionsForTopic(FlinkKafkaConsumer08.java:518) >>>> at >>>> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer08.<init>(FlinkKafkaConsumer08.java:218) >>>> at >>>> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer08.<init>(FlinkKafkaConsumer08.java:193) >>>> at com.king.deduplo.source.EventSource.readRawInput(EventSource.java:46) >>>> at com.king.deduplo.DeduploProgram.main(DeduploProgram.java:33) >>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>> at >>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >>>> at >>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>> at java.lang.reflect.Method.invoke(Method.java:497) >>>> at >>>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:505) >>>> at >>>> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:403) >>>> at org.apache.flink.client.program.Client.runBlocking(Client.java:248) >>>> at >>>> org.apache.flink.client.CliFrontend.executeProgramBlocking(CliFrontend.java:866) >>>> at org.apache.flink.client.CliFrontend.run(CliFrontend.java:333) >>>> at >>>> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:1189) >>>> at org.apache.flink.client.CliFrontend.main(CliFrontend.java:1239) >>>> >>>> Any insights? >>>> >>>> Cheers, >>>> Gyula >>>> >>> >>> >