I was using the snapshot repo in this case, let me try building my own
version...

Maybe this is interesting:
mvn dependency:tree | grep 2.11
[INFO] |  \- org.apache.kafka:kafka_2.11:jar:0.8.2.2:compile
[INFO] |     +- org.scala-lang.modules:scala-xml_2.11:jar:1.0.2:compile
[INFO] |     +-
org.scala-lang.modules:scala-parser-combinators_2.11:jar:1.0.2:compile


Robert Metzger <rmetz...@apache.org> ezt írta (időpont: 2016. febr. 26., P,
10:56):

> Are you building 1.0-SNAPSHOT yourself or are you relying on the snapshot
> repository?
>
> We had issues in the past that jars in the snapshot repo were incorrect
>
> On Fri, Feb 26, 2016 at 10:45 AM, Gyula Fóra <gyula.f...@gmail.com> wrote:
>
>> I am not sure what is happening. I tried running against a Flink cluster
>> that is definitely running the correct Scala version (2.10) and I still got
>> the error. So it might be something with the pom.xml but we just don't see
>> how it is different from the correct one.
>>
>> Gyula
>>
>> Till Rohrmann <trohrm...@apache.org> ezt írta (időpont: 2016. febr. 26.,
>> P, 10:42):
>>
>>> Hi Gyula,
>>>
>>> could it be that you compiled against a different Scala version than the
>>> one you're using for running the job? This usually happens when you compile
>>> against 2.10 and let it run with version 2.11.
>>>
>>> Cheers,
>>> Till
>>>
>>> On Fri, Feb 26, 2016 at 10:09 AM, Gyula Fóra <gyula.f...@gmail.com>
>>> wrote:
>>>
>>>> Hey,
>>>>
>>>> For one of our jobs we ran into this issue. It's probably some
>>>> dependency issue but we cant figure it out as a very similar setup works
>>>> without issues for a different program.
>>>>
>>>> java.lang.NoSuchMethodError:
>>>> scala.Predef$.ArrowAssoc(Ljava/lang/Object;)Ljava/lang/Object;
>>>> at
>>>> kafka.consumer.FetchRequestAndResponseMetrics.<init>(FetchRequestAndResponseStats.scala:32)
>>>> at
>>>> kafka.consumer.FetchRequestAndResponseStats.<init>(FetchRequestAndResponseStats.scala:46)
>>>> at
>>>> kafka.consumer.FetchRequestAndResponseStatsRegistry$$anonfun$2.apply(FetchRequestAndResponseStats.scala:59)
>>>> at
>>>> kafka.consumer.FetchRequestAndResponseStatsRegistry$$anonfun$2.apply(FetchRequestAndResponseStats.scala:59)
>>>> at kafka.utils.Pool.getAndMaybePut(Pool.scala:61)
>>>> at
>>>> kafka.consumer.FetchRequestAndResponseStatsRegistry$.getFetchRequestAndResponseStats(FetchRequestAndResponseStats.scala:63)
>>>> at kafka.consumer.SimpleConsumer.<init>(SimpleConsumer.scala:39)
>>>> at kafka.javaapi.consumer.SimpleConsumer.<init>(SimpleConsumer.scala:34)
>>>> at
>>>> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer08.getPartitionsForTopic(FlinkKafkaConsumer08.java:518)
>>>> at
>>>> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer08.<init>(FlinkKafkaConsumer08.java:218)
>>>> at
>>>> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer08.<init>(FlinkKafkaConsumer08.java:193)
>>>> at com.king.deduplo.source.EventSource.readRawInput(EventSource.java:46)
>>>> at com.king.deduplo.DeduploProgram.main(DeduploProgram.java:33)
>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>> at
>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>>> at
>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>> at java.lang.reflect.Method.invoke(Method.java:497)
>>>> at
>>>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:505)
>>>> at
>>>> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:403)
>>>> at org.apache.flink.client.program.Client.runBlocking(Client.java:248)
>>>> at
>>>> org.apache.flink.client.CliFrontend.executeProgramBlocking(CliFrontend.java:866)
>>>> at org.apache.flink.client.CliFrontend.run(CliFrontend.java:333)
>>>> at
>>>> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:1189)
>>>> at org.apache.flink.client.CliFrontend.main(CliFrontend.java:1239)
>>>>
>>>> Any insights?
>>>>
>>>> Cheers,
>>>> Gyula
>>>>
>>>
>>>
>

Reply via email to