This problem is kind of the other way around, as our 2.10 build has 2.11
dependencies pulled in by kafka. But let's see what happens. :)

Gyula

Till Rohrmann <trohrm...@apache.org> ezt írta (időpont: 2016. márc. 3., Cs,
9:46):

> Hi Gyula,
>
> we discovered yesterday that our build process for Scala 2.11 is broken
> for the Kafka connector. The reason is that a property value is not
> properly resolved and thus pulls in the 2.10 Kafka dependencies. Max
> already opened a PR to fix this problem. I hope this will also solve your
> problem.
>
> Cheers,
> Till
>
> On Mar 3, 2016 9:36 AM, "Gyula Fóra" <gyula.f...@gmail.com> wrote:
> >
> > Hey,
> >
> > Do we have any idea why this is happening in the snapshot repo? We have
> run into the same issue again...
> >
> > Cheers,
> > Gyula
> >
> > Gyula Fóra <gyula.f...@gmail.com> ezt írta (időpont: 2016. febr. 26.,
> P, 11:17):
> >>
> >> Thanks Robert, so apparently the snapshot version was screwed up
> somehow and included the 2.11 dependencies.
> >>
> >> Now it works.
> >>
> >> Cheers,
> >> Gyula
> >>
> >> Gyula Fóra <gyula.f...@gmail.com> ezt írta (időpont: 2016. febr. 26.,
> P, 11:09):
> >>>
> >>> That actually seemed to be the issue, not that I compiled my own
> version it doesnt have these wrond jars in the dependency tree...
> >>>
> >>> Gyula Fóra <gyula.f...@gmail.com> ezt írta (időpont: 2016. febr. 26.,
> P, 11:01):
> >>>>
> >>>> I was using the snapshot repo in this case, let me try building my
> own version...
> >>>>
> >>>> Maybe this is interesting:
> >>>> mvn dependency:tree | grep 2.11
> >>>> [INFO] |  \- org.apache.kafka:kafka_2.11:jar:0.8.2.2:compile
> >>>> [INFO] |     +-
> org.scala-lang.modules:scala-xml_2.11:jar:1.0.2:compile
> >>>> [INFO] |     +-
> org.scala-lang.modules:scala-parser-combinators_2.11:jar:1.0.2:compile
> >>>>
> >>>>
> >>>> Robert Metzger <rmetz...@apache.org> ezt írta (időpont: 2016. febr.
> 26., P, 10:56):
> >>>>>
> >>>>> Are you building 1.0-SNAPSHOT yourself or are you relying on the
> snapshot repository?
> >>>>>
> >>>>> We had issues in the past that jars in the snapshot repo were
> incorrect
> >>>>>
> >>>>> On Fri, Feb 26, 2016 at 10:45 AM, Gyula Fóra <gyula.f...@gmail.com>
> wrote:
> >>>>>>
> >>>>>> I am not sure what is happening. I tried running against a Flink
> cluster that is definitely running the correct Scala version (2.10) and I
> still got the error. So it might be something with the pom.xml but we just
> don't see how it is different from the correct one.
> >>>>>>
> >>>>>> Gyula
> >>>>>>
> >>>>>> Till Rohrmann <trohrm...@apache.org> ezt írta (időpont: 2016.
> febr. 26., P, 10:42):
> >>>>>>>
> >>>>>>> Hi Gyula,
> >>>>>>>
> >>>>>>> could it be that you compiled against a different Scala version
> than the one you're using for running the job? This usually happens when
> you compile against 2.10 and let it run with version 2.11.
> >>>>>>>
> >>>>>>> Cheers,
> >>>>>>> Till
> >>>>>>>
> >>>>>>> On Fri, Feb 26, 2016 at 10:09 AM, Gyula Fóra <gyula.f...@gmail.com>
> wrote:
> >>>>>>>>
> >>>>>>>> Hey,
> >>>>>>>>
> >>>>>>>> For one of our jobs we ran into this issue. It's probably some
> dependency issue but we cant figure it out as a very similar setup works
> without issues for a different program.
> >>>>>>>>
> >>>>>>>> java.lang.NoSuchMethodError:
> scala.Predef$.ArrowAssoc(Ljava/lang/Object;)Ljava/lang/Object;
> >>>>>>>> at
> kafka.consumer.FetchRequestAndResponseMetrics.<init>(FetchRequestAndResponseStats.scala:32)
> >>>>>>>> at
> kafka.consumer.FetchRequestAndResponseStats.<init>(FetchRequestAndResponseStats.scala:46)
> >>>>>>>> at
> kafka.consumer.FetchRequestAndResponseStatsRegistry$$anonfun$2.apply(FetchRequestAndResponseStats.scala:59)
> >>>>>>>> at
> kafka.consumer.FetchRequestAndResponseStatsRegistry$$anonfun$2.apply(FetchRequestAndResponseStats.scala:59)
> >>>>>>>> at kafka.utils.Pool.getAndMaybePut(Pool.scala:61)
> >>>>>>>> at
> kafka.consumer.FetchRequestAndResponseStatsRegistry$.getFetchRequestAndResponseStats(FetchRequestAndResponseStats.scala:63)
> >>>>>>>> at kafka.consumer.SimpleConsumer.<init>(SimpleConsumer.scala:39)
> >>>>>>>> at
> kafka.javaapi.consumer.SimpleConsumer.<init>(SimpleConsumer.scala:34)
> >>>>>>>> at
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer08.getPartitionsForTopic(FlinkKafkaConsumer08.java:518)
> >>>>>>>> at
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer08.<init>(FlinkKafkaConsumer08.java:218)
> >>>>>>>> at
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer08.<init>(FlinkKafkaConsumer08.java:193)
> >>>>>>>> at
> com.king.deduplo.source.EventSource.readRawInput(EventSource.java:46)
> >>>>>>>> at com.king.deduplo.DeduploProgram.main(DeduploProgram.java:33)
> >>>>>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> >>>>>>>> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> >>>>>>>> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> >>>>>>>> at java.lang.reflect.Method.invoke(Method.java:497)
> >>>>>>>> at
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:505)
> >>>>>>>> at
> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:403)
> >>>>>>>> at
> org.apache.flink.client.program.Client.runBlocking(Client.java:248)
> >>>>>>>> at
> org.apache.flink.client.CliFrontend.executeProgramBlocking(CliFrontend.java:866)
> >>>>>>>> at org.apache.flink.client.CliFrontend.run(CliFrontend.java:333)
> >>>>>>>> at
> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:1189)
> >>>>>>>> at org.apache.flink.client.CliFrontend.main(CliFrontend.java:1239)
> >>>>>>>>
> >>>>>>>> Any insights?
> >>>>>>>>
> >>>>>>>> Cheers,
> >>>>>>>> Gyula
> >>>>>>>
> >>>>>>>
> >>>>>
>

Reply via email to