Thanks Robert, so apparently the snapshot version was screwed up somehow and included the 2.11 dependencies.
Now it works. Cheers, Gyula Gyula Fóra <gyula.f...@gmail.com> ezt írta (időpont: 2016. febr. 26., P, 11:09): > That actually seemed to be the issue, not that I compiled my own version > it doesnt have these wrond jars in the dependency tree... > > Gyula Fóra <gyula.f...@gmail.com> ezt írta (időpont: 2016. febr. 26., P, > 11:01): > >> I was using the snapshot repo in this case, let me try building my own >> version... >> >> Maybe this is interesting: >> mvn dependency:tree | grep 2.11 >> [INFO] | \- org.apache.kafka:kafka_2.11:jar:0.8.2.2:compile >> [INFO] | +- org.scala-lang.modules:scala-xml_2.11:jar:1.0.2:compile >> [INFO] | +- >> org.scala-lang.modules:scala-parser-combinators_2.11:jar:1.0.2:compile >> >> >> Robert Metzger <rmetz...@apache.org> ezt írta (időpont: 2016. febr. 26., >> P, 10:56): >> >>> Are you building 1.0-SNAPSHOT yourself or are you relying on the >>> snapshot repository? >>> >>> We had issues in the past that jars in the snapshot repo were incorrect >>> >>> On Fri, Feb 26, 2016 at 10:45 AM, Gyula Fóra <gyula.f...@gmail.com> >>> wrote: >>> >>>> I am not sure what is happening. I tried running against a Flink >>>> cluster that is definitely running the correct Scala version (2.10) and I >>>> still got the error. So it might be something with the pom.xml but we just >>>> don't see how it is different from the correct one. >>>> >>>> Gyula >>>> >>>> Till Rohrmann <trohrm...@apache.org> ezt írta (időpont: 2016. febr. >>>> 26., P, 10:42): >>>> >>>>> Hi Gyula, >>>>> >>>>> could it be that you compiled against a different Scala version than >>>>> the one you're using for running the job? This usually happens when you >>>>> compile against 2.10 and let it run with version 2.11. >>>>> >>>>> Cheers, >>>>> Till >>>>> >>>>> On Fri, Feb 26, 2016 at 10:09 AM, Gyula Fóra <gyula.f...@gmail.com> >>>>> wrote: >>>>> >>>>>> Hey, >>>>>> >>>>>> For one of our jobs we ran into this issue. It's probably some >>>>>> dependency issue but we cant figure it out as a very similar setup works >>>>>> without issues for a different program. >>>>>> >>>>>> java.lang.NoSuchMethodError: >>>>>> scala.Predef$.ArrowAssoc(Ljava/lang/Object;)Ljava/lang/Object; >>>>>> at >>>>>> kafka.consumer.FetchRequestAndResponseMetrics.<init>(FetchRequestAndResponseStats.scala:32) >>>>>> at >>>>>> kafka.consumer.FetchRequestAndResponseStats.<init>(FetchRequestAndResponseStats.scala:46) >>>>>> at >>>>>> kafka.consumer.FetchRequestAndResponseStatsRegistry$$anonfun$2.apply(FetchRequestAndResponseStats.scala:59) >>>>>> at >>>>>> kafka.consumer.FetchRequestAndResponseStatsRegistry$$anonfun$2.apply(FetchRequestAndResponseStats.scala:59) >>>>>> at kafka.utils.Pool.getAndMaybePut(Pool.scala:61) >>>>>> at >>>>>> kafka.consumer.FetchRequestAndResponseStatsRegistry$.getFetchRequestAndResponseStats(FetchRequestAndResponseStats.scala:63) >>>>>> at kafka.consumer.SimpleConsumer.<init>(SimpleConsumer.scala:39) >>>>>> at >>>>>> kafka.javaapi.consumer.SimpleConsumer.<init>(SimpleConsumer.scala:34) >>>>>> at >>>>>> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer08.getPartitionsForTopic(FlinkKafkaConsumer08.java:518) >>>>>> at >>>>>> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer08.<init>(FlinkKafkaConsumer08.java:218) >>>>>> at >>>>>> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer08.<init>(FlinkKafkaConsumer08.java:193) >>>>>> at >>>>>> com.king.deduplo.source.EventSource.readRawInput(EventSource.java:46) >>>>>> at com.king.deduplo.DeduploProgram.main(DeduploProgram.java:33) >>>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>>>> at >>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >>>>>> at >>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>>>> at java.lang.reflect.Method.invoke(Method.java:497) >>>>>> at >>>>>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:505) >>>>>> at >>>>>> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:403) >>>>>> at org.apache.flink.client.program.Client.runBlocking(Client.java:248) >>>>>> at >>>>>> org.apache.flink.client.CliFrontend.executeProgramBlocking(CliFrontend.java:866) >>>>>> at org.apache.flink.client.CliFrontend.run(CliFrontend.java:333) >>>>>> at >>>>>> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:1189) >>>>>> at org.apache.flink.client.CliFrontend.main(CliFrontend.java:1239) >>>>>> >>>>>> Any insights? >>>>>> >>>>>> Cheers, >>>>>> Gyula >>>>>> >>>>> >>>>> >>>