flink-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Gyula Fóra <gyula.f...@gmail.com>
Subject Re: Kafka issue
Date Fri, 26 Feb 2016 10:09:28 GMT
That actually seemed to be the issue, not that I compiled my own version it
doesnt have these wrond jars in the dependency tree...

Gyula Fóra <gyula.fora@gmail.com> ezt írta (időpont: 2016. febr. 26., P,
11:01):

> I was using the snapshot repo in this case, let me try building my own
> version...
>
> Maybe this is interesting:
> mvn dependency:tree | grep 2.11
> [INFO] |  \- org.apache.kafka:kafka_2.11:jar:0.8.2.2:compile
> [INFO] |     +- org.scala-lang.modules:scala-xml_2.11:jar:1.0.2:compile
> [INFO] |     +-
> org.scala-lang.modules:scala-parser-combinators_2.11:jar:1.0.2:compile
>
>
> Robert Metzger <rmetzger@apache.org> ezt írta (időpont: 2016. febr. 26.,
> P, 10:56):
>
>> Are you building 1.0-SNAPSHOT yourself or are you relying on the snapshot
>> repository?
>>
>> We had issues in the past that jars in the snapshot repo were incorrect
>>
>> On Fri, Feb 26, 2016 at 10:45 AM, Gyula Fóra <gyula.fora@gmail.com>
>> wrote:
>>
>>> I am not sure what is happening. I tried running against a Flink cluster
>>> that is definitely running the correct Scala version (2.10) and I still got
>>> the error. So it might be something with the pom.xml but we just don't see
>>> how it is different from the correct one.
>>>
>>> Gyula
>>>
>>> Till Rohrmann <trohrmann@apache.org> ezt írta (időpont: 2016. febr.
>>> 26., P, 10:42):
>>>
>>>> Hi Gyula,
>>>>
>>>> could it be that you compiled against a different Scala version than
>>>> the one you're using for running the job? This usually happens when you
>>>> compile against 2.10 and let it run with version 2.11.
>>>>
>>>> Cheers,
>>>> Till
>>>>
>>>> On Fri, Feb 26, 2016 at 10:09 AM, Gyula Fóra <gyula.fora@gmail.com>
>>>> wrote:
>>>>
>>>>> Hey,
>>>>>
>>>>> For one of our jobs we ran into this issue. It's probably some
>>>>> dependency issue but we cant figure it out as a very similar setup works
>>>>> without issues for a different program.
>>>>>
>>>>> java.lang.NoSuchMethodError:
>>>>> scala.Predef$.ArrowAssoc(Ljava/lang/Object;)Ljava/lang/Object;
>>>>> at
>>>>> kafka.consumer.FetchRequestAndResponseMetrics.<init>(FetchRequestAndResponseStats.scala:32)
>>>>> at
>>>>> kafka.consumer.FetchRequestAndResponseStats.<init>(FetchRequestAndResponseStats.scala:46)
>>>>> at
>>>>> kafka.consumer.FetchRequestAndResponseStatsRegistry$$anonfun$2.apply(FetchRequestAndResponseStats.scala:59)
>>>>> at
>>>>> kafka.consumer.FetchRequestAndResponseStatsRegistry$$anonfun$2.apply(FetchRequestAndResponseStats.scala:59)
>>>>> at kafka.utils.Pool.getAndMaybePut(Pool.scala:61)
>>>>> at
>>>>> kafka.consumer.FetchRequestAndResponseStatsRegistry$.getFetchRequestAndResponseStats(FetchRequestAndResponseStats.scala:63)
>>>>> at kafka.consumer.SimpleConsumer.<init>(SimpleConsumer.scala:39)
>>>>> at
>>>>> kafka.javaapi.consumer.SimpleConsumer.<init>(SimpleConsumer.scala:34)
>>>>> at
>>>>> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer08.getPartitionsForTopic(FlinkKafkaConsumer08.java:518)
>>>>> at
>>>>> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer08.<init>(FlinkKafkaConsumer08.java:218)
>>>>> at
>>>>> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer08.<init>(FlinkKafkaConsumer08.java:193)
>>>>> at
>>>>> com.king.deduplo.source.EventSource.readRawInput(EventSource.java:46)
>>>>> at com.king.deduplo.DeduploProgram.main(DeduploProgram.java:33)
>>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>> at
>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>>>> at
>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>>> at java.lang.reflect.Method.invoke(Method.java:497)
>>>>> at
>>>>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:505)
>>>>> at
>>>>> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:403)
>>>>> at org.apache.flink.client.program.Client.runBlocking(Client.java:248)
>>>>> at
>>>>> org.apache.flink.client.CliFrontend.executeProgramBlocking(CliFrontend.java:866)
>>>>> at org.apache.flink.client.CliFrontend.run(CliFrontend.java:333)
>>>>> at
>>>>> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:1189)
>>>>> at org.apache.flink.client.CliFrontend.main(CliFrontend.java:1239)
>>>>>
>>>>> Any insights?
>>>>>
>>>>> Cheers,
>>>>> Gyula
>>>>>
>>>>
>>>>
>>

Mime
View raw message