Hi Nick,

all Flink dependencies are only compatible with the same major version.

You can workaround it by checking out the code [1] and manually set the
dependency of the respective module to your flink-core version and revert
all changes that are not compiling. But there is no guarantee that this
will ultimately work as you are pretty much backporting some changes to the
old version.

[1] https://github.com/AHeise/flink

On Thu, May 7, 2020 at 2:35 AM Nick Bendtner <buggi...@gmail.com> wrote:

> Hi guys,
> I am using flink 1.7.2 version. I have to deserialize data from kafka
> into  consumer records therefore I decided to update
> the flink-connector-kafka to 1.9.1 which provides support for consumer
> record. We use child first class loading. However it seems like I have
> compatibility issue as I get this exception, *Exception in thread "main"
> java.lang.NoClassDefFoundError:
> org/apache/flink/api/common/ExecutionConfig$ClosureCleanerLevel *.
>
> Any tricks to make this work without changing the version of flink-core ?
>
>
> Best,
> Nick.
>
>

-- 

Arvid Heise | Senior Java Developer

<https://www.ververica.com/>

Follow us @VervericaData

--

Join Flink Forward <https://flink-forward.org/> - The Apache Flink
Conference

Stream Processing | Event Driven | Real Time

--

Ververica GmbH | Invalidenstrasse 115, 10115 Berlin, Germany

--
Ververica GmbH
Registered at Amtsgericht Charlottenburg: HRB 158244 B
Managing Directors: Timothy Alexander Steinert, Yip Park Tung Jason, Ji
(Toni) Cheng

Reply via email to