Hi,

Depending how you configured your FlinkKafkaSource, but you can make the
source to commit consumed offsets back to Kafka. So one way to examine
them, would be to check those offsets in Kafka (I don't know how, but I'm
pretty sure there is a way to do it).

Secondly, if you want to examine Flink's checkpoint state you can use State
Processor API to do that [1]. As far as I know you could hook up your
checkpointed data to Table API/SQL and use SQL to query/analyse the state.

Best
Piotrek

[1]
https://ci.apache.org/projects/flink/flink-docs-stable/dev/libs/state_processor_api.html

śr., 14 kwi 2021 o 11:25 bat man <tintin0...@gmail.com> napisał(a):

> Hi All,
>
> Is there any way I can inspect/query the checkpointed data. Scenario is
> like this -
>
> We have a high volume of data coming in the data stream pipeline for which
> kafka is source, in case if fails bcoz of bad data I want to analyse the
> data which caused the issue. It could be that some data source starts
> sending bad data so I want to go in kafka to that particular offset and do
> some analysis before I start the job with checkpointed data.
>
> Can anyone suggest how this can be achieved.
>
> Thanks,
> Hemant
>
>
>

Reply via email to