Hi,

Thanks for your replies.
Yes Terry. You are right. I can try to create a custom source. 
But perhaps, according to my use case, I figured out I can use a technical 
field in my data. This is a timestamp and I think I just have to ignore late 
events with watermarks or later in the pipeline according to metadata stored in 
the Flink state. I test it now...
Thx

David

On 2020/01/03 15:44:08, Chesnay Schepler <ches...@apache.org> wrote: 
> Are you asking how to detect from within the job whether the dump is 
> complete, or how to combine these 2 jobs?
> 
> If you had a way to notice whether the dump is complete, then I would 
> suggest to create a custom source that wraps 2 kafka sources, and switch 
> between them at will based on your conditions.
> 
> 
> On 03/01/2020 03:53, Terry Wang wrote:
> > Hi,
> >
> > I’d like to share my opinion here. It seems that you need adjust the Kafka 
> > consumer to have communication each other. When your begin the dump 
> > process, you need to notify another CDC-topic consumer to wait idle.
> >
> >
> > Best,
> > Terry Wang
> >
> >
> >
> >> 2020年1月2日 16:49,David Morin <morin.david....@gmail.com> 写道:
> >>
> >> Hi,
> >>
> >> Is there a way to stop temporarily to consume one kafka source in 
> >> streaming mode ?
> >> Use case: I have to consume 2 topics but in fact one of them is more 
> >> prioritized.
> >> One of this topic is dedicated to ingest data from db (change data 
> >> capture) and one of them is dedicated to make a synchronization (a dump 
> >> i.e. a SELECT ... from db). At the moment the last one is performed by one 
> >> Flink job and we start this one after stop the previous one (CDC) manually
> >> I want to merge these 2 modes and automatically stop consumption of the 
> >> topic dedicated to the CDC mode when a dump is done.
> >> How to handle that with Flink in a streaming way ? backpressure ? ...
> >> Thx in advance for your insights
> >>
> >> David
> >
> 
> 

Reply via email to