Hi Caizhi Weng, Thanks for your input. I would explain the requirement in little more detail. Flink pipeline will be running forever (until some issue happens and we would need to restart) so It will continuously monitor if a new file comes to the *input *folder or not. In this case will your suggestion work?
Thanks & Regards, Samir Vasani On Fri, Jul 23, 2021 at 1:07 PM Caizhi Weng <tsreape...@gmail.com> wrote: > Hi! > > JobListener#onJobExecuted might help, if your job is not a forever-running > streaming job. See > https://ci.apache.org/projects/flink/flink-docs-master/api/java/org/apache/flink/core/execution/JobListener.html > > Samir Vasani <samirvas...@gmail.com> 于2021年7月23日周五 下午3:22写道: > >> Hi, >> >> I am a new bee to flink and facing some challenges to solve below use case >> >> Use Case description: >> >> I will receive a csv file with a timestamp on every single day in some >> folder say *input*.The file format would be >> *file_name_dd-mm-yy-hh-mm-ss.csv*. >> >> Now my flink pipeline will read this csv file in a row by row fashion and >> it will be written to my Kafka topic. >> >> Once the pipeline reads the entire file then this file needs to be moved >> to another folder say *historic* so that i can keep *input * folder >> empty for the new file. >> >> I googled a lot but did not find anything so can you guide me to achieve >> this. >> >> Let me know if anything else is required. >> >> >> Samir Vasani >> >