Hi all:
最近用mysql-cdc的方式,使用Flink-sql整合数据,table的部分配置如下
'debezium.min.row.count.to.stream.results'='1000',
'scan.startup.mode'='specific-offset',
'scan.startup.specific-offset.file'='mybinlog.000029',
'scan.startup.specific-offset.pos'='542607677'
在上述稳定运行接分钟之后,异常抛出了以下错误:
org.apache.kafka.connect.errors.ConnectException: The connector is trying to 
read binlog starting at GTIDs 
36b8bd0f-a435-11eb-9962-b4055d9ecb74:21820662-21926801,d713fc3c-afc6-11eb-a7b8-b4055d9ec5e6:185910-192349
 and binlog file 'mybinlog.000029', pos=808323397, skipping 6 events plus 1 
rows, but this is no longer available on the server. Reconfigure the connector 
to use a snapshot when needed.
at 
io.debezium.connector.mysql.MySqlConnectorTask.start(MySqlConnectorTask.java:133)
at io.debezium.connector.common.BaseSourceTask.start(BaseSourceTask.java:106)
at io.debezium.embedded.EmbeddedEngine.run(EmbeddedEngine.java:758)
at 
io.debezium.embedded.ConvertingEngineBuilder$2.run(ConvertingEngineBuilder.java:171)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1147)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:622)
at java.lang.Thread.run(Thread.java:834)
请问是什么情况下导致这个问题。

回复