pengyusong created FLINK-27343: ---------------------------------- Summary: flink jdbc sink with default param will lead buffer records in one batch unorder Key: FLINK-27343 URL: https://issues.apache.org/jira/browse/FLINK-27343 Project: Flink Issue Type: Improvement Components: Connectors / JDBC Affects Versions: 1.13.6 Environment: flink 1.13.6
kafka postgres jdbc sink Reporter: pengyusong * situation one when i use flink sql kafka connector re-consume a topic, the topic already has many messages. jdbc sink param with default. kafka topic is a compact topic, which contents is a mysql table cdc events. there some records with same key in one batch, buffer within one batch, finnaly sink to postgres with unorder, later record in the buffer batch are executed first. this will lead to the older message in kafka deal with after the newer message, the results are inconsistent with kafka message orders. * situation two If i set h5. sink.buffer-flush.interval = 0 h5. sink.buffer-flush.max-rows = 1 the result are inconsistent with kafka message orders. So, I have a suspicion that the order in jdbc buffer execute is non-deterministic, lead to result in jdbc unordered. -- This message was sent by Atlassian Jira (v8.20.7#820007)