Maximilian Michels created FLINK-38370:
------------------------------------------
Summary: CommitterOperator does not checkpoint all pending
committables
Key: FLINK-38370
URL: https://issues.apache.org/jira/browse/FLINK-38370
Project: Flink
Issue Type: Bug
Components: Runtime / Checkpointing
Affects Versions: 1.20.2, 1.19.3
Reporter: Maximilian Michels
Fix For: 1.19.4, 1.20.3
What is the issue?
==============
In a nutshell, the CommiterOperator (V2 Sink) does not commit all pending
committables for batch pipelines anymore, but only for current_checkpoint_id+1.
This is caused specifically by this line:
https://github.com/apache/flink/blob/5ad464d8156b5094e0aba7712e4aa22b9f44e84d/flink-runtime/src/main/java/org/apache/flink/streaming/runtime/operators/sink/CommitterOperator.java#L154
In Iceberg, we use `Long.MAX_VALUE` for the final checkpoint id:
https://github.com/apache/iceberg/blob/ee90c10e39cec0ccceb9425e03a3e0b5690daf3b/flink/v2.0/flink/src/main/java/org/apache/iceberg/flink/sink/IcebergWriteAggregator.java#L79.
This value was generally used by other connectors in Flink. This behavior was
changed in https://issues.apache.org/jira/browse/FLINK-37605 /
https://github.com/apache/flink/pull/26433.
Since committables using `Long.MAX_VALUE` won't be committed, the Iceberg table
won't be updated on shutdown, which means it will not contain any data! That
results in potential data loss.
How to fix this issue?
================
For Flink 2.1.0+, Iceberg can generate committables with `checkpoint =
last_checkpoint_id + 1`. We already took this approach in
https://github.com/apache/iceberg/pull/13714.
That is fine for the upcoming Iceberg version for Flink 2.1.0, but users who
use the current Iceberg version with 1.19.2 or 1.20.1 will be up for a surprise
when they upgrade to the latest patch release 1.19.3 or 1.20.2. We should
restore the old behavior for the Flink 1.19.4 and 1.20.3 releases.
--
This message was sent by Atlassian Jira
(v8.20.10#820010)