Jinzhong Li created FLINK-32953: ----------------------------------- Summary: [State TTL]resolve data correctness problem after ttl was changed Key: FLINK-32953 URL: https://issues.apache.org/jira/browse/FLINK-32953 Project: Flink Issue Type: Bug Components: Runtime / State Backends Reporter: Jinzhong Li
Because expired data is cleaned up in background on a best effort basis (hashmap use INCREMENTAL_CLEANUP strategy, rocksdb use ROCKSDB_COMPACTION_FILTER strategy), some expired state is often persisted into snapshots. In some scenarios, user changes the state ttl of the job and then restore job from the old state. If the user adjust the state ttl from a short value to a long value (eg, from 12 hours to 24 hours), some expired data that was not cleaned up will be alive after restore. Obviously this is unreasonable, and may break data regulatory requirements. Particularly, rocksdb stateBackend may cause data correctness problems due to level compaction in this case.(eg. One key has two versions at level-1 and level-2,both of which are ttl expired. Then level-1 version is cleaned up by compaction, and level-2 version isn't. If we adjust state ttl and restart job, the incorrect data of level-2 will become valid after restore) To solve this problem, I think we can 1) persist old state ttl into snapshot meta info; (eg. RegisteredKeyValueStateBackendMetaInfo or others) 2) During state restore, check the size between the current ttl and old ttl; 3) If current ttl is longer than old ttl, we need to iterate over all data, filter out expired data with old ttl, and wirte valid data into stateBackend. -- This message was sent by Atlassian Jira (v8.20.10#820010)