[ https://issues.apache.org/jira/browse/FLINK-26388?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Dawid Wysakowicz reassigned FLINK-26388: ---------------------------------------- Assignee: Dawid Wysakowicz > Release Testing: Repeatable Cleanup (FLINK-25433) > ------------------------------------------------- > > Key: FLINK-26388 > URL: https://issues.apache.org/jira/browse/FLINK-26388 > Project: Flink > Issue Type: Improvement > Components: Runtime / Coordination > Affects Versions: 1.15.0 > Reporter: Matthias Pohl > Assignee: Dawid Wysakowicz > Priority: Blocker > Labels: release-testing > Fix For: 1.15.0 > > > Repeatable cleanup got introduced with > [FLIP-194|https://issues.apache.org/jira/projects/FLINK/issues/FLINK-26284?filter=allopenissues] > but should be considered as an independent feature of the {{JobResultStore}} > (JRS) from a user's point of view. > Repeatable cleanup can be triggered by running into an error while cleaning > up. This can be achieved by disabling access to S3 after the job finished, > e.g.: > * Setting a reasonable enough checkpointing time (checkpointing should be > enabled to allow cleanup of s3) > * Disable s3 (removing permissions or shutting down the s3 server) > * Stop job with savepoint > Stopping the job should work but the logs should show failure with repeating > retries. Enabling S3 again should fix the issue. > Keep in mind that if testing this in with HA, you should use a different > bucket for the file-based JRS artifacts only change permissions for the > bucket that holds JRS-unrelated artifacts. Flink would fail fatally if the > JRS is not able to access it's backend storage. > Documentation and configuration is still in the process of being updated in > FLINK-26296 and FLINK-26331 -- This message was sent by Atlassian Jira (v8.20.1#820001)