@Yun Tang<mailto:myas...@live.com>,Thanks. ________________________________ From: Yun Tang <myas...@live.com> Sent: Monday, June 15, 2020 11:30 To: Thomas Huang <lyang...@hotmail.com>; Flink <user@flink.apache.org> Subject: Re: The Flink job recovered with wrong checkpoint state.
Hi Thomas The answer is yes. Without high availability, once the job manager is down and even the job manager is relaunched via YARN, the job graph and last checkpoint would not be recovered. Best Yun Tang ________________________________ From: Thomas Huang <lyang...@hotmail.com> Sent: Sunday, June 14, 2020 22:58 To: Flink <user@flink.apache.org> Subject: The Flink job recovered with wrong checkpoint state. Hi Flink Community, Currently, I'm using yarn-cluster mode to submit flink job on yarn, and I haven't set high availability configuration (zookeeper), but set restart strategy: env.getConfig.setRestartStrategy(RestartStrategies.fixedDelayRestart(10, 3000)) the attempt time is 10 and the wait time 30 seconds per failure. Today, when Infra team was rolling restart the yarn platform. Although the job manager restarted, the job hadn't recovered from the latest checkpoint, and all task managers started from the default job configuration that was not excepted. Does it mean I have to setup high availability configuration for yarn-cluster mode, or Is there any bug? Best Wish.