Hi, Yifan. Which flink version are you using ? You are using filesystem instead of rocksdb so that your checkpoint size may not be incremental IIUC.
On Thu, Sep 7, 2023 at 10:52 AM Yifan He via user <user@flink.apache.org> wrote: > Hi Shammon, > > We are using RocksDB,and the configuration is below: > execution.checkpointing.externalized-checkpoint-retention: > RETAIN_ON_CANCELLATION > execution.checkpointing.max-concurrent-checkpoints: 1 > execution.checkpointing.min-pause: 0 > execution.checkpointing.mode: EXACTLY_ONCE > execution.checkpointing.snapshot-compression: true > execution.checkpointing.timeout: 60000 > state.backend: FILESYSTEM > state.backend.incremental: true > state.backend.local-recovery: true > state.backend.rocksdb.memory.high-prio-pool-ratio: 0.1 > state.backend.rocksdb.memory.managed: true > state.backend.rocksdb.memory.write-buffer-ratio: 0.5 > state.backend.rocksdb.predefined-options: DEFAULT > state.backend.rocksdb.timer-service.factory: ROCKSDB > state.checkpoints.num-retained: 3 > > Thanks, > Yifan > > On 2023/09/06 08:00:31 Shammon FY wrote: > > Hi Yifan, > > > > Besides reading job state, I would like to know what statebackend are you > > using? Can you give the configurations about state and checkpoint for > your > > job? Maybe you can check these configuration items to confirm if they are > > correct first. > > > > Best, > > Shammon FY > > > > On Wed, Sep 6, 2023 at 3:17 PM Hang Ruan <ru...@gmail.com> wrote: > > > > > Hi, Yifan. > > > > > > I think the document[1] means to let us convert the DataStream to the > > > Table[2]. Then we could handle the state with the Table API & SQL. > > > > > > Best, > > > Hang > > > > > > [1] > > > > https://nightlies.apache.org/flink/flink-docs-release-1.17/docs/libs/state_processor_api/ > > > [2] > > > > https://nightlies.apache.org/flink/flink-docs-release-1.17/docs/dev/table/data_stream_api/#converting-between-datastream-and-table > > > > > > Yifan He via user <us...@flink.apache.org> 于2023年9月6日周三 13:10写道: > > > > > >> Hi team, > > >> > > >> We are investigating why the checkpoint size of our FlinkSQL jobs > keeps > > >> growing and we want to look into the checkpoint file to know what is > > >> causing the problem. I know we can use the state processor api to > read the > > >> state of jobs using datastream api, but how can I read the state of > jobs > > >> using table api & sql? > > >> > > >> Thanks, > > >> Yifan > > >> > > > > > > -- Best, Hangxiang.