Hi, Yifan. Unfortunately, IIUC, we could get the key and value type only by reading related sql codes currently. I think it's useful if we could support SQL semantics for the Processor API, but it indeed will take lots of effort.
On Thu, Sep 21, 2023 at 12:05 PM Yifan He via user <user@flink.apache.org> wrote: > Hi Hangxiang, > > I still have one question about this problem, when using datastream api I > know the key and value type I use in state because I > defined ValueStateDescriptor, but how can I get the ValueStateDescriptor in > flinksql? > > Thanks, > Yifan > > On 2023/09/07 06:16:41 Hangxiang Yu wrote: > > Hi, Yifan. > > Which flink version are you using ? > > You are using filesystem instead of rocksdb so that your checkpoint size > > may not be incremental IIUC. > > > > On Thu, Sep 7, 2023 at 10:52 AM Yifan He via user < > us...@flink.apache.org> > > wrote: > > > > > Hi Shammon, > > > > > > We are using RocksDB,and the configuration is below: > > > execution.checkpointing.externalized-checkpoint-retention: > > > RETAIN_ON_CANCELLATION > > > execution.checkpointing.max-concurrent-checkpoints: 1 > > > execution.checkpointing.min-pause: 0 > > > execution.checkpointing.mode: EXACTLY_ONCE > > > execution.checkpointing.snapshot-compression: true > > > execution.checkpointing.timeout: 60000 > > > state.backend: FILESYSTEM > > > state.backend.incremental: true > > > state.backend.local-recovery: true > > > state.backend.rocksdb.memory.high-prio-pool-ratio: 0.1 > > > state.backend.rocksdb.memory.managed: true > > > state.backend.rocksdb.memory.write-buffer-ratio: 0.5 > > > state.backend.rocksdb.predefined-options: DEFAULT > > > state.backend.rocksdb.timer-service.factory: ROCKSDB > > > state.checkpoints.num-retained: 3 > > > > > > Thanks, > > > Yifan > > > > > > On 2023/09/06 08:00:31 Shammon FY wrote: > > > > Hi Yifan, > > > > > > > > Besides reading job state, I would like to know what statebackend > are you > > > > using? Can you give the configurations about state and checkpoint for > > > your > > > > job? Maybe you can check these configuration items to confirm if > they are > > > > correct first. > > > > > > > > Best, > > > > Shammon FY > > > > > > > > On Wed, Sep 6, 2023 at 3:17 PM Hang Ruan <ru...@gmail.com> wrote: > > > > > > > > > Hi, Yifan. > > > > > > > > > > I think the document[1] means to let us convert the DataStream to > the > > > > > Table[2]. Then we could handle the state with the Table API & SQL. > > > > > > > > > > Best, > > > > > Hang > > > > > > > > > > [1] > > > > > > > > > https://nightlies.apache.org/flink/flink-docs-release-1.17/docs/libs/state_processor_api/ > > > > > [2] > > > > > > > > > https://nightlies.apache.org/flink/flink-docs-release-1.17/docs/dev/table/data_stream_api/#converting-between-datastream-and-table > > > > > > > > > > Yifan He via user <us...@flink.apache.org> 于2023年9月6日周三 13:10写道: > > > > > > > > > >> Hi team, > > > > >> > > > > >> We are investigating why the checkpoint size of our FlinkSQL jobs > > > keeps > > > > >> growing and we want to look into the checkpoint file to know what > is > > > > >> causing the problem. I know we can use the state processor api to > > > read the > > > > >> state of jobs using datastream api, but how can I read the state > of > > > jobs > > > > >> using table api & sql? > > > > >> > > > > >> Thanks, > > > > >> Yifan > > > > >> > > > > > > > > > > > > > > > > > > -- > > Best, > > Hangxiang. > > > -- Best, Hangxiang.