Hi Shammon, We are using RocksDB,and the configuration is below: execution.checkpointing.externalized-checkpoint-retention: RETAIN_ON_CANCELLATION execution.checkpointing.max-concurrent-checkpoints: 1 execution.checkpointing.min-pause: 0 execution.checkpointing.mode: EXACTLY_ONCE execution.checkpointing.snapshot-compression: true execution.checkpointing.timeout: 60000 state.backend: FILESYSTEM state.backend.incremental: true state.backend.local-recovery: true state.backend.rocksdb.memory.high-prio-pool-ratio: 0.1 state.backend.rocksdb.memory.managed: true state.backend.rocksdb.memory.write-buffer-ratio: 0.5 state.backend.rocksdb.predefined-options: DEFAULT state.backend.rocksdb.timer-service.factory: ROCKSDB state.checkpoints.num-retained: 3
Thanks, Yifan On 2023/09/06 08:00:31 Shammon FY wrote: > Hi Yifan, > > Besides reading job state, I would like to know what statebackend are you > using? Can you give the configurations about state and checkpoint for your > job? Maybe you can check these configuration items to confirm if they are > correct first. > > Best, > Shammon FY > > On Wed, Sep 6, 2023 at 3:17 PM Hang Ruan <ru...@gmail.com> wrote: > > > Hi, Yifan. > > > > I think the document[1] means to let us convert the DataStream to the > > Table[2]. Then we could handle the state with the Table API & SQL. > > > > Best, > > Hang > > > > [1] > > https://nightlies.apache.org/flink/flink-docs-release-1.17/docs/libs/state_processor_api/ > > [2] > > https://nightlies.apache.org/flink/flink-docs-release-1.17/docs/dev/table/data_stream_api/#converting-between-datastream-and-table > > > > Yifan He via user <us...@flink.apache.org> 于2023年9月6日周三 13:10写道: > > > >> Hi team, > >> > >> We are investigating why the checkpoint size of our FlinkSQL jobs keeps > >> growing and we want to look into the checkpoint file to know what is > >> causing the problem. I know we can use the state processor api to read the > >> state of jobs using datastream api, but how can I read the state of jobs > >> using table api & sql? > >> > >> Thanks, > >> Yifan > >> > > >