Hi, The code was already shipped in the 1.18 release, as FLINK-30113 reflects it, but the docs are still not updated. So setting `execution.checkpointing.snapshot-compression` will also compress regular and broadcast operator state.
Best, Ferenc On Monday, April 14th, 2025 at 16:02, Prasanna kumar <prasannakumarram...@gmail.com> wrote: > Hi all, > > Is Snappy Compression Enabled for broadcast state ? > https://issues.apache.org/jira/browse/FLINK-30113 and FLINK-30112 are marked > as closed. > > > Documentation link > https://nightlies.apache.org/flink/flink-docs-release-1.20/docs/ops/state/large_state_tuning/#compression > says that only > "Compression works on the granularity of key-groups in keyed state, i.e. each > key-group can be decompressed individually, which is important for rescaling." > > > And there is no mention of broadcast state. > > > Thanks, > Prasanna. > > On Mon, Jan 9, 2023 at 2:31 PM Martijn Visser <martijnvis...@apache.org> > wrote: > > > Hi Prasanna, > > There is no support for compression in operator state. This can be tracked > > under https://issues.apache.org/jira/browse/FLINK-30113 > > > > Best regards, > > > > Martijn > > > > On Fri, Jan 6, 2023 at 7:53 AM Prasanna kumar > > <prasannakumarram...@gmail.com> wrote: > > > > > Hello Flink Community , > > > > > > We are running Jobs in flink version 1.12.7 which reads from Kafka , > > > apply some rules(stored in broadcast state) and then writes to kafka. > > > This is a very low latency and high throughput and we have set up at > > > least one semantics. > > > > > > Checkpoint Configuration Used > > > > > > 1. We cannot have many duplicates during the restarts so we have set a > > > checkpoint interval of 3s. (We cannot increase it any more since , we > > > have 10s of 1000s of records processed per sec ) . > > > 2. Checkpointing target location is AWS S3. > > > 3. Max Concurrent Checkpoint is 1 > > > 4. Time Between Checkpoints is 500ms > > > > > > Earlier we had around 10 rule objects stored in broadcast state. Recently > > > we have enabled 80 rule objects. Post increase , we are seeing a lot of > > > checkpoints in progress . (Earlier we had rarely seen this in metrics > > > dashboard). The Parallelism of BroadCast Function is around 10 and the > > > present Checkpoint size is 64kb. > > > > > > Since we expect this rule objects to increase to 1000 and beyond in a > > > year's time, we are looking at ways to improve performance in checkpoint. > > > We cannot use incremental checkpoint since its supported only in RocksDB > > > and the development arc is little higher. Looking at easier solution > > > first , we tried using "SnapshotCompression" , but we did not see any > > > difference in decrease of checkpoint size. > > > > > > Have few questions on the same > > > > > > 1. Does SnapshotCompression work in version 1.12.7 ? > > > 2. If Yes , how much size reduction could we expect if this is enabled > > > and at what size does the Compression works . Is there any threshold post > > > only which the compression would work ? > > > > > > Apart from the questions above , you are welcome to suggest any config > > > changes that can be done for improvements. > > > > > > Thanks & Regards, > > > > > > Prasanna