Hi,

The code was already shipped in the 1.18 release, as FLINK-30113 reflects it, 
but the docs
are still not updated. So setting 
`execution.checkpointing.snapshot-compression` will also
compress regular and broadcast operator state.

Best,
Ferenc


On Monday, April 14th, 2025 at 16:02, Prasanna kumar 
<prasannakumarram...@gmail.com> wrote:

> Hi all,
> 
> Is Snappy Compression Enabled for broadcast state ?
> https://issues.apache.org/jira/browse/FLINK-30113 and FLINK-30112 are marked 
> as closed.
> 
> 
> Documentation link 
> https://nightlies.apache.org/flink/flink-docs-release-1.20/docs/ops/state/large_state_tuning/#compression
>  says that only
> "Compression works on the granularity of key-groups in keyed state, i.e. each 
> key-group can be decompressed individually, which is important for rescaling."
> 
> 
> And there is no mention of broadcast state.
> 
> 
> Thanks,
> Prasanna.
> 
> On Mon, Jan 9, 2023 at 2:31 PM Martijn Visser <martijnvis...@apache.org> 
> wrote:
> 
> > Hi Prasanna,
> > There is no support for compression in operator state. This can be tracked 
> > under https://issues.apache.org/jira/browse/FLINK-30113
> > 
> > Best regards,
> > 
> > Martijn
> > 
> > On Fri, Jan 6, 2023 at 7:53 AM Prasanna kumar 
> > <prasannakumarram...@gmail.com> wrote:
> > 
> > > Hello Flink Community ,
> > > 
> > > We are running Jobs in flink version 1.12.7 which reads from Kafka , 
> > > apply some rules(stored in broadcast state) and then writes to kafka. 
> > > This is a very low latency and high throughput and we have set up at 
> > > least one semantics.
> > > 
> > > Checkpoint Configuration Used
> > > 
> > > 1.  We cannot have many duplicates during the restarts so we have set a 
> > > checkpoint interval of 3s. (We cannot increase it any more since , we 
> > > have 10s of 1000s of records processed per sec ) .
> > > 2.  Checkpointing target location is AWS S3.
> > > 3.  Max Concurrent Checkpoint is 1
> > > 4.  Time Between Checkpoints is 500ms
> > > 
> > > Earlier we had around 10 rule objects stored in broadcast state. Recently 
> > > we have enabled 80 rule objects. Post increase , we are seeing a lot of 
> > > checkpoints in progress . (Earlier we had rarely seen this in metrics 
> > > dashboard). The Parallelism of BroadCast Function is around 10 and the 
> > > present Checkpoint size is 64kb.
> > > 
> > > Since we expect this rule objects to increase to 1000 and beyond in a 
> > > year's time, we are looking at ways to improve performance in checkpoint. 
> > > We cannot use incremental checkpoint since its supported only in RocksDB 
> > > and the development arc is little higher. Looking at easier solution 
> > > first , we tried using "SnapshotCompression" , but we did not see any 
> > > difference in decrease of checkpoint size.
> > > 
> > > Have few questions on the same
> > > 
> > > 1.  Does SnapshotCompression work in version 1.12.7 ?
> > > 2.  If Yes , how much size reduction could we expect if this is enabled 
> > > and at what size does the Compression works . Is there any threshold post 
> > > only which the compression would work ?
> > > 
> > > Apart from the questions above , you are welcome to suggest any config 
> > > changes that can be done for improvements.
> > > 
> > > Thanks & Regards,
> > > 
> > > Prasanna

Reply via email to