Hi, I am interested in dumping Flink state from Rockdb to datalake using state query https://ci.apache.org/projects/flink/flink-docs-master/docs/dev/datastream/fault-tolerance/queryable_state/. My map state could have 200 million key-values pairs and the total size could be 150G bytes. My batch job scheduled using airflow will have one task which uses Flink state query to dump the Flink state to datalake in parquet format so other spark tasks can use it.
Is there any scalability concern for using state query in this way? Appreciate any insight. Thanks!