[
https://issues.apache.org/jira/browse/FLINK-34558?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17824593#comment-17824593
]
Hangxiang Yu commented on FLINK-34558:
--------------------------------------
I think this mertic maybe useful, but since it's under the critical path of
every element, we should treat it very carefully.
So I'd suggest to start with:
# try to implement this just like state-latency track (just sampling and be
disabled by default).
# micro benchmark and show result firstly (maybe three results: before this
pr, disabed, enabled).
> Add RocksDB key/value size metrics
> ----------------------------------
>
> Key: FLINK-34558
> URL: https://issues.apache.org/jira/browse/FLINK-34558
> Project: Flink
> Issue Type: Improvement
> Components: Runtime / State Backends
> Affects Versions: 1.19.0
> Reporter: Jufang He
> Priority: Major
>
> In some scenarios, the poor performance of RocksDB may be caused by too large
> key/value size, but now there is a lack of metrics for key/value size. By
> adding these metrics, we can conveniently calculate the distribution of
> key/value size, such as the average size and p99 size. To reduce the
> negative impact of adding metric on RocksDB performance, we could reduce the
> impact by supporting sampling.
> The possible implementation is as follows: After the RocksDB key/value
> serialization, we could obtain the byte array and report the size of the
> array through histogram metrics.
--
This message was sent by Atlassian Jira
(v8.20.10#820010)