[
https://issues.apache.org/jira/browse/FLINK-24542?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17430043#comment-17430043
]
Qingsheng Ren commented on FLINK-24542:
---------------------------------------
Thanks [~zlzhang0122] I checked [the documentation of
Kafka|https://kafka.apache.org/documentation/#consumer_monitoring] but didn't
find the metric "freshness". According to the blog post I assume this is a
derived metric instead of a standard metric exposed by Kafka client. I think
using other tools like Burrow to monitor Kafka in the bypass would be a better
choice.
> Expose the freshness metrics for kafka connector
> ------------------------------------------------
>
> Key: FLINK-24542
> URL: https://issues.apache.org/jira/browse/FLINK-24542
> Project: Flink
> Issue Type: Improvement
> Components: Connectors / Kafka
> Affects Versions: 1.12.2, 1.14.0, 1.13.1
> Reporter: zlzhang0122
> Priority: Major
> Fix For: 1.15.0
>
>
> When we start a flink job to consume apache kafka, we usually use offsetLag,
> which can be calulated by current-offsets minus committed-offsets, but
> sometimes the offsetLag is hard to understand, we can hardly to judge wether
> the value is normal or not. Kafka have proposed a new metric: freshness(see
> [a-guide-to-kafka-consumer-freshness|https://www.jesseyates.com/2019/11/04/kafka-consumer-freshness-a-guide.html?trk=article_share_wechat&from=timeline&isappinstalled=0]).
> So we can also expose the freshness metric for kafka connector to improve the
> user experience.From this freshness metric, user can easily know wether the
> kafka message is backlog and need to deal with it.
--
This message was sent by Atlassian Jira
(v8.3.4#803005)