Hi, Peter
I have not been able to reproduce this problem. 

From your description, it is possible that the specified port 4444 has been 
listened by other processes, and PrometheusReporter failed to start. 
You can confirm it from taskmanager.log, or check if port 4444 of the host is 
being listened by the TaskManager process.


> 2022年4月20日 下午10:48,Peter Schrott <pe...@bluerootlabs.io> 写道:
> 
> Hi Flink-Users,
> 
> After upgrading to Flink 1.15 (rc3) (coming from 1.14) I noticed that there 
> is a problem with the metrics exposed through the PrometheusReporter. 
> 
> It is configured as followed in the flink-config.yml:
> metrics.reporters: prom
> metrics.reporter.prom.class: 
> org.apache.flink.metrics.prometheus.PrometheusReporter
> metrics.reporter.prom.port: 4444
> 
> My cluster is running in standalone mode with 2 taskmanagers and 2 
> jobmanagers.
> 
> More specifically: 
> 
> On the taskmanger that runs a job I get curl: (52) Empty reply from server 
> when I call curl localhost:4444. I was looking for the metrics in the 
> namespace flink_taskmanager_job_*, which are only - and obviously - exposed 
> on the taskmanager running a job.
> 
> On the other taskmanger that runs no job I get a response with a couple of 
> metrics of the namespace flink_taskmanager_Status - as expected.
> 
> When configuring the JMXReporterFactory for too. I find the desired and all 
> other metrics via VisualVM on that taskmanager running the job. Also in the 
> Flink web ui, in the "Jobs -> Overview -> Metrics" part I can select and 
> visualize metrics like flink_taskmanager_job_task_busyTimeMsPerSecond.
> 
> Does someone have any idea what's going on here? maybe even confirm my 
> findings?
> 
> Best & thanks,
> Peter
> 

Reply via email to