Brian,

Originally, I had not activated any additional collectors.  Then, I read 
somewhere that I should add the systemd and process collectors.  Still 
learning, here, so ... .  That's why you saw them in the pprof graph.  I 
then curcled back and removed them.  However, high CPU usage has *always* 
been an issue.  That goes for every system in which I have node_exporter 
running.  While a few are test machines, and I care a bit less, for 
production machines it's an issue.

Here's some time output for node_exporter, though I'm not good at 
interpreting the results:

[root@myhost1 ~]# time for ((i=1;i<=1000;i++)); do node_exporter >/dev/null 
2>&1; done

real        0m6.103s
user        0m3.658s
sys        0m3.151s

So, if the above is a good way to measure node_exporter's user versus 
system time, then they're about equal.  If you have another means to do 
such measurement, I'd appreciate your sharing it.  Once that's determined 
and, if system time versus user time is "out-of-whack", how do I remediate?

Many thanks.

On Thursday, January 20, 2022 at 3:46:35 AM UTC-5 Brian Candler wrote:

> So the systemd and process collectors aren't active.  I wonder why they 
> appeared in your pprof graph then?  Was it exactly the same binary you were 
> running?
>
> 20% CPU usage from a once-every-five-second scrape implies that it should 
> take about 1 CPU-second in total, but all the collectors seem very fast.  
> The top five use between 0.01 and 0.015 seconds - and that's wall clock 
> time, not CPU time.
>
> node_scrape_collector_duration_seconds{collector="cpu"} 0.010873961
> node_scrape_collector_duration_seconds{collector="diskstats"} 0.01727642
> node_scrape_collector_duration_seconds{collector="hwmon"} 0.014143617
> node_scrape_collector_duration_seconds{collector="netclass"} 0.013852102
> node_scrape_collector_duration_seconds{collector="thermal_zone"} 
> 0.010936983
>
> Something weird is going on.  Next you might want to drill down into 
> node_exporter's user versus system time.  Is the usage mostly system time?  
> That might point you some way, although the implication then is that the 
> high CPU usage is some part of node_exporter outside of individual 
> collectors.
>
> On Wednesday, 19 January 2022 at 23:27:40 UTC dyio...@gmail.com wrote:
>
>> [root@myhost1 ~]# curl -Ss localhost:9100/metrics | grep -i collector
>> # HELP node_scrape_collector_duration_seconds node_exporter: Duration of 
>> a collector scrape.
>> # TYPE node_scrape_collector_duration_seconds gauge
>> node_scrape_collector_duration_seconds{collector="arp"} 0.002911805
>> node_scrape_collector_duration_seconds{collector="bcache"} 1.4571e-05
>> node_scrape_collector_duration_seconds{collector="bonding"} 0.000112308
>> node_scrape_collector_duration_seconds{collector="btrfs"} 0.001308192
>> node_scrape_collector_duration_seconds{collector="conntrack"} 0.002750716
>> node_scrape_collector_duration_seconds{collector="cpu"} 0.010873961
>> node_scrape_collector_duration_seconds{collector="cpufreq"} 0.008559194
>> node_scrape_collector_duration_seconds{collector="diskstats"} 0.01727642
>> node_scrape_collector_duration_seconds{collector="dmi"} 0.000971785
>> node_scrape_collector_duration_seconds{collector="edac"} 0.006972343
>> node_scrape_collector_duration_seconds{collector="entropy"} 0.001360089
>> node_scrape_collector_duration_seconds{collector="fibrechannel"} 
>> 2.8256e-05
>> node_scrape_collector_duration_seconds{collector="filefd"} 0.000739988
>> node_scrape_collector_duration_seconds{collector="filesystem"} 0.00554684
>> node_scrape_collector_duration_seconds{collector="hwmon"} 0.014143617
>> node_scrape_collector_duration_seconds{collector="infiniband"} 1.3484e-05
>> node_scrape_collector_duration_seconds{collector="ipvs"} 7.5532e-05
>> node_scrape_collector_duration_seconds{collector="loadavg"} 0.004074291
>> node_scrape_collector_duration_seconds{collector="mdadm"} 0.000974966
>> node_scrape_collector_duration_seconds{collector="meminfo"} 0.004201816
>> node_scrape_collector_duration_seconds{collector="netclass"} 0.013852102
>> node_scrape_collector_duration_seconds{collector="netdev"} 0.006993921
>> node_scrape_collector_duration_seconds{collector="netstat"} 0.007896151
>> node_scrape_collector_duration_seconds{collector="nfs"} 0.000125062
>> node_scrape_collector_duration_seconds{collector="nfsd"} 3.6075e-05
>> node_scrape_collector_duration_seconds{collector="nvme"} 0.001064067
>> node_scrape_collector_duration_seconds{collector="os"} 0.005645435
>> node_scrape_collector_duration_seconds{collector="powersupplyclass"} 
>> 0.001394135
>> node_scrape_collector_duration_seconds{collector="pressure"} 0.001466664
>> node_scrape_collector_duration_seconds{collector="rapl"} 0.00226622
>> node_scrape_collector_duration_seconds{collector="schedstat"} 0.006677493
>> node_scrape_collector_duration_seconds{collector="sockstat"} 0.000970676
>> node_scrape_collector_duration_seconds{collector="softnet"} 0.002014497
>> node_scrape_collector_duration_seconds{collector="stat"} 0.004216999
>> node_scrape_collector_duration_seconds{collector="tapestats"} 1.0296e-05
>> node_scrape_collector_duration_seconds{collector="textfile"} 5.2573e-05
>> node_scrape_collector_duration_seconds{collector="thermal_zone"} 
>> 0.010936983
>> node_scrape_collector_duration_seconds{collector="time"} 0.00568072
>> node_scrape_collector_duration_seconds{collector="timex"} 3.3662e-05
>> node_scrape_collector_duration_seconds{collector="udp_queues"} 0.004138555
>> node_scrape_collector_duration_seconds{collector="uname"} 1.3713e-05
>> node_scrape_collector_duration_seconds{collector="vmstat"} 0.005691152
>> node_scrape_collector_duration_seconds{collector="xfs"} 0.008633677
>> node_scrape_collector_duration_seconds{collector="zfs"} 2.8179e-05
>> # HELP node_scrape_collector_success node_exporter: Whether a collector 
>> succeeded.
>> # TYPE node_scrape_collector_success gauge
>> node_scrape_collector_success{collector="arp"} 1
>> node_scrape_collector_success{collector="bcache"} 1
>> node_scrape_collector_success{collector="bonding"} 0
>> node_scrape_collector_success{collector="btrfs"} 1
>> node_scrape_collector_success{collector="conntrack"} 1
>> node_scrape_collector_success{collector="cpu"} 1
>> node_scrape_collector_success{collector="cpufreq"} 1
>> node_scrape_collector_success{collector="diskstats"} 1
>> node_scrape_collector_success{collector="dmi"} 1
>> node_scrape_collector_success{collector="edac"} 1
>> node_scrape_collector_success{collector="entropy"} 1
>> node_scrape_collector_success{collector="fibrechannel"} 0
>> node_scrape_collector_success{collector="filefd"} 1
>> node_scrape_collector_success{collector="filesystem"} 1
>> node_scrape_collector_success{collector="hwmon"} 1
>> node_scrape_collector_success{collector="infiniband"} 0
>> node_scrape_collector_success{collector="ipvs"} 0
>> node_scrape_collector_success{collector="loadavg"} 1
>> node_scrape_collector_success{collector="mdadm"} 1
>> node_scrape_collector_success{collector="meminfo"} 1
>> node_scrape_collector_success{collector="netclass"} 1
>> node_scrape_collector_success{collector="netdev"} 1
>> node_scrape_collector_success{collector="netstat"} 1
>> node_scrape_collector_success{collector="nfs"} 0
>> node_scrape_collector_success{collector="nfsd"} 0
>> node_scrape_collector_success{collector="nvme"} 0
>> node_scrape_collector_success{collector="os"} 1
>> node_scrape_collector_success{collector="powersupplyclass"} 1
>> node_scrape_collector_success{collector="pressure"} 0
>> node_scrape_collector_success{collector="rapl"} 1
>> node_scrape_collector_success{collector="schedstat"} 1
>> node_scrape_collector_success{collector="sockstat"} 1
>> node_scrape_collector_success{collector="softnet"} 1
>> node_scrape_collector_success{collector="stat"} 1
>> node_scrape_collector_success{collector="tapestats"} 0
>> node_scrape_collector_success{collector="textfile"} 1
>> node_scrape_collector_success{collector="thermal_zone"} 1
>> node_scrape_collector_success{collector="time"} 1
>> node_scrape_collector_success{collector="timex"} 1
>> node_scrape_collector_success{collector="udp_queues"} 1
>> node_scrape_collector_success{collector="uname"} 1
>> node_scrape_collector_success{collector="vmstat"} 1
>> node_scrape_collector_success{collector="xfs"} 1
>> node_scrape_collector_success{collector="zfs"} 0
>>
>> On Tuesday, January 18, 2022 at 1:12:04 PM UTC-5 Brian Candler wrote:
>>
>>> Can you show the output of:
>>>
>>> curl -Ss localhost:9100/metrics | grep -i collector
>>>
>>> On Tuesday, 18 January 2022 at 14:33:25 UTC dyio...@gmail.com wrote:
>>>
>>>> [root@myhost1 ~]# ps auxwww | grep node_exporter
>>>> node_ex+ 4143664 12.5  0.0 725828 22668 ?        Ssl  09:29   0:06 
>>>> /usr/local/bin/node_exporter --no-collector.wifi
>>>>
>>>> On Saturday, January 15, 2022 at 11:23:43 AM UTC-5 Brian Candler wrote:
>>>>
>>>>> On Friday, 14 January 2022 at 14:12:02 UTC dyio...@gmail.com wrote:
>>>>>
>>>>>> @Brian Chandler  I'm using the node_exporter defaults, as described 
>>>>>> here - https://github.com/prometheus/node_exporter.
>>>>>>
>>>>>
>>>>> Are you *really*?    Can you show the *exact* command line that 
>>>>> node_exporter is running with?  e.g.
>>>>>
>>>>> ps auxwww | grep node_exporter
>>>>>
>>>>

-- 
You received this message because you are subscribed to the Google Groups 
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to prometheus-users+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/prometheus-users/2bdae4c8-e029-49d4-a00d-11f3521c683bn%40googlegroups.com.

Reply via email to