(Looks like my previous ask of this question got spamblocked because I 
included a screenshot.  c'est la vie.)

I have alerts for when a metric's value passes above or below a threshold.  
I can ask for the minimum or maximum over a time range, I can as for a 
prediction based on the slope of a graph.

I have some resources that I know will fail soon after their metrics 
fluctuate wildly over a short period of time.  They may never exceed the 
absolute value of 85% during their fluctuations, or they may exceed this 
briefly but not long enough to cause concern if it was a smooth line.  
I.E.  If the samples over time were [30, 30, 31, 70, 5, 69, 6, 71, 5, 69, 
null, null, null]  I want to detect it before the metric goes absent 
(because the resource crashed).

Setting the threshold at ">69" doesn't work because the value drops below 
the threshold on the next scrape, closing the alert; besides, if it were at 
a steady 69 that would be healthy.
Setting the threshold at "avg(metric[interval)" doesn't work because the 
average of an oscillating metric will be well within the healthy range.
I thought of setting an alert for "max_over_time - min_over_time > 50" but 
that would trigger on a smooth ascension -- a false positive.

What's the question should I ask Prometheus to detect a metric that 
vibrates too much?

-- 
You received this message because you are subscribed to the Google Groups 
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/prometheus-users/dfdfc12b-767c-458a-b238-08d87cd3e7d1%40googlegroups.com.

Reply via email to