Did you try something like "changes(foo[30m]) > 10" ? That would alert if 
the value changed 10 times in the last 30 minutes.

On Friday, 28 February 2020 16:21:53 UTC, Moses Moore wrote:
>
> (Looks like my previous ask of this question got spamblocked because I 
> included a screenshot.  c'est la vie.)
>
> I have alerts for when a metric's value passes above or below a 
> threshold.  I can ask for the minimum or maximum over a time range, I can 
> as for a prediction based on the slope of a graph.
>
> I have some resources that I know will fail soon after their metrics 
> fluctuate wildly over a short period of time.  They may never exceed the 
> absolute value of 85% during their fluctuations, or they may exceed this 
> briefly but not long enough to cause concern if it was a smooth line.  
> I.E.  If the samples over time were [30, 30, 31, 70, 5, 69, 6, 71, 5, 69, 
> null, null, null]  I want to detect it before the metric goes absent 
> (because the resource crashed).
>
> Setting the threshold at ">69" doesn't work because the value drops below 
> the threshold on the next scrape, closing the alert; besides, if it were at 
> a steady 69 that would be healthy.
> Setting the threshold at "avg(metric[interval)" doesn't work because the 
> average of an oscillating metric will be well within the healthy range.
> I thought of setting an alert for "max_over_time - min_over_time > 50" but 
> that would trigger on a smooth ascension -- a false positive.
>
> What's the question should I ask Prometheus to detect a metric that 
> vibrates too much?
>
>

-- 
You received this message because you are subscribed to the Google Groups 
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/prometheus-users/5f7ef3b8-a311-4d98-8bcb-2594e6aaef80%40googlegroups.com.

Reply via email to