Did you try something like "changes(foo[30m]) > 10" ? That would alert if the value changed 10 times in the last 30 minutes.
On Friday, 28 February 2020 16:21:53 UTC, Moses Moore wrote: > > (Looks like my previous ask of this question got spamblocked because I > included a screenshot. c'est la vie.) > > I have alerts for when a metric's value passes above or below a > threshold. I can ask for the minimum or maximum over a time range, I can > as for a prediction based on the slope of a graph. > > I have some resources that I know will fail soon after their metrics > fluctuate wildly over a short period of time. They may never exceed the > absolute value of 85% during their fluctuations, or they may exceed this > briefly but not long enough to cause concern if it was a smooth line. > I.E. If the samples over time were [30, 30, 31, 70, 5, 69, 6, 71, 5, 69, > null, null, null] I want to detect it before the metric goes absent > (because the resource crashed). > > Setting the threshold at ">69" doesn't work because the value drops below > the threshold on the next scrape, closing the alert; besides, if it were at > a steady 69 that would be healthy. > Setting the threshold at "avg(metric[interval)" doesn't work because the > average of an oscillating metric will be well within the healthy range. > I thought of setting an alert for "max_over_time - min_over_time > 50" but > that would trigger on a smooth ascension -- a false positive. > > What's the question should I ask Prometheus to detect a metric that > vibrates too much? > > -- You received this message because you are subscribed to the Google Groups "Prometheus Users" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To view this discussion on the web visit https://groups.google.com/d/msgid/prometheus-users/5f7ef3b8-a311-4d98-8bcb-2594e6aaef80%40googlegroups.com.

