If anyone else has the same issue:
https://www.section.io/blog/beware-prometheus-counters-that-do-not-begin-at-zero/
On Thursday, February 25, 2021 at 12:05:45 PM UTC+1 Constantin Clauzel
wrote:
> Last test, looks like it's the
> `sum(increase(appbackend_alerts{metric="TEMP", type="TOO_HIGH"}[5m]))`
> query that doesn't return anything as long as the counters are at 1.
> I added the query sum(appbackend_alerts{}) which displays the attached
> graph (blue lines shows that there are indeed counter values coming in).
>
> On Thursday, February 25, 2021 at 11:48:55 AM UTC+1 Constantin Clauzel
> wrote:
>
>> Is it possible that prometheus is ignoring the counter as long as it's
>> equal to 1?
>> I can't be sure but it seems that it only started logging my counter
>> because some of them are now higher than 1.
>> This could be coherent as my system increments the value every 30min, so
>> if the condition is met twice it will take up to an hour to have one reach
>> 2.
>>
>> On Thursday, February 25, 2021 at 11:44:42 AM UTC+1 Constantin Clauzel
>> wrote:
>>
>>> And it just started appearing again.
>>>
>>> It took approximately the same time to show up again as last time,
>>> around 1h10.
>>>
>>>
>>> On Thursday, February 25, 2021 at 11:37:36 AM UTC+1 Constantin Clauzel
>>> wrote:
>>>
>>>> Attached file in previous message was broken:
>>>>
>>>> On Thursday, February 25, 2021 at 11:32:52 AM UTC+1 Constantin Clauzel
>>>> wrote:
>>>>
>>>>> Hey,
>>>>>
>>>>> Since this morning I'm experiencing some very weird behaviors with one
>>>>> of my counters.
>>>>> It randomly stays at zero for an hour, then appears again, then again
>>>>> to zero.
>>>>>
>>>>> What is strange is that all other metrics are showing up, meaning
>>>>> prometheus can reach the endpoint, and when I check the endpoint is has
>>>>> the
>>>>> missing counter in it.
>>>>>
>>>>> Is there any possible reason that could explain why a counter suddenly
>>>>> only returns zeros, and then starts working again for no apparent reason?
>>>>>
>>>>> Please find attached how the graphs look.
>>>>>
>>>>> The query that returns all zeros:
>>>>> sum(increase(appbackend_alerts{metric="TEMP", type="TOO_HIGH"}[5m]))
>>>>> sum(increase(appbackend_alerts{metric="HUMI", type="TOO_HIGH"}[5m]))
>>>>>
>>>>> The prometheus endpoint returns all those appbackend_alerts lines:
>>>>>
>>>>> appbackend_alerts{boxID="0",controllerID="xxxxxxxxx",metric="HUMI",type="TOO_HIGH"}
>>>>>
>>>>> 1
>>>>> appbackend_alerts{boxID="0",controllerID="xxxxxxxxx",metric="HUMI",type="TOO_LOW"}
>>>>>
>>>>> 1
>>>>> appbackend_alerts{boxID="0",controllerID="xxxxxxxxx",metric="TEMP",type="TOO_LOW"}
>>>>>
>>>>> 1
>>>>> appbackend_alerts{boxID="0",controllerID="xxxxxxxxx",metric="HUMI",type="TOO_LOW"}
>>>>>
>>>>> 1
>>>>> appbackend_alerts{boxID="0",controllerID="xxxxxxxxx",metric="HUMI",type="TOO_HIGH"}
>>>>>
>>>>> 1
>>>>> appbackend_alerts{boxID="0",controllerID="xxxxxxxxx",metric="HUMI",type="TOO_LOW"}
>>>>>
>>>>> 1
>>>>> appbackend_alerts{boxID="0",controllerID="xxxxxxxxx",metric="HUMI",type="TOO_LOW"}
>>>>>
>>>>> 1
>>>>> appbackend_alerts{boxID="0",controllerID="xxxxxxxxx",metric="HUMI",type="TOO_HIGH"}
>>>>>
>>>>> 1
>>>>> [ ... And may more ]
>>>>>
>>>>> Thanks,
>>>>> Constantin
>>>>>
>>>>
--
You received this message because you are subscribed to the Google Groups
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to [email protected].
To view this discussion on the web visit
https://groups.google.com/d/msgid/prometheus-users/62b45089-0304-460f-87fd-8a00d74607aan%40googlegroups.com.