The label set in the error is empty. There are still a bunch of labels on 
the metric though. The metrics don't have timestamps, just generated 
everytime prometheus scrapes them.

How do I improve the alert rule to ignore these? Looking at the docs, is 
the only way to wrap these in 2 label_replace, or otherwise sum and specify 
every relevant label?

On Thursday, 27 February 2020 14:43:05 UTC+1, Julien Pivotto wrote:
>
> On 27 Feb 05:33, Paul van der Linden wrote: 
> > I have currently a pod running in kubernetes exporting some metrics to 
> > prometheus. If we have alerts on them, they will be retriggered 
> (resolved & 
> > trigger shortly after eachother) every time I update the software (as 
> the 
> > pod would have a different name/ip). I came to the conclusing that this 
> is 
> > because of the unique pod and instance labels on these metrics. I have 
> > added a config to drop these labels, but this seems to cause the error " 
> > persist head block: write compaction: add series: out-of-order series 
> added 
> > with label set", while it solves the issue. What is the correct way to 
> > solve this issue. 
>
> Is the label set in the error empty? Are you exposing metrics with 
> timestamps? 
>
> The correct way to deal with this would probably be to ingest with the 
> instance and pod name and improve your alert rule to ignore that at this 
> point. 
>
> Thanks 
>
> -- 
>  (o-    Julien Pivotto 
>  //\    Open-Source Consultant 
>  V_/_   Inuits - https://www.inuits.eu 
>

-- 
You received this message because you are subscribed to the Google Groups 
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/prometheus-users/54ff4ef8-7154-4517-bcda-8e52411fddea%40googlegroups.com.

Reply via email to