oh nice, i read and missed that.  many thanks; i will try changing that 
setting.

On Thursday, April 16, 2020 at 9:20:13 AM UTC-4, Robert Morman wrote:
>
> Good morning!
>
> I'm new to prometheus and wondering what is causing a delay in receiving 
> resolved slack alerts.  I enabled debug logging and have tried changing 
> resolve_timeout, yet there is still a delay beyond what the configuration 
> is set.  I receive the initial InstanceDown slack alert quickly and when 
> expected, but you can see from the debug logs that prometheus is sending a 
> resolved alert to alertmanager, but alertmanager is delaying flushing and 
> sending to slack.  I tried peering into the alertmanager data, but couldn't 
> tell if prometheus is sending EndsAt.  resolve_timeout seems to have no 
> effect.  It always seems to send 5m after the active alert is triggered.
>
> ######### alertmanager version #########
>
> root@alertmanager1:/etc/alertmanager# alertmanager --version
> alertmanager, version 0.20.0 (branch: HEAD, revision: 
> f74be0400a6243d10bb53812d6fa408ad71ff32d)
>   build user:       root@00c3106655f8
>   build date:       20191211-14:13:14
>   go version:       go1.13.5
>
> ######### alertmanager.yml #########
> global:
>   resolve_timeout: 15s
>
> templates:
> - '/etc/alertmanager/*.tmpl'
>
> route:
>   repeat_interval: 1h
>   receiver: critical
>
> receivers:
> - name: 'critical'
>   slack_configs:
>   - api_url: https://hooks.slack.com/services/XXXXXXX/XXXXXXXX/XXXXXXXX
>     channel: '#alerts'
>     send_resolved: true
>     title: '{{ template "title" . }}'
>     text: '{{ template "slack_message" . }}'
>
> ######### prometheus alerts config #########
> groups:
> - name: example
>   rules:
>   # Alert for any instance that is unreachable
>   - alert: InstanceDown
>     expr: up == 0
>     labels:
>       severity: page
>     annotations:
>       summary: "Instance {{$labels.instance}} down"
>       description: "{{$labels.instance}} is down!"
>
>
> ######### alertmanager debug logs #########
> Apr 16 13:13:06 alertmanager1-mon-prd-cle alertmanager[28061]: level=debug 
> ts=2020-04-16T13:13:06.458Z caller=dispatch.go:135 component=dispatcher 
> msg="Received alert" alert=InstanceDown[bd6501b][active]
> Apr 16 13:13:06 alertmanager1-mon-prd-cle alertmanager[28061]: level=debug 
> ts=2020-04-16T13:13:06.458Z caller=dispatch.go:465 component=dispatcher 
> aggrGroup={}:{} msg=flushing alerts=[InstanceDown[bd6501b][active]]
> Apr 16 13:14:36 alertmanager1-mon-prd-cle alertmanager[28061]: level=debug 
> ts=2020-04-16T13:14:36.456Z caller=dispatch.go:135 component=dispatcher 
> msg="Received alert" alert=InstanceDown[bd6501b][active]
> Apr 16 13:15:36 alertmanager1-mon-prd-cle alertmanager[28061]: level=debug 
> ts=2020-04-16T13:15:36.456Z caller=dispatch.go:135 component=dispatcher 
> msg="Received alert" alert=InstanceDown[bd6501b][resolved]
> Apr 16 13:17:06 alertmanager1-mon-prd-cle alertmanager[28061]: level=debug 
> ts=2020-04-16T13:17:06.455Z caller=dispatch.go:135 component=dispatcher 
> msg="Received alert" alert=InstanceDown[bd6501b][resolved]
> Apr 16 13:18:06 alertmanager1-mon-prd-cle alertmanager[28061]: level=debug 
> ts=2020-04-16T13:18:06.458Z caller=dispatch.go:465 component=dispatcher 
> aggrGroup={}:{} msg=flushing alerts=[InstanceDown[bd6501b][resolved]]
>

-- 
You received this message because you are subscribed to the Google Groups 
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/prometheus-users/e2cdf7a5-b8c1-4c35-bebb-69a7a4976fdd%40googlegroups.com.

Reply via email to