Good morning!

I'm new to prometheus and wondering what is causing a delay in receiving 
resolved slack alerts.  I enabled debug logging and have tried changing 
resolve_timeout, yet there is still a delay beyond what the configuration 
is set.  I receive the initial InstanceDown slack alert quickly and when 
expected, but you can see from the debug logs that prometheus is sending a 
resolved alert to alertmanager, but alertmanager is delaying flushing and 
sending to slack.  I tried peering into the alertmanager data, but couldn't 
tell if prometheus is sending EndsAt.  resolve_timeout seems to have no 
effect.  It always seems to send 5m after the active alert is triggered.

######### alertmanager version #########

root@alertmanager1:/etc/alertmanager# alertmanager --version
alertmanager, version 0.20.0 (branch: HEAD, revision: 
f74be0400a6243d10bb53812d6fa408ad71ff32d)
  build user:       root@00c3106655f8
  build date:       20191211-14:13:14
  go version:       go1.13.5

######### alertmanager.yml #########
global:
  resolve_timeout: 15s

templates:
- '/etc/alertmanager/*.tmpl'

route:
  repeat_interval: 1h
  receiver: critical

receivers:
- name: 'critical'
  slack_configs:
  - api_url: https://hooks.slack.com/services/XXXXXXX/XXXXXXXX/XXXXXXXX
    channel: '#alerts'
    send_resolved: true
    title: '{{ template "title" . }}'
    text: '{{ template "slack_message" . }}'

######### prometheus alerts config #########
groups:
- name: example
  rules:
  # Alert for any instance that is unreachable
  - alert: InstanceDown
    expr: up == 0
    labels:
      severity: page
    annotations:
      summary: "Instance {{$labels.instance}} down"
      description: "{{$labels.instance}} is down!"


######### alertmanager debug logs #########
Apr 16 13:13:06 alertmanager1-mon-prd-cle alertmanager[28061]: level=debug 
ts=2020-04-16T13:13:06.458Z caller=dispatch.go:135 component=dispatcher 
msg="Received alert" alert=InstanceDown[bd6501b][active]
Apr 16 13:13:06 alertmanager1-mon-prd-cle alertmanager[28061]: level=debug 
ts=2020-04-16T13:13:06.458Z caller=dispatch.go:465 component=dispatcher 
aggrGroup={}:{} msg=flushing alerts=[InstanceDown[bd6501b][active]]
Apr 16 13:14:36 alertmanager1-mon-prd-cle alertmanager[28061]: level=debug 
ts=2020-04-16T13:14:36.456Z caller=dispatch.go:135 component=dispatcher 
msg="Received alert" alert=InstanceDown[bd6501b][active]
Apr 16 13:15:36 alertmanager1-mon-prd-cle alertmanager[28061]: level=debug 
ts=2020-04-16T13:15:36.456Z caller=dispatch.go:135 component=dispatcher 
msg="Received alert" alert=InstanceDown[bd6501b][resolved]
Apr 16 13:17:06 alertmanager1-mon-prd-cle alertmanager[28061]: level=debug 
ts=2020-04-16T13:17:06.455Z caller=dispatch.go:135 component=dispatcher 
msg="Received alert" alert=InstanceDown[bd6501b][resolved]
Apr 16 13:18:06 alertmanager1-mon-prd-cle alertmanager[28061]: level=debug 
ts=2020-04-16T13:18:06.458Z caller=dispatch.go:465 component=dispatcher 
aggrGroup={}:{} msg=flushing alerts=[InstanceDown[bd6501b][resolved]]

-- 
You received this message because you are subscribed to the Google Groups 
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/prometheus-users/1291a5cb-ed7e-45ff-aa69-2430dbec0208%40googlegroups.com.

Reply via email to