Hi!
I send the curl request and get;
alertmanager_notifications_failed_total{integration="opsgenie"} 0
alertmanager_notifications_total{integration="opsgenie"} 0
So in this case the first thing I control is the alerts I generate from
alerting rules have the 'severity: critical' or 'severity: warning' labels
on them right?
But how can I controll this?
17 Ocak 2022 Pazartesi tarihinde saat 17:44:11 UTC+3 itibarıyla Brian
Candler şunları yazdı:
> 1. Are you sure that the alerts you generate from alerting rules have the
> 'severity: critical' or 'severity: warning' labels on them? If not, they
> won't match any of the routes, so they'll fall back to the default you set:
>
> route:
> receiver: slack_general
>
> 2. Why do you have these empty routes?
>
> - match:
> severity: critical
> continue: true
> - match:
> severity: warning
> continue: true
>
> They don't do anything - delete them.
>
> 3. In order to see if alertmanager is attempting to send to opsgenie (and
> failing):
> * Look at the logs of the alertmanager proces (e.g. "journalctl -eu
> alertmanager" if running it under systemd)
> * Look at the notification metrics which alertmanager itself generates:
>
> curl -Ss localhost:9093/metrics | grep
> 'alertmanager_notifications.*opsgenie'
>
> If you see:
>
> alertmanager_notifications_failed_total{integration="opsgenie"} 0
> alertmanager_notifications_total{integration="opsgenie"} 0
>
> then no delivery to opsgenie has been attempted. If there are attempts
> and failures, you'll see these metrics going up.
>
> BTW, it's useful to scrape alertmanager from your prometheus, so you can
> query these metrics and get history of them (and indeed, alert on them if
> necessary):
>
> - job_name: alertmanager
> scrape_interval: 1m
> static_configs:
> - targets: ['localhost:9093']
>
> 4. If you want to deliver a particular alert to multiple destinations, a
> much cleaner way of doing it is to use a subtree of routes to list multiple
> destinations:
>
> - match:
> severity: critical
> routes: [ {receiver: slack_general, continue: true}, {receiver:
> netmera_opsgenie} ]
>
> Then you don't have to duplicate your matching logic (in this case
> "severity: critical"), and you don't get into confusion over when to use
> "continue: true".
>
> OTOH, if you want *all* alerts to go to slack regardless, then just put a
> catch-all route at the top:
>
> route:
> receiver: slack_general # this is never used because the first rule
> below always matches
> routes:
>
> * - receiver: slack_general continue: true*
> - match:
>
> severity: critical
> receiver: 'netmera_opsgenie'
> - match:
> severity: warning
> receiver: 'netmera_opsgenie'
>
>
> 5. "match" and "match_re" are deprecated, better to start using the new
> matchers syntax:
>
> - matchers:
> - 'severity =~ "warning|critical"'
> receiver: 'netmera_opsgenie'
>
> On Monday, 17 January 2022 at 12:55:49 UTC Hasene Ceren Yıkılmaz wrote:
>
>> When I restart the alertmanager it's running, but I can't see any of
>> these alerts in OpsGenie.
>> I follow this doc : https://support.atlassian.com/
>> opsgenie/docs/integrate-opsgenie-with-prometheus/ and this doc
>> https://prometheus.io/docs/alerting/latest/configuration/#opsgenie_config
>> Is there anything to controlling about Alertmanager & OpsGenie
>> integration?
>>
>> This is my alertmanager.yml file;
>>
>> global:
>> resolve_timeout: 5m
>>
>> route:
>> receiver: slack_general
>> group_by: ['instance']
>> group_wait: 1s
>> group_interval: 1s
>> routes:
>> - match:
>> severity: critical
>> continue: true
>> receiver: slack_general
>> - match:
>> severity: warning
>> continue: true
>> receiver: slack_general
>> - match:
>> severity: critical
>> continue: true
>> - match:
>> severity: warning
>> continue: true
>> # added receivers for opsgenie
>> - match:
>> severity: critical
>> receiver: 'netmera_opsgenie'
>> - match:
>> severity: warning
>> receiver: 'netmera_opsgenie'
>>
>> receivers:
>> - name: slack_general
>> slack_configs:
>> - api_url: 'slack api url'
>> channel: '#netmera-prometheus'
>> send_resolved: true
>> title: "{{ range .Alerts }}{{ .Annotations.summary }}\n{{ end }}"
>> text: "{{ range .Alerts }}{{ .Annotations.description }}\n{{ end }}"
>> # added opsgenie configs
>> - name: 'netmera_opsgenie'
>> opsgenie_configs:
>> - api_key: opsgenie api key
>> api_url: https://api.eu.opsgenie.com/
>> message: '{{ range .Alerts }}{{ .Annotations.summary }}\n{{ end }}'
>> description: '{{ range .Alerts }}{{ .Annotations.description }}\n{{
>> end }}'
>> priority: '{{ range .Alerts }}{{ if eq .Labels.severity
>> "critical"}}P2{{else}}P3{{end}}{{end}}'
>>
>> I contacted with OpsGenie support and they checked the logs, but they
>> couldn't see anything comes from alert manager.
>> Could you please help me about that?
>>
>> Thank you!
>>
>
--
You received this message because you are subscribed to the Google Groups
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to [email protected].
To view this discussion on the web visit
https://groups.google.com/d/msgid/prometheus-users/50ca5919-8d50-4b34-a63a-e7d1313dccfcn%40googlegroups.com.