>   routes:
>   - matchers:
>      
 
alertname:['HostOutOfDiskSpace','HostHighCpuLoad','HostOutOfMemory','KubeNodeNotReady']

That's invalid: alertmanager should not even start.  I tested your config, 
and I get the following error:

ts=2023-02-27T10:17:54.702Z caller=coordinator.go:118 level=error 
component=configuration msg="Loading configuration file failed" 
file=tmp.yaml err="yaml: unmarshal errors:\n  line 22: cannot unmarshal 
!!str `alertna...` into []string"

'matchers' is a list of strings, not a map.  This should work:

route:
  routes:
  - matchers:
    - 
alertname=~"HostOutOfDiskSpace|HostHighCpuLoad|HostHighCpuLoad|KubeNodeNotReady"
  receiver: elevate_alerts

See:
https://prometheus.io/docs/alerting/latest/configuration/#matcher
https://prometheus.io/docs/alerting/latest/configuration/#example

On Sunday, 26 February 2023 at 14:53:26 UTC Sampada Thorat wrote:

> Hello Everyone,
>
> I want to receive Alerts for  
> 'HostOutOfDiskSpace','HostHighCpuLoad','HostOutOfMemory','KubeNodeNotReady' 
> alertnames in "elevate_alerts" channel and rest all other alerts in 
> " default_receiver_test" channel. But for the below configuration, I'm 
> getting all the alerts in   "elevate_alerts" only.
>
> This is my ConfigMap:
>
> apiVersion: v1
> data:
>   connectors.yaml: |
>     connectors:
>       - test: 
> https://sasoffice365.webhook.office.com/webhookb2/d2415be1-2360-49c3-af48-7baf41aa1371@b1c14d5c-3625-45b3-a430-9552373a0c2f/IncomingWebhook/c7c62c1315d24c1fb5d1c731d2467dc6/5c8c1e6c-e827-4114-a893-9a1788ad41b5
>       - alertmanager: 
> https://sasoffice365.webhook.office.com/webhookb2/a7cb86de-1543-4e6d-b927-387c1f1e35ad@b1c14d5c-3625-45b3-a430-9552373a0c2f/IncomingWebhook/687a7973ffe248d081f58d94a090fb4c/05be66ae-90eb-42f5-8e0c-9c10975012ca
> kind: ConfigMap
> metadata:
>   annotations:
>     meta.helm.sh/release-name: prometheus-msteams
>     meta.helm.sh/release-namespace: monitoring
>   creationTimestamp: "2023-02-26T12:33:36Z"
>   labels:
>     app.kubernetes.io/managed-by: Helm
>   name: prometheus-msteams-config
>   namespace: monitoring
>   resourceVersion: "18040490"
>   uid: 795c96d5-8318-4885-804f-71bba707c885
>
>
> This is my alertmanager.yaml:
>
> global:
>   resolve_timeout: 5m
> receivers:
> - name: elevate_alerts
>   webhook_configs:
>   - url: "
> http://prometheus-msteams.default.svc.cluster.local:2000/alertmanager";
>     send_resolved: true
> - name: default_receiver_test
>   webhook_configs:
>   - url: "http://prometheus-msteams.default.svc.cluster.local:2000/test";
>     send_resolved: true
> route:
>   group_by:
>   - alertname
>   - severity
>   group_interval: 5m
>   group_wait: 30s
>   repeat_interval: 3h
>   receiver: default_receiver_test
>   routes:
>   - matchers:
>       
> alertname:['HostOutOfDiskSpace','HostHighCpuLoad','HostOutOfMemory','KubeNodeNotReady']
>     receiver: elevate_alerts
>
> Please help
>
>
>
>

-- 
You received this message because you are subscribed to the Google Groups 
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to prometheus-users+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/prometheus-users/bccc531d-5c21-4de3-8371-67b7256eb148n%40googlegroups.com.

Reply via email to