Correct, that is my idea

On Wed, Jun 24, 2020 at 8:28 PM [email protected] <
[email protected]> wrote:

> Hi. Thanks for such a quick response.
>
> Do you mean to say define 5 different recording rules with the same name
> for all those 5 servers?
> Also, in the recording rule now, I would have to give 2 labels correct?
> One is the instance itself and another is the component because my whole
> Memory expression is based on the component given in the recording rule.
>
> On Wednesday, June 24, 2020 at 11:52:46 PM UTC+5:30 [email protected]
> wrote:
>
>> You can redefine another recoring rule with the same metric name but a
>> different value and different label, that way one record rule can have
>> multiple values depending on labels
>>
>> On Wed, Jun 24, 2020, 20:11 [email protected] <[email protected]>
>> wrote:
>>
>>> Hi. Currently I am using a custom threshold in case of my Memory alerts.
>>> I have 2 main labels for my every node exporter target - cluster and
>>> component.
>>> My custom threshold till now has been based on the component as I had to
>>> define that particular custom threshold for all the servers of the
>>> component. But now, I have 5 instances, all from different components and I
>>> have to set the threshold as 97. How do approach this?
>>>
>>> My typical node exporter job.
>>>   - job_name: 'node_exporter_JOB-A'
>>>     static_configs:
>>>     - targets: [ 'x.x.x.x:9100' , 'x.x.x.x:9100']
>>>       labels:
>>>         cluster: 'Cluster-A'
>>>         env: 'PROD'
>>>         component: 'Comp-A'
>>>     scrape_interval: 10s
>>>
>>> Recording rule for custom thresholds.
>>>   - record: abcd_critical
>>>     expr: 99.9
>>>     labels:
>>>       component: 'Comp-A'
>>>
>>>   - record: xyz_critical
>>>     expr: 95
>>>     labels:
>>>       node: 'Comp-B'
>>>
>>> The expression for Memory Alert.
>>> ((node_memory_MemTotal_bytes - node_memory_MemFree_bytes -
>>> node_memory_Cached_bytes) / node_memory_MemTotal_bytes * 100) *
>>> on(instance) group_left(nodename) node_uname_info > on(component)
>>> group_left() (*abcd_critical* or *xyz_critical* or on(component) count
>>> by (component)((node_memory_MemTotal_bytes - node_memory_MemFree_bytes -
>>> node_memory_Cached_bytes) / node_memory_MemTotal_bytes * 100) * 0 + 90)
>>>
>>> Now, I have 5 servers with different components. How to include that in
>>> the most optimized manner?
>>>
>>> Thanks in advance.
>>>
>>> --
>>> You received this message because you are subscribed to the Google
>>> Groups "Prometheus Users" group.
>>> To unsubscribe from this group and stop receiving emails from it, send
>>> an email to [email protected].
>>> To view this discussion on the web visit
>>> https://groups.google.com/d/msgid/prometheus-users/58c7d88e-6538-4039-a5ae-4bd092cd8087n%40googlegroups.com
>>> <https://groups.google.com/d/msgid/prometheus-users/58c7d88e-6538-4039-a5ae-4bd092cd8087n%40googlegroups.com?utm_medium=email&utm_source=footer>
>>> .
>>>
>> --
> You received this message because you are subscribed to the Google Groups
> "Prometheus Users" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to [email protected].
> To view this discussion on the web visit
> https://groups.google.com/d/msgid/prometheus-users/93d1ca48-0275-4d37-b6f0-bc4eb15ffd8fn%40googlegroups.com
> <https://groups.google.com/d/msgid/prometheus-users/93d1ca48-0275-4d37-b6f0-bc4eb15ffd8fn%40googlegroups.com?utm_medium=email&utm_source=footer>
> .
>

-- 
You received this message because you are subscribed to the Google Groups 
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/prometheus-users/CALnV8Wi6SRvbmyWK_J8y3zC-3-XyFckYbVn%3DmcWk9xv%3DOcNLYg%40mail.gmail.com.

Reply via email to