I found out what the issue was.  I had the same set of rules defined in 
projects I was scraping.  I only need the rules in my main prometheus.
I remove the rules from the projects I was using federate against and the 
errors have stopped.

On Monday, February 15, 2021 at 3:45:39 PM UTC-5 Brian Beynon wrote:

> Hello,
> I recently updated our prometheus setup from using the helm chart to using 
> the prometheus-operator. 
>
> *Summary of current setup:*
> Platform:  Google Cloud
> 1.  Google Project used for monitoring:
> Prometheus Operator (prometheus,alertmanager,grafana, node-exporter, 
> kube-state-metrics...ect.)
>
> 2. Then multiple other Google Projects that now run the 
> Prometheus-Operator (node-exporter, kube-state-metrics...ect) but without 
> Alertmanager/Grafana.
>
> So the main google project (#1 above) has federate scrapes/jobs that 
> connect to each of the other google projects prometheus (#2 above).
>
> Since updating to the prometheus-operators I'm now running into these 
> errors coming from the main prometheus logs:
> *msg="Error on ingesting samples with different value but same timestamp"*
> and *msg="Error on ingesting out-of-order samples".*
>
> Below is an example of one of the federate jobs where the errors are 
> coming from. 
> When I have the job "vms" and job "node-exporter" both enabled the errors 
> occur.  If I disable either of those jobs I no longer see the errors.  
>
> - job_name: 'test-abc-123'
>   scrape_interval: 60s
>   scrape_timeout: 30s
>   honor_labels: true
>   metrics_path: '/federate'
>   scheme: 'https'
>   basic_auth:
>     username: '###################'
>     password: '###################'
>   params:
>     'match[]':
> *      - '{job="vms"} '*
> *      - '{job="node-exporter"} '*
>       - '{job="postgres"} '
>       - '{job="barman"} '
>       - '{job="apiserver"} '
>       - '{job="kube-state-metrics"} '
>   static_configs:
>     - targets:
>       - 'test-abc-123.com'
>       labels:
>         project: 'test-abc-123'
>
> Here is the node-exporter serviceMonitor from project test-abc-123:
>
> apiVersion: monitoring.coreos.com/v1
> kind: ServiceMonitor
> metadata:
>   labels:
>     app.kubernetes.io/component: exporter
>     app.kubernetes.io/name: node-exporter
>     app.kubernetes.io/part-of: kube-prometheus
>     app.kubernetes.io/version: 1.1.0
>   name: node-exporter
>   namespace: monitoring
> spec:
>   endpoints:
>   - bearerTokenFile: /var/run/secrets/kubernetes.io/serviceaccount/token
>     interval: 15s
>     port: https
>     relabelings:
>     - action: replace
>       regex: (.*)
>       replacement: $1
>       sourceLabels:
>       - __meta_kubernetes_pod_node_name
>       targetLabel: instance
>     scheme: https
>     tlsConfig:
>       insecureSkipVerify: true
>   jobLabel: app.kubernetes.io/name
>   selector:
>     matchLabels:
>       app.kubernetes.io/component: exporter
>       app.kubernetes.io/name: node-exporter
>       app.kubernetes.io/part-of: kube-prometheus
>
> Here is the "vms" job from project test-abc-123:
>     
>       - job_name: 'vms'
>         static_configs:
>           - targets: ['db-prod-1:9100','db-prod-2:9100','util-1:9100']
>             labels:
>               project: 'client-vms'
>
> I have tried updating labels but maybe not in the right way.   Any 
> suggestions or pointers would be appreciated.  
>
> Thank you
>
>
>
>
>
>
>

-- 
You received this message because you are subscribed to the Google Groups 
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/prometheus-users/26b77e64-2ae5-4364-87a0-8b05bbbe0c19n%40googlegroups.com.

Reply via email to