Is this proxying to more than one backend? And in dev, there is only one?

/MR

On Tue, Nov 17, 2020 at 2:07 PM bruno bourdolle <[email protected]>
wrote:

> hi,
> I localise the part of the conf that generate the error, but I'm not sure
> to understand how to correct it. What is strange is that the same conf on
> dev env works without these warns
>
> - job_name: 'my-proxy'
>           metrics_path: '/metrics'
>           scheme: https
>           tls_config:
>             ca_file: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt
>             insecure_skip_verify: true
>           bearer_token_file: /var/run/secrets/
> kubernetes.io/serviceaccount/token
>           static_configs:
>             - targets:
>               - '${SCRAPE_CONFIGS_TARGET}'
>
> on target , I see
>
>
> https://userprom.xxxt-intraxxx:443/metrics
> <https://userprom.kermit-eu-b.itn.intraorange:443/metrics>
>
> up
> group="production" instance="userprom.xxx.intraxxx:443" job="my-proxy"
>
> on dev no warns
> on prod a lot of warns
>
> if I remove this block, I works without warns on prod , others conf are ok
> but If I had this one, I have warns on prod
>
> Any idea to correct and explain why, I'm a newbie
>
> Le lundi 16 novembre 2020 à 16:48:25 UTC+1, [email protected] a
> écrit :
>
>> This can happen in a few ways:
>>
>> 1. whatever exports the metrics, does so with a timestamp, but actually
>> changes the value on you between scrapes without updating the timestamp.
>> This is relatively unlikely unless this is something very specialized.
>> 1.1 or it actually exposes the same metric twice within one /metrics
>> response – but the official client libraries actively prevent that
>> 2. (more likely) after all relabeling, you end up with metrics from
>> multiple targets without any distinguishing labels between them.
>>
>> Are you scraping through some kind of proxy? In general, Prometheus
>> expects to discover and access each target individually, so that it can
>> separately collect the data from all of them.
>>
>> Look at the raw metrics endpoints that Prometheus would scrape – if they
>> do not have duplicates or timestamps, it must be 2. Remove label drop or
>> replacement rules that may accidentally coalesce multiple targets into
>> having the same labels, especially if you are messing with the "instance"
>> label. Looking at Prometheus' target page may also help in identifying
>> targets that have the exact same label set.
>>
>> /MR
>>
>>
>>
>> On Mon, Nov 16, 2020 at 2:35 PM bruno bourdolle <[email protected]>
>> wrote:
>>
>>> hello,
>>>
>>> I don(t understand how to correct the WARN about error on ingesting
>>> samples...
>>> I'm into kubernetes
>>> I upgrade at last version of prometheus
>>> I cleared the datas
>>> Each time I start I have this line of warning each second
>>>
>>> Any idea ?
>>>
>>> bets
>>> bruno
>>>
>>> level=info ts=2020-11-16T14:29:36.045Z caller=main.go:353 msg="Starting
>>> Prometheus" version="(version=2.22.1, branch=HEAD,
>>> revision=00f16d1ac3a4c94561e5133b821d8e4d9ef78ec2)"
>>> level=info ts=2020-11-16T14:29:36.046Z caller=main.go:358
>>> build_context="(go=go1.15.3, user=root@516b109b1732,
>>> date=20201105-14:02:25)"
>>> level=info ts=2020-11-16T14:29:36.046Z caller=main.go:359
>>> host_details="(Linux 3.10.0-957.1.3.el7.x86_64 #1 SMP Thu Nov 15 17:36:42
>>> UTC 2018 x86_64 prometheus-0 (none))"
>>> level=info ts=2020-11-16T14:29:36.046Z caller=main.go:360
>>> fd_limits="(soft=1048576, hard=1048576)"
>>> level=info ts=2020-11-16T14:29:36.046Z caller=main.go:361
>>> vm_limits="(soft=unlimited, hard=unlimited)"
>>> level=info ts=2020-11-16T14:29:36.050Z caller=web.go:516 component=web
>>> msg="Start listening for connections" address=0.0.0.0:9090
>>> level=info ts=2020-11-16T14:29:36.050Z caller=main.go:712 msg="Starting
>>> TSDB ..."
>>> level=info ts=2020-11-16T14:29:36.058Z caller=head.go:642 component=tsdb
>>> msg="Replaying on-disk memory mappable chunks if any"
>>> level=info ts=2020-11-16T14:29:36.093Z caller=head.go:656 component=tsdb
>>> msg="On-disk memory mappable chunks replay completed" duration=35.773378ms
>>> level=info ts=2020-11-16T14:29:36.093Z caller=head.go:662 component=tsdb
>>> msg="Replaying WAL, this may take a while"
>>> level=info ts=2020-11-16T14:29:37.850Z caller=head.go:714 component=tsdb
>>> msg="WAL segment loaded" segment=0 maxSegment=17
>>> level=info ts=2020-11-16T14:29:38.244Z caller=head.go:714 component=tsdb
>>> msg="WAL segment loaded" segment=1 maxSegment=17
>>> level=info ts=2020-11-16T14:29:39.839Z caller=head.go:714 component=tsdb
>>> msg="WAL segment loaded" segment=2 maxSegment=17
>>> level=info ts=2020-11-16T14:29:39.839Z caller=head.go:714 component=tsdb
>>> msg="WAL segment loaded" segment=3 maxSegment=17
>>> level=info ts=2020-11-16T14:29:39.839Z caller=head.go:714 component=tsdb
>>> msg="WAL segment loaded" segment=4 maxSegment=17
>>> level=info ts=2020-11-16T14:29:39.839Z caller=head.go:714 component=tsdb
>>> msg="WAL segment loaded" segment=5 maxSegment=17
>>> level=info ts=2020-11-16T14:29:39.840Z caller=head.go:714 component=tsdb
>>> msg="WAL segment loaded" segment=6 maxSegment=17
>>> level=info ts=2020-11-16T14:29:39.840Z caller=head.go:714 component=tsdb
>>> msg="WAL segment loaded" segment=7 maxSegment=17
>>> level=info ts=2020-11-16T14:29:39.840Z caller=head.go:714 component=tsdb
>>> msg="WAL segment loaded" segment=8 maxSegment=17
>>> level=info ts=2020-11-16T14:29:39.840Z caller=head.go:714 component=tsdb
>>> msg="WAL segment loaded" segment=9 maxSegment=17
>>> level=info ts=2020-11-16T14:29:39.840Z caller=head.go:714 component=tsdb
>>> msg="WAL segment loaded" segment=10 maxSegment=17
>>> level=info ts=2020-11-16T14:29:39.840Z caller=head.go:714 component=tsdb
>>> msg="WAL segment loaded" segment=11 maxSegment=17
>>> level=info ts=2020-11-16T14:29:39.841Z caller=head.go:714 component=tsdb
>>> msg="WAL segment loaded" segment=12 maxSegment=17
>>> level=info ts=2020-11-16T14:29:39.841Z caller=head.go:714 component=tsdb
>>> msg="WAL segment loaded" segment=13 maxSegment=17
>>> level=info ts=2020-11-16T14:29:39.841Z caller=head.go:714 component=tsdb
>>> msg="WAL segment loaded" segment=14 maxSegment=17
>>> level=info ts=2020-11-16T14:29:39.841Z caller=head.go:714 component=tsdb
>>> msg="WAL segment loaded" segment=15 maxSegment=17
>>> level=info ts=2020-11-16T14:29:39.842Z caller=head.go:714 component=tsdb
>>> msg="WAL segment loaded" segment=16 maxSegment=17
>>> level=info ts=2020-11-16T14:29:39.842Z caller=head.go:714 component=tsdb
>>> msg="WAL segment loaded" segment=17 maxSegment=17
>>> level=info ts=2020-11-16T14:29:39.842Z caller=head.go:719 component=tsdb
>>> msg="WAL replay completed" checkpoint_replay_duration=97.404µs
>>> wal_replay_duration=3.748284098s total_replay_duration=3.78422202s
>>> level=info ts=2020-11-16T14:29:39.944Z caller=main.go:732
>>> fs_type=XFS_SUPER_MAGIC
>>> level=info ts=2020-11-16T14:29:39.944Z caller=main.go:735 msg="TSDB
>>> started"
>>> level=info ts=2020-11-16T14:29:39.944Z caller=main.go:861 msg="Loading
>>> configuration file" filename=/etc/prometheus/prometheus.yml
>>> level=info ts=2020-11-16T14:29:39.945Z caller=kubernetes.go:263
>>> component="discovery manager scrape" discovery=kubernetes msg="Using pod
>>> service account via in-cluster config"
>>> level=info ts=2020-11-16T14:29:39.947Z caller=kubernetes.go:263
>>> component="discovery manager scrape" discovery=kubernetes msg="Using pod
>>> service account via in-cluster config"
>>> level=info ts=2020-11-16T14:29:40.030Z caller=main.go:892 msg="Completed
>>> loading of configuration file" filename=/etc/prometheus/prometheus.yml
>>> totalDuration=85.979364ms remote_storage=5.031µs web_handler=653ns
>>> query_engine=1.227µs scrape=250.298µs scrape_sd=2.56867ms notify=33.201µs
>>> notify_sd=11.894µs rules=82.217333ms
>>> level=info ts=2020-11-16T14:29:40.030Z caller=main.go:684 msg="Server is
>>> ready to receive web requests."
>>> level=warn ts=2020-11-16T14:30:00.551Z caller=scrape.go:1372
>>> component="scrape manager" scrape_pool=kermit-proxy 
>>> target=https://userprom.xxx:443/metrics
>>> <https://userprom.kermit-eu-b.itn.intraorange:443/metrics> msg="Error
>>> on ingesting samples with different value but same timestamp"
>>> num_dropped=30
>>> level=warn ts=2020-11-16T14:30:11.443Z caller=scrape.go:1372
>>> component="scrape manager" scrape_pool=kermit-proxy 
>>> target=https://userprom.xxx:443/metrics
>>> <https://userprom.kermit-eu-b.itn.intraorange:443/metrics> msg="Error
>>> on ingesting samples with different value but same timestamp"
>>> num_dropped=30
>>>
>>> --
>>> You received this message because you are subscribed to the Google
>>> Groups "Prometheus Users" group.
>>> To unsubscribe from this group and stop receiving emails from it, send
>>> an email to [email protected].
>>> To view this discussion on the web visit
>>> https://groups.google.com/d/msgid/prometheus-users/16a9cc0b-b9c5-406e-8c6f-92772bf7845dn%40googlegroups.com
>>> <https://groups.google.com/d/msgid/prometheus-users/16a9cc0b-b9c5-406e-8c6f-92772bf7845dn%40googlegroups.com?utm_medium=email&utm_source=footer>
>>> .
>>>
>> --
> You received this message because you are subscribed to the Google Groups
> "Prometheus Users" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to [email protected].
> To view this discussion on the web visit
> https://groups.google.com/d/msgid/prometheus-users/c3122f62-64b8-415a-a23f-1e32c67293b3n%40googlegroups.com
> <https://groups.google.com/d/msgid/prometheus-users/c3122f62-64b8-415a-a23f-1e32c67293b3n%40googlegroups.com?utm_medium=email&utm_source=footer>
> .
>

-- 
You received this message because you are subscribed to the Google Groups 
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/prometheus-users/CAMV%3D_gbaTYqvL0-i45_%2BTmZANfvGqUHQnzrvjP0SHn6KuTXEAw%40mail.gmail.com.

Reply via email to