hi,
I localise the part of the conf that generate the error, but I'm not sure 
to understand how to correct it. What is strange is that the same conf on 
dev env works without these warns

- job_name: 'my-proxy'
          metrics_path: '/metrics'
          scheme: https
          tls_config:
            ca_file: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt
            insecure_skip_verify: true
          bearer_token_file: 
/var/run/secrets/kubernetes.io/serviceaccount/token
          static_configs:
            - targets:
              - '${SCRAPE_CONFIGS_TARGET}'
             
on target , I see


https://userprom.xxxt-intraxxx:443/metrics 
<https://userprom.kermit-eu-b.itn.intraorange:443/metrics>

up 
group="production" instance="userprom.xxx.intraxxx:443" job="my-proxy"

on dev no warns
on prod a lot of warns 

if I remove this block, I works without warns on prod , others conf are ok 
but If I had this one, I have warns on prod

Any idea to correct and explain why, I'm a newbie

Le lundi 16 novembre 2020 à 16:48:25 UTC+1, [email protected] a écrit :

> This can happen in a few ways:
>
> 1. whatever exports the metrics, does so with a timestamp, but actually 
> changes the value on you between scrapes without updating the timestamp. 
> This is relatively unlikely unless this is something very specialized.
> 1.1 or it actually exposes the same metric twice within one /metrics 
> response – but the official client libraries actively prevent that
> 2. (more likely) after all relabeling, you end up with metrics from 
> multiple targets without any distinguishing labels between them.
>
> Are you scraping through some kind of proxy? In general, Prometheus 
> expects to discover and access each target individually, so that it can 
> separately collect the data from all of them.
>
> Look at the raw metrics endpoints that Prometheus would scrape – if they 
> do not have duplicates or timestamps, it must be 2. Remove label drop or 
> replacement rules that may accidentally coalesce multiple targets into 
> having the same labels, especially if you are messing with the "instance" 
> label. Looking at Prometheus' target page may also help in identifying 
> targets that have the exact same label set.
>
> /MR
>
>
>
> On Mon, Nov 16, 2020 at 2:35 PM bruno bourdolle <[email protected]> 
> wrote:
>
>> hello,
>>
>> I don(t understand how to correct the WARN about error on ingesting 
>> samples...
>> I'm into kubernetes
>> I upgrade at last version of prometheus
>> I cleared the datas
>> Each time I start I have this line of warning each second
>>
>> Any idea ?
>>
>> bets
>> bruno
>>
>> level=info ts=2020-11-16T14:29:36.045Z caller=main.go:353 msg="Starting 
>> Prometheus" version="(version=2.22.1, branch=HEAD, 
>> revision=00f16d1ac3a4c94561e5133b821d8e4d9ef78ec2)" 
>> level=info ts=2020-11-16T14:29:36.046Z caller=main.go:358 
>> build_context="(go=go1.15.3, user=root@516b109b1732, 
>> date=20201105-14:02:25)" 
>> level=info ts=2020-11-16T14:29:36.046Z caller=main.go:359 
>> host_details="(Linux 3.10.0-957.1.3.el7.x86_64 #1 SMP Thu Nov 15 17:36:42 
>> UTC 2018 x86_64 prometheus-0 (none))" 
>> level=info ts=2020-11-16T14:29:36.046Z caller=main.go:360 
>> fd_limits="(soft=1048576, hard=1048576)" 
>> level=info ts=2020-11-16T14:29:36.046Z caller=main.go:361 
>> vm_limits="(soft=unlimited, hard=unlimited)" 
>> level=info ts=2020-11-16T14:29:36.050Z caller=web.go:516 component=web 
>> msg="Start listening for connections" address=0.0.0.0:9090 
>> level=info ts=2020-11-16T14:29:36.050Z caller=main.go:712 msg="Starting 
>> TSDB ..." 
>> level=info ts=2020-11-16T14:29:36.058Z caller=head.go:642 component=tsdb 
>> msg="Replaying on-disk memory mappable chunks if any" 
>> level=info ts=2020-11-16T14:29:36.093Z caller=head.go:656 component=tsdb 
>> msg="On-disk memory mappable chunks replay completed" duration=35.773378ms 
>> level=info ts=2020-11-16T14:29:36.093Z caller=head.go:662 component=tsdb 
>> msg="Replaying WAL, this may take a while" 
>> level=info ts=2020-11-16T14:29:37.850Z caller=head.go:714 component=tsdb 
>> msg="WAL segment loaded" segment=0 maxSegment=17 
>> level=info ts=2020-11-16T14:29:38.244Z caller=head.go:714 component=tsdb 
>> msg="WAL segment loaded" segment=1 maxSegment=17 
>> level=info ts=2020-11-16T14:29:39.839Z caller=head.go:714 component=tsdb 
>> msg="WAL segment loaded" segment=2 maxSegment=17 
>> level=info ts=2020-11-16T14:29:39.839Z caller=head.go:714 component=tsdb 
>> msg="WAL segment loaded" segment=3 maxSegment=17 
>> level=info ts=2020-11-16T14:29:39.839Z caller=head.go:714 component=tsdb 
>> msg="WAL segment loaded" segment=4 maxSegment=17 
>> level=info ts=2020-11-16T14:29:39.839Z caller=head.go:714 component=tsdb 
>> msg="WAL segment loaded" segment=5 maxSegment=17 
>> level=info ts=2020-11-16T14:29:39.840Z caller=head.go:714 component=tsdb 
>> msg="WAL segment loaded" segment=6 maxSegment=17 
>> level=info ts=2020-11-16T14:29:39.840Z caller=head.go:714 component=tsdb 
>> msg="WAL segment loaded" segment=7 maxSegment=17 
>> level=info ts=2020-11-16T14:29:39.840Z caller=head.go:714 component=tsdb 
>> msg="WAL segment loaded" segment=8 maxSegment=17 
>> level=info ts=2020-11-16T14:29:39.840Z caller=head.go:714 component=tsdb 
>> msg="WAL segment loaded" segment=9 maxSegment=17 
>> level=info ts=2020-11-16T14:29:39.840Z caller=head.go:714 component=tsdb 
>> msg="WAL segment loaded" segment=10 maxSegment=17 
>> level=info ts=2020-11-16T14:29:39.840Z caller=head.go:714 component=tsdb 
>> msg="WAL segment loaded" segment=11 maxSegment=17 
>> level=info ts=2020-11-16T14:29:39.841Z caller=head.go:714 component=tsdb 
>> msg="WAL segment loaded" segment=12 maxSegment=17 
>> level=info ts=2020-11-16T14:29:39.841Z caller=head.go:714 component=tsdb 
>> msg="WAL segment loaded" segment=13 maxSegment=17 
>> level=info ts=2020-11-16T14:29:39.841Z caller=head.go:714 component=tsdb 
>> msg="WAL segment loaded" segment=14 maxSegment=17 
>> level=info ts=2020-11-16T14:29:39.841Z caller=head.go:714 component=tsdb 
>> msg="WAL segment loaded" segment=15 maxSegment=17 
>> level=info ts=2020-11-16T14:29:39.842Z caller=head.go:714 component=tsdb 
>> msg="WAL segment loaded" segment=16 maxSegment=17 
>> level=info ts=2020-11-16T14:29:39.842Z caller=head.go:714 component=tsdb 
>> msg="WAL segment loaded" segment=17 maxSegment=17 
>> level=info ts=2020-11-16T14:29:39.842Z caller=head.go:719 component=tsdb 
>> msg="WAL replay completed" checkpoint_replay_duration=97.404µs 
>> wal_replay_duration=3.748284098s total_replay_duration=3.78422202s 
>> level=info ts=2020-11-16T14:29:39.944Z caller=main.go:732 
>> fs_type=XFS_SUPER_MAGIC 
>> level=info ts=2020-11-16T14:29:39.944Z caller=main.go:735 msg="TSDB 
>> started" 
>> level=info ts=2020-11-16T14:29:39.944Z caller=main.go:861 msg="Loading 
>> configuration file" filename=/etc/prometheus/prometheus.yml 
>> level=info ts=2020-11-16T14:29:39.945Z caller=kubernetes.go:263 
>> component="discovery manager scrape" discovery=kubernetes msg="Using pod 
>> service account via in-cluster config" 
>> level=info ts=2020-11-16T14:29:39.947Z caller=kubernetes.go:263 
>> component="discovery manager scrape" discovery=kubernetes msg="Using pod 
>> service account via in-cluster config" 
>> level=info ts=2020-11-16T14:29:40.030Z caller=main.go:892 msg="Completed 
>> loading of configuration file" filename=/etc/prometheus/prometheus.yml 
>> totalDuration=85.979364ms remote_storage=5.031µs web_handler=653ns 
>> query_engine=1.227µs scrape=250.298µs scrape_sd=2.56867ms notify=33.201µs 
>> notify_sd=11.894µs rules=82.217333ms 
>> level=info ts=2020-11-16T14:29:40.030Z caller=main.go:684 msg="Server is 
>> ready to receive web requests." 
>> level=warn ts=2020-11-16T14:30:00.551Z caller=scrape.go:1372 
>> component="scrape manager" scrape_pool=kermit-proxy 
>> target=https://userprom.xxx:443/metrics 
>> <https://userprom.kermit-eu-b.itn.intraorange:443/metrics> msg="Error on 
>> ingesting samples with different value but same timestamp" num_dropped=30 
>> level=warn ts=2020-11-16T14:30:11.443Z caller=scrape.go:1372 
>> component="scrape manager" scrape_pool=kermit-proxy 
>> target=https://userprom.xxx:443/metrics 
>> <https://userprom.kermit-eu-b.itn.intraorange:443/metrics> msg="Error on 
>> ingesting samples with different value but same timestamp" num_dropped=30 
>>
>> -- 
>> You received this message because you are subscribed to the Google Groups 
>> "Prometheus Users" group.
>> To unsubscribe from this group and stop receiving emails from it, send an 
>> email to [email protected].
>> To view this discussion on the web visit 
>> https://groups.google.com/d/msgid/prometheus-users/16a9cc0b-b9c5-406e-8c6f-92772bf7845dn%40googlegroups.com
>>  
>> <https://groups.google.com/d/msgid/prometheus-users/16a9cc0b-b9c5-406e-8c6f-92772bf7845dn%40googlegroups.com?utm_medium=email&utm_source=footer>
>> .
>>
>

-- 
You received this message because you are subscribed to the Google Groups 
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/prometheus-users/c3122f62-64b8-415a-a23f-1e32c67293b3n%40googlegroups.com.

Reply via email to