maybe It 's an explication but I'm not the owner of xxx-proxy......How to add a distinct labl/annotation to differentiate them ?
Le mercredi 18 novembre 2020 à 10:00:12 UTC+1, [email protected] a écrit : > Is this proxying to more than one backend? And in dev, there is only one? > > /MR > > On Tue, Nov 17, 2020 at 2:07 PM bruno bourdolle <[email protected]> > wrote: > >> hi, >> I localise the part of the conf that generate the error, but I'm not sure >> to understand how to correct it. What is strange is that the same conf on >> dev env works without these warns >> >> - job_name: 'my-proxy' >> metrics_path: '/metrics' >> scheme: https >> tls_config: >> ca_file: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt >> insecure_skip_verify: true >> bearer_token_file: /var/run/secrets/ >> kubernetes.io/serviceaccount/token >> static_configs: >> - targets: >> - '${SCRAPE_CONFIGS_TARGET}' >> >> on target , I see >> >> >> https://userprom.xxxt-intraxxx:443/metrics >> <https://userprom.kermit-eu-b.itn.intraorange:443/metrics> >> >> up >> group="production" instance="userprom.xxx.intraxxx:443" job="my-proxy" >> >> on dev no warns >> on prod a lot of warns >> >> if I remove this block, I works without warns on prod , others conf are >> ok but If I had this one, I have warns on prod >> >> Any idea to correct and explain why, I'm a newbie >> >> Le lundi 16 novembre 2020 à 16:48:25 UTC+1, [email protected] a >> écrit : >> >>> This can happen in a few ways: >>> >>> 1. whatever exports the metrics, does so with a timestamp, but actually >>> changes the value on you between scrapes without updating the timestamp. >>> This is relatively unlikely unless this is something very specialized. >>> 1.1 or it actually exposes the same metric twice within one /metrics >>> response – but the official client libraries actively prevent that >>> 2. (more likely) after all relabeling, you end up with metrics from >>> multiple targets without any distinguishing labels between them. >>> >>> Are you scraping through some kind of proxy? In general, Prometheus >>> expects to discover and access each target individually, so that it can >>> separately collect the data from all of them. >>> >>> Look at the raw metrics endpoints that Prometheus would scrape – if they >>> do not have duplicates or timestamps, it must be 2. Remove label drop or >>> replacement rules that may accidentally coalesce multiple targets into >>> having the same labels, especially if you are messing with the "instance" >>> label. Looking at Prometheus' target page may also help in identifying >>> targets that have the exact same label set. >>> >>> /MR >>> >>> >>> >>> On Mon, Nov 16, 2020 at 2:35 PM bruno bourdolle <[email protected]> >>> wrote: >>> >>>> hello, >>>> >>>> I don(t understand how to correct the WARN about error on ingesting >>>> samples... >>>> I'm into kubernetes >>>> I upgrade at last version of prometheus >>>> I cleared the datas >>>> Each time I start I have this line of warning each second >>>> >>>> Any idea ? >>>> >>>> bets >>>> bruno >>>> >>>> level=info ts=2020-11-16T14:29:36.045Z caller=main.go:353 msg="Starting >>>> Prometheus" version="(version=2.22.1, branch=HEAD, >>>> revision=00f16d1ac3a4c94561e5133b821d8e4d9ef78ec2)" >>>> level=info ts=2020-11-16T14:29:36.046Z caller=main.go:358 >>>> build_context="(go=go1.15.3, user=root@516b109b1732, >>>> date=20201105-14:02:25)" >>>> level=info ts=2020-11-16T14:29:36.046Z caller=main.go:359 >>>> host_details="(Linux 3.10.0-957.1.3.el7.x86_64 #1 SMP Thu Nov 15 17:36:42 >>>> UTC 2018 x86_64 prometheus-0 (none))" >>>> level=info ts=2020-11-16T14:29:36.046Z caller=main.go:360 >>>> fd_limits="(soft=1048576, hard=1048576)" >>>> level=info ts=2020-11-16T14:29:36.046Z caller=main.go:361 >>>> vm_limits="(soft=unlimited, hard=unlimited)" >>>> level=info ts=2020-11-16T14:29:36.050Z caller=web.go:516 component=web >>>> msg="Start listening for connections" address=0.0.0.0:9090 >>>> level=info ts=2020-11-16T14:29:36.050Z caller=main.go:712 msg="Starting >>>> TSDB ..." >>>> level=info ts=2020-11-16T14:29:36.058Z caller=head.go:642 >>>> component=tsdb msg="Replaying on-disk memory mappable chunks if any" >>>> level=info ts=2020-11-16T14:29:36.093Z caller=head.go:656 >>>> component=tsdb msg="On-disk memory mappable chunks replay completed" >>>> duration=35.773378ms >>>> level=info ts=2020-11-16T14:29:36.093Z caller=head.go:662 >>>> component=tsdb msg="Replaying WAL, this may take a while" >>>> level=info ts=2020-11-16T14:29:37.850Z caller=head.go:714 >>>> component=tsdb msg="WAL segment loaded" segment=0 maxSegment=17 >>>> level=info ts=2020-11-16T14:29:38.244Z caller=head.go:714 >>>> component=tsdb msg="WAL segment loaded" segment=1 maxSegment=17 >>>> level=info ts=2020-11-16T14:29:39.839Z caller=head.go:714 >>>> component=tsdb msg="WAL segment loaded" segment=2 maxSegment=17 >>>> level=info ts=2020-11-16T14:29:39.839Z caller=head.go:714 >>>> component=tsdb msg="WAL segment loaded" segment=3 maxSegment=17 >>>> level=info ts=2020-11-16T14:29:39.839Z caller=head.go:714 >>>> component=tsdb msg="WAL segment loaded" segment=4 maxSegment=17 >>>> level=info ts=2020-11-16T14:29:39.839Z caller=head.go:714 >>>> component=tsdb msg="WAL segment loaded" segment=5 maxSegment=17 >>>> level=info ts=2020-11-16T14:29:39.840Z caller=head.go:714 >>>> component=tsdb msg="WAL segment loaded" segment=6 maxSegment=17 >>>> level=info ts=2020-11-16T14:29:39.840Z caller=head.go:714 >>>> component=tsdb msg="WAL segment loaded" segment=7 maxSegment=17 >>>> level=info ts=2020-11-16T14:29:39.840Z caller=head.go:714 >>>> component=tsdb msg="WAL segment loaded" segment=8 maxSegment=17 >>>> level=info ts=2020-11-16T14:29:39.840Z caller=head.go:714 >>>> component=tsdb msg="WAL segment loaded" segment=9 maxSegment=17 >>>> level=info ts=2020-11-16T14:29:39.840Z caller=head.go:714 >>>> component=tsdb msg="WAL segment loaded" segment=10 maxSegment=17 >>>> level=info ts=2020-11-16T14:29:39.840Z caller=head.go:714 >>>> component=tsdb msg="WAL segment loaded" segment=11 maxSegment=17 >>>> level=info ts=2020-11-16T14:29:39.841Z caller=head.go:714 >>>> component=tsdb msg="WAL segment loaded" segment=12 maxSegment=17 >>>> level=info ts=2020-11-16T14:29:39.841Z caller=head.go:714 >>>> component=tsdb msg="WAL segment loaded" segment=13 maxSegment=17 >>>> level=info ts=2020-11-16T14:29:39.841Z caller=head.go:714 >>>> component=tsdb msg="WAL segment loaded" segment=14 maxSegment=17 >>>> level=info ts=2020-11-16T14:29:39.841Z caller=head.go:714 >>>> component=tsdb msg="WAL segment loaded" segment=15 maxSegment=17 >>>> level=info ts=2020-11-16T14:29:39.842Z caller=head.go:714 >>>> component=tsdb msg="WAL segment loaded" segment=16 maxSegment=17 >>>> level=info ts=2020-11-16T14:29:39.842Z caller=head.go:714 >>>> component=tsdb msg="WAL segment loaded" segment=17 maxSegment=17 >>>> level=info ts=2020-11-16T14:29:39.842Z caller=head.go:719 >>>> component=tsdb msg="WAL replay completed" >>>> checkpoint_replay_duration=97.404µs wal_replay_duration=3.748284098s >>>> total_replay_duration=3.78422202s >>>> level=info ts=2020-11-16T14:29:39.944Z caller=main.go:732 >>>> fs_type=XFS_SUPER_MAGIC >>>> level=info ts=2020-11-16T14:29:39.944Z caller=main.go:735 msg="TSDB >>>> started" >>>> level=info ts=2020-11-16T14:29:39.944Z caller=main.go:861 msg="Loading >>>> configuration file" filename=/etc/prometheus/prometheus.yml >>>> level=info ts=2020-11-16T14:29:39.945Z caller=kubernetes.go:263 >>>> component="discovery manager scrape" discovery=kubernetes msg="Using pod >>>> service account via in-cluster config" >>>> level=info ts=2020-11-16T14:29:39.947Z caller=kubernetes.go:263 >>>> component="discovery manager scrape" discovery=kubernetes msg="Using pod >>>> service account via in-cluster config" >>>> level=info ts=2020-11-16T14:29:40.030Z caller=main.go:892 >>>> msg="Completed loading of configuration file" >>>> filename=/etc/prometheus/prometheus.yml totalDuration=85.979364ms >>>> remote_storage=5.031µs web_handler=653ns query_engine=1.227µs >>>> scrape=250.298µs scrape_sd=2.56867ms notify=33.201µs notify_sd=11.894µs >>>> rules=82.217333ms >>>> level=info ts=2020-11-16T14:29:40.030Z caller=main.go:684 msg="Server >>>> is ready to receive web requests." >>>> level=warn ts=2020-11-16T14:30:00.551Z caller=scrape.go:1372 >>>> component="scrape manager" scrape_pool=kermit-proxy >>>> target=https://userprom.xxx:443/metrics >>>> <https://userprom.kermit-eu-b.itn.intraorange:443/metrics> msg="Error >>>> on ingesting samples with different value but same timestamp" >>>> num_dropped=30 >>>> level=warn ts=2020-11-16T14:30:11.443Z caller=scrape.go:1372 >>>> component="scrape manager" scrape_pool=kermit-proxy >>>> target=https://userprom.xxx:443/metrics >>>> <https://userprom.kermit-eu-b.itn.intraorange:443/metrics> msg="Error >>>> on ingesting samples with different value but same timestamp" >>>> num_dropped=30 >>>> >>>> -- >>>> You received this message because you are subscribed to the Google >>>> Groups "Prometheus Users" group. >>>> To unsubscribe from this group and stop receiving emails from it, send >>>> an email to [email protected]. >>>> To view this discussion on the web visit >>>> https://groups.google.com/d/msgid/prometheus-users/16a9cc0b-b9c5-406e-8c6f-92772bf7845dn%40googlegroups.com >>>> >>>> <https://groups.google.com/d/msgid/prometheus-users/16a9cc0b-b9c5-406e-8c6f-92772bf7845dn%40googlegroups.com?utm_medium=email&utm_source=footer> >>>> . >>>> >>> -- >> You received this message because you are subscribed to the Google Groups >> "Prometheus Users" group. >> To unsubscribe from this group and stop receiving emails from it, send an >> email to [email protected]. >> > To view this discussion on the web visit >> https://groups.google.com/d/msgid/prometheus-users/c3122f62-64b8-415a-a23f-1e32c67293b3n%40googlegroups.com >> >> <https://groups.google.com/d/msgid/prometheus-users/c3122f62-64b8-415a-a23f-1e32c67293b3n%40googlegroups.com?utm_medium=email&utm_source=footer> >> . >> > -- You received this message because you are subscribed to the Google Groups "Prometheus Users" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To view this discussion on the web visit https://groups.google.com/d/msgid/prometheus-users/215904ef-5523-4eac-bcb8-51a2641913ecn%40googlegroups.com.

