That's 7 billion metrics, which would require approximately 30-50TiB of ram.
On Thu, Jul 27, 2023 at 5:50 PM Brian Candler <[email protected]> wrote: > As Stuart says, that looks correct, assuming your metrics don't have any > labels other than the ones you've excluded. You'd save a lot of typing just > by doing: > > sum(scrape_samples_scraped) > > which is expected to return a single value, with no labels (as it's summed > across all timeseries of this metric). > > The value 7,525,871,918 does seem quite high - what was it before? You > can set an execution time for this query in the PromQL browser, or draw a > graph this expression over time, to see historical values. > > You could also look at > count(scrape_samples_scraped) > > or more simply > count(up) > > and see if that has jumped up: it would imply that lots more targets have > been added (e.g. more pods are being monitored). > > If not, then as well as Stuart's suggestion of graphing > "scrape_samples_scraped" by itself to see if one particular target is > generating way more metrics than usual, you could try different summary > variants like > > sum by (instance,job) (scrape_samples_scraped) > sum by (clusterName) (scrape_samples_scraped) > ... etc > > and see if there's a spike in any of these. This may help you drill down > to the offending item(s). > > On Thursday, 27 July 2023 at 15:51:24 UTC+1 Uvais Ibrahim wrote: > >> Hi Brain, >> >> This is the query that I have used. >> >> sum(scrape_samples_scraped)without(app,app_kubernetes_io_managed_by,clusterName,release,environment,instance,job,k8s_cluster,kubernetes_name,kubernetes_namespace,ou,app_kubernetes_io_component,app_kubernetes_io_name,app_kubernetes_io_version,kustomize_toolkit_fluxcd_io_name,kustomize_toolkit_fluxcd_io_namespace,application,name,role,app_kubernetes_io_instance,app_kubernetes_io_part_of,control_plane,beta_kubernetes_io_arch,beta_kubernetes_io_instance_type, >> beta_kubernetes_io_os, failure_domain_beta_kubernetes_io_region, >> failure_domain_beta_kubernetes_io_zone,kubernetes_io_arch, >> kubernetes_io_hostname, kubernetes_io_os, node_kubernetes_io_instance_type, >> nodegroup, topology_kubernetes_io_region, >> topology_kubernetes_io_zone,chart,heritage,revised,transit,component,namespace, >> pod_name, pod_template_hash, security_istio_io_tlsMode, >> service_istio_io_canonical_name, >> service_istio_io_canonical_revision,k8s_app,kubernetes_io_cluster_service,kubernetes_io_name,route_reflector) >> >> Which simply excluded every label but still I am getting a result like >> this >> >> {} 7525871918 >> >> >> It shouldn't return any results right? >> >> Prometheus version: 2.36.2 >> >> By increased traffic I meant that, the prometheus servers are getting >> high traffic from a specific point of time. Currently prometheus is getting >> 13 million packets earlier it was like 2 to 3 M packets on an average. And >> the prometheus endpoint is not public. >> >> >> On Thursday, July 27, 2023 at 6:06:10 PM UTC+5:30 Brian Candler wrote: >> >>> scrape_samples_scraped always has the labels which prometheus itself >>> adds (i.e. job and instance). >>> >>> Extraordinary claims require extraordinary evidence. Are you saying that >>> the PromQL query *scrape_samples_scraped{job="",instance=""}* returns a >>> result? If so, what's the number? What do you mean by "with increased >>> size" - increased as compared to what? And what version of prometheus are >>> you running? >>> >>> In any case, what you see with scrape_samples_scraped may be completely >>> unrelated to the "high traffic" issue. Is your prometheus server exposed >>> to the Internet? Maybe someone is accessing it remotely. Even if not, you >>> can use packet capture to work out where the traffic is going to and from. >>> A tool like https://www.sniffnet.net/ may be helpful. >>> >>> On Thursday, 27 July 2023 at 13:14:25 UTC+1 Uvais Ibrahim wrote: >>> >>>> Hi, >>>> >>>> Since last night, my Prometheus EC2 servers are getting high traffic >>>> unusually. When I was checking in Prometheus I can see this >>>> metric scrape_samples_scraped with with increased size but without any >>>> labels. What could be the reason? >>>> >>>> >>>> Thanks, >>>> Uvais Ibrahim >>>> >>>> >>>> >>>> -- > You received this message because you are subscribed to the Google Groups > "Prometheus Users" group. > To unsubscribe from this group and stop receiving emails from it, send an > email to [email protected]. > To view this discussion on the web visit > https://groups.google.com/d/msgid/prometheus-users/811fba5c-1bd3-4677-b276-84116180a1acn%40googlegroups.com > <https://groups.google.com/d/msgid/prometheus-users/811fba5c-1bd3-4677-b276-84116180a1acn%40googlegroups.com?utm_medium=email&utm_source=footer> > . > -- You received this message because you are subscribed to the Google Groups "Prometheus Users" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To view this discussion on the web visit https://groups.google.com/d/msgid/prometheus-users/CABbyFmrvtq8QH_jVeuFdC%3DoiowNa8gVzW9CJzqQnPBRv1eA49Q%40mail.gmail.com.

