Another query to try:
topk(10, scrape_samples_scraped)

On Friday, 28 July 2023 at 09:53:00 UTC+1 Ben Kochie wrote:

> That's 7 billion metrics, which would require approximately  30-50TiB of 
> ram.
>
> On Thu, Jul 27, 2023 at 5:50 PM Brian Candler <[email protected]> wrote:
>
>> As Stuart says, that looks correct, assuming your metrics don't have any 
>> labels other than the ones you've excluded. You'd save a lot of typing just 
>> by doing:
>>
>>     sum(scrape_samples_scraped)
>>
>> which is expected to return a single value, with no labels (as it's 
>> summed across all timeseries of this metric).
>>
>> The value 7,525,871,918 does seem quite high - what was it before?  You 
>> can set an execution time for this query in the PromQL browser, or draw a 
>> graph this expression over time, to see historical values.
>>
>> You could also look at
>>     count(scrape_samples_scraped)
>>
>> or more simply
>>     count(up)
>>
>> and see if that has jumped up: it would imply that lots more targets have 
>> been added (e.g. more pods are being monitored).
>>
>> If not, then as well as Stuart's suggestion of graphing 
>> "scrape_samples_scraped" by itself to see if one particular target is 
>> generating way more metrics than usual, you could try different summary 
>> variants like
>>
>> sum by (instance,job) (scrape_samples_scraped)
>> sum by (clusterName) (scrape_samples_scraped)
>> ... etc
>>
>> and see if there's a spike in any of these.  This may help you drill down 
>> to the offending item(s).
>>
>> On Thursday, 27 July 2023 at 15:51:24 UTC+1 Uvais Ibrahim wrote:
>>
>>> Hi Brain,
>>>
>>> This is the query that I have used.
>>>
>>> sum(scrape_samples_scraped)without(app,app_kubernetes_io_managed_by,clusterName,release,environment,instance,job,k8s_cluster,kubernetes_name,kubernetes_namespace,ou,app_kubernetes_io_component,app_kubernetes_io_name,app_kubernetes_io_version,kustomize_toolkit_fluxcd_io_name,kustomize_toolkit_fluxcd_io_namespace,application,name,role,app_kubernetes_io_instance,app_kubernetes_io_part_of,control_plane,beta_kubernetes_io_arch,beta_kubernetes_io_instance_type,
>>>  
>>> beta_kubernetes_io_os, failure_domain_beta_kubernetes_io_region, 
>>> failure_domain_beta_kubernetes_io_zone,kubernetes_io_arch, 
>>> kubernetes_io_hostname, kubernetes_io_os, node_kubernetes_io_instance_type, 
>>> nodegroup, topology_kubernetes_io_region, 
>>> topology_kubernetes_io_zone,chart,heritage,revised,transit,component,namespace,
>>>  
>>> pod_name, pod_template_hash, security_istio_io_tlsMode, 
>>> service_istio_io_canonical_name, 
>>> service_istio_io_canonical_revision,k8s_app,kubernetes_io_cluster_service,kubernetes_io_name,route_reflector)
>>>
>>> Which simply excluded every label but still I am getting a result like 
>>> this
>>>
>>> {}  7525871918
>>>
>>>
>>> It shouldn't return any results right?
>>>
>>> Prometheus version: 2.36.2
>>>
>>> By increased traffic I meant that, the prometheus servers are getting 
>>> high traffic from a specific point of time. Currently prometheus is getting 
>>> 13 million packets earlier it was like 2 to 3 M packets on an average. And 
>>> the prometheus endpoint is not public.
>>>
>>>
>>> On Thursday, July 27, 2023 at 6:06:10 PM UTC+5:30 Brian Candler wrote:
>>>
>>>> scrape_samples_scraped always has the labels which prometheus itself 
>>>> adds (i.e. job and instance).
>>>>
>>>> Extraordinary claims require extraordinary evidence. Are you saying 
>>>> that the PromQL query *scrape_samples_scraped{job="",instance=""}* 
>>>> returns a result?  If so, what's the number?  What do you mean by "with 
>>>> increased size" - increased as compared to what? And what version of 
>>>> prometheus are you running?
>>>>
>>>> In any case, what you see with scrape_samples_scraped may be completely 
>>>> unrelated to the "high traffic" issue.  Is your prometheus server exposed 
>>>> to the Internet? Maybe someone is accessing it remotely.  Even if not, you 
>>>> can use packet capture to work out where the traffic is going to and from. 
>>>>  
>>>> A tool like https://www.sniffnet.net/ may be helpful.
>>>>
>>>> On Thursday, 27 July 2023 at 13:14:25 UTC+1 Uvais Ibrahim wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> Since last night, my Prometheus EC2 servers are getting high traffic 
>>>>> unusually. When I was checking in Prometheus I can see this 
>>>>> metric scrape_samples_scraped with with increased size but without any 
>>>>> labels. What could be the reason?
>>>>>
>>>>>
>>>>> Thanks,
>>>>> Uvais Ibrahim
>>>>>
>>>>>
>>>>>
>>>>> -- 
>> You received this message because you are subscribed to the Google Groups 
>> "Prometheus Users" group.
>> To unsubscribe from this group and stop receiving emails from it, send an 
>> email to [email protected].
>> To view this discussion on the web visit 
>> https://groups.google.com/d/msgid/prometheus-users/811fba5c-1bd3-4677-b276-84116180a1acn%40googlegroups.com
>>  
>> <https://groups.google.com/d/msgid/prometheus-users/811fba5c-1bd3-4677-b276-84116180a1acn%40googlegroups.com?utm_medium=email&utm_source=footer>
>> .
>>
>

-- 
You received this message because you are subscribed to the Google Groups 
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/prometheus-users/96aeb5fc-068d-4264-9c5b-f017a44188dan%40googlegroups.com.

Reply via email to