On Tuesday, 15 September 2020 06:11:48 UTC+1, Nick wrote:
>
> Keeping cardinality explosion in mind, what's a decent maximum number of
> exported metrics that can be considered performant for scraping and
> time-series processing?
It depends on how much resource you're prepared to throw at it. If you
build a powerful prometheus server then a total of 2 million timeseries is
doable; beyond that you ought to look at sharding across multiple servers.
As I mainly need the counter total, I can split the web analytics to reduce
> the number of possible label combinations, for example:
>
> { domain, page }
> { domain, browser }
>
Yes that's fine, but you still want to limit the number of values for each
label. As Stuart said: in the case of browser you don't want the raw
User-Agent header, but pick between a small pre-determined set of values
like "Firefox", "Chrome", "IE", "Other". In the case of "page" strip out
any query string, and ideally also limit to known pages or "Other".
If you also want to record the raw User-Agent values for every request then
do that in a separate logging system (e.g. loki, elasticsearch, a SQL
database, or even just plain text files)
--
You received this message because you are subscribed to the Google Groups
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to [email protected].
To view this discussion on the web visit
https://groups.google.com/d/msgid/prometheus-users/1dd2911d-2347-4e21-928c-ce05ffc93a4ao%40googlegroups.com.