3. Add labels that identify metrics that belong to each deployment. Have separate remote_write sections for each remote_write target with its own bearer token. Use write_relabel_configs in each section to filter the writes to each target, so that they only include metrics with the appropriate label.
On Friday, 20 November 2020 at 14:11:14 UTC [email protected] wrote: > I'm setting up a kubernetes cluster for user. They need to deploy > different deployment, each of them have their own metrics for > pods/deployment/service, and also have its own bear_token to write to > remote storage. It's something like a big company user, own the cluster, > with multiple department , each department owns a seprate bear_token and > remote storage. > > What's the best way to set a prometheus cluster to scrape the metrics and > send them to remote storage? > > I can think of two way: > > > 1. set up a prometheus pod for each deployment > 1. pro: easy and safe, different deployments have different token > and remote storage > 2. con: bad performance?Not quite sure. Each of prometheus will > collect all metrics from cadvisors. If there are many deployments, with > many prometheus, each of them will scrape the whole cluster metrics, it > may > lead to high pressure of cluster > 2. set up an HA prometheus for a whole cluster and wrote a own program > to scrape its own metrics and send it to remote storage > 1. pro: seems lighter? not quite sure. It seems also does > something similar to prometheus > 2. con: need a lot of code to implement the scrape and send metrics > part > > Does anyone have any suggestions? Of are there any other better solutions? > -- You received this message because you are subscribed to the Google Groups "Prometheus Users" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To view this discussion on the web visit https://groups.google.com/d/msgid/prometheus-users/789bf088-484f-47b4-947b-b7d6d0e22e3cn%40googlegroups.com.

