On 02/06/2021 11:01, nina guo wrote:
So the better solution would be mount another storage rather than NFS separately to each Pod.
Yes.
For example, 2 Prometheus Pods are running with 2 separate volumes, if one of the Pod goes down(but the data is still in memory), according to k8s mechanism, another Pod will be started automatically. Currently the data which was in memeroy will be lost. It will cause data inconsistency. Because the other running Pod probably already have written the data to persistent volume.

The two instances will record different data to each other, as they will be scraping the common set of targets at slightly different times. The usual way to handle this (as well as deal with gaps due to restarts/errors) is to use a "sticky" load balancer or something like Thanos or promxy.

Anything which is purely in memory and not written to disk when Prometheus crashes (or is forceable destroyed) will be lost, but Prometheus tries to write to the WAL regularly to reduce that risk. When restarted (assuming the same storage is reattached) the WAL is read back into memory.

--
Stuart Clark

--
You received this message because you are subscribed to the Google Groups 
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/prometheus-users/e202144c-c1e5-f11e-6221-c4bd88791227%40Jahingo.com.

Reply via email to