On Thu, Nov 21, 2019 at 03:09:30PM +0500, Илья Шипицин wrote:
> I understand. However, those patches add complexity (which might be moved
> to another dedicated tool)

those patch makes sense for heavy haproxy instances. As you might have
seen above, we are talking about > 130MB of data. So for a full scraping
every 60s or less, this is not realistic. Even the data loading might
take too much time. We had cases where loading the data on exporter side
was taking more time than the frequency of scraping, generating a
snowball effect.
It's a good pratice to avoid exporting data you know
you won't use instead of: scraping -> deleting -> aggregate
In our case we do not use most of the server metrics. It represents a
factor of 10 in terms of exported data.


Reply via email to