Thanks, Ben.

What happens to Alerting in case of HA Prometheus while using 
Thanos/VictoriaMetrics/Cortex on top of 2 Prometheus instances?

On Monday, October 12, 2020 at 12:57:17 AM UTC+5:30 [email protected] wrote:

> I don't think this is something you can or should be optimizing for. You 
> are on the edge of needing to shard, which means you will need to manage 
> many individual instance disks.
>
> But if you really want to have a single big disk for storage, you can use 
> Minio[0] for simple object storage if you aren't already on a platform that 
> provides object storage.
>
> [0]: https://min.io/
>
> On Sun, Oct 11, 2020, 18:29 [email protected] <[email protected]> 
> wrote:
>
>> Thanks, Ben.
>>
>> One thing. I don't want to maintain 2 5TB disks for Prometheus HA, i.e 1 
>> 5TB disk on each instance, that is why I want to put a single huge disk in 
>> my VictoriaMetrics instance and maintain a single persistent disk rather 
>> than 2. Can Thanos also store the data in a persistent disk rather than 
>> Object disk? Because from the docs I have seen till now, I haven't found 
>> this feature in Thanos yet. This is the sole reason I am inclined towards 
>> VictoriaMetrics.
>>
>> On Sunday, October 11, 2020 at 6:46:56 PM UTC+5:30 [email protected] 
>> wrote:
>>
>>> On Sun, Oct 11, 2020 at 10:45 AM [email protected] <
>>> [email protected]> wrote:
>>>
>>>> On Saturday, October 10, 2020 at 2:32:14 PM UTC+5:30 [email protected] 
>>>> wrote:
>>>> 4.6TB for 50 days seems like a lot. How many metrics and how many 
>>>> samples per second are you collecting? Just estimating based on the data, 
>>>> it sounds like you might have more than 10 million series and 600-700 
>>>> samples per second. This might be the time to start thinking about 
>>>> sharding.
>>>> You can check for sure with these queries:
>>>> prometheus_tsdb_head_series
>>>> rate(prometheus_tsdb_head_samples_appended_total[1h]
>>>> >>>>
>>>> Hi Ben, my time series collection hasn't touched 10 million yet, its 
>>>> around 5.5 million as of now, but my sampling rate is quite steep, sitting 
>>>> at approximately 643522. Since my time series are quite manageable by a 
>>>> single Prometheus instance I am avoiding sharding as of now because it 
>>>> would complicate the entire setup. What is your thought on this? 
>>>>
>>>
>>> I usually recommend thinking about a sharding plan when you hit this 
>>> level. You don't need to yet, but it's worth thinking about how you would.
>>>  
>>>
>>>>
>>>> For handling HA clustering and sharding, I recommend looking into 
>>>> Thanos. It can be added to your existing Prometheus and rolled out 
>>>> incrementally.
>>>> >>>>
>>>> Yes, I looked at Thanos but my only problem is that Thanos will use 
>>>> Object Storage for long time retention which will have latency while 
>>>> extracting old data. That is why I am inclined towards VictoriaMetrics. 
>>>> What's your view on going with VictoriaMetrics?
>>>>
>>>
>>> You don't need to use Thanos for long-term storage. It works just fine 
>>> as a query-proxy only setup.  This is how we got into using Thanos. We had 
>>> an existing sharded fleet of Prometheus HA instances. We had been using 
>>> multiple Grafana data sources and simple nginx reverse proxy for HA 
>>> querying. We added Thanos Query/Sidecar just to provide a single query 
>>> interface. It wasn't until some time later that we started to use object 
>>> storage.
>>>
>>> Thanos object storage is optional, it can use Prometheus TSDB as the 
>>> backend.
>>>
>>> That said, Thanos object storage latency isn't a huge problem. It does 
>>> depend a bit on what object storage provider/software you use. But it works 
>>> just fine.
>>>
>>> I don't recommend VictoriaMetrics. I would go with Thanos or Cortex, as 
>>> these are maintained by core Prometheus community contributors.
>>>  
>>>
>>>>
>>>> > d) d) If we do use 2 separate disks for the 2 instances, how will we 
>>>> manage the config files?
>>>> If you don't have any configuration management, I recommend using 
>>>> https://github.com/cloudalchemy/ansible-prometheus. It's very easy to 
>>>> get going.
>>>> >>>>
>>>> Thanks. I'll check it out.
>>>>
>>>> -- 
>>>>
>>> You received this message because you are subscribed to the Google 
>>>> Groups "Prometheus Users" group.
>>>> To unsubscribe from this group and stop receiving emails from it, send 
>>>> an email to [email protected].
>>>>
>>> To view this discussion on the web visit 
>>>> https://groups.google.com/d/msgid/prometheus-users/6cfb736c-ba38-4e8d-8468-cdc84f2971f2n%40googlegroups.com
>>>>  
>>>> <https://groups.google.com/d/msgid/prometheus-users/6cfb736c-ba38-4e8d-8468-cdc84f2971f2n%40googlegroups.com?utm_medium=email&utm_source=footer>
>>>> .
>>>>
>>> -- 
>> You received this message because you are subscribed to the Google Groups 
>> "Prometheus Users" group.
>> To unsubscribe from this group and stop receiving emails from it, send an 
>> email to [email protected].
>>
> To view this discussion on the web visit 
>> https://groups.google.com/d/msgid/prometheus-users/cb4928de-bb1f-4769-828b-6d9e10a3f532n%40googlegroups.com
>>  
>> <https://groups.google.com/d/msgid/prometheus-users/cb4928de-bb1f-4769-828b-6d9e10a3f532n%40googlegroups.com?utm_medium=email&utm_source=footer>
>> .
>>
>

-- 
You received this message because you are subscribed to the Google Groups 
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/prometheus-users/f2dc0f42-8e7c-4a26-9dc0-94ecefb32e19n%40googlegroups.com.

Reply via email to