On 18/01/2021 11:28, Brock Palen wrote:
Disclaimer I have not used s3 with bareos but done many cloud calculations.

Few things to think about using cloud.
Are you running your SD in the cloud?
Are your backup clients in the cloud?
If not what’s your bandwidth? It will impact your backup and restore times significantly if you have modest WAN capacity for local  clients servers.

No, no. I was thinking about keeping an extra copy "off-site". I'm mostly cloud-free at the moment and I do not wish to change it significantly. I was thinking whether S3 could be an option for extending my home backup setup.

Of course I understand the impact of bandwidth on the backup/restore times. :-)

As for s3 pricing read this carefully

https://aws.amazon.com/s3/pricing/ <https://aws.amazon.com/s3/pricing/>

You have three components to pricing with s3 and I expect only two move the needle on cost.

Data stored
Bandwidth and retrieval
Operations

Opts rations are so cheap and guessing how bareos uses virtual tape volumes it’s prob not a big issue. Someone who has used it though can speak.

That's a good observation. Thanks!

Data stored its straight $/gb/month. So you need to estimate your total data stored for all your fulls and incrementals. Your right these costs decline when you look at glacier but there is a trade off. The cheaper to store the more expensive to access.

Retrieval fees come in two forms. The first is bandwidth. Which fit most people is .09$/gb (unless your clients and servers are in the same aws region) for my cloud activities this is 50% of my monthly bill. It’s the thing that messes most cloud calculators for budget.  That said if your sever is on prem you likely will never pay this if you don’t use always incremental or do any restores. So if your ok paying for restores maybe it’s ok.

The cold tiers like glacier charge to access data. Again maybe fine if you almost never read it. Glacier runs $10/tb or more for transfer vs nothing for regular s3. With bandwidth your at ~$100/tb  This is something to avoid deep archive. Their sla is many hours to get data.  I don’t think deep archive is a backup replacement but a compliance archive replacement


Well, that's what I'm counting on - it's better to have backup copy and not need to use it than not having it ;-)

What I was also interested in was also how to approach the long SLA regarding Bareos SD operation. Would I have to firstly request access to the glacier data independently of the SD and after receiving confirmation of data availability would have to run a restore job? Or would I just run a restore job from storage using cold-tiered bucket and the job would simply wait for data availability (similar to mounting tape)?

Also be aware glacier and deep archive have minimum retention times of 90 and 180 days. So you will always pay that at a minimum. Ok if your keeping fulls for a long time.  Look at the auto tier options to manage aging volumes.

Yes, I noticed that



So YMMV. If you are 100% in the cloud or you don’t use always incremental or have small data volumes or just a dr copy it works great.

Personally I run my servers in aws and my full bareos setup on prem with a $400 tape library from eBay. This gives me diversity and most of the data in the cloud is small (websites email text) while the on prem is video photos and road warriors using always incremental.


So it all comes to "try the free tier and see for yourself" :-) I'll have to do it anyway when I get some spare time just to see how it works and get some understanding about achievable througputs, needed space and so on.


Thanks for valuable insight!

--
You received this message because you are subscribed to the Google Groups 
"bareos-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/bareos-users/f17e6555-1f48-3d1a-1ba6-f9681d1e8b73%40gmail.com.

Reply via email to