Hello! The answer is pretty much depends on your fears. If you afraid of hardware failures you could have more then standard 3 copies, configure your failure domain properly and so on. If you afraid of some big disaster that can hurt all of your hardware - you could consider making an async replica to a cluster in an another datacenter on another content. If you afraid of some kind of cluster software issues - then you can build an another cluster and use third-party tools to backup data there, but as you correctly noticed it will not be too convenient.
As a common sollution I would offer you to use the same cluster for backups as well (may be just a different pool\OSD tree with less expensive drives) - in most cases it's enough. Best regards, Vladimir 2017-02-14 14:15 GMT+05:00 Götz Reinicke <[email protected]>: > Hi, > > I guess that's a question that pops up in different places, but I could > not find any which fits to my thoughts. > > Currently we start to use ceph for file shares of our films produced by > our students and some xen/vmware VMs. Thd VM data is already backed up; the > fils original footage is stored in other places. > > We start with some 100TB rbd and mount smb/NFS shares from the clients. > May be we look into ceph fs soon. > > The question is: How would someone handle a backup of 100 TB data? > Rsyncing that to an other system or having a commercial backup solution > looks not that good e.g. regarding the price. > > One thought is, is there some sort of best practice in the ceph world e.g. > replicating to an other physical independent cluster? Or use more replicas, > odds, nodes and do snapshots in one cluster? > > Having productive data and backup on the same hardware currently makes me > feel not that good too….But the world changes :) > > Long story short: How do you do backup hundreds of TB? > > Curious for suggestions and thoughts .. Thanks and Regards . Götz > > > _______________________________________________ > ceph-users mailing list > [email protected] > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
