Raw or qcow doesn't change anything about the backup.
Georep always have to sync the whole file

Additionally, raw images has much less features than qcow

Il 23 mar 2017 8:40 PM, "Joe Julian" <[email protected]> ha scritto:

> I always use raw images. And yes, sharding would also be good.
>
> On 03/23/17 12:36, Gandalf Corvotempesta wrote:
>
> Georep expose to another problem:
> When using gluster as storage for VM, the VM file is saved as qcow.
> Changes are inside the qcow, thus rsync has to sync the whole file every
> time
>
> A little workaround would be sharding, as rsync has to sync only the
> changed shards, but I don't think this is a good solution
>
> Il 23 mar 2017 8:33 PM, "Joe Julian" <[email protected]> ha scritto:
>
>> In many cases, a full backup set is just not feasible. Georep to the same
>> or different DC may be an option if the bandwidth can keep up with the
>> change set. If not, maybe breaking the data up into smaller more manageable
>> volumes where you only keep a smaller set of critical data and just back
>> that up. Perhaps an object store (swift?) might handle fault tolerance
>> distribution better for some workloads.
>>
>> There's no one right answer.
>>
>> On 03/23/17 12:23, Gandalf Corvotempesta wrote:
>>
>> Backing up from inside each VM doesn't solve the problem
>> If you have to backup 500VMs you just need more than 1 day and what if
>> you have to restore the whole gluster storage?
>>
>> How many days do you need to restore 1PB?
>>
>> Probably the only solution should be a georep in the same datacenter/rack
>> with a similiar cluster,
>> ready to became the master storage.
>> In this case you don't need to restore anything as data are already
>> there,
>> only a little bit back in time but this double the TCO
>>
>> Il 23 mar 2017 6:39 PM, "Serkan Çoban" <[email protected]> ha
>> scritto:
>>
>>> Assuming a backup window of 12 hours, you need to send data at 25GB/s
>>> to backup solution.
>>> Using 10G Ethernet on hosts you need at least 25 host to handle 25GB/s.
>>> You can create an EC gluster cluster that can handle this rates, or
>>> you just backup valuable data from inside VMs using open source backup
>>> tools like borg,attic,restic , etc...
>>>
>>> On Thu, Mar 23, 2017 at 7:48 PM, Gandalf Corvotempesta
>>> <[email protected]> wrote:
>>> > Let's assume a 1PB storage full of VMs images with each brick over ZFS,
>>> > replica 3, sharding enabled
>>> >
>>> > How do you backup/restore that amount of data?
>>> >
>>> > Backing up daily is impossible, you'll never finish the backup that the
>>> > following one is starting (in other words, you need more than 24 hours)
>>> >
>>> > Restoring is even worse. You need more than 24 hours with the whole
>>> cluster
>>> > down
>>> >
>>> > You can't rely on ZFS snapshot due to sharding (the snapshot took from
>>> one
>>> > node is useless without all other node related at the same shard) and
>>> you
>>> > still have the same restore speed
>>> >
>>> > How do you backup this?
>>> >
>>> > Even georep isn't enough, if you have to restore the whole storage in
>>> case
>>> > of disaster
>>> >
>>> > _______________________________________________
>>> > Gluster-users mailing list
>>> > [email protected]
>>> > http://lists.gluster.org/mailman/listinfo/gluster-users
>>>
>>
>>
>> _______________________________________________
>> Gluster-users mailing 
>> [email protected]http://lists.gluster.org/mailman/listinfo/gluster-users
>>
>> _______________________________________________ Gluster-users mailing
>> list [email protected] http://lists.gluster.org/mailm
>> an/listinfo/gluster-users
>
>
_______________________________________________
Gluster-users mailing list
[email protected]
http://lists.gluster.org/mailman/listinfo/gluster-users

Reply via email to