Thanks, Krutika,

Sorry, got busy with some other work.
But, better late than never.

Regards,


Indivar Nair

On Thu, Mar 28, 2019 at 12:26 PM Krutika Dhananjay <kdhan...@redhat.com>
wrote:

> Right. So Gluster stores what are called "indices" for each modified file
> (or shard)
> under a special hidden directory of the "good" bricks at
> $BRICK_PATH/.glusterfs/indices/xattrop.
> When the offline brick comes back up, the file corresponding to each index
> is healed, and then the index deleted
> to mark the fact that the file has been healed.
>
> You can try this and see it for yourself. Just create a 1x3 plain
> replicate volume, and enable shard on it.
> Create a big file (big enough to have multiple shards). Check that the
> shards are created under $BRICK_PATH/.shard.
> Now kill a brick. Modify a small portion of the file. Hit `ls` on
> $BRICK_PATH/.glusterfs/indices/xattrop of the online bricks.
> You'll notice there will be entries named after the gfid (unique
> identifier in gluster for each file) of the shards.
> And only for those shards that the write modified, and not ALL shards of
> this really big file.
> And then when you bring the brick back up using `gluster volume start $VOL
> force`, the
> shards get healed and the directory eventually becomes empty.
>
> -Krutika
>
>
> On Thu, Mar 28, 2019 at 12:14 PM Indivar Nair <indivar.n...@techterra.in>
> wrote:
>
>> Hi Krutika,
>>
>> So how does the Gluster node know which shards were modified after it
>> went down?
>> Do the other Gluster nodes keep track of it?
>>
>> Regards,
>>
>>
>> Indivar Nair
>>
>>
>> On Thu, Mar 28, 2019 at 9:45 AM Krutika Dhananjay <kdhan...@redhat.com>
>> wrote:
>>
>>> Each shard is a separate file of size equal to value of
>>> "features.shard-block-size".
>>> So when a brick/node was down, only those shards belonging to the VM
>>> that were modified will be sync'd later when the brick's back up.
>>> Does that answer your question?
>>>
>>> -Krutika
>>>
>>> On Wed, Mar 27, 2019 at 7:48 PM Sahina Bose <sab...@redhat.com> wrote:
>>>
>>>> On Wed, Mar 27, 2019 at 7:40 PM Indivar Nair <indivar.n...@techterra.in>
>>>> wrote:
>>>> >
>>>> > Hi Strahil,
>>>> >
>>>> > Ok. Looks like sharding should make the resyncs faster.
>>>> >
>>>> > I searched for more info on it, but couldn't find much.
>>>> > I believe it will still have to compare each shard to determine
>>>> whether there are any changes that need to be replicated.
>>>> > Am I right?
>>>>
>>>> +Krutika Dhananjay
>>>> >
>>>> > Regards,
>>>> >
>>>> > Indivar Nair
>>>> >
>>>> >
>>>> >
>>>> > On Wed, Mar 27, 2019 at 4:34 PM Strahil <hunter86...@yahoo.com>
>>>> wrote:
>>>> >>
>>>> >> By default ovirt uses 'sharding' which splits the files into logical
>>>> chunks. This greatly reduces healing time, as VM's disk is not always
>>>> completely overwritten and only the shards that are different will be
>>>> healed.
>>>> >>
>>>> >> Maybe you should change the default shard size.
>>>> >>
>>>> >> Best Regards,
>>>> >> Strahil Nikolov
>>>> >>
>>>> >> On Mar 27, 2019 08:24, Indivar Nair <indivar.n...@techterra.in>
>>>> wrote:
>>>> >>
>>>> >> Hi All,
>>>> >>
>>>> >> We are planning a 2 + 1 arbitrated mirrored Gluster setup.
>>>> >> We would have around 50 - 60 VMs, with an average 500GB disk size.
>>>> >>
>>>> >> Now in case one of the Gluster Nodes go completely out of sync,
>>>> roughly, how long would it take to resync? (as per your experience)
>>>> >> Will it impact the working of VMs in any way?
>>>> >> Is there anything to be taken care of, in advance, to prepare for
>>>> such a situation?
>>>> >>
>>>> >> Regards,
>>>> >>
>>>> >>
>>>> >> Indivar Nair
>>>> >>
>>>> > _______________________________________________
>>>> > Users mailing list -- users@ovirt.org
>>>> > To unsubscribe send an email to users-le...@ovirt.org
>>>> > Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>> > oVirt Code of Conduct:
>>>> https://www.ovirt.org/community/about/community-guidelines/
>>>> > List Archives:
>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/WZW5RRVHFRMAIBUZDUSTXTIF4Z4WW5Y5/
>>>>
>>>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/HU6ACV2QJVNGMWJHAVFJ63Y67GB3NPQU/

Reply via email to