Hi Krutika,
I have noticed some performance penalties (10%-15%) when using sharing in
v3.12 .
What is the situation now with 5.5 ?
Best Regards,
Strahil NikolovOn Mar 28, 2019 08:56, Krutika Dhananjay <[email protected]>
wrote:
>
> Right. So Gluster stores what are called "indices" for each modified file (or
> shard)
> under a special hidden directory of the "good" bricks at
> $BRICK_PATH/.glusterfs/indices/xattrop.
> When the offline brick comes back up, the file corresponding to each index is
> healed, and then the index deleted
> to mark the fact that the file has been healed.
>
> You can try this and see it for yourself. Just create a 1x3 plain replicate
> volume, and enable shard on it.
> Create a big file (big enough to have multiple shards). Check that the shards
> are created under $BRICK_PATH/.shard.
> Now kill a brick. Modify a small portion of the file. Hit `ls` on
> $BRICK_PATH/.glusterfs/indices/xattrop of the online bricks.
> You'll notice there will be entries named after the gfid (unique identifier
> in gluster for each file) of the shards.
> And only for those shards that the write modified, and not ALL shards of this
> really big file.
> And then when you bring the brick back up using `gluster volume start $VOL
> force`, the
> shards get healed and the directory eventually becomes empty.
>
> -Krutika
>
>
> On Thu, Mar 28, 2019 at 12:14 PM Indivar Nair <[email protected]>
> wrote:
>>
>> Hi Krutika,
>>
>> So how does the Gluster node know which shards were modified after it went
>> down?
>> Do the other Gluster nodes keep track of it?
>>
>> Regards,
>>
>>
>> Indivar Nair
>>
>>
>> On Thu, Mar 28, 2019 at 9:45 AM Krutika Dhananjay <[email protected]>
>> wrote:
>>>
>>> Each shard is a separate file of size equal to value of
>>> "features.shard-block-size".
>>> So when a brick/node was down, only those shards belonging to the VM that
>>> were modified will be sync'd later when the brick's back up.
>>> Does that answer your question?
>>>
>>> -Krutika
>>>
>>> On Wed, Mar 27, 2019 at 7:48 PM Sahina Bose <[email protected]> wrote:
>>>>
>>>> On Wed, Mar 27, 2019 at 7:40 PM Indivar Nair <[email protected]>
>>>> wrote:
>>>> >
>>>> > Hi Strahil,
>>>> >
>>>> > Ok. Looks like sharding should make the resyncs faster.
>>>> >
>>>> > I searched for more info on it, but couldn't find much.
>>>> > I believe it will still have to compare each shard to determine whether
>>>> > there are any changes that need to be replicated.
>>>> > Am I right?
>>>>
>>>> +Krutika Dhananjay
>>>> >
>>>> > Regards,
>>>> >
>>>> > Indivar Nair
>>>> >
>>>> >
>>>> >
>>>> > On Wed, Mar 27, 2019 at 4:34 PM Strahil <[email protected]> wrote:
>>>> >>
>>>> >> By default ovirt uses 'sharding' which splits the files into logical
>>>> >> chunks. This greatly reduces healing time, as VM's disk is not always
>>>> >> completely overwritten and only the shards that are different will be
>>>> >> healed.
>>>> >>
>>>> >> Maybe you should change the default shard size.
>>>> >>
>>>> >> Best Regards,
>>>> >> Strahil Nikolov
>>>> >>
>>>> >> On Mar 27, 2019 08:24, Indivar Nair <[email protected]> wrote:
>>>> >>
>>>> >> Hi All,
>>>> >>
>>>> >> We are planning a 2 + 1 arbitrated mirrored Gluster setup.
>>>> >> We would have around 50 - 60 VMs, with an average 500GB disk size.
>>>> >>
>>>> >> Now in case one of the Gluster Nodes go completely out of sync,
>>>> >> roughly, how long would it take to resync? (as per your experience)
>>>> >> Will it impact the working of VMs in any way?
>>>> >> Is there anything to be taken care of, in advance, to prepare for such
>>>> >> a situation?
>>>> >>
>>>> >> Regards,
>>>> >>
>>>> >>
>>>> >> Indivar Nair
>>>> >>
>>>> > ______________
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
List Archives:
https://lists.ovirt.org/archives/list/[email protected]/message/RJEPBXW2CTKZIQB5X4PJKMGRQTJ7SU6Q/