Thanks, Krutika, Sorry, got busy with some other work. But, better late than never.
Regards, Indivar Nair On Thu, Mar 28, 2019 at 12:26 PM Krutika Dhananjay <kdhan...@redhat.com> wrote: > Right. So Gluster stores what are called "indices" for each modified file > (or shard) > under a special hidden directory of the "good" bricks at > $BRICK_PATH/.glusterfs/indices/xattrop. > When the offline brick comes back up, the file corresponding to each index > is healed, and then the index deleted > to mark the fact that the file has been healed. > > You can try this and see it for yourself. Just create a 1x3 plain > replicate volume, and enable shard on it. > Create a big file (big enough to have multiple shards). Check that the > shards are created under $BRICK_PATH/.shard. > Now kill a brick. Modify a small portion of the file. Hit `ls` on > $BRICK_PATH/.glusterfs/indices/xattrop of the online bricks. > You'll notice there will be entries named after the gfid (unique > identifier in gluster for each file) of the shards. > And only for those shards that the write modified, and not ALL shards of > this really big file. > And then when you bring the brick back up using `gluster volume start $VOL > force`, the > shards get healed and the directory eventually becomes empty. > > -Krutika > > > On Thu, Mar 28, 2019 at 12:14 PM Indivar Nair <indivar.n...@techterra.in> > wrote: > >> Hi Krutika, >> >> So how does the Gluster node know which shards were modified after it >> went down? >> Do the other Gluster nodes keep track of it? >> >> Regards, >> >> >> Indivar Nair >> >> >> On Thu, Mar 28, 2019 at 9:45 AM Krutika Dhananjay <kdhan...@redhat.com> >> wrote: >> >>> Each shard is a separate file of size equal to value of >>> "features.shard-block-size". >>> So when a brick/node was down, only those shards belonging to the VM >>> that were modified will be sync'd later when the brick's back up. >>> Does that answer your question? >>> >>> -Krutika >>> >>> On Wed, Mar 27, 2019 at 7:48 PM Sahina Bose <sab...@redhat.com> wrote: >>> >>>> On Wed, Mar 27, 2019 at 7:40 PM Indivar Nair <indivar.n...@techterra.in> >>>> wrote: >>>> > >>>> > Hi Strahil, >>>> > >>>> > Ok. Looks like sharding should make the resyncs faster. >>>> > >>>> > I searched for more info on it, but couldn't find much. >>>> > I believe it will still have to compare each shard to determine >>>> whether there are any changes that need to be replicated. >>>> > Am I right? >>>> >>>> +Krutika Dhananjay >>>> > >>>> > Regards, >>>> > >>>> > Indivar Nair >>>> > >>>> > >>>> > >>>> > On Wed, Mar 27, 2019 at 4:34 PM Strahil <hunter86...@yahoo.com> >>>> wrote: >>>> >> >>>> >> By default ovirt uses 'sharding' which splits the files into logical >>>> chunks. This greatly reduces healing time, as VM's disk is not always >>>> completely overwritten and only the shards that are different will be >>>> healed. >>>> >> >>>> >> Maybe you should change the default shard size. >>>> >> >>>> >> Best Regards, >>>> >> Strahil Nikolov >>>> >> >>>> >> On Mar 27, 2019 08:24, Indivar Nair <indivar.n...@techterra.in> >>>> wrote: >>>> >> >>>> >> Hi All, >>>> >> >>>> >> We are planning a 2 + 1 arbitrated mirrored Gluster setup. >>>> >> We would have around 50 - 60 VMs, with an average 500GB disk size. >>>> >> >>>> >> Now in case one of the Gluster Nodes go completely out of sync, >>>> roughly, how long would it take to resync? (as per your experience) >>>> >> Will it impact the working of VMs in any way? >>>> >> Is there anything to be taken care of, in advance, to prepare for >>>> such a situation? >>>> >> >>>> >> Regards, >>>> >> >>>> >> >>>> >> Indivar Nair >>>> >> >>>> > _______________________________________________ >>>> > Users mailing list -- users@ovirt.org >>>> > To unsubscribe send an email to users-le...@ovirt.org >>>> > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>> > oVirt Code of Conduct: >>>> https://www.ovirt.org/community/about/community-guidelines/ >>>> > List Archives: >>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/WZW5RRVHFRMAIBUZDUSTXTIF4Z4WW5Y5/ >>>> >>>
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/HU6ACV2QJVNGMWJHAVFJ63Y67GB3NPQU/