Recommended would be creating a new storage domain with shard size as 64 MB and migrating all the disks from 4MB storagedomain
On Mon, Sep 18, 2017 at 12:01 PM, Ravishankar N <ravishan...@redhat.com> wrote: > Possibly. I don't think changing shard size on the fly is supported, > especially when there are files on the volume that are sharded with a > different size. > > -Ravi > > > On 09/18/2017 11:40 AM, Alex K wrote: > > The heal status is showing that no pending files need healing (also shown > at GUI). > When checking the bricks on the file system I see that what is different > between the server is the .shard folder of the volume. One server reports > 835GB while the other 1.1 TB. > I recall to have changed the shard size at some point from 4 MB to 64MB. > Could this be the cause? > > Thanx, > Alex > > On Mon, Sep 18, 2017 at 8:14 AM, Ravishankar N <ravishan...@redhat.com> > wrote: > >> >> On 09/18/2017 10:08 AM, Alex K wrote: >> >> Hi Ravishankar, >> >> I am not referring to the arbiter volume(which is showing 0% usage). I am >> referring to the other 2 volumes which are replicas and should have the >> exact same data. Checking the status of other bricks in ovirt (bricks used >> from iso and export domain) I see that they all report same usage of data >> on the data volumes, except the "vms" volume used for storing vms. >> >> >> Ah, okay. Some of the things that can cause a variation in disk usage: >> - Pending self-heals in gluster (check if `gluster volume heal <volname> >> info` doesn't show any entries. Also if there is anything under >> `.glusterfs/landfill` folder of the bricks). >> - XFS speculative preallocation >> - Possibly some bug in self-healing of sparse files by gluster (although >> we fixed known bugs a long time back in this area). >> >> Regards >> Ravi >> >> >> Thanx, >> Alex >> >> On Sep 18, 2017 07:00, "Ravishankar N" <ravishan...@redhat.com> wrote: >> >>> >>> >>> On 09/17/2017 08:41 PM, Alex K wrote: >>> >>> Hi all, >>> >>> I have replica 3 with 1 arbiter. >>> When checking the gluster volume bricks they are reported as using >>> different space, as per attached. How come they use different space? One >>> would expect to use exactly the same space since they are replica. >>> >>> The 3rd brick (arbiter ) only holds meta data, so it would not consume >>> as much space as the other 2 data bricks. So what you are seeing is >>> expected behaviour. >>> Regards, >>> Ravi >>> >>> Thanx, >>> Alex >>> >>> >>> _______________________________________________ >>> Users mailing >>> listUsers@ovirt.orghttp://lists.ovirt.org/mailman/listinfo/users >>> >>> >>> >> > > > _______________________________________________ > Users mailing list > Users@ovirt.org > http://lists.ovirt.org/mailman/listinfo/users > >
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users