Can you give volume info? Looks like you are using 2 way replica. On Wed, Sep 19, 2018 at 9:39 AM, Johan Karlsson <[email protected]> wrote:
> I have two servers setup with glusterFS in replica mode, a single volume > exposed via a mountpoint. The servers are running Ubuntu 16.04 LTS > > After a package upgrade + reboot of both servers, it was discovered that > the data was completely gone. New data written on the volume via the > mountpoint is replicated correctly, and gluster status/info commands states > that everything is ok (no split brain scenario or any healing needed etc). > But the previous data is completely gone, not even present on any of the > bricks. > > The following upgrade was done: > > glusterfs-server:amd64 (4.1.0-ubuntu1~xenial3 -> 4.1.4-ubuntu1~xenial1) > glusterfs-client:amd64 (4.1.0-ubuntu1~xenial3 -> 4.1.4-ubuntu1~xenial1) > glusterfs-common:amd64 (4.1.0-ubuntu1~xenial3 -> 4.1.4-ubuntu1~xenial1) > > The logs only show that connection between the servers was lost, which is > expected. > > I can't even determine if it was the package upgrade or the reboot that > caused this issue, but I've tried to recreate the issue without success. > > Any idea what could have gone wrong, or if I have done some wrong during > the setup. For reference, this is how I've done the setup: > > --- > Add a separate disk with a single partition on both servers (/dev/sdb1) > > Add gfs hostnames for direct communication without DNS, on both servers: > > /etc/hosts > > 192.168.4.45 gfs01 > 192.168.4.46 gfs02 > > On gfs01, create a new LVM Volume Group: > vgcreate gfs01-vg /dev/sdb1 > > And on the gfs02: > vgcreate gfs02-vg /dev/sdb1 > > Create logical volumes named "brick" on the servers: > > gfs01: > lvcreate -l 100%VG -n brick1 gfs01-vg > gfs02: > lvcreate -l 100%VG -n brick2 gfs02-vg > > Format the volumes with ext4 filesystem: > > gfs01: > mkfs.ext4 /dev/gfs01-vg/brick1 > gfs02: > mkfs.ext4 /dev/gfs02-vg/brick2 > > Create a mountpoint for the bricks on the servers: > > gfs01: > mkdir -p /glusterdata/brick1 > gds02: > mkdir -p /glusterdata/brick2 > > Make a permanent mount on the servers: > > gfs01: > /dev/gfs01-vg/brick1 /glusterdata/brick1 ext4 defaults 0 > 0 > gfs02: > /dev/gfs02-vg/brick2 /glusterdata/brick2 ext4 defaults 0 > 0 > > Mount it: > mount -a > > Create a gluster volume mount point on the bricks on the servers: > > gfs01: > mkdir -p /glusterdata/brick1/gvol0 > gfs02: > mkdir -p /glusterdata/brick2/gvol0 > > From each server, peer probe the other one: > > gluster peer probe gfs01 > peer probe: success > > gluster peer probe gfs02 > peer probe: success > > From any single server, create the gluster volume as a "replica" with two > nodes; gfs01 and gfs02: > > gluster volume create gvol0 replica 2 gfs01:/glusterdata/brick1/gvol0 > gfs02:/glusterdata/brick2/gvol0 > > Start the volume: > > gluster volume start gvol0 > > On each server, mount the gluster filesystem on the /filestore mount point: > > gfs01: > mount -t glusterfs gfs01:/gvol0 /filestore > gfs02: > mount -t glusterfs gfs02:/gvol0 /filestore > > Make the mount permanent on the servers: > > /etc/fstab > > gfs01: > gfs01:/gvol0 /filestore glusterfs defaults,_netdev 0 0 > gfs02: > gfs02:/gvol0 /filestore glusterfs defaults,_netdev 0 0 > --- > > Regards, > > Johan Karlsson > _______________________________________________ > Gluster-users mailing list > [email protected] > https://lists.gluster.org/mailman/listinfo/gluster-users >
_______________________________________________ Gluster-users mailing list [email protected] https://lists.gluster.org/mailman/listinfo/gluster-users
