So glusterp3 is in a reject state, [root@glusterp1 /]# gluster peer status Number of Peers: 2
Hostname: glusterp2.graywitch.co.nz Uuid: 93eebe2c-9564-4bb0-975f-2db49f12058b State: Peer in Cluster (Connected) Other names: glusterp2 Hostname: glusterp3.graywitch.co.nz Uuid: 5d59b704-e42f-46c6-8c14-cf052c489292 State: Peer Rejected (Connected) Other names: glusterp3 [root@glusterp1 /]# ======== [root@glusterp2 /]# gluster peer status Number of Peers: 2 Hostname: glusterp1.graywitch.co.nz Uuid: 4ece8509-033e-48d1-809f-2079345caea2 State: Peer in Cluster (Connected) Other names: glusterp1 Hostname: glusterp3.graywitch.co.nz Uuid: 5d59b704-e42f-46c6-8c14-cf052c489292 State: Peer Rejected (Connected) Other names: glusterp3 [root@glusterp2 /]# ======== [root@glusterp3 /]# gluster peer status Number of Peers: 2 Hostname: glusterp1.graywitch.co.nz Uuid: 4ece8509-033e-48d1-809f-2079345caea2 State: Peer Rejected (Connected) Other names: glusterp1 Hostname: glusterp2.graywitch.co.nz Uuid: 93eebe2c-9564-4bb0-975f-2db49f12058b State: Peer Rejected (Connected) Other names: glusterp2 ========== on glusterp3 gluster is dead and will not start, [root@glusterp3 /]# systemctl status gluster ● gluster.service Loaded: not-found (Reason: No such file or directory) Active: inactive (dead) [root@glusterp3 /]# systemctl restart gluster Failed to restart gluster.service: Unit gluster.service failed to load: No such file or directory. [root@glusterp3 /]# systemctl enable gluster Failed to execute operation: Access denied [root@glusterp3 /]# systemctl enable gluster.service Failed to execute operation: Access denied [root@glusterp3 /]# systemctl start gluster.service Failed to start gluster.service: Unit gluster.service failed to load: No such file or directory. ========== [root@glusterp3 /]# rpm -qa |grep gluster glusterfs-client-xlators-3.8.4-1.el7.x86_64 glusterfs-server-3.8.4-1.el7.x86_64 nfs-ganesha-gluster-2.3.3-1.el7.x86_64 glusterfs-cli-3.8.4-1.el7.x86_64 glusterfs-api-3.8.4-1.el7.x86_64 glusterfs-fuse-3.8.4-1.el7.x86_64 glusterfs-ganesha-3.8.4-1.el7.x86_64 glusterfs-3.8.4-1.el7.x86_64 centos-release-gluster38-1.0-1.el7.centos.noarch glusterfs-libs-3.8.4-1.el7.x86_64 [root@glusterp3 /]# ? On 14 October 2016 at 12:31, Thing <[email protected]> wrote: > Hmm seem I have something rather inconsistent, > > [root@glusterp1 /]# gluster volume create gv1 replica 3 > glusterp1:/brick1/gv1 glusterp2:/brick1/gv1 glusterp3:/brick1/gv1 > volume create: gv1: failed: Host glusterp3 is not in 'Peer in Cluster' > state > [root@glusterp1 /]# gluster peer probe glusterp3 > peer probe: success. Host glusterp3 port 24007 already in peer list > [root@glusterp1 /]# gluster peer probe glusterp2 > peer probe: success. Host glusterp2 port 24007 already in peer list > [root@glusterp1 /]# gluster volume create gv1 replica 3 > glusterp1:/brick1/gv1 glusterp2:/brick1/gv1 glusterp3:/brick1/gv1 > volume create: gv1: failed: /brick1/gv1 is already part of a volume > [root@glusterp1 /]# gluster volume show > unrecognized word: show (position 1) > [root@glusterp1 /]# gluster volume > add-brick delete info quota > reset status > barrier geo-replication list rebalance > set stop > clear-locks heal log remove-brick > start sync > create help profile replace-brick > statedump top > [root@glusterp1 /]# gluster volume list > volume1 > [root@glusterp1 /]# gluster volume start gv0 > volume start: gv0: failed: Volume gv0 does not exist > [root@glusterp1 /]# gluster volume start gv1 > volume start: gv1: failed: Volume gv1 does not exist > [root@glusterp1 /]# gluster volume status > Status of volume: volume1 > Gluster process TCP Port RDMA Port Online > Pid > ------------------------------------------------------------ > ------------------ > Brick glusterp1.graywitch.co.nz:/data1 49152 0 Y > 2958 > Brick glusterp2.graywitch.co.nz:/data1 49152 0 Y > 2668 > NFS Server on localhost N/A N/A N > N/A > Self-heal Daemon on localhost N/A N/A Y > 1038 > NFS Server on glusterp2.graywitch.co.nz N/A N/A N > N/A > Self-heal Daemon on glusterp2.graywitch.co. > nz N/A N/A Y > 676 > > Task Status of Volume volume1 > ------------------------------------------------------------ > ------------------ > There are no active volume tasks > > [root@glusterp1 /]# > > On 14 October 2016 at 12:20, Thing <[email protected]> wrote: > >> I deleted a gluster volume gv0 as I wanted to make it thin provisioned. >> >> I have rebuilt "gv0" but I am getting a failure, >> >> ========== >> [root@glusterp1 /]# df -h >> Filesystem Size Used Avail Use% Mounted on >> /dev/mapper/centos-root 20G 3.9G 17G 20% / >> devtmpfs 1.8G 0 1.8G 0% /dev >> tmpfs 1.8G 12K 1.8G 1% /dev/shm >> tmpfs 1.8G 8.9M 1.8G 1% /run >> tmpfs 1.8G 0 1.8G 0% /sys/fs/cgroup >> /dev/mapper/centos-tmp 3.9G 33M 3.9G 1% /tmp >> /dev/mapper/centos-home 50G 41M 50G 1% /home >> /dev/mapper/centos-data1 120G 33M 120G 1% /data1 >> /dev/sda1 997M 312M 685M 32% /boot >> /dev/mapper/centos-var 20G 401M 20G 2% /var >> tmpfs 368M 0 368M 0% /run/user/1000 >> /dev/mapper/vol_brick1-brick1 100G 33M 100G 1% /brick1 >> [root@glusterp1 /]# mkdir /brick1/gv0 >> [root@glusterp1 /]# gluster volume create gv0 replica 3 >> glusterp1:/brick1/gv0 glusterp2:/brick1/gv0 glusterp3:/brick1/gv0 >> volume create: gv0: failed: Host glusterp3 is not in 'Peer in Cluster' >> state >> [root@glusterp1 /]# gluster peer probe glusterp3 >> peer probe: success. Host glusterp3 port 24007 already in peer list >> [root@glusterp1 /]# gluster volume create gv0 replica 3 >> glusterp1:/brick1/gv0 glusterp2:/brick1/gv0 glusterp3:/brick1/gv0 >> volume create: gv0: failed: /brick1/gv0 is already part of a volume >> [root@glusterp1 /]# gluster volume start gv0 >> volume start: gv0: failed: Volume gv0 does not exist >> [root@glusterp1 /]# gluster volume create gv0 replica 3 >> glusterp1:/brick1/gv0 glusterp2:/brick1/gv0 glusterp3:/brick1/gv0 --force >> unrecognized option --force >> [root@glusterp1 /]# gluster volume create gv0 replica 3 >> glusterp1:/brick1/gv0 glusterp2:/brick1/gv0 glusterp3:/brick1/gv0 >> volume create: gv0: failed: /brick1/gv0 is already part of a volume >> [root@glusterp1 /]# >> ========== >> >> Obviously something isnt happy here but I have no idea what....... >> >> how to fix this please? >> > >
_______________________________________________ Gluster-users mailing list [email protected] http://www.gluster.org/mailman/listinfo/gluster-users
