Hi, First of all , following command is not for disperse volume - gluster volume heal elastic-volume info split-brain
This is applicable for replicate volumes only. Could you please let us know what exactly do you want to test? If you want to test disperse volume against failure of bricks or servers, you can kill some of the brick process. Maximum redundant number of brick process. In 4+2, 2 is the redundancy count. After killing two brick process, by using kill command, you can write some data on volume and the do force start of the volume. gluster v <volname> start force This will start all the killed brick processes also. At the end you can see tha heal should be done by self heal daemon and volume should become healthy again. --- Ashish ----- Original Message ----- From: "fusillator" <[email protected]> To: [email protected] Sent: Friday, June 7, 2019 2:09:01 AM Subject: [Gluster-users] healing of disperse volume Hi all, I'm pretty new to glusterfs, I managed to setup a dispersed volume (4+2) using the release 6.1 from centos repository.. Is it a stable release? Then I forced the volume stop when the application were writing on the mount point.. getting a wanted inconsistent state, I'm wondering what are the best practice to solve this kinds of situation...I just found a detailed explanation about how to solve splitting-head state of replicated volume at https://docs.gluster.org/en/latest/Troubleshooting/resolving-splitbrain/ but it seems to be not applicable to the disperse volume type. Do I miss to read some important piece of documentation? Please point me to some reference. Here's some command detail: #gluster volume info elastic-volume Volume Name: elastic-volume Type: Disperse Volume ID: 96773fef-c443-465b-a518-6630bcf83397 Status: Started Snapshot Count: 0 Number of Bricks: 1 x (4 + 2) = 6 Transport-type: tcp Bricks: Brick1: dev-netflow01.fineco.it:/data/gfs/lv_elastic/brick1/brick Brick2: dev-netflow02.fineco.it:/data/gfs/lv_elastic/brick1/brick Brick3: dev-netflow03.fineco.it:/data/gfs/lv_elastic/brick1/brick Brick4: dev-netflow04.fineco.it:/data/gfs/lv_elastic/brick1/brick Brick5: dev-netflow05.fineco.it:/data/gfs/lv_elastic/brick1/brick Brick6: dev-netflow06.fineco.it:/data/gfs/lv_elastic/brick1/brick Options Reconfigured: performance.io-cache: off performance.io-thread-count: 64 performance.write-behind-window-size: 100MB performance.cache-size: 1GB nfs.disable: on transport.address-family: inet # gluster volume heal elastic-volume info Brick dev01:/data/gfs/lv_elastic/brick1/brick <gfid:7d7d7292-76eb-430a-ac10-b4f5e9311a17> /data/logs/20190606/ns-coreiol-iol-app-listini.2019060615.log /data/logs/20190606/ns-coreiol-iol-app-fns.2019060615.log /data/logs/20190606/ns-coreiol-iol-app-news.2019060615.log /data/logs/20190606/ns-coreiol-iol-app-trkd.2019060615.log <gfid:5c577478-9a2c-4d99-9189-36e9afed1039> <gfid:813ccd43-1578-4275-a342-416a658cd714> <gfid:60c74f7e-bed3-44a1-9129-99541a83e71b> <gfid:9417e4db-5c68-4812-9ab1-77b4f5ad7174> /data/logs/20190606/ns-coreiol-iol-lib-managers.2019060615.log /data/logs/20190606/ns-coreiol-iol-lib-httpwrapper.2019060615.log /data/logs/20190606/ns-coreiol-iol-app-chart.2019060615.log Status: Connected Number of entries: 12 Brick dev02:/data/gfs/lv_elastic/brick1/brick /data/logs/20190606/ns-coreiol-iol-lib-managers.2019060615.log /data/logs/20190606/ns-coreiol-iol-app-news.2019060615.log /data/logs/20190606/ns-coreiol-iol-app-chart.2019060615.log <gfid:5c577478-9a2c-4d99-9189-36e9afed1039> <gfid:813ccd43-1578-4275-a342-416a658cd714> <gfid:60c74f7e-bed3-44a1-9129-99541a83e71b> <gfid:9417e4db-5c68-4812-9ab1-77b4f5ad7174> <gfid:7d7d7292-76eb-430a-ac10-b4f5e9311a17> /data/logs/20190606/ns-coreiol-iol-app-listini.2019060615.log /data/logs/20190606/ns-coreiol-iol-app-fns.2019060615.log /data/logs/20190606/ns-coreiol-iol-app-trkd.2019060615.log /data/logs/20190606/ns-coreiol-iol-lib-httpwrapper.2019060615.log Status: Connected Number of entries: 12 Brick dev03:/data/gfs/lv_elastic/brick1/brick /data/logs/20190606/ns-coreiol-iol-app-listini.2019060615.log /data/logs/20190606/ns-coreiol-iol-app-fns.2019060615.log /data/logs/20190606/ns-coreiol-iol-app-news.2019060615.log /data/logs/20190606/ns-coreiol-iol-app-trkd.2019060615.log <gfid:5c577478-9a2c-4d99-9189-36e9afed1039> <gfid:813ccd43-1578-4275-a342-416a658cd714> <gfid:60c74f7e-bed3-44a1-9129-99541a83e71b> <gfid:9417e4db-5c68-4812-9ab1-77b4f5ad7174> <gfid:7d7d7292-76eb-430a-ac10-b4f5e9311a17> /data/logs/20190606/ns-coreiol-iol-lib-managers.2019060615.log /data/logs/20190606/ns-coreiol-iol-lib-httpwrapper.2019060615.log /data/logs/20190606/ns-coreiol-iol-app-chart.2019060615.log Status: Connected Number of entries: 12 Brick dev04:/data/gfs/lv_elastic/brick1/brick <gfid:7d7d7292-76eb-430a-ac10-b4f5e9311a17> /data/logs/20190606/ns-coreiol-iol-app-news.2019060615.log /data/logs/20190606/ns-coreiol-iol-app-trkd.2019060615.log /data/logs/20190606/ns-coreiol-iol-lib-managers.2019060615.log /data/logs/20190606/ns-coreiol-iol-lib-httpwrapper.2019060615.log <gfid:5c577478-9a2c-4d99-9189-36e9afed1039> <gfid:813ccd43-1578-4275-a342-416a658cd714> <gfid:60c74f7e-bed3-44a1-9129-99541a83e71b> <gfid:9417e4db-5c68-4812-9ab1-77b4f5ad7174> /data/logs/20190606/ns-coreiol-iol-app-listini.2019060615.log /data/logs/20190606/ns-coreiol-iol-app-fns.2019060615.log /data/logs/20190606/ns-coreiol-iol-app-chart.2019060615.log Status: Connected Number of entries: 12 Brick dev05:/data/gfs/lv_elastic/brick1/brick /data/logs/20190606/ns-coreiol-iol-app-news.2019060615.log /data/logs/20190606/ns-coreiol-iol-app-trkd.2019060615.log <gfid:5c577478-9a2c-4d99-9189-36e9afed1039> <gfid:813ccd43-1578-4275-a342-416a658cd714> <gfid:60c74f7e-bed3-44a1-9129-99541a83e71b> <gfid:9417e4db-5c68-4812-9ab1-77b4f5ad7174> <gfid:7d7d7292-76eb-430a-ac10-b4f5e9311a17> /data/logs/20190606/ns-coreiol-iol-app-listini.2019060615.log /data/logs/20190606/ns-coreiol-iol-lib-managers.2019060615.log /data/logs/20190606/ns-coreiol-iol-app-fns.2019060615.log /data/logs/20190606/ns-coreiol-iol-lib-httpwrapper.2019060615.log /data/logs/20190606/ns-coreiol-iol-app-chart.2019060615.log Status: Connected Number of entries: 12 Brick dev06:/data/gfs/lv_elastic/brick1/brick /data/logs/20190606/ns-coreiol-iol-app-chart.2019060615.log <gfid:5c577478-9a2c-4d99-9189-36e9afed1039> <gfid:813ccd43-1578-4275-a342-416a658cd714> <gfid:60c74f7e-bed3-44a1-9129-99541a83e71b> <gfid:9417e4db-5c68-4812-9ab1-77b4f5ad7174> <gfid:7d7d7292-76eb-430a-ac10-b4f5e9311a17> /data/logs/20190606/ns-coreiol-iol-lib-managers.2019060615.log /data/logs/20190606/ns-coreiol-iol-app-news.2019060615.log /data/logs/20190606/ns-coreiol-iol-app-trkd.2019060615.log /data/logs/20190606/ns-coreiol-iol-app-listini.2019060615.log /data/logs/20190606/ns-coreiol-iol-app-fns.2019060615.log /data/logs/20190606/ns-coreiol-iol-lib-httpwrapper.2019060615.log Status: Connected Number of entries: 12 # gluster volume heal elastic-volume info split-brain Volume elastic-volume is not of type replicate Any advice? Best regards Luca _______________________________________________ Gluster-users mailing list [email protected] https://lists.gluster.org/mailman/listinfo/gluster-users
_______________________________________________ Gluster-users mailing list [email protected] https://lists.gluster.org/mailman/listinfo/gluster-users
