Thanks, this looks ok to me, I will reset brick because I don't have any data anymore on failed node so I can use same path / brick name.
Is reseting brick dangerous command? Should I be worried about some possible failure that will impact remaining two nodes? I am running really old 3.7.6 but stable version. Thanks, BR! Martin > On 10 Apr 2019, at 12:20, Karthik Subrahmanya <[email protected]> wrote: > > Hi Martin, > > After you add the new disks and creating raid array, you can run the > following command to replace the old brick with new one: > > - If you are going to use a different name to the new brick you can run > gluster volume replace-brick <volname> <old-brick> <new-brick> commit force > > - If you are planning to use the same name for the new brick as well then you > can use > gluster volume reset-brick <volname> <old-brick> <new-brick> commit force > Here old-brick & new-brick's hostname & path should be same. > > After replacing the brick, make sure the brick comes online using volume > status. > Heal should automatically start, you can check the heal status to see all the > files gets replicated to the newly added brick. If it does not start > automatically, you can manually start that by running gluster volume heal > <volname>. > > HTH, > Karthik > > On Wed, Apr 10, 2019 at 3:13 PM Martin Toth <[email protected] > <mailto:[email protected]>> wrote: > Hi all, > > I am running replica 3 gluster with 3 bricks. One of my servers failed - all > disks are showing errors and raid is in fault state. > > Type: Replicate > Volume ID: 41d5c283-3a74-4af8-a55d-924447bfa59a > Status: Started > Number of Bricks: 1 x 3 = 3 > Transport-type: tcp > Bricks: > Brick1: node1.san:/tank/gluster/gv0imagestore/brick1 > Brick2: node2.san:/tank/gluster/gv0imagestore/brick1 <— this brick is down > Brick3: node3.san:/tank/gluster/gv0imagestore/brick1 > > So one of my bricks is totally failed (node2). It went down and all data are > lost (failed raid on node2). Now I am running only two bricks on 2 servers > out from 3. > This is really critical problem for us, we can lost all data. I want to add > new disks to node2, create new raid array on them and try to replace failed > brick on this node. > > What is the procedure of replacing Brick2 on node2, can someone advice? I > can’t find anything relevant in documentation. > > Thanks in advance, > Martin > _______________________________________________ > Gluster-users mailing list > [email protected] <mailto:[email protected]> > https://lists.gluster.org/mailman/listinfo/gluster-users > <https://lists.gluster.org/mailman/listinfo/gluster-users>
_______________________________________________ Gluster-users mailing list [email protected] https://lists.gluster.org/mailman/listinfo/gluster-users
