Is this normal? Brick3: sfsccl03:/data/brick-sdc2/glusterfs/dht Brick4: sfsccl03:/data/brick-sdd2/glusterfs/dht
Both are pointing to the same dir location. Could that be confusing gluster? On Fri, Dec 2, 2011 at 10:24 AM, Joe Landman <[email protected]> wrote: > Can't start a volume. > > [root@sfsccl03 ~]# gluster volume start brick1 > brick: sfsccl03:/data/brick-sdc2/glusterfs/dht, path creation failed, > reason: No such file or directory > > But ... > > > [root@sfsccl03 ~]# ls -alF /data/brick-sdc2/glusterfs > total 0 > drwxr-xr-x 4 root root 27 Dec 2 13:00 ./ > drwxr-xr-x 4 root root 107 Jul 5 11:55 ../ > drwxrwxrwt 7 root root 61 Sep 15 11:35 dht/ > drwxr-xr-x 2 root root 6 Dec 2 13:00 dht2/ > > So it is there. > > [root@sfsccl03 ~]# ls -alF /data/brick-sdc2/glusterfs/dht > total 128 > drwxrwxrwt 7 root root 61 Sep 15 11:35 ./ > drwxr-xr-x 4 root root 27 Dec 2 13:00 ../ > drwxr-xr-x 1230 root root 65536 Oct 24 09:14 equity/ > drwxr-xr-x 1740 oracle root 65536 Nov 30 23:33 opra/ > drwxr-xr-x 35 oracle oinstall 501 Jul 9 17:07 tag/ > drwxr-xr-x 11 root root 126 Jul 1 08:51 taq/ > drwxr-xr-x 2 root root 34 Jul 11 19:44 test/ > > > and it is readable. > > More info: > > [root@sfsccl03 ~]# gluster volume info brick1 > > Volume Name: brick1 > Type: Distribute > Status: Stopped > Number of Bricks: 4 > Transport-type: tcp > Bricks: > Brick1: sfsccl01:/data/glusterfs/dht > Brick2: sfsccl02:/data/glusterfs/dht > Brick3: sfsccl03:/data/brick-sdc2/glusterfs/dht > Brick4: sfsccl03:/data/brick-sdd2/glusterfs/dht > > [root@sfsccl03 ~]# gluster peer status > Number of Peers: 2 > > Hostname: sfsccl02 > Uuid: 6e72d1a8-bdeb-4bfb-806c-7fa8b98cb697 > State: Peer in Cluster (Connected) > > Hostname: sfsccl01 > Uuid: 116197cd-5dfe-4881-85ad-5de2be484ba6 > State: Peer in Cluster (Connected) > > a volume reset doesn't help. > > [root@sfsccl03 ~]# gluster volume reset brick1 > reset volume successful > > [root@sfsccl03 ~]# gluster volume start brick1 > brick: sfsccl03:/data/brick-sdc2/glusterfs/dht, path creation failed, > reason: No such file or directory > > New volume creation also fails. > > [root@sfsccl03 ~]# gluster volume create brick2 transport tcp > sfsccl01:/data/glusterfs/dht2 sfsccl03:/data/brick-sdc2/glusterfs/dht2 > sfsccl02:/data/glusterfs/dht2 sfsccl03:/data/brick-sdd2/glusterfs/dht2 > brick: sfsccl03:/data/brick-sdc2/glusterfs/dht2, path creation failed, > reason: No such file or directory > > Not good. > > Taking out the 03 machine > > [root@sfsccl03 ~]# gluster volume create brick2 transport tcp > sfsccl01:/data/glusterfs/dht2 sfsccl02:/data/glusterfs/dht2Creation of > volume brick2 has been successful. Please start the volume to access data. > > I am wondering if I should remove the 03 machine from the volume, start it > up with 01 and 02, and then add the 03 machine in, after forcing the volume > back up. Any thoughts? > > -- > Joseph Landman, Ph.D > Founder and CEO > Scalable Informatics Inc. > email: [email protected] > web : http://scalableinformatics.com > http://scalableinformatics.com/sicluster > phone: +1 734 786 8423 x121 > fax : +1 866 888 3112 > cell : +1 734 612 4615 > _______________________________________________ > Gluster-users mailing list > [email protected] > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users _______________________________________________ Gluster-users mailing list [email protected] http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
