On 19 May 2015 17:10, "Varadharajan S" <[email protected]> wrote: > > Hi, > > We are using Ubuntu 14.04 server and for storage purpose we configured gluster 3.5 as distributed volume and find the below details, > > 1).4 Servers - 14.04 Ubuntu Server and each server disks free spaces are configured as ZFS raiddz2 volume > > 2). Each server has /pool/gluster zfs volume and capacity as - 5 TB,8 TB,6 TB and 10 TB > > 3). Bricks are - rep1,rep2,rep3 and st1 and all the bricks are connected as Distributed Volume and mounted on each system as, > > For E.x in rep1 -> mount -t glusterfs rep1:/glustervol /data. > rep2 -> mount -t glusterfs rep2:/glustervol /data > rep3 -> mount -t glusterfs rep3:/glustervol /data > st1 -> mount -t glusterfs st1:/glustervol /data > > So we get /data is having around 29 TB and all our applications data's are stored in /data mount point. > > Details about volume: > > volume glustervol-client-0 > type protocol/client > option send-gids true > option password b217da9d1d8b-bb55 > option username 9d76-4553-8c75 > option transport-type tcp > option remote-subvolume /pool/gluster > option remote-host rep1 > option ping-timeout 42 > end-volume > > volume glustervol-client-1 > type protocol/client > option send-gids true > option password b217da9d1d8b-bb55 > option username jkd76-4553-5347 > option transport-type tcp > option remote-subvolume /pool/gluster > option remote-host rep2 > option ping-timeout 42 > end-volume > > volume glustervol-client-2 > type protocol/client > option send-gids true > option password b217da9d1d8b-bb55 > option username 19d7-5a190c2 > option transport-type tcp > option remote-subvolume /pool/gluster > option remote-host rep3 > option ping-timeout 42 > end-volume > > volume glustervol-client-3 > type protocol/client > option send-gids true > option password b217da9d1d8b-bb55 > option username c75-5436b5a168347 > option transport-type tcp > option remote-subvolume /pool/gluster > option remote-host st1 > > option ping-timeout 42 > end-volume > > volume glustervol-dht > type cluster/distribute > subvolumes glustervol-client-0 glustervol-client-1 glustervol-client-2 glustervol-client-3 > end-volume > > volume glustervol-write-behind > type performance/write-behind > subvolumes glustervol-dht > end-volume > > volume glustervol-read-ahead > type performance/read-ahead > subvolumes glustervol-write-behind > end-volume > > volume glustervol-io-cache > type performance/io-cache > subvolumes glustervol-read-ahead > end-volume > > volume glustervol-quick-read > type performance/quick-read > subvolumes glustervol-io-cache > end-volume > > volume glustervol-open-behind > type performance/open-behind > subvolumes glustervol-quick-read > end-volume > > volume glustervol-md-cache > type performance/md-cache > subvolumes glustervol-open-behind > end-volume > > volume glustervol > type debug/io-stats > option count-fop-hits off > option latency-measurement off > subvolumes glustervol-md-cache > end-volume > > > ap@rep3:~$ sudo gluster volume info > > Volume Name: glustervol > Type: Distribute > Volume ID: 165b-XXXXX > Status: Started > Number of Bricks: 4 > Transport-type: tcp > Bricks: > Brick1: rep1:/pool/gluster > Brick2: rep2:/pool/gluster > Brick3: rep3:/pool/gluster > Brick4: st1:/pool/gluster > > Problem: > > If we shutdown any of the bricks , the volume size is reduced (this is ok) but from the other servers , i can see my mount point /data but it's only listing contents and i can't write or edit any single files/folders. > > Solution Required: > > If anyone brick is not available, From other servers should allow for Write and edit functions This is expected since you are using distributed volume. You wouldn't be able to write/edit files belonging to the brick which is down. Solution would be to migrate to distributed replicate volume. > > Please let us know, what can i try further ? > > Regards, > Varad > > > _______________________________________________ > Gluster-users mailing list > [email protected] > http://www.gluster.org/mailman/listinfo/gluster-users
_______________________________________________ Gluster-users mailing list [email protected] http://www.gluster.org/mailman/listinfo/gluster-users
