Someone who knows pipe up here, but I think that I am confused.
Lots of things look wrong to me, but I am not sure what you are trying to do,
and what is different between 2.x and 3.x.
I think what you want:
1 (or more) mount points/directories exported from your 2 servers server002 and
bee003 to be mounted by your N nodes.
What it looks like you have:
1 mount points/directories exported from servers raidvol-0 and raidvol-1
1 mount points/directories exported from servers server002 and bee003
A distributed volume across all your client nodes (mirrors 2-15...)
Are you exporting 2 mounts from 4 different servers?
Are you really trying to create a raid0 array across all your client nodes?
If you want what I think you want your config files would/should be much
simpler.
Here is my guess:
Here's glusterfsd.vol:
**********************************
<snip of cmd line from volgen>
# TRANSPORT-TYPE tcp
# PORT 6996
volume posix
type storage/posix
option directory /glusterfs1
end-volume
volume locks
type features/locks
subvolumes posix
end-volume
volume brick
type performance/io-threads
option thread-count 8
subvolumes locks
end-volume
volume server
type protocol/server
option transport-type tcp
option auth.addr.brick.allow *
option listen-port 6996
subvolumes brick
end-volume
*************************
glusterfs.vol is
# RAID 1
# TRANSPORT-TYPE tcp
# PORT 6996
volume raidvol-0
type protocol/client
option transport-type tcp
option remote-host server002
option remote-port 6996
option remote-subvolume brick
end-volume
volume raidvol-1
type protocol/client
option transport-type tcp
option remote-host bee003
option remote-port 6996
option remote-subvolume brick
end-volume
volume mirror-0
type cluster/replicate
subvolumes raidvol-0 raidvol-1
end-volume
volume writebehind
type performance/write-behind
option cache-size 4MB
subvolumes mirror-0
end-volume
volume io-cache
type performance/io-cache
option cache-size 1GB
subvolumes writebehind
end-volume
**********************************************
Finally your mount point in /etc/fstab on client nodes 1-N would look like:
/etc/glusterfs/glusterfs.vol /remote_glusterfs1 glusterfs
defaults 0 0
^C
[email protected] wrote:
Chad wrote:
[email protected] wrote:
Chad wrote:
I never used 2.x so I am not sure if this it the issue or not, but do
you have an allow line in your glusterfsd.vol file?
That is how glusterfs authenticates clients, if the server can not
authenticate the client the mount will fail.
My allow line looks like this:
option auth.addr.tcb_brick.allow *
Obviously your auth line would need to change "tcb_brick" to the name
of your volume export.
Oh. Um, sorry, this is completely unobvious to me. Also, I would have
assumed that glusterfs-volgen would have put the correct name in which
it gleaned from the command line.
So, if the volume is /glusterfsr1, and the export for mounting is
/export/glusterfs1, tcb_brick should be changed to, um,
/export/glusterfs1?
If you just post your .vol files we can read them and tell you if anything
is wrong.
One other thing - is "brick" a reserved name, or does there have to be an
actual subdirectory called brick?
Here's glusterfsd.vol:
**********************************
<snip of cmd line from volgen>
# TRANSPORT-TYPE tcp
# PORT 6996
volume posix
type storage/posix
option directory /glusterfs1
end-volume
volume locks
type features/locks
subvolumes posix
end-volume
volume brick
type performance/io-threads
option thread-count 8
subvolumes locks
end-volume
volume server
type protocol/server
option transport-type tcp
option auth.addr.brick.allow *
option listen-port 6996
subvolumes brick
end-volume
*************************
And the short version of glusterfs.vol is
# RAID 1
# TRANSPORT-TYPE tcp
# PORT 6996
volume raidvol-0
type protocol/client
option transport-type tcp
option remote-host raidvol-0
option remote-port 6996
option remote-subvolume brick
end-volume
volume raidvol-1
type protocol/client
option transport-type tcp
option remote-host raidvol-1
option remote-port 6996
option remote-subvolume brick
end-volume
volume server002:/glusterfs1
type protocol/client
option transport-type tcp
option remote-host server002:/glusterfs1
option remote-port 6996
option remote-subvolume brick
end-volume
<servers 003-34, minus the several dead ones...>
volume mirror-0
type cluster/replicate
subvolumes raidvol-0 raidvol-1
end-volume
volume mirror-1
type cluster/replicate
subvolumes server002:/glusterfs1 bee003:/glusterfs1
end-volume
<mirrors 2-15...>
volume distribute
type cluster/distribute
subvolumes mirror-0 mirror-1 mirror-2 mirror-3 mirror-4 mirror-5
mirror-6 mirror-7 mirror-8 mirror-9 mirror-10 mirror-11 mirror-12
mirror-13 mirror-14 mirror-15
end-volume
volume writebehind
type performance/write-behind
option cache-size 4MB
subvolumes distribute
end-volume
volume io-cache
type performance/io-cache
option cache-size 1GB
subvolumes writebehind
end-volume
**********************************************
mark
_______________________________________________
Gluster-users mailing list
[email protected]
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users