Dear all,
I just subscribed and started reading archives, and I have two points where
I'm unsure of my conf, and I don't want to start wrong...

I used glusterfs-volgen to generate my conf, which is at this point :
* glusterfsd.vol :

volume posix1
  type storage/posix
  option directory /srv/glus
end-volume
volume locks1
    type features/locks
    subvolumes posix1
end-volume
volume brick1
    type performance/io-threads
    option thread-count 8
    subvolumes locks1
end-volume

volume posix2
  type storage/posix
  option directory /srv/glus2
end-volume
volume locks2
    type features/locks
    subvolumes posix2
end-volume
volume brick2
    type performance/io-threads
    option thread-count 8
    subvolumes locks2
end-volume

volume server-tcp
    type protocol/server
    option transport-type tcp
    option auth.addr.brick1.allow *
    option auth.addr.brick2.allow *
    option transport.socket.listen-port 6996
    option transport.socket.nodelay on
    subvolumes brick1 brick2
end-volume
*************************************************** (2 bricks on 1 node, and 1 brick on 2 others nodes not shown)

* glusterfs.vol :
volume 192.168.108.83-1
    type protocol/client
    option transport-type tcp
    option remote-host 192.168.108.83
    option transport.socket.nodelay on
    option remote-port 6996
    option remote-subvolume brick1
end-volume
volume 192.168.108.83-2
    type protocol/client
    option transport-type tcp
    option remote-host 192.168.108.83
    option transport.socket.nodelay on
    option remote-port 6996
    option remote-subvolume brick2
end-volume

volume 192.168.108.13-1
    type protocol/client
    option transport-type tcp
    option remote-host 192.168.108.13
    option transport.socket.nodelay on
    option remote-port 6996
    option remote-subvolume brick1
end-volume
volume 192.168.106.8-1
    type protocol/client
    option transport-type tcp
    option remote-host 192.168.106.8
    option transport.socket.nodelay on
    option remote-port 6996
    option remote-subvolume brick1
end-volume

volume mirror-0
    type cluster/replicate
    subvolumes 192.168.108.83-1 192.168.108.13-1 192.168.106.8-1
end-volume
volume readahead
    type performance/read-ahead
    option page-count 4
    subvolumes mirror-0
end-volume
volume iocache
    type performance/io-cache
option cache-size `echo $(( $(grep 'MemTotal' /proc/meminfo | sed 's/[^0-9]//g') / 5120 ))`MB
    option cache-timeout 1
    subvolumes readahead
end-volume
volume quickread
    type performance/quick-read
    option cache-timeout 1
    option max-file-size 64kB
    subvolumes iocache
end-volume
volume writebehind
    type performance/write-behind
    option cache-size 4MB
    subvolumes quickread
end-volume
volume statprefetch
    type performance/stat-prefetch
    subvolumes writebehind
end-volume

volume readahead2
    type performance/read-ahead
    option page-count 4
    subvolumes 192.168.108.83-2
end-volume
volume iocache2
    type performance/io-cache
option cache-size `echo $(( $(grep 'MemTotal' /proc/meminfo | sed 's/[^0-9]//g') / 5120 ))`MB
    option cache-timeout 1
    subvolumes readahead2
end-volume
volume quickread2
    type performance/quick-read
    option cache-timeout 1
    option max-file-size 64kB
    subvolumes iocache2
end-volume
volume writebehind2
    type performance/write-behind
    option cache-size 4MB
    subvolumes quickread2
end-volume
volume statprefetch2
    type performance/stat-prefetch
    subvolumes writebehind2
end-volume
*************************************************** (one brick replicated 3 times, one brick on just a node)

I've tried this on 4 lenny box, ok (3 servers and 1 client)

So my questions :
- when I mount the node with 2 bricks, I use :
# mount -t glusterfs -o volume-name=statprefetch 192.168.108.83 /mnt/glus
# mount -t glusterfs -o volume-name=statprefetch2 192.168.108.83 /mnt/glus2
ok, but I am right to choose statprefetch ? I guess because it seems on "top of stack",
but maybe a weird thing ? If I try "volume-name=brick1", it doesn't work.
If I say "volume-name=192.168.108.83-1" it's ok, but will I get the performances
enhancements by choosing the "bottom of stack" ?

- when I use glusterfs-volgen asking for a 3 nodes mirror, it complains about "wrong number of nodes for replication" - however, making conf by hand like the one above, replication seems to work ok on the 3 nodes (and I have find somewhere a 3-nodes conf on
internet) : again, doing something weird ?

And a little "wish-list" :
I would be very pleased if some userquota capabilities would be included
on a future glusterfs version (I saw a post about this on the ML, so I join my voice !)


Many thanks for any answer !

Cheers
DB

_______________________________________________
Gluster-users mailing list
[email protected]
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

Reply via email to