hi!

I'm almost absolutely new to glusterfs.

Until now we used Storage Platform (3.0.5).
Today we installed Ubuntu 11.04 and glusterfs 3.2.1.

$ cat w-vol-fuse.vol
volume w-vol-client-0
    type protocol/client
    option remote-host gl0
    option remote-subvolume /mnt/brick1
    option transport-type tcp
end-volume

volume w-vol-client-1
    type protocol/client
    option remote-host gl1
    option remote-subvolume /mnt/brick1
    option transport-type tcp
end-volume

volume w-vol-client-2
    type protocol/client
    option remote-host gl2
    option remote-subvolume /mnt/brick1
    option transport-type tcp
end-volume

volume w-vol-client-3
    type protocol/client
    option remote-host gl3
    option remote-subvolume /mnt/brick1
    option transport-type tcp
end-volume

volume w-vol-client-4
    type protocol/client
    option remote-host gl4
    option remote-subvolume /mnt/brick1
    option transport-type tcp
end-volume

volume w-vol-dht
    type cluster/distribute
subvolumes w-vol-client-0 w-vol-client-1 w-vol-client-2 w-vol-client-3 w-vol-client-4
end-volume

volume w-vol-write-behind
    type performance/write-behind
    option cache-size 4MB
    subvolumes w-vol-dht
end-volume

volume w-vol-read-ahead
    type performance/read-ahead
    subvolumes w-vol-write-behind
end-volume

volume w-vol-io-cache
    type performance/io-cache
    option cache-size 128MB
    subvolumes w-vol-read-ahead
end-volume

volume w-vol-quick-read
    type performance/quick-read
    option cache-size 128MB
    subvolumes w-vol-io-cache
end-volume

volume w-vol-stat-prefetch
    type performance/stat-prefetch
    subvolumes w-vol-quick-read
end-volume

volume w-vol
    type debug/io-stats
    option latency-measurement off
    option count-fop-hits off
    subvolumes w-vol-stat-prefetch
end-volume


$ cat w-vol.gl0.mnt-brick1.vol
volume w-vol-posix
    type storage/posix
    option directory /mnt/brick1
end-volume

volume w-vol-access-control
    type features/access-control
    subvolumes w-vol-posix
end-volume

volume w-vol-locks
    type features/locks
    subvolumes w-vol-access-control
end-volume

volume w-vol-io-threads
    type performance/io-threads
    subvolumes w-vol-locks
end-volume

volume w-vol-marker
    type features/marker
    option volume-uuid ad362448-7ef0-49ae-b13c-74cb82ce9be5
    option timestamp-file /etc/glusterd/vols/w-vol/marker.tstamp
    option xtime off
    option quota off
    subvolumes w-vol-io-threads
end-volume

volume /mnt/brick1
    type debug/io-stats
    option latency-measurement off
    option count-fop-hits off
    subvolumes w-vol-marker
end-volume

volume w-vol-server
    type protocol/server
    option transport-type tcp
    option auth.addr./mnt/brick1.allow *
    subvolumes /mnt/brick1
end-volume


There is 5 nodes. 3 have 8 disks in RAID 6 (supermicro server, are controller), 2 have 8 disks in raid5+spare (DL180). Filesystem of datas was created via this command (of course a bit different on HPs):

mkfs.xfs -b size=4096 -d sunit=256,swidth=1536 -L gluster /dev/sda4


The performance is far away that was before. I tried to modify

performance.write-behind-window-size 4MB
gluster volume set w-vol performance.cache-size 128MB
gluster volume set w-vol nfs.disable on

echo 512 > /sys/block/sda/queue/nr_requests
blockdev --setra 16384 /dev/sda
sysctl -w vm.swappiness=5
sysctl -w vm.dirty_background_ratio=3
sysctl -w vm.dirty_ratio=40
sysctl -w kernel.sysrq=0

Nothing really helped.
Can somebody give some instructions?


Thank you,

tamas
_______________________________________________
Gluster-users mailing list
[email protected]
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

Reply via email to