btw- I'm using Gluster 3.0.5 and centos55 x86_64 in this case. (updated
subject line too)
Jeremy
On 11/19/2010 7:27 PM, Craig Carl wrote:
Jeremy -
We use ramfs to test with here, it should work. If you want to
troubleshoot that issue please let me know.
Craig
On 11/19/2010 05:12 PM, Jeremy Enos wrote:
Looks like RAMFS has the same issue TMPFS does... I'm looking into
the RNA networks. Thanks-
Jeremy
On 11/18/2010 6:55 PM, Craig Carl wrote:
On 11/18/2010 04:33 PM, Jeremy Enos wrote:
Post is almost a year old... ever any response here? Is it
possible to export tmpfs locations w/ gluster?
thx-
Jeremy
On 12/1/2009 8:14 AM, Alexander Beregalov wrote:
Hi
Is it possible to start server on tmpfs ?
It is announced that stripe can be used over tmpfs, but stripe is
client plugin and server cannot start on tmpfs because lack of
xattrs.
I am trying to setup a small fast storage for small files
(compiling purpose).
I made ext2 with xattr on ramdisk on 4 hosts, joined them with
replicate plugin and mounted it on one client. Also io-cache,
write-behind, quick-read and io threads were used on client side.
I compiled linux kernel, performance was 10 times worse than tmpfs
exported by NFS on one node.
Any ideas?
Servers:
volume posix
type storage/posix # POSIX FS translator
option directory /mnt/ost # Export this directory
end-volume
volume locks
type features/locks
option mandatory-locks on
subvolumes posix
end-volume
volume brick
type performance/io-threads
option thread-count 4 # Four CPUs
subvolumes locks
end-volume
volume server
type protocol/server
option transport-type tcp
option transport.socket.nodelay on
subvolumes brick
option auth.addr.brick.allow * # Allow access to "brick" volume
end-volume
Client:
volume server1
type protocol/client
option transport-type tcp
option remote-host<IP>
option transport.socket.nodelay on
option remote-subvolume brick # name of the remote volume
end-volume
<the same for server[2-4]>
volume replicated
type cluster/replicate
subvolumes server1 server2 server3 server4
end-volume
volume iocache
type performance/io-cache
option cache-size 1000MB # default is 32MB
option priority *.h:3,*.o:2,*:1 # default is '*:0'
option cache-timeout 1 # default is 1 second
subvolumes replicated
end-volume
volume writeback
type performance/write-behind
option cache-size 500MB # default is equal to aggregate-size
option flush-behind off # default is 'off'
subvolumes iocache
end-volume
volume quickread
type performance/quick-read
option cache-timeout 1 # default 1 second
option max-file-size 256KB # default 64Kb
subvolumes iocache
end-volume
volume iothreads
type performance/io-threads
option thread-count 16 # default is 16
subvolumes quickread
end-volume
_______________________________________________
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
Jeremy -
xattrs are required for any Gluster setup regardless of the
volume design, tmpfs doesn't support user xattrs. RamFS works well,
if you have the budget FisionIO cards are very fast and work well
with Gluster, so does the solution from RNA Networks.
(http://www.rnanetworks.com/)
Thanks,
Craig
-->
Craig Carl
Senior Systems Engineer
Gluster
_______________________________________________
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
Gluster-users@gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users