Hi,

Would my setup bennefit from the next release too?
I am using CentOS 5.5 with 32 bit gluster and fuse patch on 2 servers with 1 GB memory and 2 Xeon cpu's but maildir performance is so slow. Sending an email is like watching 10 seconds to my Thunderbird saying: 'Copying message to sent folder' The strange thing is that the email sent to myself, is already in my inbox (read) but the writing of a plain text file to the Sent folder is taking ages. I tried the default gen script settings, some versions on some howto pages but none really helped making my write performance faster. Btw I am on 2 VPS systems with kernel 2.6.31 per compiled by the hosting party (with fuse compiled in).

I am looking forward for a faster small file performance hit (thinking that that is the issue with my setup).

Regards,

Michiel

On 5/19/10 3:06 PM, Tejas N. Bhise wrote:
Roberto,

We recently made some code changes we think will considerably help small file 
performance -

selective readdirp - http://patches.gluster.com/patch/3203/
dht lookup revalidation optimization - http://patches.gluster.com/patch/3204/
updated write-behind default values - http://patches.gluster.com/patch/3223/

These are tentatively scheduled to go into 3.0.5.
If its possible for you, I would suggest you test them in a non-production 
environment
and see if it  helps with distribute config itself.

Please do not use in production, for that wait for the release which these 
patches go in.

Do let me know if you have any questions about this.

Regards,
Tejas.


----- Original Message -----
From: "Roberto Franchini"<[email protected]>
To: "gluster-users"<[email protected]>
Sent: Wednesday, May 19, 2010 5:29:47 PM
Subject: Re: [Gluster-users] Running Gluster client/server on single process

On Sat, May 15, 2010 at 10:06 PM, Craig Carl<[email protected]>  wrote:
Robert -
       NUFA has been deprecated and doesn't apply to any recent version of
Gluster. What version are you running? ('glusterfs --version')

We run 3.0.4 on ubuntu 9.10 and 10.04 server.
Is there a way to mimic NUFA behaviour?

We are using gluster to store Lucene indexes. Indexes are created
locally from milions of small files and then copied to the storage.
I tried read this little files from gluster but was too slow.
So maybe a NUFA way, e.g. prefer local disk for read, could improve performance.
Let me know :)

At the moment we use dht/replicate:


#CLIENT

volume remote1
  type protocol/client
  option transport-type tcp
  option remote-host zeus
  option remote-subvolume brick
end-volume

volume remote2
  type protocol/client
  option transport-type tcp
  option remote-host hera
  option remote-subvolume brick
end-volume

volume remote3
  type protocol/client
  option transport-type tcp
  option remote-host apollo
  option remote-subvolume brick
end-volume

volume remote4
  type protocol/client
  option transport-type tcp
  option remote-host demetra
  option remote-subvolume brick
end-volume

volume remote5
  type protocol/client
  option transport-type tcp
  option remote-host ade
  option remote-subvolume brick
end-volume

volume remote6
  type protocol/client
  option transport-type tcp
  option remote-host athena
  option remote-subvolume brick
end-volume

volume replicate1
  type cluster/replicate
  subvolumes remote1 remote2
end-volume

volume replicate2
  type cluster/replicate
  subvolumes remote3 remote4
end-volume

volume replicate3
  type cluster/replicate
  subvolumes remote5 remote6
end-volume

volume distribute
  type cluster/distribute
  subvolumes replicate1 replicate2 replicate3
end-volume

volume writebehind
  type performance/write-behind
  option window-size 1MB
  subvolumes distribute
end-volume

volume quickread
  type performance/quick-read
  option cache-timeout 1         # default 1 second
#  option max-file-size 256KB        # default 64Kb
  subvolumes writebehind
end-volume

### Add io-threads for parallel requisitions
volume iothreads
  type performance/io-threads
  option thread-count 16 # default is 16
  subvolumes quickread
end-volume


#SERVER

volume posix
  type storage/posix
  option directory /data/export
end-volume

volume locks
  type features/locks
  subvolumes posix
end-volume

volume brick
  type performance/io-threads
  option thread-count 8
  subvolumes locks
end-volume

volume server
  type protocol/server
  option transport-type tcp
  option auth.addr.brick.allow *
  subvolumes brick
end-volume
_______________________________________________
Gluster-users mailing list
[email protected]
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

Reply via email to