Done..... but now have this error in /var/log/glusterfs/mnt-mail.log
[2010-05-11 20:40:21] E [quick-read.c:2194:init] quickread: FATAL:
volume (quickread) not configured with exactly one child
[2010-05-11 20:40:21] E [xlator.c:839:xlator_init_rec] quickread:
Initialization of volume 'quickread' failed, review your volfile again
[2010-05-11 20:40:21] E [glusterfsd.c:591:_xlator_graph_init]
glusterfs: initializing translator failed
[2010-05-11 20:40:21] E [glusterfsd.c:1394:main] glusterfs: translator
initialization failed. exiting
I did change the option max-file-size but when I received the errors I
put it back to 64kb.
New glusterfs.vol
## file auto generated by /bin/glusterfs-volgen (mount.vol)
# Cmd line:
# $ /bin/glusterfs-volgen --name repstore1 --raid 1 192.168.1.15:/fs
192.168.1.16:/fs
# RAID 1
# TRANSPORT-TYPE tcp
volume 192.168.1.16-1
type protocol/client
option transport-type tcp
option remote-host 192.168.1.16
option transport.socket.nodelay on
option transport.remote-port 6996
option remote-subvolume brick1
end-volume
volume 192.168.1.15-1
type protocol/client
option transport-type tcp
option remote-host 192.168.1.15
option transport.socket.nodelay on
option transport.remote-port 6996
option remote-subvolume brick1
end-volume
volume mirror-0
type cluster/replicate
subvolumes 192.168.1.15-1 192.168.1.16-1
end-volume
#volume readahead
# type performance/read-ahead
# option page-count 4
# subvolumes mirror-0
#end-volume
#volume iocache
# type performance/io-cache
# option cache-size `echo $(( $(grep 'MemTotal' /proc/meminfo | sed
's/[^0-9]//g') / 5120 ))`MB
# option cache-timeout 1
# subvolumes readahead
#end-volume
volume quickread
type performance/quick-read
option cache-timeout 10
option max-file-size 64kB
# subvolumes iocache
end-volume
volume writebehind
type performance/write-behind
option cache-size 4MB
subvolumes quickread
end-volume
volume statprefetch
type performance/stat-prefetch
subvolumes writebehind
end-volume
Bryan McGuire
On May 11, 2010, at 7:37 PM, Craig Carl wrote:
Bryan -
Your server vol file isn't perfect for large rsync operations. The
changes I'm recommending will improve your rsync performance if you
are moving a lot of small files. Please backup the current file
before making any changes. You should comment out "readahead" and
"iocache". In the quickread section change the "option cache-
timeout" to 10 and change the "max-file-size" to the size of the
largest file of which you have many, rounded up to the nearest
factor of 4.
After you have made the changes across all the storage nodes
please restart Gluster and measure the throughput again.
Thanks,
Craig
----- Original Message -----
From: "Bryan McGuire" <[email protected]>
To: "Craig Carl" <[email protected]>
Cc: [email protected]
Sent: Tuesday, May 11, 2010 2:31:48 PM GMT -08:00 US/Canada Pacific
Subject: Re: [Gluster-users] mnt-mail.log
Here they are,
For msvr1 - 192.168.1.15
## file auto generated by /bin/glusterfs-volgen (export.vol)
# Cmd line:
# $ /bin/glusterfs-volgen --name repstore1 --raid 1 192.168.1.15:/fs
192.168.1.16:/fs
volume posix1
type storage/posix
option directory /fs
end-volume
volume locks1
type features/locks
subvolumes posix1
end-volume
volume brick1
type performance/io-threads
option thread-count 8
subvolumes locks1
end-volume
volume server-tcp
type protocol/server
option transport-type tcp
option auth.addr.brick1.allow *
option transport.socket.listen-port 6996
option transport.socket.nodelay on
subvolumes brick1
end-volume
## file auto generated by /bin/glusterfs-volgen (mount.vol)
# Cmd line:
# $ /bin/glusterfs-volgen --name repstore1 --raid 1 192.168.1.15:/fs
192.168.1.16:/fs
# RAID 1
# TRANSPORT-TYPE tcp
volume 192.168.1.16-1
type protocol/client
option transport-type tcp
option remote-host 192.168.1.16
option transport.socket.nodelay on
option transport.remote-port 6996
option remote-subvolume brick1
end-volume
volume 192.168.1.15-1
type protocol/client
option transport-type tcp
option remote-host 192.168.1.15
option transport.socket.nodelay on
option transport.remote-port 6996
option remote-subvolume brick1
end-volume
volume mirror-0
type cluster/replicate
subvolumes 192.168.1.15-1 192.168.1.16-1
end-volume
volume readahead
type performance/read-ahead
option page-count 4
subvolumes mirror-0
end-volume
volume iocache
type performance/io-cache
option cache-size `echo $(( $(grep 'MemTotal' /proc/meminfo | sed
's/[^0-9]//g') / 5120 ))`MB
option cache-timeout 1
subvolumes readahead
end-volume
volume quickread
type performance/quick-read
option cache-timeout 1
option max-file-size 64kB
subvolumes iocache
end-volume
volume writebehind
type performance/write-behind
option cache-size 4MB
subvolumes quickread
end-volume
volume statprefetch
type performance/stat-prefetch
subvolumes writebehind
end-volume
For msvr2 192.168.1.16
## file auto generated by /bin/glusterfs-volgen (export.vol)
# Cmd line:
# $ /bin/glusterfs-volgen --name repstore1 --raid 1 192.168.1.15:/fs
192.168.1.16:/fs
volume posix1
type storage/posix
option directory /fs
end-volume
volume locks1
type features/locks
subvolumes posix1
end-volume
volume brick1
type performance/io-threads
option thread-count 8
subvolumes locks1
end-volume
volume server-tcp
type protocol/server
option transport-type tcp
option auth.addr.brick1.allow *
option transport.socket.listen-port 6996
option transport.socket.nodelay on
subvolumes brick1
end-volume
## file auto generated by /bin/glusterfs-volgen (mount.vol)
# Cmd line:
# $ /bin/glusterfs-volgen --name repstore1 --raid 1 192.168.1.15:/fs
192.168.1.16:/fs
# RAID 1
# TRANSPORT-TYPE tcp
volume 192.168.1.16-1
type protocol/client
option transport-type tcp
option remote-host 192.168.1.16
option transport.socket.nodelay on
option transport.remote-port 6996
option remote-subvolume brick1
end-volume
volume 192.168.1.15-1
type protocol/client
option transport-type tcp
option remote-host 192.168.1.15
option transport.socket.nodelay on
option transport.remote-port 6996
option remote-subvolume brick1
end-volume
volume mirror-0
type cluster/replicate
subvolumes 192.168.1.15-1 192.168.1.16-1
end-volume
volume readahead
type performance/read-ahead
option page-count 4
subvolumes mirror-0
end-volume
volume iocache
type performance/io-cache
option cache-size `echo $(( $(grep 'MemTotal' /proc/meminfo | sed
's/[^0-9]//g') / 5120 ))`MB
option cache-timeout 1
subvolumes readahead
end-volume
volume quickread
type performance/quick-read
option cache-timeout 1
option max-file-size 64kB
subvolumes iocache
end-volume
volume writebehind
type performance/write-behind
option cache-size 4MB
subvolumes quickread
end-volume
volume statprefetch
type performance/stat-prefetch
subvolumes writebehind
end-volume
Bryan McGuire
Senior Network Engineer
NewNet 66
918.231.8063
[email protected]
On May 11, 2010, at 4:26 PM, Craig Carl wrote:
Bryan -
Can you send your client and server vol files?
Thanks,
Craig
--
Craig Carl
Sales Engineer
Gluster, Inc.
Cell - (408) 829-9953 (California, USA)
Office - (408) 770-1884
Gtalk - [email protected]
Twitter - @gluster
----- Original Message -----
From: "Bryan McGuire" <[email protected]>
To: [email protected]
Sent: Tuesday, May 11, 2010 2:12:13 PM GMT -08:00 US/Canada Pacific
Subject: [Gluster-users] mnt-mail.log
Hello,
I have Glusterfs 3.0.4 setup in a two node replication. It appears to
be working just fine. Although I am using rsync to move over 350 Gig
of email files and the process is very slow.
I have noticed the following in the file /var/log/gluserfs/mntl-
mail.log....... could someone explain what the lines mean. Thanks
[2010-05-11 15:41:51] W [fuse-bridge.c:491:fuse_entry_cbk] glusterfs-
fuse: LOOKUP(/_outgoing/retry/201005100854105298-1273614110_8.tmp)
inode (ptr=0xa235c70, ino=808124434, gen=5468694309383486383) found
conflict (ptr=0x2aaaea26c290, ino=808124434, gen=5468694309383486383)
[2010-05-11 15:46:53] W [fuse-bridge.c:491:fuse_entry_cbk] glusterfs-
fuse: LOOKUP(/_outgoing/retry/201005101016464462-1273614395_8.tmp)
inode (ptr=0x2aaaf07f5550, ino=808124438, gen=5468694309383486385)
found conflict (ptr=0x82e4420, ino=808124438,
gen=5468694309383486385)
[2010-05-11 15:46:53] W [fuse-bridge.c:491:fuse_entry_cbk] glusterfs-
fuse: LOOKUP(/_outgoing/retry/201005100830599960-1273614395_8.tmp)
inode (ptr=0x2aaac01da520, ino=808124430, gen=5468694309383486381)
found conflict (ptr=0x60d7a90, ino=808124430,
gen=5468694309383486381)
[2010-05-11 15:46:53] W [fuse-bridge.c:491:fuse_entry_cbk] glusterfs-
fuse: LOOKUP(/_outgoing/retry/201005101417175132-1273614396_8.tmp)
inode (ptr=0x2aaaf07f5550, ino=808124446, gen=5468694309383486389)
found conflict (ptr=0x8eb16e0, ino=808124446,
gen=5468694309383486389)
[2010-05-11 15:51:53] W [fuse-bridge.c:491:fuse_entry_cbk] glusterfs-
fuse: LOOKUP(/_outgoing/retry/201005100749045904-1273614665_8.tmp)
inode (ptr=0x1ec11ee0, ino=808124420, gen=5468694309383486379) found
conflict (ptr=0x2aaaea26bd30, ino=808124420, gen=5468694309383486379)
Bryan
_______________________________________________
Gluster-users mailing list
[email protected]
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
--
Craig Carl
Join us for a Webinar on May 26;
Case Studies: Deploying Open Source Storage Clouds
Sales Engineer
Gluster, Inc.
Cell - (408) 829-9953 (California, USA)
Office - (408) 770-1884
Gtalk - [email protected]
Twitter - @gluster
http://www.gluster.com/files/installation-demo/demo.html
_______________________________________________
Gluster-users mailing list
[email protected]
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users