You should really upgrade to gluster 2.0.6, there has been many bug fixes.

ls



On Sep 9, 2009, at 4:36 AM, Roland Rabben <[email protected]> wrote:

Hi
I am using GlusterFS 2.02 on Ubuntu 9.04 64 bit. I have 4 data-nodes and 3
clients. Se my vol files at the end of this email.

After adding more disks to my data-nodes for more capacity and reconfiguring
GlusterFS to include those drives I am experiencing problems.

I am getting "No such file or directory" if I try to copy a new file into an
existing directory.
However, if I copy a new file into a new directory everyting works fine.

It seems that if I create the folderstructure from the old data- nodes on the
new disks, everything works fine.

So my questions are?

1. Am I doing somthing wrong in the upgrade process?
2. Do I need to manually create the existing folders on the new hard drives?
3. Self heal does not fix this. Shouldn't it?
4. Is there a tool that will create the folderstructure on the new disks for
me?


Client vol file example:
=================
# DN-000
volume dn-000-01
       type protocol/client
       option transport-type tcp
       option remote-host dn-000
       option remote-subvolume brick-01
end-volume

volume dn-000-02
       type protocol/client
       option transport-type tcp
       option remote-host dn-000
       option remote-subvolume brick-02
end-volume

volume dn-000-03
       type protocol/client
       option transport-type tcp
       option remote-host dn-000
       option remote-subvolume brick-03
end-volume

volume dn-000-04
       type protocol/client
       option transport-type tcp
       option remote-host dn-000
       option remote-subvolume brick-04
end-volume


volume dn-000-ns
       type protocol/client
       option transport-type tcp
       option remote-host dn-000
       option remote-subvolume brick-ns
end-volume

# DN-001
volume dn-001-01
       type protocol/client
       option transport-type tcp
       option remote-host dn-001
       option remote-subvolume brick-01
end-volume

volume dn-001-02
       type protocol/client
       option transport-type tcp
       option remote-host dn-001
       option remote-subvolume brick-02
end-volume

volume dn-001-03
       type protocol/client
       option transport-type tcp
       option remote-host dn-001
       option remote-subvolume brick-03
end-volume

volume dn-001-04
       type protocol/client
       option transport-type tcp
       option remote-host dn-001
       option remote-subvolume brick-04
end-volume

volume dn-001-ns
       type protocol/client
       option transport-type tcp
       option remote-host dn-001
       option remote-subvolume brick-ns
end-volume

# DN-002
volume dn-002-01
       type protocol/client
       option transport-type tcp
       option remote-host dn-002
       option remote-subvolume brick-01
end-volume

volume dn-002-02
       type protocol/client
       option transport-type tcp
       option remote-host dn-002
       option remote-subvolume brick-02
end-volume

volume dn-002-03
       type protocol/client
       option transport-type tcp
       option remote-host dn-002
       option remote-subvolume brick-03
end-volume

volume dn-002-04
       type protocol/client
       option transport-type tcp
       option remote-host dn-002
       option remote-subvolume brick-04
end-volume

# DN-003
volume dn-003-01
       type protocol/client
       option transport-type tcp
       option remote-host dn-003
       option remote-subvolume brick-01
end-volume

volume dn-003-02
       type protocol/client
       option transport-type tcp
       option remote-host dn-003
       option remote-subvolume brick-02
end-volume

volume dn-003-03
       type protocol/client
       option transport-type tcp
       option remote-host dn-003
       option remote-subvolume brick-03
end-volume

volume dn-003-04
       type protocol/client
       option transport-type tcp
       option remote-host dn-003
       option remote-subvolume brick-04
end-volume

# Replicate data between the servers
# Use pairs, but swtich the order to distribute read load
volume repl-000-001-01
       type cluster/replicate
       subvolumes dn-000-01 dn-001-01
end-volume

volume repl-000-001-02
       type cluster/replicate
       subvolumes dn-001-02 dn-000-02
end-volume

volume repl-000-001-03
       type cluster/replicate
       subvolumes dn-000-03 dn-001-03
end-volume

volume repl-000-001-04
       type cluster/replicate
       subvolumes dn-001-04 dn-000-04
end-volume


volume repl-002-003-01
       type cluster/replicate
       subvolumes dn-002-01 dn-003-01
end-volume

volume repl-002-003-02
       type cluster/replicate
       subvolumes dn-003-02 dn-002-02
end-volume

volume repl-002-003-03
       type cluster/replicate
       subvolumes dn-002-03 dn-003-03
end-volume

volume repl-002-003-04
       type cluster/replicate
       subvolumes dn-003-04 dn-002-04
end-volume


# Also replicate the namespace
volume repl-ns
       type cluster/replicate
       subvolumes dn-000-ns dn-001-ns
end-volume

# Distribute the data using the "adaptive least usage" scheduler
# We have a 5GB treshold for disk-usage first, then we look at write- usage,
and finanly read-usage
volume dfs
       type cluster/unify
       option namespace repl-ns
       option scheduler alu
       option scheduler.limits.min-free-disk 5%
       option scheduler.alu.order disk-usage:write-usage:read-usage
       option scheduler.alu.disk-usage.entry-threshold 5GB
       option scheduler.alu.disk-usage.exit-threshold 1GB
       option scheduler.alu.write-usage.entry-threshold 25
       option scheduler.alu.write-usage.exit-threshold 5
       option scheduler.alu.read-usage.entry-threshold 25
       option scheduler.alu.read-usage.exit-threshold 5
       subvolumes repl-000-001-01 repl-000-001-02 repl-000-001-03
repl-000-001-04 repl-002-003-01 repl-002-003-02 repl-002-003-03
repl-002-003-04

end-volume

# Enable write-behind to decrease write latency
volume wb
       type performance/write-behind
       option flush-behind off
       option cache-size 128MB
       subvolumes dfs
end-volume

volume cache
       type performance/io-cache
       option cache-size 1024MB
       subvolumes wb
end-volume





Server vol file example:
==================
# The posix volumes
volume posix-01
   type storage/posix
   option directory /mnt/data01
end-volume

volume posix-02
   type storage/posix
   option directory /mnt/data02
end-volume

volume posix-03
       type storage/posix
       option directory /mnt/data03
end-volume

volume posix-04
       type storage/posix
       option directory /mnt/data04
end-volume


volume posix-ns
   type storage/posix
   option directory /var/lib/glusterfs/ns
end-volume

# Add locking capabilities
volume locks-01
   type features/locks
   subvolumes posix-01
end-volume

volume locks-02
   type features/locks
   subvolumes posix-02
end-volume

volume locks-03
       type features/locks
       subvolumes posix-03
end-volume

volume locks-04
       type features/locks
       subvolumes posix-04
end-volume


volume locks-ns
   type features/locks
   subvolumes posix-ns
end-volume

# Finally add threads to the briks
volume brick-01
   type performance/io-threads
   option thread-count 8
   subvolumes locks-01
end-volume

volume brick-02
   type performance/io-threads
   option thread-count 8
   subvolumes locks-02
end-volume

volume brick-03
       type performance/io-threads
       option thread-count 8
       subvolumes locks-03
end-volume

volume brick-04
       type performance/io-threads
       option thread-count 8
       subvolumes locks-04
end-volume


volume brick-ns
   type performance/io-threads
   option thread-count 8
   subvolumes locks-ns
end-volume

# Mount the posix drives as a network drive
volume server
   type protocol/server
   option transport-type tcp
   subvolumes brick-01 brick-02 brick-03 brick-04 brick-ns
   option auth.addr.brick-01.allow 10.0.*
   option auth.addr.brick-02.allow 10.0.*
   option auth.addr.brick-03.allow 10.0.*
   option auth.addr.brick-04.allow 10.0.*
   option auth.addr.brick-ns.allow 10.0.*
end-volume



Regards

Roland Rabben
Founder & CEO Jotta AS
Cell: +47 90 85 85 39
Phone: +47 21 04 29 00
Email: [email protected]
_______________________________________________
Gluster-users mailing list
[email protected]
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
[email protected]
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

Reply via email to