We use ceph on a 3 server Debian proxmox cluster that has 4 x4TB disks each
in a shared Ceph cluster.  The io wait is much too high (around 9%).  The
default file system is XFS and I found a suggestion that by adding this line

osd mount options xfs = rw,noatime,inode64

into the global area of our ceph.conf file we can expect significant
performance improvement.  Should we have any concerns about adding this into
our live ceph.conf file?  Here is our current config file.  All suggestions
welcome.

[global]
         auth client required = cephx
         auth cluster required = cephx
         auth service required = cephx
         cluster network = 10.10.10.0/24
         filestore xattr use omap = true
         fsid = a1ee9e98-3b8d-4929-816d-ed15576efaa9
         keyring = /etc/pve/priv/$cluster.$name.keyring
         osd journal size = 20480
         osd pool default min size = 1
         public network = 10.10.10.0/24

mon_pg_warn_max_per_osd = 0

osd_op_threads = 5
osd_op_num_threads_per_shard = 1
osd_op_num_shards = 25
#osd_op_num_sharded_pool_threads = 25
filestore_op_threads = 4

ms_nocrc = true
filestore_fd_cache_size = 64
filestore_fd_cache_shards = 32

[client]
#rbd cache = true
rbd cache size = 67108864 # (64MB)
rbd cache max dirty = 50331648 # (48MB)
rbd cache target dirty = 33554432 # (32MB)
rbd cache max dirty age = 2
rbd cache writethrough until flush = true

[osd]
         keyring = /var/lib/ceph/osd/ceph-$id/keyring


_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to