We use ceph on a 3 server Debian proxmox cluster that has 4 x4TB disks each
in a shared Ceph cluster. The io wait is much too high (around 9%). The
default file system is XFS and I found a suggestion that by adding this line
osd mount options xfs = rw,noatime,inode64
into the global area of our ceph.conf file we can expect significant
performance improvement. Should we have any concerns about adding this into
our live ceph.conf file? Here is our current config file. All suggestions
welcome.
[global]
auth client required = cephx
auth cluster required = cephx
auth service required = cephx
cluster network = 10.10.10.0/24
filestore xattr use omap = true
fsid = a1ee9e98-3b8d-4929-816d-ed15576efaa9
keyring = /etc/pve/priv/$cluster.$name.keyring
osd journal size = 20480
osd pool default min size = 1
public network = 10.10.10.0/24
mon_pg_warn_max_per_osd = 0
osd_op_threads = 5
osd_op_num_threads_per_shard = 1
osd_op_num_shards = 25
#osd_op_num_sharded_pool_threads = 25
filestore_op_threads = 4
ms_nocrc = true
filestore_fd_cache_size = 64
filestore_fd_cache_shards = 32
[client]
#rbd cache = true
rbd cache size = 67108864 # (64MB)
rbd cache max dirty = 50331648 # (48MB)
rbd cache target dirty = 33554432 # (32MB)
rbd cache max dirty age = 2
rbd cache writethrough until flush = true
[osd]
keyring = /var/lib/ceph/osd/ceph-$id/keyring
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com