Hi all,
I just finished setting up a new Ceph cluster (Luminous 12.2.7, 3xMON
nodes and 6xOSDs nodes, BlueStore OSD on sata hdd with WAL/DB on
separated NVMe devices, 2x10 Gbs network per node, 3 replicas by pool)
I created a CephFS pool : data pool uses hdd OSDs and metadata pool uses
dedicated NVMe OSDs. I deployed 3 MDS demons (2 active + 1 failover).
My Ceph cluster is in 'HEALTH_OK' state, for now everything seems to be
working perfectly.
My question is regarding the cephfs client, and especially the huge
performance gap between the fuse client and the kernel one.
On the same writing test, done one after the other, I find a factor 55
between the 2 !
Here is an example from a client (connected to 10 Gbs on the same LAN) :
CephFS Fuse client :
# ceph-fuse -m /FIRST_MON_NODE_IP/:6789 /mnt/ceph_newhome/
# time sh -c "dd if=/dev/zero
of=/mnt/ceph_newhome/test_io_fuse_mount.tmp bs=4k count=2000000 && sync"
2000000+0 records in
2000000+0 records out
8192000000 bytes (8.2 GB, 7.6 GiB) copied, 305.57 s, 26.8 MB/s
real 5m5.607s
user 0m1.784s
sys 0m28.584s
CephFS Kernel driver :
# umount /mnt/ceph_newhome
# mount -t ceph /FIRST_MON_NODE_IP/:6789:/ /mnt/ceph_newhome -o
name=admin,secret=`ceph-authtool -p /etc/ceph/ceph.client.admin.keyring`
# time sh -c "dd if=/dev/zero
of=/mnt/ceph_newhome/test_io_kernel_mount.tmp bs=4k count=2000000 && sync"
2000000+0 records in
2000000+0 records out
8192000000 bytes (8.2 GB, 7.6 GiB) copied, 5.47228 s, 1.5 GB/s
real 0m15.161s
user 0m0.444s
sys 0m5.024s
I'm impressed by the write speed with the kernel driver and, as I must
be able to use this kernel driver on my client systems, I'm
statisfied...but I would like to know if such a difference is normal, or
are there options/optimizations that improve the IO speed with the fuse
client ? (I'm thinking in particular of the recovery scenario where the
kernel driver is no longer mounted following a system update/upgrade and
where I have to use the fuse client as a temporary replacement...)
Thanks for your suggestions,
Hervé
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com