Oops - not paying close enough attention here - my bad

Here is the actual performance of the NFS mounted drive:
[root@o01 ~]# dd if=/dev/sdc of=/dev/null bs=4k status=progress
57427484672 bytes (57 GB, 53 GiB) copied, 200 s, 287 MB/s

It's still running but this is more indicative of how thing are running...

Ron Gage

On 8/2/2025 12:45 PM, Ron Gage wrote:
Re: [ceph-users] Performance issues

All:

On 8/2/2025 12:20 PM, Anthony D'Atri wrote:
I suspect that the NVMe drives are client-class located on external systems, 
each with a conventional filesystem with big files that are exported to VMs 
that mount them as block devices.  A lot of layers and media that aren’t up to 
a sustained workload.

The NVME is a Samsung 990 Plus.  Not exactly entrprise grade, but it should do fairly well.  It is also fairly new having picked it up yesterday.  It's not going to be 100 MBit - for sure.

The NVME connection path:
NVME -> USB C Interface -> Nas Server (ubuntu 24.04) -> 2.5 GBit Ethernet (NFS) -> ProxMox vmbr1

Raw device performance from my "nas" system:
root@nas:~# hdparm -t /dev/sdb

/dev/sdb:
 Timing buffered disk reads: 1310 MB in  3.00 seconds = 436.21 MB/sec
root@nas:~#

On Aug 2, 2025, at 12:13 PM,dar...@soothill.com wrote:

As you rightly point out the 110MB/s it sounds very much like the traffic is 
going through the wrong interface or being limited.
Limited - I agree.  Wrong interface is unlikely - main network (192.168.0.0/23) has no route to the ceph network and vice versa - no gateway defined on the ceph network.
So am I correct in my reading of this that this a virtual Ceph environment 
running on Proxmox?

What do you mean by this statement? " All Ceph drives are exposed and an NFS 
mounted NVME drive. “

Do I take this to mean that you 4 servers are all mounting the same NVME device 
over NFS? Just a bit confused as to the exact hardware setup here.

What is the performance you can get from a single Ceph OSD? Just do a simple dd 
to read not write to an OSD drive.

[root@o01 ~]# dd if=/dev/sdb of=10M bs=4k status=progress
5117947904 bytes (5.1 GB, 4.8 GiB) copied, 23 s, 223 MB/s
1310720+0 records in
1310720+0 records out
5368709120 bytes (5.4 GB, 5.0 GiB) copied, 23.9114 s, 225 MB/s
[root@o01 ~]#

Also, `ceph tell osd.X bench`

[root@c01 ~]# ceph tell osd.2 bench
{
     "bytes_written": 1073741824,
     "blocksize": 4194304,
     "elapsed_sec": 3.7980597760000001,
     "bytes_per_sec": 282707984.42536151,
     "iops": 67.402835947361353
}
[root@c01 ~]#

Ron Gage



Darren


On 2 Aug 2025, at 15:25, Ron Gage<r...@rongage.org> wrote:

Hello from Detroit MI:

I have been doing some limited benchmarking of a Squid cluster. The arrangement 
of the cluster:
Server        Function
c01             MGR, MON
c02             MGR, MON
o01            OSD
o02            OSD
o03            OSD
o04            OSD

Each OSD has 2 x NVME disks for Ceph, each at 370 Gig

The backing network is as follows:
ens18        Gigabit, mon-ip (192.168.0.0/23) regular MTU (1500)
ens19        2.5 Gigabit, Cluster Network (10.0.0.0/24) Jumbo MTU (9000)

Behind all this is a small ProxMox cluster.  All Ceph machines are running on a 
single node.  All Ceph drives are exposed and an NFS mounted NVME drive.  All 
Ceph OSD drives are mounted with no cache and single controller per drive.  
Networking bridges are all set to either MTU 9000 or MTU 1500 as appropriate.

iPerf3 is showing 2.46 Gbit/sec between servers c01 and o01 on the ens19 
network.  Firewall is off all the way around.  OS is CentOS 10.  SELinux of 
disabled.  No network enhancements have been performed (increasing send/rcv 
buffer size, queue length, etc).

The concern given all this: rados bench can't exceed 110 MB/s in all tests.  In 
fact if I didn't know better I would swear that the traffic is being either 
throttled or is somehow routing through a 1Gbit network.  The numbers that are 
returning from rados bench are acting like saturation at Gigabit and not 
exhibiting any evidence of being on a 2.5 Gbit network.  Monitoring at both 
Ceph and ProxMox consoles confirm the same.  Cluster traffic is confirmed to be 
going out ens19 - tested via tcpdump.

Typical command line used for rados bench: rados bench -p s3block 20 write

What the heck am I doing wrong here?

Ron Gage

_______________________________________________
ceph-users mailing list --ceph-users@ceph.io
To unsubscribe send an email toceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list --ceph-users@ceph.io
To unsubscribe send an email toceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to