Sorry for the delay Here are the results when using bs=16k and rw=write ( Note: I am running the command directly on a OSD host as root)
fio /home/cephuser/write.fio write-4M: (g=0): rw=write, bs=16K-16K/16K-16K/16K-16K, ioengine=rbd, iodepth=32 fio-2.2.8 Starting 1 process rbd engine: RBD version: 1.12.0 Jobs: 1 (f=1): [W(1)] [100.0% done] [0KB/172.6MB/0KB /s] [0/11.5K/0 iops] [eta 00m:00s] Here are the results when runnnig with bs=4k and rw=randwrite [root@osd03 ~]# fio /home/cephuser/write.fio write-4M: (g=0): rw=randwrite, bs=4K-4K/4K-4K/4K-4K, ioengine=rbd, iodepth=32 fio-2.2.8 Starting 1 process rbd engine: RBD version: 1.12.0 Jobs: 1 (f=0): [w(1)] [100.0% done] [0KB/54056KB/0KB /s] [0/13.6K/0 iops] [eta 00m:00s] On 3 January 2018 at 15:28, <ceph.nov...@habmalnefrage.de> wrote: > Hi Steven. > > interesting... 'm quite curious after your post now. > > I've migrated our prod. CEPH cluster to 12.2.2 and Bluestore just today > and haven't heard back anything "bad" from the applications/users so far. > performance tests on our test cluster were good before, but we use S3/RGW > only anyhow ;) > > there are two things I would like to know/learn... could you try/test and > feed back?! > > - change all your tests to use >=16k block size, see also BStore comments > here (https://www.mail-archive.com/ceph-users@lists.ceph.com/msg43023.html > ) > - change your "write.fio" file profile from "rw=randwrite" to "rw=write" > (or something similar :O ) to compare apples with apples ;) > > thanks for your efforts and looking forward for those results ;) > > best regards > Notna > > ------------------------------------------------------------ > ---------------------- > > Gesendet: Mittwoch, 03. Januar 2018 um 16:20 Uhr > Von: "Steven Vacaroaia" <ste...@gmail.com> > An: "Brady Deetz" <bde...@gmail.com> > Cc: ceph-users <ceph-users@lists.ceph.com> > Betreff: Re: [ceph-users] ceph luminous - performance issue > > Thanks for your willingness to help > > DELL R620, 1 CPU, 8 cores, 64 GB RAM > cluster network is using 2 bonded 10 GB NICs ( mode=4), MTU=9000 > > SSD drives are Enterprise grade - 400 GB SSD Toshiba PX04SHB040 > HDD drives are - 10k RPM, 600 GB Toshiba AL13SEB600 > > Steven > > > On 3 January 2018 at 09:41, Brady Deetz <bde...@gmail.com[mailto:bdeet > z...@gmail.com]> wrote: > Can you provide more detail regarding the infrastructure backing this > environment? What hard drive, ssd, and processor are you using? Also, what > is providing networking? > > I'm seeing 4k blocksize tests here. Latency is going to destroy you. > > > On Jan 3, 2018 8:11 AM, "Steven Vacaroaia" <ste...@gmail.com[mailto:stef9 > 7...@gmail.com]> wrote: > > Hi, > > I am doing a PoC with 3 DELL R620 and 12 OSD , 3 SSD drives ( one on each > server), bluestore > > I configured the OSD using the following ( /dev/sda is my SSD drive) > ceph-disk prepare --zap-disk --cluster ceph --bluestore /dev/sde > --block.wal /dev/sda --block.db /dev/sda > > Unfortunately both fio and bench tests show much worse performance for the > pools than for the individual disks > > Example: > DISKS > fio --filename=/dev/sda --direct=1 --sync=1 --rw=write --bs=4k > --numjobs=14 --iodepth=1 --runtime=60 --time_based --group_reporting > --name=journal-test > > SSD drive > Jobs: 14 (f=14): [W(14)] [100.0% done] [0KB/465.2MB/0KB /s] [0/119K/0 > iops] [eta 00m:00s] > > HD drive > Jobs: 14 (f=14): [W(14)] [100.0% done] [0KB/179.2MB/0KB /s] [0/45.9K/0 > iops] [eta 00m:00s] > > POOL > > fio write.fio > Jobs: 1 (f=0): [w(1)] [100.0% done] [0KB/51428KB/0KB /s] [0/12.9K/0 iops] > > > cat write.fio > [write-4M] > description="write test with 4k block" > ioengine=rbd > clientname=admin > pool=scbench > rbdname=image01 > iodepth=32 > runtime=120 > rw=randwrite > bs=4k > > > rados bench -p scbench 12 write > > > Max bandwidth (MB/sec): 224 > Min bandwidth (MB/sec): 0 > Average IOPS: 26 > Stddev IOPS: 24 > Max IOPS: 56 > Min IOPS: 0 > Average Latency(s): 0.59819 > Stddev Latency(s): 1.64017 > Max latency(s): 10.8335 > Min latency(s): 0.00475139 > > > > > I must be missing something - any help/suggestions will be greatly > appreciated > > Here are some specific info > > > ceph -s > cluster: > id: 91118dde-f231-4e54-a5f0-a1037f3d5142 > health: HEALTH_OK > > services: > mon: 1 daemons, quorum mon01 > mgr: mon01(active) > osd: 12 osds: 12 up, 12 in > > data: > pools: 4 pools, 484 pgs > objects: 70082 objects, 273 GB > usage: 570 GB used, 6138 GB / 6708 GB avail > pgs: 484 active+clean > > io: > client: 2558 B/s rd, 2 op/s rd, 0 op/s wr > > > ceph osd pool ls detail > pool 1 'test-replicated' replicated size 2 min_size 1 crush_rule 0 > object_hash rjenkins pg_num 128 pgp_num 128 last_change 157 flags > hashpspool stripe_width 0 application rbd > removed_snaps [1~3] > pool 2 'test-erasure' erasure size 3 min_size 3 crush_rule 1 object_hash > rjenkins pg_num 128 pgp_num 128 last_change 334 flags hashpspool > stripe_width 8192 application rbd > removed_snaps [1~5] > pool 3 'rbd' replicated size 2 min_size 1 crush_rule 0 object_hash > rjenkins pg_num 128 pgp_num 128 last_change 200 flags hashpspool > stripe_width 0 application rbd > removed_snaps [1~3] > pool 4 'scbench' replicated size 2 min_size 1 crush_rule 0 object_hash > rjenkins pg_num 100 pgp_num 100 last_change 330 flags hashpspool > stripe_width 0 > removed_snaps [1~3] > > [cephuser@ceph ceph-config]$ ceph osd df tree > ID CLASS WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR PGS TYPE NAME > -1 6.55128 - 2237G 198G 2038G 0 0 - root default > -7 0 - 0 0 0 0 0 - host ods03 > -3 2.18475 - 2237G 181G 2055G 8.12 0.96 - host osd01 > 3 hdd 0.54619 1.00000 559G 53890M 506G 9.41 1.11 90 osd.3 > 4 hdd 0.54619 1.00000 559G 30567M 529G 5.34 0.63 89 osd.4 > 5 hdd 0.54619 1.00000 559G 59385M 501G 10.37 1.22 93 osd.5 > 6 hdd 0.54619 1.00000 559G 42156M 518G 7.36 0.87 93 osd.6 > -5 2.18178 - 2234G 189G 2044G 8.50 1.00 - host osd02 > 0 hdd 0.54520 1.00000 558G 32460M 526G 5.68 0.67 90 osd.0 > 1 hdd 0.54520 1.00000 558G 54578M 504G 9.55 1.12 89 osd.1 > 2 hdd 0.54520 1.00000 558G 47761M 511G 8.35 0.98 93 osd.2 > 7 hdd 0.54619 1.00000 559G 59584M 501G 10.40 1.22 92 osd.7 > -9 2.18475 - 2237G 198G 2038G 8.88 1.04 - host osd03 > 8 hdd 0.54619 1.00000 559G 52462M 508G 9.16 1.08 99 osd.8 > 10 hdd 0.54619 1.00000 559G 35284M 524G 6.16 0.73 88 osd.10 > 11 hdd 0.54619 1.00000 559G 71739M 489G 12.53 1.47 87 osd.11 > 12 hdd 0.54619 1.00000 559G 43832M 516G 7.65 0.90 93 osd.12 > TOTAL 6708G 570G 6138G 8.50 > MIN/MAX VAR: 0.63/1.47 STDDEV: 2.06 > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com[mailto:ceph-users@lists.ceph.com] > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ ceph-users mailing list > ceph-users@lists.ceph.com http://lists.ceph.com/ > listinfo.cgi/ceph-users-ceph.com[http://lists.ceph.com/ > listinfo.cgi/ceph-users-ceph.com] > > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com