Hi Steven.

interesting... 'm quite curious after your post now.

I've migrated our prod. CEPH cluster to 12.2.2 and Bluestore just today and 
haven't heard back anything "bad" from the applications/users so far. 
performance tests on our test cluster were good before, but we use S3/RGW only 
anyhow ;)

there are two things I would like to know/learn... could you try/test and feed 
back?!

- change all your tests to use >=16k block size, see also BStore comments here 
(https://www.mail-archive.com/ceph-users@lists.ceph.com/msg43023.html)
- change your "write.fio" file profile from "rw=randwrite" to "rw=write" (or 
something similar :O ) to compare apples with apples ;)

thanks for your efforts and looking forward for those results ;)

best regards
 Notna

----------------------------------------------------------------------------------

Gesendet: Mittwoch, 03. Januar 2018 um 16:20 Uhr
Von: "Steven Vacaroaia" <ste...@gmail.com>
An: "Brady Deetz" <bde...@gmail.com>
Cc: ceph-users <ceph-users@lists.ceph.com>
Betreff: Re: [ceph-users] ceph luminous - performance issue

Thanks for your willingness to help
 
DELL R620, 1 CPU, 8 cores, 64 GB RAM 
cluster network is using 2 bonded 10 GB NICs ( mode=4), MTU=9000
 
SSD drives are Enterprise grade  - 400 GB SSD  Toshiba PX04SHB040
HDD drives are  - 10k RPM, 600 GB  Toshiba AL13SEB600
 
Steven

 
On 3 January 2018 at 09:41, Brady Deetz 
<bde...@gmail.com[mailto:bde...@gmail.com]> wrote:
Can you provide more detail regarding the infrastructure backing this 
environment? What hard drive, ssd, and processor are you using? Also, what is 
providing networking?
 
I'm seeing 4k blocksize tests here. Latency is going to destroy you. 
 

On Jan 3, 2018 8:11 AM, "Steven Vacaroaia" 
<ste...@gmail.com[mailto:ste...@gmail.com]> wrote:

Hi,
 
I am doing a PoC with 3 DELL R620 and 12 OSD , 3 SSD drives ( one on each 
server), bluestore 
 
I configured the OSD using the following ( /dev/sda is my SSD drive)
ceph-disk prepare --zap-disk --cluster ceph  --bluestore /dev/sde --block.wal 
/dev/sda --block.db /dev/sda
 
Unfortunately both fio and bench tests show much worse performance for the 
pools than for the individual disks 
 
Example:
DISKS
fio --filename=/dev/sda --direct=1 --sync=1 --rw=write --bs=4k --numjobs=14 
--iodepth=1 --runtime=60 --time_based --group_reporting --name=journal-test
 
SSD drive
Jobs: 14 (f=14): [W(14)] [100.0% done] [0KB/465.2MB/0KB /s] [0/119K/0 iops] 
[eta 00m:00s]
 
HD drive
Jobs: 14 (f=14): [W(14)] [100.0% done] [0KB/179.2MB/0KB /s] [0/45.9K/0 iops] 
[eta 00m:00s]
 
POOL
 
fio write.fio
Jobs: 1 (f=0): [w(1)] [100.0% done] [0KB/51428KB/0KB /s] [0/12.9K/0 iops] 
 

 cat write.fio
[write-4M]
description="write test with 4k block"
ioengine=rbd
clientname=admin
pool=scbench
rbdname=image01
iodepth=32
runtime=120
rw=randwrite
bs=4k
 
 
rados bench -p scbench 12 write
 

Max bandwidth (MB/sec): 224
Min bandwidth (MB/sec): 0
Average IOPS:           26
Stddev IOPS:            24
Max IOPS:               56
Min IOPS:               0
Average Latency(s):     0.59819
Stddev Latency(s):      1.64017
Max latency(s):         10.8335
Min latency(s):         0.00475139
 
 
 
 
I must be missing something - any help/suggestions will be greatly appreciated 
 
Here are some specific info 
 

ceph -s
  cluster:
    id:     91118dde-f231-4e54-a5f0-a1037f3d5142
    health: HEALTH_OK
 
  services:
    mon: 1 daemons, quorum mon01
    mgr: mon01(active)
    osd: 12 osds: 12 up, 12 in
 
  data:
    pools:   4 pools, 484 pgs
    objects: 70082 objects, 273 GB
    usage:   570 GB used, 6138 GB / 6708 GB avail
    pgs:     484 active+clean
 
  io:
    client:   2558 B/s rd, 2 op/s rd, 0 op/s wr
 

ceph osd pool ls detail
pool 1 'test-replicated' replicated size 2 min_size 1 crush_rule 0 object_hash 
rjenkins pg_num 128 pgp_num 128 last_change 157 flags hashpspool stripe_width 0 
application rbd
        removed_snaps [1~3]
pool 2 'test-erasure' erasure size 3 min_size 3 crush_rule 1 object_hash 
rjenkins pg_num 128 pgp_num 128 last_change 334 flags hashpspool stripe_width 
8192 application rbd
        removed_snaps [1~5]
pool 3 'rbd' replicated size 2 min_size 1 crush_rule 0 object_hash rjenkins 
pg_num 128 pgp_num 128 last_change 200 flags hashpspool stripe_width 0 
application rbd
        removed_snaps [1~3]
pool 4 'scbench' replicated size 2 min_size 1 crush_rule 0 object_hash rjenkins 
pg_num 100 pgp_num 100 last_change 330 flags hashpspool stripe_width 0
        removed_snaps [1~3]
 
[cephuser@ceph ceph-config]$ ceph osd df tree
ID CLASS WEIGHT  REWEIGHT SIZE  USE    AVAIL %USE  VAR  PGS TYPE NAME
-1       6.55128        - 2237G   198G 2038G     0    0   - root default
-7             0        -     0      0     0     0    0   -     host ods03
-3       2.18475        - 2237G   181G 2055G  8.12 0.96   -     host osd01
 3   hdd 0.54619  1.00000  559G 53890M  506G  9.41 1.11  90         osd.3
 4   hdd 0.54619  1.00000  559G 30567M  529G  5.34 0.63  89         osd.4
 5   hdd 0.54619  1.00000  559G 59385M  501G 10.37 1.22  93         osd.5
 6   hdd 0.54619  1.00000  559G 42156M  518G  7.36 0.87  93         osd.6
-5       2.18178        - 2234G   189G 2044G  8.50 1.00   -     host osd02
 0   hdd 0.54520  1.00000  558G 32460M  526G  5.68 0.67  90         osd.0
 1   hdd 0.54520  1.00000  558G 54578M  504G  9.55 1.12  89         osd.1
 2   hdd 0.54520  1.00000  558G 47761M  511G  8.35 0.98  93         osd.2
 7   hdd 0.54619  1.00000  559G 59584M  501G 10.40 1.22  92         osd.7
-9       2.18475        - 2237G   198G 2038G  8.88 1.04   -     host osd03
 8   hdd 0.54619  1.00000  559G 52462M  508G  9.16 1.08  99         osd.8
10   hdd 0.54619  1.00000  559G 35284M  524G  6.16 0.73  88         osd.10
11   hdd 0.54619  1.00000  559G 71739M  489G 12.53 1.47  87         osd.11
12   hdd 0.54619  1.00000  559G 43832M  516G  7.65 0.90  93         osd.12
                    TOTAL 6708G   570G 6138G  8.50
MIN/MAX VAR: 0.63/1.47  STDDEV: 2.06
 
 _______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com[mailto:ceph-users@lists.ceph.com]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
 _______________________________________________ ceph-users mailing list 
ceph-users@lists.ceph.com 
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com[http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com]
 
 
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to