Thanks for your willingness to help

DELL R620, 1 CPU, 8 cores, 64 GB RAM
cluster network is using 2 bonded 10 GB NICs ( mode=4), MTU=9000

SSD drives are Enterprise grade  - 400 GB SSD  Toshiba PX04SHB040

HDD drives are  - 10k RPM, 600 GB  Toshiba AL13SEB600


Steven

On 3 January 2018 at 09:41, Brady Deetz <bde...@gmail.com> wrote:

> Can you provide more detail regarding the infrastructure backing this
> environment? What hard drive, ssd, and processor are you using? Also, what
> is providing networking?
>
> I'm seeing 4k blocksize tests here. Latency is going to destroy you.
>
> On Jan 3, 2018 8:11 AM, "Steven Vacaroaia" <ste...@gmail.com> wrote:
>
>> Hi,
>>
>> I am doing a PoC with 3 DELL R620 and 12 OSD , 3 SSD drives ( one on each
>> server), bluestore
>>
>> I configured the OSD using the following ( /dev/sda is my SSD drive)
>> ceph-disk prepare --zap-disk --cluster ceph  --bluestore /dev/sde
>> --block.wal /dev/sda --block.db /dev/sda
>>
>> Unfortunately both fio and bench tests show much worse performance for
>> the pools than for the individual disks
>>
>> Example:
>> DISKS
>> fio --filename=/dev/sda --direct=1 --sync=1 --rw=write --bs=4k
>> --numjobs=14 --iodepth=1 --runtime=60 --time_based --group_reporting
>> --name=journal-test
>>
>> SSD drive
>> Jobs: 14 (f=14): [W(14)] [100.0% done] [0KB/465.2MB/0KB /s] [0/119K/0
>> iops] [eta 00m:00s]
>>
>> HD drive
>> Jobs: 14 (f=14): [W(14)] [100.0% done] [0KB/179.2MB/0KB /s] [0/45.9K/0
>> iops] [eta 00m:00s]
>>
>> POOL
>>
>> fio write.fio
>> Jobs: 1 (f=0): [w(1)] [100.0% done] [0KB/51428KB/0KB /s] [0/12.9K/0 iops]
>>
>>  cat write.fio
>> [write-4M]
>> description="write test with 4k block"
>> ioengine=rbd
>> clientname=admin
>> pool=scbench
>> rbdname=image01
>> iodepth=32
>> runtime=120
>> rw=randwrite
>> bs=4k
>>
>>
>> rados bench -p scbench 12 write
>>
>> Max bandwidth (MB/sec): 224
>> Min bandwidth (MB/sec): 0
>> Average IOPS:           26
>> Stddev IOPS:            24
>> Max IOPS:               56
>> Min IOPS:               0
>> Average Latency(s):     0.59819
>> Stddev Latency(s):      1.64017
>> Max latency(s):         10.8335
>> Min latency(s):         0.00475139
>>
>>
>>
>>
>> I must be missing something - any help/suggestions will be greatly
>> appreciated
>>
>> Here are some specific info
>>
>> ceph -s
>>   cluster:
>>     id:     91118dde-f231-4e54-a5f0-a1037f3d5142
>>     health: HEALTH_OK
>>
>>   services:
>>     mon: 1 daemons, quorum mon01
>>     mgr: mon01(active)
>>     osd: 12 osds: 12 up, 12 in
>>
>>   data:
>>     pools:   4 pools, 484 pgs
>>     objects: 70082 objects, 273 GB
>>     usage:   570 GB used, 6138 GB / 6708 GB avail
>>     pgs:     484 active+clean
>>
>>   io:
>>     client:   2558 B/s rd, 2 op/s rd, 0 op/s wr
>>
>> ceph osd pool ls detail
>> pool 1 'test-replicated' replicated size 2 min_size 1 crush_rule 0
>> object_hash rjenkins pg_num 128 pgp_num 128 last_change 157 flags
>> hashpspool stripe_width 0 application rbd
>>         removed_snaps [1~3]
>> pool 2 'test-erasure' erasure size 3 min_size 3 crush_rule 1 object_hash
>> rjenkins pg_num 128 pgp_num 128 last_change 334 flags hashpspool
>> stripe_width 8192 application rbd
>>         removed_snaps [1~5]
>> pool 3 'rbd' replicated size 2 min_size 1 crush_rule 0 object_hash
>> rjenkins pg_num 128 pgp_num 128 last_change 200 flags hashpspool
>> stripe_width 0 application rbd
>>         removed_snaps [1~3]
>> pool 4 'scbench' replicated size 2 min_size 1 crush_rule 0 object_hash
>> rjenkins pg_num 100 pgp_num 100 last_change 330 flags hashpspool
>> stripe_width 0
>>         removed_snaps [1~3]
>>
>> [cephuser@ceph ceph-config]$ ceph osd df tree
>> ID CLASS WEIGHT  REWEIGHT SIZE  USE    AVAIL %USE  VAR  PGS TYPE NAME
>> -1       6.55128        - 2237G   198G 2038G     0    0   - root default
>> -7             0        -     0      0     0     0    0   -     host ods03
>> -3       2.18475        - 2237G   181G 2055G  8.12 0.96   -     host osd01
>>  3   hdd 0.54619  1.00000  559G 53890M  506G  9.41 1.11  90         osd.3
>>  4   hdd 0.54619  1.00000  559G 30567M  529G  5.34 0.63  89         osd.4
>>  5   hdd 0.54619  1.00000  559G 59385M  501G 10.37 1.22  93         osd.5
>>  6   hdd 0.54619  1.00000  559G 42156M  518G  7.36 0.87  93         osd.6
>> -5       2.18178        - 2234G   189G 2044G  8.50 1.00   -     host osd02
>>  0   hdd 0.54520  1.00000  558G 32460M  526G  5.68 0.67  90         osd.0
>>  1   hdd 0.54520  1.00000  558G 54578M  504G  9.55 1.12  89         osd.1
>>  2   hdd 0.54520  1.00000  558G 47761M  511G  8.35 0.98  93         osd.2
>>  7   hdd 0.54619  1.00000  559G 59584M  501G 10.40 1.22  92         osd.7
>> -9       2.18475        - 2237G   198G 2038G  8.88 1.04   -     host osd03
>>  8   hdd 0.54619  1.00000  559G 52462M  508G  9.16 1.08  99         osd.8
>> 10   hdd 0.54619  1.00000  559G 35284M  524G  6.16 0.73  88         osd.10
>> 11   hdd 0.54619  1.00000  559G 71739M  489G 12.53 1.47  87         osd.11
>> 12   hdd 0.54619  1.00000  559G 43832M  516G  7.65 0.90  93         osd.12
>>                     TOTAL 6708G   570G 6138G  8.50
>> MIN/MAX VAR: 0.63/1.47  STDDEV: 2.06
>>
>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to