Hi Udo and Irek,
Good day to you, and thank you for your emails.
>perhaps due IOs from the journal?
>You can test with iostat (like "iostat -dm 5 sdg").
Yes, I have shared the iostat result earlier on this same thread. At times
the utilisation of the 2 journal drives will hit 100%, especially when I
simulate writing data using rados bench command. Any suggestions what could
be the cause of the I/O issue?
====
avg-cpu: %user %nice %system %iowait %steal %idle
1.85 0.00 1.65 3.14 0.00 93.36
Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz
avgqu-sz await r_await w_await svctm %util
sdg 0.00 0.00 0.00 55.00 0.00 25365.33
922.38 34.22 568.90 0.00 568.90 17.82 98.00
sdf 0.00 0.00 0.00 55.67 0.00 25022.67
899.02 29.76 500.57 0.00 500.57 17.60 98.00
avg-cpu: %user %nice %system %iowait %steal %idle
2.10 0.00 1.37 2.07 0.00 94.46
Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz
avgqu-sz await r_await w_await svctm %util
sdg 0.00 0.00 0.00 56.67 0.00 25220.00
890.12 23.60 412.14 0.00 412.14 17.62 99.87
sdf 0.00 0.00 0.00 52.00 0.00 24637.33
947.59 33.65 587.41 0.00 587.41 19.23 100.00
avg-cpu: %user %nice %system %iowait %steal %idle
2.21 0.00 1.77 6.75 0.00 89.27
Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz
avgqu-sz await r_await w_await svctm %util
sdg 0.00 0.00 0.00 54.33 0.00 24802.67
912.98 25.75 486.36 0.00 486.36 18.40 100.00
sdf 0.00 0.00 0.00 53.00 0.00 24716.00
932.68 35.26 669.89 0.00 669.89 18.87 100.00
avg-cpu: %user %nice %system %iowait %steal %idle
1.87 0.00 1.67 5.25 0.00 91.21
Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz
avgqu-sz await r_await w_await svctm %util
sdg 0.00 0.00 0.00 94.33 0.00 26257.33
556.69 18.29 208.44 0.00 208.44 10.50 99.07
sdf 0.00 0.00 0.00 51.33 0.00 24470.67
953.40 32.75 684.62 0.00 684.62 19.51 100.13
avg-cpu: %user %nice %system %iowait %steal %idle
1.51 0.00 1.34 7.25 0.00 89.89
Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz
avgqu-sz await r_await w_await svctm %util
sdg 0.00 0.00 0.00 52.00 0.00 22565.33
867.90 24.73 446.51 0.00 446.51 19.10 99.33
sdf 0.00 0.00 0.00 64.67 0.00 24892.00
769.86 19.50 330.02 0.00 330.02 15.32 99.07
====
>You what model SSD?
For this one, I am using Seagate 100GB SSD, model: HDS-2TM-ST100FM0012
>Which version of the kernel?
Ubuntu 13.04, Linux kernel version: 3.8.0-19-generic #30-Ubuntu SMP Wed May
1 16:35:23 UTC 2013 x86_64 x86_64 x86_64 GNU/Linux
Looking forward to your reply, thank you.
Cheers.
On Mon, Apr 28, 2014 at 4:45 PM, Irek Fasikhov <[email protected]> wrote:
> You what model SSD?
> Which version of the kernel?
>
>
>
> 2014-04-28 12:35 GMT+04:00 Udo Lembke <[email protected]>:
>
>> Hi,
>> perhaps due IOs from the journal?
>> You can test with iostat (like "iostat -dm 5 sdg").
>>
>> on debian iostat is in the package sysstat.
>>
>> Udo
>>
>> Am 28.04.2014 07:38, schrieb Indra Pramana:
>> > Hi Craig,
>> >
>> > Good day to you, and thank you for your enquiry.
>> >
>> > As per your suggestion, I have created a 3rd partition on the SSDs and
>> did
>> > the dd test directly into the device, and the result is very slow.
>> >
>> > ====
>> > root@ceph-osd-08:/mnt# dd bs=1M count=128 if=/dev/zero of=/dev/sdg3
>> > conv=fdatasync oflag=direct
>> > 128+0 records in
>> > 128+0 records out
>> > 134217728 bytes (134 MB) copied, 19.5223 s, 6.9 MB/s
>> >
>> > root@ceph-osd-08:/mnt# dd bs=1M count=128 if=/dev/zero of=/dev/sdf3
>> > conv=fdatasync oflag=direct
>> > 128+0 records in
>> > 128+0 records out
>> > 134217728 bytes (134 MB) copied, 5.34405 s, 25.1 MB/s
>> > ====
>> >
>> > I did a test onto another server with exactly similar specification and
>> > similar SSD drive (Seagate SSD 100 GB) but not added into the cluster
>> yet
>> > (thus no load), and the result is fast:
>> >
>> > ====
>> > root@ceph-osd-09:/home/indra# dd bs=1M count=128 if=/dev/zero
>> of=/dev/sdf1
>> > conv=fdatasync oflag=direct
>> > 128+0 records in
>> > 128+0 records out
>> > 134217728 bytes (134 MB) copied, 0.742077 s, 181 MB/s
>> > ====
>> >
>> > Is the Ceph journal load really takes up a lot of the SSD resources? I
>> > don't understand how come the performance can drop significantly.
>> > Especially since the two Ceph journals are only taking the first 20 GB
>> out
>> > of the 100 GB of the SSD total capacity.
>> >
>> > Any advice is greatly appreciated.
>> >
>> > Looking forward to your reply, thank you.
>> >
>> > Cheers.
>> >
>> >
>> >
>> _______________________________________________
>> ceph-users mailing list
>> [email protected]
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>
>
>
> --
> С уважением, Фасихов Ирек Нургаязович
> Моб.: +79229045757
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com