Turn off write cache on the controller. Your probably seeing the flush to disk. 





        

Tyler Bishop 
Chief Executive Officer 
513-299-7108 x10 



tyler.bis...@beyondhosting.net 


If you are not the intended recipient of this transmission you are notified 
that disclosing, copying, distributing or taking any action in reliance on the 
contents of this information is strictly prohibited. 




From: "Franck Allouis" <franck.allo...@stef.com> 
To: "ceph-users" <ceph-us...@ceph.com> 
Sent: Friday, May 29, 2015 8:54:41 AM 
Subject: [ceph-users] TR: High apply latency on OSD causes poor performance on 
VM 



Hi, 



Could you take a look on my problem. 

It’s about high latency on my OSDs on HP G8 servers (ceph01, ceph02 and 
ceph03). 

When I run a rados bench for 60 sec, the results are surprising : after a few 
seconds, there is no traffic, then it’s resume, etc. 

Finally, the maximum latency is high and VM’s disks freeze lot. 



#rados bench -p pool-test-g8 60 write 

Maintaining 16 concurrent writes of 4194304 bytes for up to 60 seconds or 0 
objects 

Object prefix: benchmark_data_ceph02_56745 

sec Cur ops started finished avg MB/s cur MB/s last lat avg lat 

0 0 0 0 0 0 - 0 

1 16 82 66 263.959 264 0.0549584 0.171148 

2 16 134 118 235.97 208 0.344873 0.232103 

3 16 189 173 230.639 220 0.015583 0.24581 

4 16 248 232 231.973 236 0.0704699 0.252504 

5 16 306 290 231.974 232 0.0229872 0.258343 

6 16 371 355 236.64 260 0.27183 0.255469 

7 16 419 403 230.26 192 0.0503492 0.263304 

8 16 460 444 221.975 164 0.0157241 0.261779 

9 16 506 490 217.754 184 0.199418 0.271501 

10 16 518 502 200.778 48 0.0472324 0.269049 

11 16 518 502 182.526 0 - 0.269049 

12 16 556 540 179.981 76 0.100336 0.301616 

13 16 607 591 181.827 204 0.173912 0.346105 

14 16 655 639 182.552 192 0.0484904 0.339879 

15 16 683 667 177.848 112 0.0504184 0.349929 

16 16 746 730 182.481 252 0.276635 0.347231 

17 16 807 791 186.098 244 0.391491 0.339275 

18 16 845 829 184.203 152 0.188608 0.342021 

19 16 850 834 175.561 20 0.960175 0.342717 

2015-05-28 17:09:48.397376min lat: 0.013532 max lat: 6.28387 avg lat: 0.346987 

sec Cur ops started finished avg MB/s cur MB/s last lat avg lat 

20 16 859 843 168.582 36 0.0182246 0.346987 

21 16 863 847 161.316 16 3.18544 0.355051 

22 16 897 881 160.165 136 0.0811037 0.371209 

23 16 901 885 153.897 16 0.0482124 0.370793 

24 16 943 927 154.484 168 0.63064 0.397204 

25 15 997 982 157.104 220 0.0933448 0.392701 

26 16 1058 1042 160.291 240 0.166463 0.385943 

27 16 1088 1072 158.798 120 1.63882 0.388568 

28 16 1125 1109 158.412 148 0.0511479 0.38419 

29 16 1155 1139 157.087 120 0.162266 0.385898 

30 16 1163 1147 152.917 32 0.0682181 0.383571 

31 16 1190 1174 151.468 108 0.0489185 0.386665 

32 16 1196 1180 147.485 24 2.95263 0.390657 

33 16 1213 1197 145.076 68 0.0467788 0.389299 

34 16 1265 1249 146.926 208 0.0153085 0.420687 

35 16 1332 1316 150.384 268 0.0157061 0.42259 

36 16 1374 1358 150.873 168 0.251626 0.417373 

37 16 1402 1386 149.822 112 0.0475302 0.413886 

38 16 1444 1428 150.3 168 0.0507577 0.421055 

39 16 1500 1484 152.189 224 0.0489163 0.416872 

2015-05-28 17:10:08.399434min lat: 0.013532 max lat: 9.26596 avg lat: 0.415296 

sec Cur ops started finished avg MB/s cur MB/s last lat avg lat 

40 16 1530 1514 151.384 120 0.951713 0.415296 

41 16 1551 1535 149.741 84 0.0686787 0.416571 

42 16 1606 1590 151.413 220 0.0826855 0.41684 

43 16 1656 1640 152.542 200 0.0706539 0.409974 

44 16 1663 1647 149.712 28 0.046672 0.408476 

45 16 1685 1669 148.34 88 0.0989566 0.424918 

46 16 1707 1691 147.028 88 0.0490569 0.421116 

47 16 1707 1691 143.9 0 - 0.421116 

48 16 1707 1691 140.902 0 - 0.421116 

49 16 1720 1704 139.088 17.3333 0.0480335 0.428997 

50 16 1752 1736 138.866 128 0.053219 0.4416 

51 16 1786 1770 138.809 136 0.602946 0.440357 

52 16 1810 1794 137.986 96 0.0472518 0.438376 

53 16 1831 1815 136.967 84 0.0148999 0.446801 

54 16 1831 1815 134.43 0 - 0.446801 

55 16 1853 1837 133.586 44 0.0499486 0.455561 

56 16 1898 1882 134.415 180 0.0566593 0.461019 

57 16 1932 1916 134.442 136 0.0162902 0.454385 

58 16 1948 1932 133.227 64 0.62188 0.464403 

59 16 1966 1950 132.19 72 0.563613 0.472147 

2015-05-28 17:10:28.401525min lat: 0.013532 max lat: 12.4828 avg lat: 0.472084 

sec Cur ops started finished avg MB/s cur MB/s last lat avg lat 

60 16 1983 1967 131.12 68 0.030789 0.472084 

61 16 1984 1968 129.036 4 0.0519125 0.471871 

62 16 1984 1968 126.955 0 - 0.471871 

63 16 1984 1968 124.939 0 - 0.471871 

64 14 1984 1970 123.112 2.66667 4.20878 0.476035 

Total time run: 64.823355 

Total writes made: 1984 

Write size: 4194304 

Bandwidth (MB/sec): 122.425 

Stddev Bandwidth: 85.3816 

Max bandwidth (MB/sec): 268 

Min bandwidth (MB/sec): 0 

Average Latency: 0.520956 

Stddev Latency: 1.17678 

Max latency: 12.4828 

Min latency: 0.013532 





I have installed a new ceph06 box which has best latencies but hardware is 
different (RAID card, disks, …). 

All OSDs are formatted in XFS but are mounted differently : 

- On ceph box with high latency (strip size of raid logical disk is 256KB) : 

/dev/sdd1 on /var/lib/ceph/osd/ceph-4 type xfs 
(rw,noatime,attr2,inode64,logbsize=256k,sunit=512,swidth=512,noquota) 

- On ceph box with low latency (strip size of raid logical disk is 128KB) : 

/dev/sdc1 on /var/lib/ceph/osd/ceph-9 type xfs 
(rw,noatime,attr2,inode64,noquota) 



Thread on Proxmox Forum : 
http://forum.proxmox.com/threads/22206-Ceph-High-apply-latency-on-OSD-causes-poor-performance-on-VM
 



If you have any idea ... 



Thanks. 



Franck ALLOUIS 

franck.allo...@stef.com 






De : Franck Allouis 
Envoyé : jeudi 21 mai 2015 17:12 
À : 'ceph-commun...@lists.ceph.com' 
Objet : High apply latency on OSD causes poor performance on VM 




Hi, 



Since we have installing our new Ceph Cluster, we have frequently high apply 
latency on OSDs (near 200 ms to 1500 ms), while commit latency is continuously 
at 0 ms ! 



In Ceph documentation, when you run the command "ceph osd perf", the 
fs_commit_latency is generally higher than fs_apply_latency. For us it's the 
opposite. 

The phenomenon has increased since we changed the Ceph version (migrate Giant 
0.87.1 to Hammer 0.94.1) 

The consequence is that our Windows VMs are very slow. 

Does anyone could tell us if our configuration is good or not, and in what 
direction investigate ? 



# ceph osd perf 

osd fs_commit_latency(ms) fs_apply_latency(ms) 

0 0 62 

1 0 193 

2 0 88 

3 0 269 

4 0 1055 

5 0 322 

6 0 272 

7 0 116 

8 0 653 

9 0 4 

10 0 1 

11 0 7 

12 0 4 



Different informations in our configuration : 



- Proxmox 3.4-6 

- kernel : 3.10.0-10-pve 

- CEPH : 

- Hammer 0.94.1 

- 3 hosts with 3 OSDs of 4 TB (9 OSDs) + 1 SSD of 500 GB per host for journals 

- 1 host with 4 OSDs of 300 GB (4 OSDs) + 1 SSD of 500 GB for journals 



- OSD Tree : 



# ceph osd tree 

ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY 

-1 33.83995 root default 

-6 22.91995 room salle-dr 

-2 10.92000 host ceph01 

0 3.64000 osd.0 up 1.00000 1.00000 

2 3.64000 osd.2 up 1.00000 1.00000 

1 3.64000 osd.1 up 1.00000 1.00000 

-3 10.92000 host ceph02 

3 3.64000 osd.3 up 1.00000 1.00000 

4 3.64000 osd.4 up 1.00000 1.00000 

5 3.64000 osd.5 up 1.00000 1.00000 

-5 1.07996 host ceph06 

9 0.26999 osd.9 up 1.00000 1.00000 

10 0.26999 osd.10 up 1.00000 1.00000 

11 0.26999 osd.11 up 1.00000 1.00000 

12 0.26999 osd.12 up 1.00000 1.00000 

-7 10.92000 room salle-log 

-4 10.92000 host ceph03 

6 3.64000 osd.6 up 1.00000 1.00000 

7 3.64000 osd.7 up 1.00000 1.00000 

8 3.64000 osd.8 up 1.00000 1.00000 



- ceph.conf : 



[global] 

auth client required = cephx 

auth cluster required = cephx 

auth service required = cephx 

auth supported = cephx 

cluster network = 10.10.1.0/24 

filestore xattr use omap = true 

fsid = 2dbbec32-a464-4bc5-bb2b-983695d1d0c6 

keyring = /etc/pve/priv/$cluster.$name.keyring 

mon osd adjust heartbeat grace = true 

mon osd down out subtree limit = host 

osd disk threads = 24 

osd heartbeat grace = 10 

osd journal size = 5120 

osd max backfills = 1 

osd op threads = 24 

osd pool default min size = 1 

osd recovery max active = 1 

public network = 192.168.80.0/24 

[osd] 

keyring = /var/lib/ceph/osd/ceph-$id/keyring 

[mon.0] 

host = ceph01 

mon addr = 192.168.80.41:6789 

[mon.1] 

host = ceph02 

mon addr = 192.168.80.42:6789 

[mon.2] 

host = ceph03 

mon addr = 192.168.80.43:6789 



Thanks. 

Best regards 



Franck ALLOUIS 

franck.allo...@stef.com 






NOTICE: This e-mail (including any attachments) may contain information that is 
private, confidential or legally privileged information or material and is 
intended solely for the use of the addressee(s). If you receive this e-mail in 
error, please delete it from your system without copying it and immediately 
notify the sender(s) by reply e-mail. Any unauthorized use or disclosure of 
this message is strictly prohibited. STEF does not guarantee the integrity of 
this transmission and may therefore never be liable if the message is altered 
or falsified nor for any virus, interception or damage to your system. 

AVIS : Ce message (y compris toutes pièces jointes) peut contenir des 
informations privées, confidentielles et est pour l'usage du(es) seul(s) 
destinataire(s). Si vous avez reçu ce message par erreur, merci d'en avertir 
l'expéditeur par retour d'email immédiatement et de procéder à la destruction 
de l'ensemble des éléments reçus, dont vous ne devez garder aucune copie. Toute 
diffusion, utilisation ou copie de ce message ou des renseignements qu'il 
contient par une personne autre que le(les) destinataire(s) désigné(s) est 
interdite. STEF ne garantit pas l'intégrité de cette transmission et ne saurait 
être tenu responsable du message, de son contenu, de toute modification ou 
falsification, d’une interception ou de dégâts à votre système. 

_______________________________________________ 
ceph-users mailing list 
ceph-users@lists.ceph.com 
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to