Hi Alfredo Deza, We have 5 node platforms with lvm osd created from scratch and another 5 node platform migrated from kraken which is ceph volume simple. Both has same issue . Both platform has only hdd for osd.
We also noticed 2 times disk iops more compare to kraken , this causes less read performance. During rocksdb compaction the situation is worse. Meanwhile we are building another platform creating osd using ceph-disk and analyse on this. Thanks, Muthu On Tuesday, February 20, 2018, Alfredo Deza <ad...@redhat.com> wrote: > > > On Mon, Feb 19, 2018 at 2:01 PM, nokia ceph <nokiacephus...@gmail.com> > wrote: > >> Hi All, >> >> We have 5 node clusters with EC 4+1 and use bluestore since last year >> from Kraken. >> Recently we migrated all our platforms to luminous 12.2.2 and finally all >> OSDs migrated to ceph-volume simple type and on few platforms installed >> ceph using ceph-volume . >> >> Now we see two times more traffic in read compare to client traffic on >> migrated platform and newly created platforms . This was not the case in >> older releases where ceph status read B/W will be same as client read >> traffic. >> >> Some network graphs : >> >> *Client network interface* towards ceph public interface : shows >> *4.3Gbps* read >> >> >> [image: Inline image 2] >> >> *Ceph Node Public interface* : Each node around 960Mbps * 5 node =* 4.6 >> Gbps *- this matches. >> [image: Inline image 3] >> >> Ceph status output : show 1032 MB/s =* 8.06 Gbps* >> >> cn6.chn6us1c1.cdn ~# ceph status >> cluster: >> id: abda22db-3658-4d33-9681-e3ff10690f88 >> health: HEALTH_OK >> >> services: >> mon: 5 daemons, quorum cn6,cn7,cn8,cn9,cn10 >> mgr: cn6(active), standbys: cn7, cn9, cn10, cn8 >> osd: 340 osds: 340 up, 340 in >> >> data: >> pools: 1 pools, 8192 pgs >> objects: 270M objects, 426 TB >> usage: 581 TB used, 655 TB / 1237 TB avail >> pgs: 8160 active+clean >> 32 active+clean+scrubbing >> >> io: >> client: *1032 MB/s rd*, 168 MB/s wr, 1908 op/s rd, 1594 op/s wr >> >> >> Write operation we don't see this issue. Client traffic and this matches. >> Is this expected behavior in Luminous and ceph-volume lvm or a bug ? >> Wrong calculation in ceph status read B/W ? >> > > You mentioned `ceph-volume simple` but here you say lvm. With LVM > ceph-volume will create the OSDs from scratch, while "simple" will keep > whatever OSD was created before. > > Have you created the OSDs from scratch with ceph-volume? or is it just > using "simple" , managing a previously deployed OSD? > >> >> Please provide your feedback. >> >> Thanks, >> Muthu >> >> >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com