On Mon, Feb 19, 2018 at 2:01 PM, nokia ceph <nokiacephus...@gmail.com>
wrote:

> Hi All,
>
> We have 5 node clusters with EC 4+1 and use bluestore since last year from
> Kraken.
> Recently we migrated all our platforms to luminous 12.2.2 and finally all
> OSDs migrated to ceph-volume simple type and on few platforms installed
> ceph using ceph-volume .
>
> Now we see two times more traffic in read compare to client traffic on
> migrated platform and newly created platforms . This was not the case in
> older releases where ceph status read B/W will be same as client read
> traffic.
>
> Some network graphs :
>
> *Client network interface* towards ceph public interface : shows *4.3Gbps*
> read
>
>
> [image: Inline image 2]
>
> *Ceph Node Public interface* : Each node around 960Mbps * 5 node =* 4.6
> Gbps *- this matches.
> [image: Inline image 3]
>
> Ceph status output : show  1032 MB/s =* 8.06 Gbps*
>
> cn6.chn6us1c1.cdn ~# ceph status
>   cluster:
>     id:     abda22db-3658-4d33-9681-e3ff10690f88
>     health: HEALTH_OK
>
>   services:
>     mon: 5 daemons, quorum cn6,cn7,cn8,cn9,cn10
>     mgr: cn6(active), standbys: cn7, cn9, cn10, cn8
>     osd: 340 osds: 340 up, 340 in
>
>   data:
>     pools:   1 pools, 8192 pgs
>     objects: 270M objects, 426 TB
>     usage:   581 TB used, 655 TB / 1237 TB avail
>     pgs:     8160 active+clean
>              32   active+clean+scrubbing
>
>   io:
>     client:   *1032 MB/s rd*, 168 MB/s wr, 1908 op/s rd, 1594 op/s wr
>
>
> Write operation we don't see this issue. Client traffic and this matches.
> Is this expected behavior in Luminous and ceph-volume lvm or a bug ?
> Wrong calculation in ceph status read B/W ?
>

You mentioned `ceph-volume simple` but here you say lvm. With LVM
ceph-volume will create the OSDs from scratch, while "simple" will keep
whatever OSD was created before.

Have you created the OSDs from scratch with ceph-volume? or is it just
using "simple" , managing a previously deployed OSD?

>
> Please provide your feedback.
>
> Thanks,
> Muthu
>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to