Hello,

I'm seeing something that seems to be odd behavior when reweighting OSDs. I've just upgraded to 12.2.5 and am adding in a new osd server to the cluster. I gradually weight the 10TB OSDs into the cluster by doing a +1, letting things backfill for a while, then +1 until I reach my desired weight. This hasn't been a problem in the past, a proportionate amount of PGs would get remapped, peer and activate across this cluster. Now on 12.2.5 when I do this, almost all PGs peer and reactivate. Sometimes it recovers within a minute, other times it takes longer, this last time actually saw some OSDs on the new node crash and caused a longer time for peering/activating. Regardless of recovery time, this is a fairly violent reaction to reweighting.

Has anyone else seen behavior like this or have any ideas what's going on?

For example...

[root@sephmon1 ~]# ceph -s
  cluster:
    id:     bc2a1488-74f8-4d87-b2f6-615ae26bf7c9
    health: HEALTH_OK

  services:
    mon: 3 daemons, quorum sephmon1,sephmon2,sephmon3
    mgr: sephmon2(active), standbys: sephmon1, sephmon3
    mds: cephfs1-1/1/1 up  {0=sephmon1=up:active}, 1 up:standby
    osd: 789 osds: 789 up, 789 in

  data:
    pools:   7 pools, 39168 pgs
    objects: 74046k objects, 2989 TB
    usage:   3756 TB used, 1517 TB / 5273 TB avail
    pgs:     39137 active+clean
             26    active+clean+scrubbing+deep
             5     active+clean+scrubbing

  io:
    client:   3522 MB/s rd, 118 MB/s wr, 1295 op/s rd, 833 op/s wr

[root@sephmon1 ~]# for i in {771..779}; do ceph osd crush reweight osd.${i} 6.5; done
reweighted item id 771 name 'osd.771' to 6.5 in crush map
reweighted item id 772 name 'osd.772' to 6.5 in crush map
reweighted item id 773 name 'osd.773' to 6.5 in crush map
reweighted item id 774 name 'osd.774' to 6.5 in crush map
reweighted item id 775 name 'osd.775' to 6.5 in crush map
reweighted item id 776 name 'osd.776' to 6.5 in crush map
reweighted item id 777 name 'osd.777' to 6.5 in crush map
reweighted item id 778 name 'osd.778' to 6.5 in crush map
reweighted item id 779 name 'osd.779' to 6.5 in crush map
[root@sephmon1 ~]# ceph -s
  cluster:
    id:     bc2a1488-74f8-4d87-b2f6-615ae26bf7c9
    health: HEALTH_WARN
            2 osds down
            78219/355096089 objects misplaced (0.022%)
            Reduced data availability: 668 pgs inactive, 1920 pgs down, 551 pgs peering, 29 pgs incomplete             Degraded data redundancy: 803425/355096089 objects degraded (0.226%), 204 pgs degraded
            3 slow requests are blocked > 32 sec

  services:
    mon: 3 daemons, quorum sephmon1,sephmon2,sephmon3
    mgr: sephmon2(active), standbys: sephmon1, sephmon3
    mds: cephfs1-1/1/1 up  {0=sephmon1=up:active}, 1 up:standby
    osd: 789 osds: 787 up, 789 in; 257 remapped pgs

  data:
    pools:   7 pools, 39168 pgs
    objects: 73964k objects, 2985 TB
    usage:   3756 TB used, 1517 TB / 5273 TB avail
    pgs:     0.028% pgs unknown
             94.904% pgs not active
             803425/355096089 objects degraded (0.226%)
             78219/355096089 objects misplaced (0.022%)
             20215 peering
             14335 activating
             1882  active+clean
             1788  down
             205   remapped+peering
             167   stale+peering
             142   activating+undersized+degraded
             127   activating+undersized
             126   stale+down
             57    active+undersized+degraded
             39    stale+active+clean
             27    incomplete
             17    activating+remapped
             11    unknown
             7     stale+activating
             6     down+remapped
             3     stale+activating+undersized
             2     stale+incomplete
             2     active+undersized
             2     stale+activating+undersized+degraded
             1     activating+undersized+degraded+remapped
             1     stale+active+undersized+degraded
             1     remapped
             1     active+clean+scrubbing
             1 active+undersized+degraded+remapped+backfill_wait
             1     stale+remapped+peering
             1     active+clean+remapped
             1     active+remapped+backfilling

  io:
    client:   3896 GB/s rd, 339 GB/s wr, 8004 kop/s rd, 320 kop/s wr
    recovery: 726 GB/s, 11172 objects/s


Thanks,
Kevin

_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to