Hi,

I'm trying to update my cluster (19.2.2 -> 19.2.3), mon and mgr upgrade goes 
well but I had some issue with OSD :


Upgrade: unsafe to stop osd(s) at this time (165 PGs are or would become 
offline)


Cluster is in health_ok

All pools are replica 3 and pgs active+clean

autoscaler is off following the ceph docs


Does ceph osd ok-to-stop lead to lost data ?


The only rules used in the cluster is replicated_rule :


root@ceph-monitor-1:/# ceph osd crush rule dump replicated_rule

{
    "rule_id": 0,
    "rule_name": "replicated_rule",
    "type": 1,
    "steps": [
        {
            "op": "take",
            "item": -1,
            "item_name": "inist"
        },
        {
            "op": "chooseleaf_firstn",
            "num": 0,
            "type": "host"
        },
        {
            "op": "emit"
        }
    ]
}

root@ceph-monitor-1:/# ceph osd tree
ID   CLASS  WEIGHT   TYPE NAME                             STATUS  REWEIGHT  
PRI-AFF
 -1         6.63879  root inist
-15         0.90970      host ceph-monitor-1
  6    hdd  0.90970          osd.6                             up   1.00000  
1.00000
-21         5.72910      datacenter bat1
-20         2.72910          room room01
-19         2.72910              row left
-18         2.72910                  rack 10
 -3         0.90970                      host ceph-node-1
  2    hdd  0.90970                          osd.2             up   1.00000  
1.00000
 -5         0.90970                      host ceph-node-2
  1    hdd  0.90970                          osd.1             up   1.00000  
1.00000
 -9         0.90970                      host ceph-node-3
  5    hdd  0.90970                          osd.5             up   1.00000  
1.00000
-36         3.00000          room room03
-35         3.00000              row left06
-34         3.00000                  rack 08
 -7         1.00000                      host ceph-node-4
  0    hdd  1.00000                          osd.0             up   1.00000  
1.00000
-13         1.00000                      host ceph-node-5
  3    hdd  1.00000                          osd.3             up   1.00000  
1.00000
-11         1.00000                      host ceph-node-6
  4    hdd  1.00000                          osd.4             up   1.00000  
1.00000

Thanks !

Vivien



_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to