I should have 10 OSDs, below is the output:

root@ceph-mon1:~# ceph osd tree
ID  CLASS  WEIGHT   TYPE NAME           STATUS  REWEIGHT  PRI-AFF
-1         1.95297  root default
-5         0.78119      host ceph-mon1
 2    hdd  0.19530          osd.2         down         0  1.00000
 4    hdd  0.19530          osd.4         down         0  1.00000
 8    hdd  0.19530          osd.8           up   1.00000  1.00000
 9    hdd  0.19530          osd.9           up   1.00000  1.00000
-7         0.58589      host ceph-mon2
 1    hdd  0.19530          osd.1           up   1.00000  1.00000
 5    hdd  0.19530          osd.5           up   1.00000  1.00000
 7    hdd  0.19530          osd.7           up   1.00000  1.00000
-3         0.58589      host ceph-mon3
 0    hdd  0.19530          osd.0         down   1.00000  1.00000
 3    hdd  0.19530          osd.3           up   1.00000  1.00000
 6    hdd  0.19530          osd.6         down         0  1.00000

Try to restart these are down but failed

On Tue, Feb 22, 2022 at 4:42 PM <ash...@amerrick.co.uk> wrote:

> What does
>
> ‘ceph osd tree’ show?
>
> How many OSD’s should you have 7 or 10?
>
> On 22 Feb 2022, at 14:40, Michel Niyoyita <mico...@gmail.com> wrote:
>
> Actually one of my colleagues tried to reboot all nodes and he did not
> prepare the node like setting noout , norecover ......, once all node are
> up the cluster is no longer accessible and above are messages we are
> getting. I did not remove any osd . except are marked down.
> below is my ceph.conf:
>
> mon initial members = ceph-mon1,ceph-mon2,ceph-mon3
> mon_allow_pool_delete = True
> mon_clock_drift_allowed = 0.5
> mon_max_pg_per_osd = 400
> mon_osd_allow_primary_affinity = 1
> mon_pg_warn_max_object_skew = 0
> mon_pg_warn_max_per_osd = 0
> mon_pg_warn_min_per_osd = 0
> osd pool default crush rule = -1
> osd_pool_default_min_size = 1
> osd_pool_default_size = 2
> public network = 0.0.0.0/0
>
> On Tue, Feb 22, 2022 at 4:32 PM <ash...@amerrick.co.uk> wrote:
>
>> You have 1 OSD offline, has this disk failed or you aware of what has
>> caused this to go offline?
>> Shows you have 10 OSD’s but only 7in, have you removed the other 3? Was
>> the data fully drained off these first?
>>
>> I see you have 11 Pool’s what are these setup as, type and min/max size?
>>
>> > On 22 Feb 2022, at 14:15, Michel Niyoyita <mico...@gmail.com> wrote:
>> >
>> > Dear Ceph Users,
>> >
>> > Kindly help me to repair my cluster is down from yesterday up to now I
>> am
>> > not able to make it up and running . below are some findings:
>> >
>> >    id:     6ad86187-2738-42d8-8eec-48b2a43c298f
>> >    health: HEALTH_ERR
>> >            mons are allowing insecure global_id reclaim
>> >            1/3 mons down, quorum ceph-mon1,ceph-mon3
>> >            10/32332 objects unfound (0.031%)
>> >            1 osds down
>> >            3 scrub errors
>> >            Reduced data availability: 124 pgs inactive, 60 pgs down, 411
>> > pgs stale
>> >            Possible data damage: 9 pgs recovery_unfound, 1 pg
>> > backfill_unfound, 1 pg inconsistent
>> >            Degraded data redundancy: 6009/64664 objects degraded
>> (9.293%),
>> > 55 pgs degraded, 80 pgs undersized
>> >            11 pgs not deep-scrubbed in time
>> >            5 slow ops, oldest one blocked for 1638 sec, osd.9 has slow
>> ops
>> >
>> >  services:
>> >    mon: 3 daemons, quorum ceph-mon1,ceph-mon3 (age 3h), out of quorum:
>> > ceph-mon2
>> >    mgr: ceph-mon1(active, since 9h), standbys: ceph-mon2
>> >    osd: 10 osds: 6 up (since 7h), 7 in (since 9h); 43 remapped pgs
>> >
>> >  data:
>> >    pools:   11 pools, 560 pgs
>> >    objects: 32.33k objects, 159 GiB
>> >    usage:   261 GiB used, 939 GiB / 1.2 TiB avail
>> >    pgs:     11.429% pgs unknown
>> >             10.714% pgs not active
>> >             6009/64664 objects degraded (9.293%)
>> >             1384/64664 objects misplaced (2.140%)
>> >             10/32332 objects unfound (0.031%)
>> >             245 stale+active+clean
>> >             70  active+clean
>> >             64  unknown
>> >             48  stale+down
>> >             45  stale+active+undersized+degraded
>> >             37  stale+active+clean+remapped
>> >             28  stale+active+undersized
>> >             12  down
>> >             2   stale+active+recovery_unfound+degraded
>> >             2   stale+active+recovery_unfound+undersized+degraded
>> >             2
>>  stale+active+recovery_unfound+undersized+degraded+remapped
>> >             2   active+recovery_unfound+undersized+degraded+remapped
>> >             1   active+clean+inconsistent
>> >             1   stale+active+recovery_unfound+degraded+remapped
>> >             1
>>  stale+active+backfill_unfound+undersized+degraded+remapped
>> >
>> > If someone faced same issue please help me.
>> >
>> > Best Regards.
>> >
>> > Michel
>> > _______________________________________________
>> > ceph-users mailing list -- ceph-users@ceph.io
>> > To unsubscribe send an email to ceph-users-le...@ceph.io
>>
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@ceph.io
>> To unsubscribe send an email to ceph-users-le...@ceph.io
>>
>
>
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to