Re: [ceph-users] 12.2.8: 1 node comes up (noout set), from a 6 nodes cluster -> I/O stuck (rbd usage)

2018-10-19 Thread Eugen Block
No, you do not need to set nobackfill and norecover if you only shut down one server. The guide you are referencing is about shutting down everything. It will not recover degraded PGs if you shut down one server with noout. You are right, I must have confused something in my memory with the

Re: [ceph-users] 12.2.8: 1 node comes up (noout set), from a 6 nodes cluster -> I/O stuck (rbd usage)

2018-10-19 Thread Paul Emmerich
No, you do not need to set nobackfill and norecover if you only shut down one server. The guide you are referencing is about shutting down everything. It will not recover degraded PGs if you shut down one server with noout. Paul Am Fr., 19. Okt. 2018 um 11:37 Uhr schrieb Eugen Block : > > Hi

Re: [ceph-users] 12.2.8: 1 node comes up (noout set), from a 6 nodes cluster -> I/O stuck (rbd usage)

2018-10-19 Thread Eugen Block
Hi Denny, the recommendation for ceph maintenance is to set three flags if you need to shutdown a node (or the entire cluster): ceph osd set noout ceph osd set nobackfill ceph osd set norecover Although the 'noout' flag seems to be enough for many maintenance tasks it doesn't prevent the

[ceph-users] 12.2.8: 1 node comes up (noout set), from a 6 nodes cluster -> I/O stuck (rbd usage)

2018-10-18 Thread Denny Fuchs
Hi, today we had an issue with our 6 node Ceph cluster. We had to shutdown one node (Ceph-03), to replace a disk (because, we did now know the slot). We set the noout flag and did a graceful shutdown. All was O.K. After the disk was replaced, the node comes up and our VMs had a big I/O