Hey John. Just to be sure; by "deleting the pools" you mean the *cephfs_metadata* and *cephfs_metadata* pools, right? Does it have any impact over radosgw? Thanks.
On Tue, Nov 15, 2016 at 10:10 AM John Spray <jsp...@redhat.com> wrote: > On Tue, Nov 15, 2016 at 11:58 AM, Webert de Souza Lima > <webert.b...@gmail.com> wrote: > > Hi, > > > > after running a cephfs on my ceph cluster I got stuck with the following > > heath status: > > > > # ceph status > > cluster ac482f5b-dce7-410d-bcc9-7b8584bd58f5 > > health HEALTH_WARN > > 128 pgs degraded > > 128 pgs stuck unclean > > 128 pgs undersized > > recovery 24/40282627 <(24)%204028-2627> objects degraded > (0.000%) > > monmap e3: 3 mons at > > {dc1-master-ds01= > 10.2.0.1:6789/0,dc1-master-ds02=10.2.0.2:6789/0,dc1-master-ds03=10.2.0.3:6789/0 > } > > election epoch 140, quorum 0,1,2 > > dc1-master-ds01,dc1-master-ds02,dc1-master-ds03 > > fsmap e18: 1/1/1 up {0=b=up:active}, 1 up:standby > > osdmap e15851: 10 osds: 10 up, 10 in > > flags sortbitwise > > pgmap v11924989: 1088 pgs, 18 pools, 11496 GB data, 19669 kobjects > > 23325 GB used, 6349 GB / 29675 GB avail > > 24/40282627 <(24)%204028-2627> objects degraded (0.000%) > > 958 active+clean > > 128 active+undersized+degraded > > 2 active+clean+scrubbing > > client io 1968 B/s rd, 1 op/s rd, 0 op/s wr > > > > # ceph health detail > > -> https://paste.debian.net/895825/ > > > > # ceph osd lspools > > 2 .rgw.root,3 master.rgw.control,4 master.rgw.data.root,5 master.rgw.gc,6 > > master.rgw.log,7 master.rgw.intent-log,8 master.rgw.usage,9 > > master.rgw.users.keys,10 master.rgw.users.email,11 > master.rgw.users.swift,12 > > master.rgw.users.uid,13 master.rgw.buckets.index,14 > > master.rgw.buckets.data,15 master.rgw.meta,16 > master.rgw.buckets.non-ec,22 > > rbd,23 cephfs_metadata,24 cephfs_data, > > > > on this cluster I run cephfs, which is empty atm, and a radosgw service. > > How can I clean this? > > Stop your MDS daemons > Run "ceph mds fail <id>" for each MDS daemon > Use "ceph fs rm <your fs name>" > Then you can delete the pools. > > John > > > > > _______________________________________________ > > ceph-users mailing list > > ceph-users@lists.ceph.com > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com