All; Since I haven't heard otherwise, I have to assume that the only way to get this to go away is to dump the contents of the RGW bucket(s), and recreate it (them)?
How did this get past release approval? A change which makes a valid cluster state in-valid, with no mitigation other than downtime, in a minor release. Thank you, Dominic L. Hilsbos, MBA Director – Information Technology Perform Air International Inc. [email protected] www.PerformAir.com -----Original Message----- From: ceph-users [mailto:[email protected]] On Behalf Of [email protected] Sent: Friday, November 15, 2019 9:13 AM To: [email protected] Cc: Stephen Self Subject: Re: [ceph-users] Large OMAP Object Wido; Ok, yes, I have tracked it down to the index for one of our buckets. I missed the ID in the ceph df output previously. Next time I'll wait to read replies until I've finished my morning coffee. How would I go about correcting this? The content for this bucket is basically just junk, as we're still doing production qualification, and workflow planning. Moving from Windows file shares to self-hosted cloud storage is a significant undertaking. Thank you, Dominic L. Hilsbos, MBA Director – Information Technology Perform Air International Inc. [email protected] www.PerformAir.com -----Original Message----- From: ceph-users [mailto:[email protected]] On Behalf Of Wido den Hollander Sent: Friday, November 15, 2019 8:40 AM To: [email protected] Subject: Re: [ceph-users] Large OMAP Object On 11/15/19 4:35 PM, [email protected] wrote: > All; > > Thank you for your help so far. I have found the log entries from when the > object was found, but don't see a reference to the pool. > > Here the logs: > 2019-11-14 03:10:16.508601 osd.1 (osd.1) 21 : cluster [DBG] 56.7 deep-scrub > starts > 2019-11-14 03:10:18.325881 osd.1 (osd.1) 22 : cluster [WRN] Large omap object > found. Object: > 56:f7d15b13:::.dir.f91aeff8-a365-47b4-a1c8-928cd66134e8.44130.1:head Key > count: 380425 Size (bytes): 82896978 > In this case it's in pool 56, check 'ceph df' to see which pool that is. To me this seems like a RGW bucket which index grew too big. Use: $ radosgw-admin bucket list $ radosgw-admin metadata get bucket:<BUCKET> And match that UUID back to the bucket. Wido > Thank you, > > Dominic L. Hilsbos, MBA > Director – Information Technology > Perform Air International Inc. > [email protected] > www.PerformAir.com > > > > -----Original Message----- > From: Wido den Hollander [mailto:[email protected]] > Sent: Friday, November 15, 2019 1:56 AM > To: Dominic Hilsbos; [email protected] > Cc: Stephen Self > Subject: Re: [ceph-users] Large OMAP Object > > Did you check /var/log/ceph/ceph.log on one of the Monitors to see which > pool and Object the large Object is in? > > Wido > > On 11/15/19 12:23 AM, [email protected] wrote: >> All; >> >> We had a warning about a large OMAP object pop up in one of our clusters >> overnight. The cluster is configured for CephFS, but nothing mounts a >> CephFS, at this time. >> >> The cluster mostly uses RGW. I've checked the cluster log, the MON log, and >> the MGR log on one of the mons, with no useful references to the pool / pg >> where the large OMAP objects resides. >> >> Is my only option to find this large OMAP object to go through the OSD logs >> for the individual OSDs in the cluster? >> >> Thank you, >> >> Dominic L. Hilsbos, MBA >> Director - Information Technology >> Perform Air International Inc. >> [email protected] >> www.PerformAir.com >> _______________________________________________ >> ceph-users mailing list >> [email protected] >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > _______________________________________________ > ceph-users mailing list > [email protected] > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
