Hello Christopher, Keep in mind that the PGs per OSD (and per pool) calculations take into account the replica count ( pool size= parameter ). So, for example.. if you're using a default of 3 replicas.. 16 * 3 = 48 PGs which allows for at least one PG per OSD on that pool. Even with a size=2, 32 PGs total still gives very close to 1 PG per OSD. Being that it's such a low utilization pool, this is still sufficient.
Thanks, Michael J. Kidd Sr. Storage Consultant Inktank Professional Services - by Red Hat On Wed, Jan 7, 2015 at 3:17 PM, Christopher O'Connell <[email protected]> wrote: > Hi, > > I"m playing with this with a modest sized ceph cluster (36x6TB disks). > Based on this it says that small pools (such as .users) would have just 16 > PGs. Is this correct? I've historically always made even these small pools > have at least as many PGs as the next power of 2 over my number of OSDs (64 > in this case). > > All the best, > > ~ Christopher > > On Wed, Jan 7, 2015 at 3:08 PM, Michael J. Kidd <[email protected]> > wrote: > >> Hello all, >> Just a quick heads up that we now have a PG calculator to help >> determine the proper PG per pool numbers to achieve a target PG per OSD >> ratio. >> >> http://ceph.com/pgcalc >> >> Please check it out! Happy to answer any questions, and always welcome >> any feedback on the tool / verbiage, etc... >> >> As an aside, we're also working to update the documentation to reflect >> the best practices. See Ceph.com tracker for this at: >> http://tracker.ceph.com/issues/9867 >> >> Thanks! >> Michael J. Kidd >> Sr. Storage Consultant >> Inktank Professional Services >> - by Red Hat >> >> _______________________________________________ >> ceph-users mailing list >> [email protected] >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> >
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
