Re: [ceph-users] CephFS: number of PGs for metadata pool
Good point. Thanks! Triple-failure is essentially what I've faced about a months ago. So now I want to make sure that the new cephfs setup I am deploying at the moment will handle this kind of things better. On Wed, Dec 9, 2015 at 2:41 PM, John Spray wrote: On Wed, Dec 9, 2015 at 1:25 PM, Mykola Dvornik wrote: Hi Jan, Thanks for the reply. I see your point about replicas. However my motivation was a bit different. Consider some given amount of objects that are stored in the metadata pool. If I understood correctly ceph data placement approach, the number of objects per PG should decrease with the amount of PGs per pool. So my concern is that in catastrophic event of some PG(s) being lost I will loose more objects if the amount of PGs per pool is small. At the same time I don't want to have too few objects per PG to keep things disk IO, but not CPU bounded. If you are especially concerned about triple-failures (i.e. permanent PG loss), I would suggest you look at doing things like a size=4 pool for your metadata (maybe on SSDs). You could also look at simply segregating your size=3 metadata on to separate spinning drives, so that these comparatively less loaded OSDs will be able to undergo recovery faster in the event of a failure than an ordinary data drive that's full of terabytes of data, and have a lower probability of a triple failure. John ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] CephFS: number of PGs for metadata pool
On Wed, Dec 9, 2015 at 1:25 PM, Mykola Dvornik wrote: > Hi Jan, > > Thanks for the reply. I see your point about replicas. However my motivation > was a bit different. > > Consider some given amount of objects that are stored in the metadata pool. > If I understood correctly ceph data placement approach, the number of > objects per PG should decrease with the amount of PGs per pool. > > So my concern is that in catastrophic event of some PG(s) being lost I will > loose more objects if the amount of PGs per pool is small. At the same time > I don't want to have too few objects per PG to keep things disk IO, but not > CPU bounded. If you are especially concerned about triple-failures (i.e. permanent PG loss), I would suggest you look at doing things like a size=4 pool for your metadata (maybe on SSDs). You could also look at simply segregating your size=3 metadata on to separate spinning drives, so that these comparatively less loaded OSDs will be able to undergo recovery faster in the event of a failure than an ordinary data drive that's full of terabytes of data, and have a lower probability of a triple failure. John ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] CephFS: number of PGs for metadata pool
Hi Jan, Thanks for the reply. I see your point about replicas. However my motivation was a bit different. Consider some given amount of objects that are stored in the metadata pool. If I understood correctly ceph data placement approach, the number of objects per PG should decrease with the amount of PGs per pool. So my concern is that in catastrophic event of some PG(s) being lost I will loose more objects if the amount of PGs per pool is small. At the same time I don't want to have too few objects per PG to keep things disk IO, but not CPU bounded. So I thought maybe somebody did some research in this direction? On Wed, Dec 9, 2015 at 1:13 PM, Jan Schermer wrote: Number of PGs doesn't affect the number of replicas, so don't worry about it. Jan On 09 Dec 2015, at 13:03, Mykola Dvornik wrote: Hi guys, I am creating a 4-node/16OSD/32TB CephFS from scratch. According to the ceph documentation the metadata pool should have small amount of PGs since it contains some negligible amount of data compared to data pool. This makes me feel it might not be safe. So I was wondering how to chose the number of PGs per metadata pool to maintain its performance and reliability? Regards, Mykola ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] CephFS: number of PGs for metadata pool
Number of PGs doesn't affect the number of replicas, so don't worry about it. Jan > On 09 Dec 2015, at 13:03, Mykola Dvornik wrote: > > Hi guys, > > I am creating a 4-node/16OSD/32TB CephFS from scratch. > > According to the ceph documentation the metadata pool should have small > amount of PGs since it contains some negligible amount of data compared to > data pool. This makes me feel it might not be safe. > > So I was wondering how to chose the number of PGs per metadata pool to > maintain its performance and reliability? > > Regards, > > Mykola > ___ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
[ceph-users] CephFS: number of PGs for metadata pool
Hi guys, I am creating a 4-node/16OSD/32TB CephFS from scratch. According to the ceph documentation the metadata pool should have small amount of PGs since it contains some negligible amount of data compared to data pool. This makes me feel it might not be safe. So I was wondering how to chose the number of PGs per metadata pool to maintain its performance and reliability? Regards, Mykola ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com