Re: [ceph-users] CephFS: number of PGs for metadata pool

2015-12-09 Thread Mykola Dvornik

Good point. Thanks!

Triple-failure is essentially what I've faced about a months ago. So 
now I want to make sure that the new cephfs setup I am deploying at the 
moment will handle this kind of things better.


On Wed, Dec 9, 2015 at 2:41 PM, John Spray  wrote:
On Wed, Dec 9, 2015 at 1:25 PM, Mykola Dvornik 
 wrote:

 Hi Jan,

 Thanks for the reply. I see your point about replicas. However my 
motivation

 was a bit different.

 Consider some given amount of objects that are stored in the 
metadata pool.
 If I understood correctly ceph data placement approach, the number 
of

 objects per PG should decrease with the amount of PGs per pool.

 So my concern is that in catastrophic event of some PG(s) being 
lost I will
 loose more objects if the amount of PGs per pool is small. At the 
same time
 I don't want to have too few objects per PG to keep things disk IO, 
but not

 CPU bounded.


If you are especially concerned about triple-failures (i.e. permanent
PG loss), I would suggest you look at doing things like a size=4 pool
for your metadata (maybe on SSDs).

You could also look at simply segregating your size=3 metadata on to
separate spinning drives, so that these comparatively less loaded OSDs
will be able to undergo recovery faster in the event of a failure than
an ordinary data drive that's full of terabytes of data, and have a
lower probability of a triple failure.

John
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] CephFS: number of PGs for metadata pool

2015-12-09 Thread John Spray
On Wed, Dec 9, 2015 at 1:25 PM, Mykola Dvornik  wrote:
> Hi Jan,
>
> Thanks for the reply. I see your point about replicas. However my motivation
> was a bit different.
>
> Consider some given amount of objects that are stored in the metadata pool.
> If I understood correctly ceph data placement approach, the number of
> objects per PG should decrease with the amount of PGs per pool.
>
> So my concern is that in catastrophic event of some PG(s) being lost I will
> loose more objects if the amount of PGs per pool is small. At the same time
> I don't want to have too few objects per PG to keep things disk IO, but not
> CPU bounded.

If you are especially concerned about triple-failures (i.e. permanent
PG loss), I would suggest you look at doing things like a size=4 pool
for your metadata (maybe on SSDs).

You could also look at simply segregating your size=3 metadata on to
separate spinning drives, so that these comparatively less loaded OSDs
will be able to undergo recovery faster in the event of a failure than
an ordinary data drive that's full of terabytes of data, and have a
lower probability of a triple failure.

John
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] CephFS: number of PGs for metadata pool

2015-12-09 Thread Mykola Dvornik

Hi Jan,

Thanks for the reply. I see your point about replicas. However my 
motivation was a bit different.


Consider some given amount of objects that are stored in the metadata 
pool.
If I understood correctly ceph data placement approach, the number of 
objects per PG should decrease with the amount of PGs per pool.


So my concern is that in catastrophic event of some PG(s) being lost I 
will loose more objects if the amount of PGs per pool is small. At the 
same time I don't want to have too few objects per PG to keep things 
disk IO, but not CPU bounded.


So I thought maybe somebody did some research in this direction?













On Wed, Dec 9, 2015 at 1:13 PM, Jan Schermer  wrote:
Number of PGs doesn't affect the number of replicas, so don't worry 
about it.


Jan

 On 09 Dec 2015, at 13:03, Mykola Dvornik  
wrote:


 Hi guys,

 I am creating a 4-node/16OSD/32TB CephFS from scratch.

 According to the ceph documentation the metadata pool should have 
small amount of PGs since it contains some negligible amount of data 
compared to data pool. This makes me feel it might not be safe.


 So I was wondering how to chose the number of PGs per metadata pool 
to maintain its performance and reliability?


 Regards,

 Mykola
 ___
 ceph-users mailing list
 ceph-users@lists.ceph.com
 http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] CephFS: number of PGs for metadata pool

2015-12-09 Thread Jan Schermer
Number of PGs doesn't affect the number of replicas, so don't worry about it.

Jan

> On 09 Dec 2015, at 13:03, Mykola Dvornik  wrote:
> 
> Hi guys,
> 
> I am creating a 4-node/16OSD/32TB CephFS from scratch. 
> 
> According to the ceph documentation the metadata pool should have small 
> amount of PGs since it contains some negligible amount of data compared to 
> data pool. This makes me feel it might not be safe.
> 
> So I was wondering how to chose the number of PGs per metadata pool to 
> maintain its performance and reliability?
> 
> Regards,
> 
> Mykola
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] CephFS: number of PGs for metadata pool

2015-12-09 Thread Mykola Dvornik

Hi guys,

I am creating a 4-node/16OSD/32TB CephFS from scratch.

According to the ceph documentation the metadata pool should have small 
amount of PGs since it contains some negligible amount of data compared 
to data pool. This makes me feel it might not be safe.


So I was wondering how to chose the number of PGs per metadata pool to 
maintain its performance and reliability?


Regards,

Mykola
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com