Yes, the reasoning is the number of PGs.  I currently have all of my data 
stored in various RBDs in a pool and am planning to move most of it out of the 
RBDs into CephFS.  The pool would have the exact same use case that it does 
now, just with 90% of it's data in CephFS rather than RBDs.  My osds aren't to 
a point of having too many PGs on them, I just wanted to mitigate the memory 
need of the osd processes.

________________________________

[cid:[email protected]]<https://storagecraft.com>       David 
Turner | Cloud Operations Engineer | StorageCraft Technology 
Corporation<https://storagecraft.com>
380 Data Drive Suite 300 | Draper | Utah | 84020
Office: 801.871.2760 | Mobile: 385.224.2943

________________________________

If you are not the intended recipient of this message or received it 
erroneously, please notify the sender and delete it, together with any 
attachments, and be advised that any dissemination or copying of this message 
is prohibited.

________________________________

________________________________
From: Nick Fisk [[email protected]]
Sent: Saturday, January 07, 2017 3:21 PM
To: David Turner; [email protected]
Subject: RE: cephfs AND rbds

Technically I think there is no reason why you couldn’t do this, but I think it 
is unadvisable. There was a similar thread a while back where somebody had done 
this and it caused problems when he was trying to do maintenance/recovery 
further down the line.

I’m assuming you want to do this because you have already created a pool with 
the max number of PG’s per OSD and extra pools would take you further over this 
limit? If it’s the case I would just bump up the limit, it’s not worth the risk.

From: ceph-users [mailto:[email protected]] On Behalf Of David 
Turner
Sent: 07 January 2017 00:54
To: [email protected]
Subject: [ceph-users] cephfs AND rbds

Can cephfs and rbds use the same pool to store data?  I know you would need a 
separate metadata pool for cephfs, but could they share the same data pool?
________________________________
[cid:[email protected]]<http://xo4t.mj.am/lnk/AEMAHNQ6k5QAAAAAAAAAAF3gdtwAADNJBWwAAAAAAACRXwBYcWnMqBmri5aAT3OC_B5ECZEPkQAAlBI/1/U0BGpa1QpeR7MvfKmkwuvg/aHR0cHM6Ly9zdG9yYWdlY3JhZnQuY29t>

David Turner | Cloud Operations Engineer | StorageCraft Technology 
Corporation<http://xo4t.mj.am/lnk/AEMAHNQ6k5QAAAAAAAAAAF3gdtwAADNJBWwAAAAAAACRXwBYcWnMqBmri5aAT3OC_B5ECZEPkQAAlBI/2/0YzmPgAjmKPtnBOdSLpwug/aHR0cHM6Ly9zdG9yYWdlY3JhZnQuY29t>
380 Data Drive Suite 300 | Draper | Utah | 84020
Office: 801.871.2760 | Mobile: 385.224.2943

________________________________
If you are not the intended recipient of this message or received it 
erroneously, please notify the sender and delete it, together with any 
attachments, and be advised that any dissemination or copying of this message 
is prohibited.

________________________________


_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to