Hello Eugen, Below is the o/p:- ceph osd df:-
[image: image.png] ceph osd tree:- [image: image.png] On Thu, 28 Nov 2019 at 14:54, Eugen Block <[email protected]> wrote: > Hi, > > can you share the output of `ceph osd df` and `ceph osd tree`? > The smallest of your OSDs will be the bottleneck. Since ceph tries to > distribute the data evenly across all OSDs you won't be able to use > the large OSDs, at least not without adjusting your setup. > > Regards, > Eugen > > > Zitat von Alokkumar Mahajan <[email protected]>: > > > Thanks Wido. > > @ But normally CephFS will be able to use all the space inside your > > Cephcluster. > > So, you are saying even if i see the size for CephFS pools as 55 GB, it > can > > still use whole 600GB (or the available disk) from Cluster? > > > > This is what i have with PGNum = 150 (for Data) and 32 (for Metadata) in > my > > cluster. > > > > Pool > > Type > > Size > > Usage > > cephfs_data > > data > > 55.5366GiB > > 4% > > cephfs_meta > > metadata > > 55.7469GiB > > > > Thanks > > > > > > On Thu, 28 Nov 2019 at 13:49, Wido den Hollander <[email protected]> wrote: > > > >> > >> > >> On 11/28/19 6:41 AM, Alokkumar Mahajan wrote: > >> > Hello, > >> > I am new to Ceph and currently i am working on setting up CephFs and > RBD > >> > environment. I have successfully setup Ceph Cluster with 4 OSD's (2 > >> > OSD's with size 50GB and 2 OSD's with size 300GB). > >> > > >> > But while setting up CephFs the size which i see allocated for CephFs > >> > Data and metadata pools is 55GB. But i want to have 300GB assigned for > >> > CephFs. > >> > > >> > I tried using "target_size_bytes" flag while creating pool but it is > >> > not working (it saus invalid command). Same result when i > >> > use target_size_bytes with (ceph osd pool set) after creating pool. > >> > > >> > I am not sure if i am doing something silly here. > >> > > >> > Can someone please guide me on this? > >> > > >> > >> > >> You can set quotas on CephFS or on the RADOS pool for the CephFS 'data' > >> (haven't tried the last one though). > >> > >> But normally CephFS will be able to use all the space inside your Ceph > >> cluster. > >> > >> It's not that you can easily just allocate X GB/TB to CephFS. > >> > >> Wido > >> > >> > Thanks in adv.! > >> > > >> > _______________________________________________ > >> > ceph-users mailing list -- [email protected] > >> > To unsubscribe send an email to [email protected] > >> > > >> _______________________________________________ > >> ceph-users mailing list -- [email protected] > >> To unsubscribe send an email to [email protected] > >> > > > _______________________________________________ > ceph-users mailing list -- [email protected] > To unsubscribe send an email to [email protected] >
_______________________________________________ ceph-users mailing list -- [email protected] To unsubscribe send an email to [email protected]
