And we may be interested in your cluster's configuration. # ceph --show-config > $(hostname).$(date +%Y%m%d).ceph_conf.txt
On Fri, Dec 30, 2016 at 7:48 AM, David Turner <[email protected] > wrote: > Another thing that I need to make sure on is that your number of PGs in > the pool with 90% of the data is a power of 2 (256, 512, 1024, 2048, etc). > If that is the case, then I need the following information. > > 1) Pool replica size > 2) The number of the pool with the data > 3) A copy of your osdmap (ceph osd getmap -o osd_map.bin) > 4) Full output of (ceph osd tree) > 5) Full output of (ceph osd df) > > With that I can generate a new crushmap that is balanced for your cluster > to equalize all of the osds % used. > > Our clusters have more than 1k osds and the difference between the top > used osd and the least used osd is within 2% in those clusters. We have > 99.9% of our data in 1 pool. > > ------------------------------ > > <https://storagecraft.com> David Turner | Cloud Operations Engineer | > StorageCraft > Technology Corporation <https://storagecraft.com> > 380 Data Drive Suite 300 | Draper | Utah | 84020 > Office: 801.871.2760 <(801)%20871-2760> | Mobile: 385.224.2943 > <(385)%20224-2943> > > ------------------------------ > > If you are not the intended recipient of this message or received it > erroneously, please notify the sender and delete it, together with any > attachments, and be advised that any dissemination or copying of this > message is prohibited. > > ------------------------------ > > ________________________________________ > From: ceph-users [[email protected]] on behalf of Bryan > Henderson [[email protected]] > Sent: Thursday, December 29, 2016 3:31 PM > To: [email protected] > Subject: [ceph-users] ceph program uses lots of memory > > > Does anyone know why the 'ceph' program uses so much memory? If I run it > with > an address space rlimit of less than 300M, it usually dies with messages > about > not being able to allocate memory. > > I'm curious as to what it could be doing that requires so much address > space. > > It doesn't matter what specific command I'm doing and it does this even > with > there is no ceph cluster running, so it must be something pretty basic. > > -- > Bryan Henderson San Jose, California > _______________________________________________ > ceph-users mailing list > [email protected] > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > _______________________________________________ > ceph-users mailing list > [email protected] > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
