Please review http://docs.ceph.com/docs/master/rados/operations/crush-map/ regarding weights
Best regards, Alex On Wed, Sep 23, 2015 at 3:08 AM, wikison <wiki...@163.com> wrote: > Hi, > I have four storage machines to build a ceph storage cluster as > storage nodes. Each of them is attached a 120 GB HDD and a 1 TB HDD. Is it > OK to think that those storage devices are same when write a ceph.conf? > For example, when setting *osd pool default pg num* , I thought: *osd > pool default pg num* = (100 * 8 ) / 3 = 266, where *osd pool default > size* = 3 and the number of OSDs is 8 (one Daemon per device). > > And, when Add the OSD to the CRUSH map so that it can begin > receiving data. You may also decompile the CRUSH map, add the OSD to the > device list, add the host as a bucket (if it’s not already in the CRUSH > map), add the device as an item in the host, assign it a weight, recompile > it and set it. > ceph [--cluster {cluster-name}] osd crush add {id-or-name} > {weight} [{bucket-type}={bucket-name} ...] > What is the meaning of weight? How should I set it to satisfy my > hardware condition ? > > > > -- > Zhen Wang > > > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com