Hello,
Probably this is anti-pattern, but I have to get answer how this
will work / not work.
Input:
I have single host for tests with ceph 0.80.1 and 2 OSD:
OSD.0 – 1000 Gb
OSD.1 – 750 Gb
Recompiled CRUSH map to set „step chooseleaf firstn 0 type osd“
I am expecting, that part of PG-s will have status „active+clean“
(with size of ~750Gb) another part of PG-s will have „active+degradated“ (with
size of ~250Gb), because there is not enough place to replicate data on the
second OSD.
Instead I have ALL PG-s „active + degradated“
Output:
health HEALTH_WARN 192 pgs degraded; 192 pgs stuck unclean
monmap e1: 1 mons at {storage=172.16.3.2:6789/0}, election epoch 2, quorum
0 storage
osdmap e15: 2 osds: 2 up, 2 in
pgmap v29: 192 pgs, 3 pools, 0 bytes data, 0 objects
71496 kB used, 1619 GB / 1619 GB avail
192 active+degraded
What is the logic behind this?? Can I use different hard drives
successfully? If yes – how?
Thank you for explanation,
Vadim
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com