I see the Problem, as your OSD's are only 8GB they have a zero weight, I think 
the minimum size you can get away with is 10GB in Ceph as the size is measured 
in TB and only has 2 decimal places.

For a work around try running :-

ceph osd crush reweight osd.X 1     

for each osd, this will reweight the OSD's. Assuming this is a test cluster and 
you won't be adding any larger OSD's in the future this shouldn't cause any 
problems.

> 
> admin@cp-admin:~/safedrive$ ceph osd tree
> # id    weight    type name    up/down    reweight
> -1    0    root default
> -2    0        host osd-001
> 0    0            osd.0    up    1
> 1    0            osd.1    up    1
> -3    0        host osd-002
> 2    0            osd.2    up    1
> 3    0            osd.3    up    1
> -4    0        host osd-003
> 4    0            osd.4    up    1
> 5    0            osd.5    up    1




_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to