I was curious if anyone has filled ceph storage beyond 75%. Admitedly we
lost a single host due to power failure and are down 1 host until the
replacement parts arrive but outside of that I am seeing disparity between
the most and least full osd::

ID  WEIGHT  REWEIGHT SIZE  USE   AVAIL %USE  VAR
MIN/MAX VAR: 0/1.26  STDDEV: 7.12
               TOTAL 2178T 1625T  552T 74.63

559 4.54955  1.00000 3724G 2327G 1396G 62.50 0.84
193 2.48537  1.00000 3724G 3406G  317G 91.47 1.23

The crush weights are really off right now but even with a default crush
map I am seeing a similar spread::

# osdmaptool --test-map-pgs --pool 1 /tmp/osdmap
 avg 82 stddev 10.54 (0.128537x) (expected 9.05095 0.110377x))
 min osd.336 55
 max osd.54 115

That's with a default weight of 3.000 across all osds. I was wondering if
anyone can give me any tips on how to reach closer to 80% full.

We have 630 osds (down one host right now but it will be back in in a week
or so) spread across 3 racks of 7 hosts (30 osds each). Our data
replication scheme is by rack and we only use S3 (so 98% of our data is in
.rgw.buckets pool). We are on hammer (94.7) and using the hammer tunables.




-- 
- Sean:  I wrote this. -
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to