Hi all,
I'm running a 6 node 24 OSD cluster, Jewel 10.2.5 with kernel 4.8.
I put about 1TB of data in the cluster, with all pools having size 3. Yet
about 5TB of raw disk is used as opposed to the expected 3TB.
result of ceph -s:
pgmap v1057361: 2400 pgs, 3 pools, 984 GB data, 125 Mobjects
5039 GB used, 12353 GB / 17393 GB avail
2398 active+clean
2 active+clean+scrubbing
Result of ceph df:
GLOBAL:
SIZE AVAIL RAW USED %RAW USED
17393G 12353G 5039G 28.97
POOLS:
NAME ID USED %USED MAX AVAIL OBJECTS
rbd 1 801G 17.12 3880G 206299
cephfs_data 2 182G 4.49 3880G 130875410
cephfs_metadata 3 32540k 0 3880G 201555
Result of ceph osd dump:
pool 1 'rbd' replicated size 3 min_size 1 crush_ruleset 0 object_hash
rjenkins pg_num 800 pgp_num 800 last_change 482 flags hashpspool
stripe_width 0
pool 2 'cephfs_data' replicated size 3 min_size 1 crush_ruleset 0
object_hash rjenkins pg_num 800 pgp_num 800 last_change 410 flags
hashpspool crash_replay_interval 45 stripe_width 0
pool 3 'cephfs_metadata' replicated size 3 min_size 1 crush_ruleset 0
object_hash rjenkins pg_num 800 pgp_num 800 last_change 408 flags
hashpspool stripe_width 0
The cluster was set up with ceph-deploy. So each OSD drive is formatted
with xfs. One more thing I should mention is that I'm using the
experimental directory fragmentation feature in cephfs.
After scouring the mailing list I've found this other post that seems to be
related or the same issue:
http://lists.ceph.com/pipermail/ceph-users-ceph.com/2013-October/004901.html
Does anyone know if this is a bug or is it legitimate overhead that I
failed to account for?
Thanks,
Pavel
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com