Hi All
I'm observing some weird behavior in the amount of space ceph reports
while copying files from an rbd image in one pool to an rbd image in another.
The AVAIL number reported by 'ceph df' goes up as the copy proceeds rather
than goes down!
The output of 'ceph df' shows that the AVAIL space is 9219G initially
and then after the copy has proceeded for some time shows 9927G. (See below.)
Ceph is somehow reclaiming ~700GB of space when it should only be
losing
space! I like it! :)
Some details:
I ran fstrim on the mount points of both source and destinatino mount
points.
The data is being copied from tibs/tibs-ecpool to
3-replica/3-replica-ec.
tibs is a 3 replica pool which is backed by tibs-ecpool which is a k2m2
erasure coded pool. 3-replica/3-replica-ec is the same arrangement, but with
fewer PGs.
No data is being deleted.
I see that USED in 3-replica/3-replica-ec is going up as expected.
But the USED in both tibs/tibs-ecpool is going down. This appears to
be
from where the space is being reclaimed.
Question: why is a read of data causing it to take up less space?
Thanks!
Chad.
# ceph df
GLOBAL:
SIZE AVAIL RAW USED %RAW USED
22908G 9219G 13689G 59.76
POOLS:
NAME ID USED %USED MAX AVAIL OBJECTS
rbd 13 281 0 1990G 3
tibs 22 72724M 0.31 1990G 612278
tibs-ecpool 23 4555G 19.88 2985G 1166644
cephfs_data 27 8 0 2985G 2
cephfs_metadata 28 34800k 0 1990G 28
3-replica 31 745G 3.25 1990G 5516809
3-replica-ec 32 942G 4.12 2985G 241626
---------------------------- copy progresses ---------------------------------
# ceph df
GLOBAL:
SIZE AVAIL RAW USED %RAW USED
22908G 9927G 12980G 56.66
POOLS:
NAME ID USED %USED MAX AVAIL OBJECTS
rbd 13 281 0 2284G 3
tibs 22 68456M 0.29 2284G 88561
tibs-ecpool 23 4227G 18.45 3427G 1082734
cephfs_data 27 8 0 3427G 2
cephfs_metadata 28 34832k 0 2284G 28
3-replica 31 745G 3.25 2284G 2676207
3-replica-ec 32 945G 4.13 3427G 242194
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com