Hi all,
Description of problem: [RGW] Buckets/objects deletion is causing large
quantities orphan raods objects
The cluster was running a cosbench workload, then remove the partial data
by deleting objects from the cosbench client, then we have deleted all the
buckets with the help of `s3cmd rb --recursive --force` command that
removed all the buckets, but that did not help in the space reclaimation.
```
[root@node01 /]# rgw-orphan-list
Available pools:
device_health_metrics
.rgw.root
os-test.rgw.buckets.non-ec
os-test.rgw.log
os-test.rgw.control
os-test.rgw.buckets.index
os-test.rgw.meta
os-test.rgw.buckets.data
deeproute-replica-hdd-pool
deeproute-replica-ssd-pool
cephfs-metadata
cephfs-replicated-pool
.nfs
Which pool do you want to search for orphans (for multiple, use
space-separated list)? os-test.rgw.buckets.data
Pool is "os-test.rgw.buckets.data".
Note: output files produced will be tagged with the current timestamp --
20221008062356.
running 'rados ls' at Sat Oct 8 06:24:05 UTC 2022
running 'rados ls' on pool os-test.rgw.buckets.data.
running 'radosgw-admin bucket radoslist' at Sat Oct 8 06:43:21 UTC 2022
computing delta at Sat Oct 8 06:47:17 UTC 2022
39662551 potential orphans found out of a possible 39844453 (99%).
The results can be found in './orphan-list-20221008062356.out'.
Intermediate files are './rados-20221008062356.intermediate' and
'./radosgw-admin-20221008062356.intermediate'.
***
*** WARNING: This is EXPERIMENTAL code and the results should be used
*** only with CAUTION!
***
Done at Sat Oct 8 06:48:07 UTC 2022.
[root@node01 /]# radosgw-admin gc list
[]
[root@node01 /]# cat orphan-list-20221008062356.out | wc -l
39662551
[root@node01 /]# rados df
POOL_NAME USED OBJECTS CLONES COPIES
MISSING_ON_PRIMARY UNFOUND DEGRADED RD_OPS RD WR_OPS
WR USED COMPR UNDER COMPR
.nfs 4.3 MiB 4 0 12
0 0 0 77398 76 MiB 146 79
KiB 0 B 0 B
.rgw.root 180 KiB 16 0 48
0 0 0 28749 28 MiB 0
0 B 0 B 0 B
cephfs-metadata 932 MiB 14772 0 44316
0 0 0 1569690 3.8 GiB 1258651 3.4
GiB 0 B 0 B
cephfs-replicated-pool 738 GiB 300962 0 902886
0 0 0 794612 470 GiB 770689 245
GiB 0 B 0 B
deeproute-replica-hdd-pool 1016 GiB 104276 0 312828
0 0 0 18176216 298 GiB 441783780 6.7
TiB 0 B 0 B
deeproute-replica-ssd-pool 30 GiB 3691 0 11073
0 0 0 2466079 2.1 GiB 8416232 221
GiB 0 B 0 B
device_health_metrics 50 MiB 108 0 324
0 0 0 1836 1.8 MiB 1944 18
MiB 0 B 0 B
os-test.rgw.buckets.data 5.6 TiB 39844453 0 239066718
0 0 0 552896177 3.0 TiB 999441015 60
TiB 0 B 0 B
os-test.rgw.buckets.index 1.8 GiB 33 0 99
0 0 0 153600295 154 GiB 110916573 62
GiB 0 B 0 B
os-test.rgw.buckets.non-ec 2.1 MiB 45 0 135
0 0 0 574240 349 MiB 153725 139
MiB 0 B 0 B
os-test.rgw.control 0 B 8 0 24
0 0 0 0 0 B 0
0 B 0 B 0 B
os-test.rgw.log 3.7 MiB 346 0 1038
0 0 0 83877803 80 GiB 6306730 7.6
GiB 0 B 0 B
os-test.rgw.meta 220 KiB 23 0 69
0 0 0 640854 506 MiB 108229 53
MiB 0 B 0 B
total_objects 40268737
total_used 7.8 TiB
total_avail 1.1 PiB
total_space 1.1 PiB
```
ceph verison:
```
[root@node01 /]# ceph versions
{
"mon": {
"ceph version 16.2.10 (45fa1a083152e41a408d15505f594ec5f1b4fe17)
pacific (stable)": 3
},
"mgr": {
"ceph version 16.2.10 (45fa1a083152e41a408d15505f594ec5f1b4fe17)
pacific (stable)": 2
},
"osd": {
"ceph version 16.2.10 (45fa1a083152e41a408d15505f594ec5f1b4fe17)
pacific (stable)": 108
},
"mds": {
"ceph version 16.2.10 (45fa1a083152e41a408d15505f594ec5f1b4fe17)
pacific (stable)": 2
},
"rgw": {
"ceph version 16.2.10 (45fa1a083152e41a408d15505f594ec5f1b4fe17)
pacific (stable)": 9
},
"overall": {
"ceph version 16.2.10 (45fa1a083152e41a408d15505f594ec5f1b4fe17)
pacific (stable)": 124
}
}
```
Thanks,
Best regards
Liang Zheng
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]