Hi all, Description of problem: [RGW] Buckets/objects deletion is causing large quantities orphan raods objects
The cluster was running a cosbench workload, then remove the partial data by deleting objects from the cosbench client, then we have deleted all the buckets with the help of `s3cmd rb --recursive --force` command that removed all the buckets, but that did not help in the space reclaimation. ``` [root@node01 /]# rgw-orphan-list Available pools: device_health_metrics .rgw.root os-test.rgw.buckets.non-ec os-test.rgw.log os-test.rgw.control os-test.rgw.buckets.index os-test.rgw.meta os-test.rgw.buckets.data deeproute-replica-hdd-pool deeproute-replica-ssd-pool cephfs-metadata cephfs-replicated-pool .nfs Which pool do you want to search for orphans (for multiple, use space-separated list)? os-test.rgw.buckets.data Pool is "os-test.rgw.buckets.data". Note: output files produced will be tagged with the current timestamp -- 20221008062356. running 'rados ls' at Sat Oct 8 06:24:05 UTC 2022 running 'rados ls' on pool os-test.rgw.buckets.data. running 'radosgw-admin bucket radoslist' at Sat Oct 8 06:43:21 UTC 2022 computing delta at Sat Oct 8 06:47:17 UTC 2022 39662551 potential orphans found out of a possible 39844453 (99%). The results can be found in './orphan-list-20221008062356.out'. Intermediate files are './rados-20221008062356.intermediate' and './radosgw-admin-20221008062356.intermediate'. *** *** WARNING: This is EXPERIMENTAL code and the results should be used *** only with CAUTION! *** Done at Sat Oct 8 06:48:07 UTC 2022. [root@node01 /]# radosgw-admin gc list [] [root@node01 /]# cat orphan-list-20221008062356.out | wc -l 39662551 [root@node01 /]# rados df POOL_NAME USED OBJECTS CLONES COPIES MISSING_ON_PRIMARY UNFOUND DEGRADED RD_OPS RD WR_OPS WR USED COMPR UNDER COMPR .nfs 4.3 MiB 4 0 12 0 0 0 77398 76 MiB 146 79 KiB 0 B 0 B .rgw.root 180 KiB 16 0 48 0 0 0 28749 28 MiB 0 0 B 0 B 0 B cephfs-metadata 932 MiB 14772 0 44316 0 0 0 1569690 3.8 GiB 1258651 3.4 GiB 0 B 0 B cephfs-replicated-pool 738 GiB 300962 0 902886 0 0 0 794612 470 GiB 770689 245 GiB 0 B 0 B deeproute-replica-hdd-pool 1016 GiB 104276 0 312828 0 0 0 18176216 298 GiB 441783780 6.7 TiB 0 B 0 B deeproute-replica-ssd-pool 30 GiB 3691 0 11073 0 0 0 2466079 2.1 GiB 8416232 221 GiB 0 B 0 B device_health_metrics 50 MiB 108 0 324 0 0 0 1836 1.8 MiB 1944 18 MiB 0 B 0 B os-test.rgw.buckets.data 5.6 TiB 39844453 0 239066718 0 0 0 552896177 3.0 TiB 999441015 60 TiB 0 B 0 B os-test.rgw.buckets.index 1.8 GiB 33 0 99 0 0 0 153600295 154 GiB 110916573 62 GiB 0 B 0 B os-test.rgw.buckets.non-ec 2.1 MiB 45 0 135 0 0 0 574240 349 MiB 153725 139 MiB 0 B 0 B os-test.rgw.control 0 B 8 0 24 0 0 0 0 0 B 0 0 B 0 B 0 B os-test.rgw.log 3.7 MiB 346 0 1038 0 0 0 83877803 80 GiB 6306730 7.6 GiB 0 B 0 B os-test.rgw.meta 220 KiB 23 0 69 0 0 0 640854 506 MiB 108229 53 MiB 0 B 0 B total_objects 40268737 total_used 7.8 TiB total_avail 1.1 PiB total_space 1.1 PiB ``` ceph verison: ``` [root@node01 /]# ceph versions { "mon": { "ceph version 16.2.10 (45fa1a083152e41a408d15505f594ec5f1b4fe17) pacific (stable)": 3 }, "mgr": { "ceph version 16.2.10 (45fa1a083152e41a408d15505f594ec5f1b4fe17) pacific (stable)": 2 }, "osd": { "ceph version 16.2.10 (45fa1a083152e41a408d15505f594ec5f1b4fe17) pacific (stable)": 108 }, "mds": { "ceph version 16.2.10 (45fa1a083152e41a408d15505f594ec5f1b4fe17) pacific (stable)": 2 }, "rgw": { "ceph version 16.2.10 (45fa1a083152e41a408d15505f594ec5f1b4fe17) pacific (stable)": 9 }, "overall": { "ceph version 16.2.10 (45fa1a083152e41a408d15505f594ec5f1b4fe17) pacific (stable)": 124 } } ``` Thanks, Best regards Liang Zheng _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io