Ah you are right. [root@s3db1 ~]# ceph daemon osd.23 config get bluestore_min_alloc_size_hdd { "bluestore_min_alloc_size_hdd": "65536" } But I also checked how many objects our s3 hold and the numbers just do not add up. There are only 26509200 objects, which would result in around 1TB "waste" if every object would be empty.
I think the problem began when I updated the PG count from 1024 to 2048. Could there be an issue where the data is written twice? Am Do., 15. Apr. 2021 um 16:48 Uhr schrieb Amit Ghadge <amitg....@gmail.com >: > verify those two parameter values ,bluestore_min_alloc_size_hdd & > bluestore_min_alloc_size_sdd, If you are using hdd disk then > bluestore_min_alloc_size_hdd are applicable. > > On Thu, Apr 15, 2021 at 8:06 PM Boris Behrens <b...@kervyn.de> wrote: > >> So, I need to live with it? A value of zero leads to use the default? >> [root@s3db1 ~]# ceph daemon osd.23 config get bluestore_min_alloc_size >> { >> "bluestore_min_alloc_size": "0" >> } >> >> I also checked the fragmentation on the bluestore OSDs and it is around >> 0.80 - 0.89 on most OSDs. yikes. >> [root@s3db1 ~]# ceph daemon osd.23 bluestore allocator score block >> { >> "fragmentation_rating": 0.85906054329923576 >> } >> >> The problem I currently have is, that I barely keep up with adding OSD >> disks. >> >> Am Do., 15. Apr. 2021 um 16:18 Uhr schrieb Amit Ghadge < >> amitg....@gmail.com>: >> >>> size_kb_actual are actually bucket object size but on OSD level the >>> bluestore_min_alloc_size default 64KB and SSD are 16KB >>> >>> >>> https://access.redhat.com/documentation/en-us/red_hat_ceph_storage/3/html/administration_guide/osd-bluestore >>> >>> -AmitG >>> >>> On Thu, Apr 15, 2021 at 7:29 PM Boris Behrens <b...@kervyn.de> wrote: >>> >>>> Hi, >>>> >>>> maybe it is just a problem in my understanding, but it looks like our s3 >>>> requires twice the space it should use. >>>> >>>> I ran "radosgw-admin bucket stats", and added all "size_kb_actual" >>>> values >>>> up and divided to TB (/1024/1024/1024). >>>> The resulting space is 135,1636733 TB. When I tripple it because of >>>> replication I end up with around 405TB which is nearly half the space of >>>> what ceph df tells me. >>>> >>>> Hope someone can help me. >>>> >>>> ceph df shows >>>> RAW STORAGE: >>>> CLASS SIZE AVAIL USED RAW USED %RAW >>>> USED >>>> hdd 1009 TiB 189 TiB 820 TiB 820 TiB >>>> 81.26 >>>> TOTAL 1009 TiB 189 TiB 820 TiB 820 TiB >>>> 81.26 >>>> >>>> POOLS: >>>> POOL ID PGS STORED >>>> OBJECTS >>>> USED %USED MAX AVAIL >>>> rbd 0 64 0 B >>>> 0 >>>> 0 B 0 18 TiB >>>> .rgw.root 1 64 99 KiB >>>> 119 >>>> 99 KiB 0 18 TiB >>>> eu-central-1.rgw.control 2 64 0 B >>>> 8 >>>> 0 B 0 18 TiB >>>> eu-central-1.rgw.data.root 3 64 1.0 MiB >>>> 3.15k >>>> 1.0 MiB 0 18 TiB >>>> eu-central-1.rgw.gc 4 64 71 MiB >>>> 32 >>>> 71 MiB 0 18 TiB >>>> eu-central-1.rgw.log 5 64 267 MiB >>>> 564 >>>> 267 MiB 0 18 TiB >>>> eu-central-1.rgw.users.uid 6 64 2.8 MiB >>>> 6.91k >>>> 2.8 MiB 0 18 TiB >>>> eu-central-1.rgw.users.keys 7 64 263 KiB >>>> 6.73k >>>> 263 KiB 0 18 TiB >>>> eu-central-1.rgw.meta 8 64 384 KiB >>>> 1k >>>> 384 KiB 0 18 TiB >>>> eu-central-1.rgw.users.email 9 64 40 B >>>> 1 >>>> 40 B 0 18 TiB >>>> eu-central-1.rgw.buckets.index 10 64 10 GiB >>>> 67.61k >>>> 10 GiB 0.02 18 TiB >>>> eu-central-1.rgw.buckets.data 11 2048 264 TiB >>>> 138.31M >>>> 264 TiB 83.37 18 TiB >>>> eu-central-1.rgw.buckets.non-ec 12 64 297 MiB >>>> 11.32k >>>> 297 MiB 0 18 TiB >>>> eu-central-1.rgw.usage 13 64 536 MiB >>>> 32 >>>> 536 MiB 0 18 TiB >>>> eu-msg-1.rgw.control 56 64 0 B >>>> 8 >>>> 0 B 0 18 TiB >>>> eu-msg-1.rgw.data.root 57 64 72 KiB >>>> 227 >>>> 72 KiB 0 18 TiB >>>> eu-msg-1.rgw.gc 58 64 300 KiB >>>> 32 >>>> 300 KiB 0 18 TiB >>>> eu-msg-1.rgw.log 59 64 835 KiB >>>> 242 >>>> 835 KiB 0 18 TiB >>>> eu-msg-1.rgw.users.uid 60 64 56 KiB >>>> 104 >>>> 56 KiB 0 18 TiB >>>> eu-msg-1.rgw.usage 61 64 37 MiB >>>> 25 >>>> 37 MiB 0 18 TiB >>>> eu-msg-1.rgw.users.keys 62 64 3.8 KiB >>>> 97 >>>> 3.8 KiB 0 18 TiB >>>> eu-msg-1.rgw.meta 63 64 607 KiB >>>> 1.60k >>>> 607 KiB 0 18 TiB >>>> eu-msg-1.rgw.buckets.index 64 64 71 MiB >>>> 119 >>>> 71 MiB 0 18 TiB >>>> eu-msg-1.rgw.users.email 65 64 0 B >>>> 0 >>>> 0 B 0 18 TiB >>>> eu-msg-1.rgw.buckets.data 66 64 2.9 TiB >>>> 1.16M >>>> 2.9 TiB 5.30 18 TiB >>>> eu-msg-1.rgw.buckets.non-ec 67 64 2.2 MiB >>>> 354 >>>> 2.2 MiB 0 18 TiB >>>> default.rgw.control 69 32 0 B >>>> 8 >>>> 0 B 0 18 TiB >>>> default.rgw.data.root 70 32 0 B >>>> 0 >>>> 0 B 0 18 TiB >>>> default.rgw.gc 71 32 0 B >>>> 0 >>>> 0 B 0 18 TiB >>>> default.rgw.log 72 32 0 B >>>> 0 >>>> 0 B 0 18 TiB >>>> default.rgw.users.uid 73 32 0 B >>>> 0 >>>> 0 B 0 18 TiB >>>> fra-1.rgw.control 74 32 0 B >>>> 8 >>>> 0 B 0 18 TiB >>>> fra-1.rgw.meta 75 32 0 B >>>> 0 >>>> 0 B 0 18 TiB >>>> fra-1.rgw.log 76 32 50 B >>>> 28 >>>> 50 B 0 18 TiB >>>> >>>> >>>> -- >>>> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im >>>> groüen Saal. >>>> _______________________________________________ >>>> ceph-users mailing list -- ceph-users@ceph.io >>>> To unsubscribe send an email to ceph-users-le...@ceph.io >>>> >>> >> >> -- >> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im >> groüen Saal. >> > -- Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im groüen Saal. _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io