[ceph-users] Why is my cephfs almostfull?

2023-04-05 Thread Jorge Garcia
We have a ceph cluster with a cephfs filesystem that we use mostly for backups. When I do a "ceph -s" or a "ceph df", it reports lots of space:     data:   pools:   3 pools, 4104 pgs   objects: 1.09 G objects, 944 TiB   usage:   1.5 PiB used, 1.0 PiB / 2.5 PiB avail   GLOBAL:    

[ceph-users] Mysteriously dead OSD process

2023-04-05 Thread J-P Methot
Hi, We currently use Ceph Pacific 16.2.10 deployed with Cephadm on this storage cluster. Last night, one of our OSD died. However, since its storage is a SSD, we ran hardware checks and found no issue with the SSD itself. However, if we try starting the service again, the container just

[ceph-users] Re: Misplaced objects greater than 100%

2023-04-05 Thread Johan Hattne
I think this is resolved—and you're right about the 0-weight of the root bucket being strange. I had created the rack buckets with # ceph osd crush add-bucket rack-0 rack whereas I should have used something like # ceph osd crush add-bucket rack-0 rack root=default There's a bit in the

[ceph-users] Re: Misplaced objects greater than 100%

2023-04-05 Thread ceph
I guess this is related to your crush rules.. Unfortunaly i dont know much about creating the rules... But someone cloud give more insights when you also provide crush rule dump your "-1 0 root default" is a bit strange Am 1. April 2023 01:01:39 MESZ schrieb Johan Hattne : >Here goes: >

[ceph-users] Re: quincy v17.2.6 QE Validation status

2023-04-05 Thread Crown Upholstery
ceph-users@ceph.io stop From: Thomas Widhalm Sent: Wednesday, April 5, 2023 7:26 AM To: ceph-users@ceph.io Subject: [ceph-users] Re: quincy v17.2.6 QE Validation status Sorry for interfereing, but: Wh!! Thank you so much for the great work. Can't wait

[ceph-users] Re: quincy v17.2.6 QE Validation status

2023-04-05 Thread Crown Upholstery
d...@ceph.io stop From: Thomas Widhalm Sent: Wednesday, April 5, 2023 7:26 AM To: ceph-users@ceph.io Subject: [ceph-users] Re: quincy v17.2.6 QE Validation status Sorry for interfereing, but: Wh!! Thank you so much for the great work. Can't wait for

[ceph-users] Ceph Object Gateway and lua scripts

2023-04-05 Thread Thomas Bennett
Hi, We're currently testing out lua scripting in the Ceph Object Gateway (Radosgw). Ceph version: 17.2.5 We've tried a simple experiment with the simple lua script which is based on the documentation (see fixed width text below). However, the issue we're having is that we can't find the log

[ceph-users] Re: quincy v17.2.6 QE Validation status

2023-04-05 Thread Yuri Weinstein
Thx, Josh! We will start releasing now. Release notes - https://github.com/ceph/ceph/pull/50721 On Wed, Apr 5, 2023 at 7:16 AM Josh Durgin wrote: > The LRC upgraded with no problems, so this release is good to go! > > Josh > > On Mon, Apr 3, 2023 at 3:36 PM Yuri Weinstein wrote: > >> Josh,

[ceph-users] Re: CephFS thrashing through the page cache

2023-04-05 Thread Gregory Farnum
On Fri, Mar 17, 2023 at 1:56 AM Ashu Pachauri wrote: > > Hi Xiubo, > > As you have correctly pointed out, I was talking about the stipe_unit > setting in the file layout configuration. Here is the documentation for > that for anyone else's reference: >

[ceph-users] Re: Upgrading to 16.2.11 timing out on ceph-volume due to raw list performance bug, downgrade isn't possible due to new OP code in bluestore

2023-04-05 Thread Mikael Öhman
Thank you for the suggestion Frank. We've managed to avoid patches so far, but I guess that convenience ends now :( With # lsblk -P -p -o 'NAME' | wc -l 137 it takes about 10 minutes to run. 70 probably would also bring you over the 2 minute timeout window, so I certainly wouldn't consider

[ceph-users] Re: quincy v17.2.6 QE Validation status

2023-04-05 Thread Thomas Widhalm
Sorry for interfereing, but: Wh!! Thank you so much for the great work. Can't wait for the release with a good chance to get access to my data again. On 05.04.23 16:15, Josh Durgin wrote: The LRC upgraded with no problems, so this release is good to go! Josh On Mon, Apr 3, 2023 at

[ceph-users] Re: quincy v17.2.6 QE Validation status

2023-04-05 Thread Josh Durgin
The LRC upgraded with no problems, so this release is good to go! Josh On Mon, Apr 3, 2023 at 3:36 PM Yuri Weinstein wrote: > Josh, the release is ready for your review and approval. > > Adam, can you please update the LRC upgrade to 17.2.6 RC? > > Thx > > > On Wed, Mar 29, 2023 at 3:07 PM

[ceph-users] Re: RGW can't create bucket

2023-04-05 Thread Kamil Madac
Hi Boris, debug log showed that the problem was that the customer accidentally misconfigured placement_targets and default_placement in zonegroup configuration which caused access denied issues during bucket creation. This is what was found in debug logs: s3:create_bucket user not permitted to

[ceph-users] RGW don't use .rgw.root multisite configuration

2023-04-05 Thread Guillaume Morin
Hello, for my multisite configuration, i create and use an root pool other than .rgw.root to store realm and zone configuration with the follow option: rgw_realm_root_pool=myzone.rgw.root rgw_zonegroup_root_pool=myzone.rgw.root rgw_zone_root_pool=myzone.rgw.root But , i can see during the

[ceph-users] Re: Upgrading to 16.2.11 timing out on ceph-volume due to raw list performance bug, downgrade isn't possible due to new OP code in bluestore

2023-04-05 Thread Frank Schilder
Hi Mikael, thanks for sharing this (see also https://www.stroustrup.com/whitespace98.pdf, python ha ha ha). We would probably have observed the same problem (70+ OSDs per host). You might want to consider configuring deployment against a local registry and use a patched image. Local container

[ceph-users] Re: Upgrading to 16.2.11 timing out on ceph-volume due to raw list performance bug, downgrade isn't possible due to new OP code in bluestore

2023-04-05 Thread Chris Dunlop
On Wed, Apr 05, 2023 at 01:18:57AM +0200, Mikael Öhman wrote: Trying to upgrade a containerized setup from 16.2.10 to 16.2.11 gave us two big surprises, I wanted to share in case anyone else encounters the same. I don't see any nice solution to this apart from a new release that fixes the

[ceph-users] Re: Crushmap rule for multi-datacenter erasure coding

2023-04-05 Thread Frédéric Nass
Hello Michel, What you need is: step choose indep 0 type datacenter step chooseleaf indep 2 type host step emit I think you're right about the need to tweak the crush rule by editing the crushmap directly. Regards Frédéric. - Le 3 Avr 23, à 18:34, Michel Jouvin