[ceph-users] Re: ceph rgw zone create fails EINVAL

2024-06-25 Thread Matthew Vernon
On 24/06/2024 21:18, Matthew Vernon wrote: 2024-06-24T17:33:26.880065+00:00 moss-be2001 ceph-mgr[129346]: [rgw ERROR root] Non-zero return from ['radosgw-admin', '-k', '/var/lib/ceph/mgr/ceph-moss-be2001.qvwcaq/keyring', '-n', 'mgr.moss-be2001.qvwcaq', 'realm', 'pull', '--url', 'https

[ceph-users] Re: ceph rgw zone create fails EINVAL

2024-06-24 Thread Matthew Vernon
On 24/06/2024 20:49, Matthew Vernon wrote: On 19/06/2024 19:45, Adam King wrote: I think this is at least partially a code bug in the rgw module. Where ...the code path seems to have a bunch of places it might raise an exception; are those likely to result in some entry in a log-file? I've

[ceph-users] Re: ceph rgw zone create fails EINVAL

2024-06-24 Thread Matthew Vernon
out what the problem is quite challenging... Thanks, Matthew ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] ceph rgw zone create fails EINVAL

2024-06-19 Thread Matthew Vernon
in the docs[0]. Can anyone point me in the right direction, please? [if the underlying command emits anything useful, I can't find it in the logs] Thanks, Matthew [0] https://docs.ceph.com/en/reef/mgr/rgw/#realm-credentials-token ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Setting hostnames for zonegroups via cephadm / rgw mgr module?

2024-06-04 Thread Matthew Vernon
up (and thus control what hostname(s) the rgws are expecting to serve)? Have I missed something, or do I need to set up the realm/zonegroup/zone, extract the zonegroup json and edit hostnames by hand? Thanks, Matthew ___ ceph-users mailing list -

[ceph-users] rgw mgr module not shipped? (in reef at least)

2024-05-31 Thread Matthew Vernon
already, and the rgw one is effectively one small python file, I think... I'm using 18.2.2. Thanks, Matthew ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: ceph orch osd rm --zap --replace leaves cluster in odd state

2024-05-28 Thread Matthew Vernon
; echo $? OSD(s) 35 are safe to destroy without reducing data durability. 0 I should have said - this is a reef 18.2.2 cluster, cephadm deployed. Regards, Matthew ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] ceph orch osd rm --zap --replace leaves cluster in odd state

2024-05-28 Thread Matthew Vernon
do wrong? I don't much care about the OSD id (but obviously it's neater to not just incrementally increase OSD numbers every time a disk died), but I thought that telling ceph orch not to make new OSDs then using ceph orch osd rm to zap the disk and NVME lv would have been the way to g

[ceph-users] Re: cephadm bootstraps cluster with bad CRUSH map(?)

2024-05-22 Thread Matthew Vernon
at least useful as a starter-for-ten: https://github.com/ceph/ceph/pull/57633 Thanks, Matthew ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: cephadm bootstraps cluster with bad CRUSH map(?)

2024-05-21 Thread Matthew Vernon
nd similar for the others, but is there a way to have what I want done by cephadm bootstrap? Thanks, Matthew ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: cephadm bootstraps cluster with bad CRUSH map(?)

2024-05-20 Thread Matthew Vernon
vgdisplay on the vg that pvs tells me the nvme device is in shows 24 LVs... Thanks, Matthew ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: cephadm bootstraps cluster with bad CRUSH map(?)

2024-05-20 Thread Matthew Vernon
Hi, On 20/05/2024 17:29, Anthony D'Atri wrote: On May 20, 2024, at 12:21 PM, Matthew Vernon wrote: This has left me with a single sad pg: [WRN] PG_AVAILABILITY: Reduced data availability: 1 pg inactive pg 1.0 is stuck inactive for 33m, current state unknown, last acting [] .mgr pool

[ceph-users] cephadm bootstraps cluster with bad CRUSH map(?)

2024-05-20 Thread Matthew Vernon
thing to want to do with cephadm? I'm running ceph version 18.2.2 (531c0d11a1c5d39fbfe6aa8a521f023abf3bf3e2) reef (stable) Thanks, Matthew ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] cephadm basic questions: image config, OS reimages

2024-05-16 Thread Matthew Vernon
; how does one do this in a cephadm cluster? [I presume involves telling cephadm to download a new image for podman to use and suchlike] Would the process be smoother if we arranged to leave /var/lib/ceph intact between reimages? Thanks, Matthew

[ceph-users] Re: Ceph reef and (slow) backfilling - how to speed it up

2024-05-10 Thread Matthew Darwin
We have had pgs get stuck  in quincy (17.2.7).  After changing to wpq, no such problems were observed.  We're using a replicated (x3) pool. On 2024-05-02 10:02, Wesley Dillingham wrote: In our case it was with a EC pool as well. I believe the PG state was degraded+recovering / recovery_wait

[ceph-users] How to define a read-only sub-user?

2024-05-08 Thread Matthew Darwin
Hi, I'm new to bucket policies. I'm trying to create a sub-user that has only read-only access to all the buckets of the main user. I created the below policy, I can't create or delete files, but I can still create buckets using "rclone mkdir".  Any idea what I'm doing wrong? I'm using ceph

[ceph-users] Re: Reconstructing an OSD server when the boot OS is corrupted

2024-05-02 Thread Matthew Vernon
On 24/04/2024 13:43, Bailey Allison wrote: A simple ceph-volume lvm activate should get all of the OSDs back up and running once you install the proper packages/restore the ceph config file/etc., What's the equivalent procedure in a cephadm-managed cluster? Thanks, Matthew

[ceph-users] Linux Laptop Losing CephFS mounts on Sleep/Hibernate

2024-03-25 Thread matthew
Hi All, So I've got a Ceph Reef Cluster (latest version) with a CephFS system set up with a number of directories on it. On a Laptop (running Rocky Linux (latest version)) I've used fstab to mount a number of those directories - all good, everything works, happy happy joy joy! :-) However,

[ceph-users] Mounting A RBD Image via Kernal Modules

2024-03-25 Thread matthew
Hi All, I'm looking for a bit of advice on the subject of this post. I've been "staring at the trees so long I can't see the forest any more". :-) Rocky Linux Client latest version. Ceph Reef latest version. I have read *all* the doco on the Ceph website. I have created a pool (my_pool) and

[ceph-users] Re: Ceph Cluster Config File Locations?

2024-03-06 Thread matthew
Thanks Eugen, you pointed me in the right direction :-) Yes, the config files I mentioned were the ones in `/var/lib/ceph/{FSID}/mgr.{MGR}/config` - I wasn't aware there were others (well, I suspected their was, hence my Q). The `global public-network` was (re-)set to the old subnet, while

[ceph-users] Re: Ceph-storage slack access

2024-03-06 Thread Matthew Vernon
://docs.ceph.com/en/latest/start/get-involved/ which lacks the registration link. Regards, Matthew ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Ceph-storage slack access

2024-03-06 Thread Matthew Vernon
Hi, How does one get an invite to the ceph-storage slack, please? Thanks, Matthew ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: [Urgent] Ceph system Down, Ceph FS volume in recovering

2024-02-26 Thread Matthew Leonard (BLOOMBERG/ 120 PARK)
Glad to hear it all worked out for you! From: nguyenvand...@baoviet.com.vn At: 02/26/24 05:32:32 UTC-5:00To: ceph-users@ceph.io Subject: [ceph-users] Re: [Urgent] Ceph system Down, Ceph FS volume in recovering Dear Mr Eugen, Mr Matthew, Mr David, Mr Anthony My System is UP. Thank you so

[ceph-users] Re: [Urgent] Ceph system Down, Ceph FS volume in recovering

2024-02-24 Thread Matthew Leonard (BLOOMBERG/ 120 PARK)
Once recovery is underway way simply restarting the RGWs should be enough to reset them and get your object store back up. Bloomberg doesn’t use cephfs so hopefully David’s suggestions work or if anyone else in the community can chip in for that part. Sent from Bloomberg Professional for

[ceph-users] Re: [Urgent] Ceph system Down, Ceph FS volume in recovering

2024-02-24 Thread Matthew Leonard (BLOOMBERG/ 120 PARK)
: nguyenvand...@baoviet.com.vn To: ceph-users@ceph.io At: 02/24/24 16:14:12 UTC Thank you Matthew. Im following guidance from Mr Anthony and now my recovery progress speed is much faster. I will update my case day by day. Thank you so much ___ ceph-users mailing

[ceph-users] Re: [Urgent] Ceph system Down, Ceph FS volume in recovering

2024-02-24 Thread Matthew Leonard (BLOOMBERG/ 120 PARK)
Anthony is correct, this is what I was getting at as well when seeing your ceph -s output. More details in the Ceph docs here if you want to understand the details of why you need to balance your nodes. https://docs.ceph.com/en/quincy/rados/operations/monitoring-osd-pg/ But you need to get

[ceph-users] Re: [Urgent] Ceph system Down, Ceph FS volume in recovering

2024-02-24 Thread Matthew Leonard (BLOOMBERG/ 120 PARK)
It looks like you have quite a few problems I’ll try and address them one by one. 1) Looks like you had a bunch of crashes, from the ceph -s it looks like you don’t have enough MDS daemons running for a quorum. So you’ll need to restart the crashed containers. 2) It looks like you might

[ceph-users] Re: [Urgent] Ceph system Down, Ceph FS volume in recovering

2024-02-23 Thread Matthew Leonard (BLOOMBERG/ 120 PARK)
Can you send sudo ceph -s and sudo ceph health detail Sent from Bloomberg Professional for iPhone - Original Message - From: nguyenvand...@baoviet.com.vn To: ceph-users@ceph.io At: 02/23/24 20:27:53 UTC-05:00 Could you pls guide me more detail :( im very newbie in Ceph :(

[ceph-users] Re: Issue with Setting Public/Private Permissions for Bucket

2024-02-23 Thread Matthew Leonard (BLOOMBERG/ 120 PARK)
https://docs.aws.amazon.com/AmazonS3/latest/userguide/acl-overview.html From: asad.siddi...@rapidcompute.com At: 02/23/24 09:42:29 UTC-5:00To: ceph-users@ceph.io Subject: [ceph-users] Issue with Setting Public/Private Permissions for Bucket Hi Team, I'm currently working with Ceph object

[ceph-users] Re: Debian 12 (bookworm) / Reef 18.2.1 problems

2024-02-21 Thread Matthew Vernon
); there is a MR to fix just the dashboard issue which got merged into main. I've opened a MR to backport that change to Reef: https://github.com/ceph/ceph/pull/55689 I don't know what the devs' plans are for dealing with the broader pyO3 issue, but I'll ask on the dev list... Regards, Matthew

[ceph-users] Re: Debian 12 (bookworm) / Reef 18.2.1 problems

2024-02-02 Thread Matthew Darwin
Chris, Thanks for all the investigations you are doing here. We're on quincy/debian11.  Is there any working path at this point to reef/debian12?  Ideally I want to go in two steps.  Upgrade ceph first or upgrade debian first, then do the upgrade to the other one. Most of our infra is

[ceph-users] Understanding subvolumes

2024-01-31 Thread Matthew Melendy
~]# ceph fs subvolume ls cephfs csvg [ { "name": "staff" } ] -- Sincerely, Matthew Melendy IT Services Specialist CS System Services Group FEC 3550, University of New Mexico ___ ceph-users mailing list -- ceph-users@ce

[ceph-users] Re: v18.2.1 Reef released

2023-12-19 Thread Matthew Vernon
18.2.1 (whereas the reporter is still on 18.2.0)? i.e. one has to upgrade to 18.2.1 before this bug will be fixed and so the upgrade _to_ 18.2.1 is still affected. Regards, Matthew ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send

[ceph-users] Re: Debian 12 support

2023-11-13 Thread Matthew Vernon
is that the next point release of Reef (due soon!) will have Debian packages built as part of it. Regards, Matthew ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: v17.2.7 Quincy released

2023-11-12 Thread Matthew Darwin
It would be nice if the dashboard changes which are very big would have been covered in the release notes, especially since they are not really backwards compatible. (See my previous messages on this topic) On 2023-10-30 10:50, Yuri Weinstein wrote: We're happy to announce the 7th backport

[ceph-users] Re: Debian 12 support

2023-11-12 Thread Matthew Darwin
We are still waiting on debian 12 support.  Currently our ceph is stuck on debian 11 due to lack of debian 12 releases. On 2023-11-01 03:23, nessero karuzo wrote: Hi to all ceph community. I have a question about Debian 12 support for ceph 17. I didn’t find repo for that release

[ceph-users] Re: OSD fails to start after 17.2.6 to 17.2.7 update

2023-11-07 Thread Matthew Booth
I just discovered that rook is tracking this here: https://github.com/rook/rook/issues/13136 On Tue, 7 Nov 2023 at 18:09, Matthew Booth wrote: > On Tue, 7 Nov 2023 at 16:26, Matthew Booth wrote: > >> FYI I left rook as is and reverted to ceph 17.2.6 and the issue is >> resolv

[ceph-users] Re: OSD fails to start after 17.2.6 to 17.2.7 update

2023-11-07 Thread Matthew Booth
On Tue, 7 Nov 2023 at 16:26, Matthew Booth wrote: > FYI I left rook as is and reverted to ceph 17.2.6 and the issue is > resolved. > > The code change was added by > commit 2e52c029bc2b052bb96f4731c6bb00e30ed209be: > ceph-volume: fix broken workaround for atari partiti

[ceph-users] Re: OSD fails to start after 17.2.6 to 17.2.7 update

2023-11-07 Thread Matthew Booth
that regression. Fixes: https://tracker.ceph.com/issues/62001 Signed-off-by: Guillaume Abrioux (cherry picked from commit b3fd5b513176fb9ba1e6e0595ded4b41d401c68e) It feels like a regression to me. Matt On Tue, 7 Nov 2023 at 16:13, Matthew Booth wrote: > Firstly I'm rolling

[ceph-users] OSD fails to start after 17.2.6 to 17.2.7 update

2023-11-07 Thread Matthew Booth
enerate info_device = [info for info in info_devices if info['NAME'] == dev][0] IndexError: list index out of range So it has failed executing `ceph-volume raw list /dev/sdc`. It looks like this code is new in 17.2.7. Is this a regression? What would be the simplest way to back out of it?

[ceph-users] Re: Many pgs inactive after node failure

2023-11-06 Thread Matthew Booth
> > [1] https://docs.ceph.com/en/reef/cephadm/services/osd/#activate-existing-osds > > Zitat von Matthew Booth : > > > I have a 3 node ceph cluster in my home lab. One of the pools spans 3 > > hdds, one on each node, and has size 2, min size 1. One of my nodes is

[ceph-users] Many pgs inactive after node failure

2023-11-04 Thread Matthew Booth
will most likely rebuild it. I'm running rook, and I will most likely delete the old node and create a new one with the same name. AFAIK, the OSDs are fine. When rook rediscovers the OSDs, will it add them back with data intact? If not, is there any way I can make it so it will? Thanks! -- Matthew Booth

[ceph-users] Re: 17.2.7 quincy dashboard issues

2023-11-02 Thread Matthew Darwin
Gomez <mailto:pegon...@redhat.com> @Ankush Behl <mailto:anb...@redhat.com> @Aashish Sharma <mailto:aasha...@redhat.com> Regards, Nizam On Mon, Oct 30, 2023 at 11:05 PM Matthew Darwin wrote: Ok, so I tried the new ceph dashboard by "set-prometheus-api-host"

[ceph-users] Re: 17.2.7 quincy dashboard issues

2023-10-30 Thread Matthew Darwin
t's why the utilization charts are empty because it relies on the prometheus info. And I raised a PR to disable the new dashboard in quincy. https://github.com/ceph/ceph/pull/54250 Regards, Nizam On Mon, Oct 30, 2023 at 6:09 PM Matthew Darwin wrote: Hello, We're not using prometheu

[ceph-users] Re: 17.2.7 quincy

2023-10-30 Thread Matthew Darwin
by default. "ceph dashboard feature disable dashboard" works to put the old dashboard back.  Thanks. On 2023-10-30 00:09, Nizamudeen A wrote: Hi Matthew, Is the prometheus configured in the cluster? And also the PROMETHUEUS_API_URL is set? You can set it manually by ceph dashboard set-prom

[ceph-users] 17.2.7 quincy

2023-10-29 Thread Matthew Darwin
Hi all, I see17.2.7 quincy is published as debian-bullseye packages.  So I tried it on a test cluster. I must say I was not expecting the big dashboard change in a patch release.  Also all the "cluster utilization" numbers are all blank now (any way to fix it?), so the dashboard is much

[ceph-users] Re: radosgw-admin sync error trim seems to do nothing

2023-10-03 Thread Matthew Darwin
On 2023-08-22 08:00, Matthew Darwin wrote: Thanks Rich, On quincy it seems that provding an end-date is an error.  Any other ideas from anyone? $ radosgw-admin sync error trim --end-date="2023-08-20 23:00:00" end-date not allowed. On 2023-08-20 19:00, Richard Bade wrote: Hi Matthew

[ceph-users] Re: Debian/bullseye build for reef

2023-09-07 Thread Matthew Vernon
, Matthew ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Debian/bullseye build for reef

2023-09-04 Thread Matthew Vernon
much time] Regards, Matthew ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: radosgw-admin sync error trim seems to do nothing

2023-08-22 Thread Matthew Darwin
Thanks Rich, On quincy it seems that provding an end-date is an error.  Any other ideas from anyone? $ radosgw-admin sync error trim --end-date="2023-08-20 23:00:00" end-date not allowed. On 2023-08-20 19:00, Richard Bade wrote: Hi Matthew, At least for nautilus (14.2.22) i have

[ceph-users] Re: Debian/bullseye build for reef

2023-08-21 Thread Matthew Darwin
Last few upgrades we upgraded ceph, then upgraded the O/S... it worked great... I was hoping we could do the same again this time. On 2023-08-21 12:18, Chris Palmer wrote: Ohhh.. so if I read that correctly we can't upgrade either debian or ceph until the dependency problem is

[ceph-users] radosgw-admin sync error trim seems to do nothing

2023-08-19 Thread Matthew Darwin
Hello all, "radosgw-admin sync error list" returns errors from 2022.  I want to clear those out. I tried "radosgw-admin sync error trim" but it seems to do nothing.  The man page seems to offer no suggestions https://docs.ceph.com/en/quincy/man/8/radosgw-admin/ Any ideas what I need to do

[ceph-users] Re: 1 PG stucked in "active+undersized+degraded for long time

2023-07-20 Thread Matthew Leonard (BLOOMBERG/ 120 PARK)
Assuming you're running systemctl OSDs you can run the following command on the host that OSD 343 resides on. systemctl restart ceph-osd@343 From: siddhit.ren...@nxtgen.com At: 07/20/23 13:44:36 UTC-4:00To: ceph-users@ceph.io Subject: [ceph-users] Re: 1 PG stucked in

[ceph-users] Re: RBD with PWL cache shows poor performance compared to cache device

2023-07-10 Thread Matthew Booth
On Thu, 6 Jul 2023 at 12:54, Mark Nelson wrote: > > > On 7/6/23 06:02, Matthew Booth wrote: > > On Wed, 5 Jul 2023 at 15:18, Mark Nelson wrote: > >> I'm sort of amazed that it gave you symbols without the debuginfo > >> packages installed. I'll ne

[ceph-users] Re: RBD with PWL cache shows poor performance compared to cache device

2023-07-06 Thread Matthew Booth
of tp_pwl > threads from 4 to 1 and see if that changes anything. Will do. Any idea how to do that? I don't see an obvious rbd config option. Thanks for looking into this, Matt -- Matthew Booth ___ ceph-users mailing list -- ceph-users@ceph.io To

[ceph-users] Re: RBD with PWL cache shows poor performance compared to cache device

2023-07-04 Thread Matthew Booth
On Tue, 4 Jul 2023 at 10:00, Matthew Booth wrote: > > On Mon, 3 Jul 2023 at 18:33, Ilya Dryomov wrote: > > > > On Mon, Jul 3, 2023 at 6:58 PM Mark Nelson wrote: > > > > > > > > > On 7/3/23 04:53, Matthew Booth wrote: > > &

[ceph-users] Re: RBD with PWL cache shows poor performance compared to cache device

2023-07-04 Thread Matthew Booth
On Tue, 4 Jul 2023 at 14:24, Matthew Booth wrote: > On Tue, 4 Jul 2023 at 10:45, Yin, Congmin wrote: > > > > Hi , Matthew > > > > I see "rbd with pwl cache: 5210112 ns", This latency is beyond my > > expectations and I believe it is unlikely

[ceph-users] Re: RBD with PWL cache shows poor performance compared to cache device

2023-07-04 Thread Matthew Booth
On Tue, 4 Jul 2023 at 10:45, Yin, Congmin wrote: > > Hi , Matthew > > I see "rbd with pwl cache: 5210112 ns", This latency is beyond my > expectations and I believe it is unlikely to occur. In theory, this value > should be around a few hundred microseconds. But

[ceph-users] Re: RBD with PWL cache shows poor performance compared to cache device

2023-07-04 Thread Matthew Booth
On Mon, 3 Jul 2023 at 18:33, Ilya Dryomov wrote: > > On Mon, Jul 3, 2023 at 6:58 PM Mark Nelson wrote: > > > > > > On 7/3/23 04:53, Matthew Booth wrote: > > > On Thu, 29 Jun 2023 at 14:11, Mark Nelson wrote: > > >>>>> This contain

[ceph-users] Re: RBD with PWL cache shows poor performance compared to cache device

2023-07-03 Thread Matthew Booth
On Fri, 30 Jun 2023 at 08:50, Yin, Congmin wrote: > > Hi Matthew, > > Due to the latency of rbd layers, the write latency of the pwl cache is more > than ten times that of the Raw device. > I replied directly below the 2 questions. > > Best regards. > Congmin Yin

[ceph-users] Re: RBD with PWL cache shows poor performance compared to cache device

2023-07-03 Thread Matthew Booth
ntime=60 --time_based=1 > >>> > >>> And extracts sync.lat_ns.percentile["99.00"] > >> > >> Matthew, do you have the rest of the fio output captured? It would be > >> interesting to see if it's just the 99th percentile that is bad or the P

[ceph-users] Re: RBD with PWL cache shows poor performance compared to cache device

2023-06-29 Thread Matthew Booth
ing with fio. Specifically I am running a containerised test, >> executed with: >>podman run --volume .:/var/lib/etcd:Z quay.io/openshift-scale/etcd-perf >> >> This container runs: >>fio --rw=write --ioengine=sync --fdatasync=1 >> --directory=/var/lib/et

[ceph-users] Re: RBD with PWL cache shows poor performance compared to cache device

2023-06-27 Thread Matthew Booth
On Tue, 27 Jun 2023 at 18:20, Josh Baergen wrote: > > Hi Matthew, > > We've done a limited amount of work on characterizing the pwl and I think it > suffers the classic problem of some writeback caches in that, once the cache > is saturated, it's actually worse than just bei

[ceph-users] RBD with PWL cache shows poor performance compared to cache device

2023-06-27 Thread Matthew Booth
d: 180 MiB cached: 135 MiB dirty: 0 B free: 844 MiB hits_full: 1 / 0% hits_partial: 3 / 0% misses: 21952 hit_bytes: 6 KiB / 0% miss_bytes: 349 MiB -- Matthew Booth ___ ceph-users mailing li

[ceph-users] Re: Bucket sync policy

2023-04-24 Thread Matthew Darwin
I have basically given up relying on bucket sync to work properly in quincy.  I have been running a cron job to manually sync files between datacentres to catch the files that don't get replicated.  It's pretty inefficient, but at least all the files get to the backup datacentre. Would love

[ceph-users] Re: Do not use SSDs with (small) SLC cache

2023-02-28 Thread Matthew Stroud
A bit late to the game, but I'm not sure if it is your drives. I had a very similar issue to yours on enterprise drives (not that means much outside of support). What I was seeing is that a rebuild would kick off, PGs would instantly start to become laggy and then our clients (openstack rbd)

[ceph-users] Re: Debian update to 16.2.11-1~bpo11+1 failing

2023-01-31 Thread Matthew Booth
d an email to ceph-users-le...@ceph.io > -- Matthew Booth ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] PSA: Potential problems in a recent kernel?

2023-01-27 Thread Matthew Booth
7.4-2.fc37 selinux-policy 37.16-1.fc37 -> 37.17-1.fc37 selinux-policy-targeted 37.16-1.fc37 -> 37.17-1.fc37 tpm2-tss 3.2.0-3.fc37 -> 3.2.1-1.fc37 Removed: cracklib-dicts-2.9.7-30.fc37.x86_64 -- Matthew Booth ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Building Ceph containers

2023-01-16 Thread Matthew Vernon
... Thanks, Matthew [0] https://docs.ceph.com/en/quincy/install/build-ceph/ ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Mysterious HDD-Space Eating Issue

2023-01-16 Thread matthew
Hi Guys, I've got a funny one I'm hoping someone can point me in the right direction with: We've got three identical(?) Ceph nodes running 4 OSDs, Mon, Man, and iSCSI G/W each (we're only a small shop) on Rocky Linux 8 / Ceph Quincy. Everything is running fine, no bottle-necks (as far as we

[ceph-users] Laggy PGs on a fairly high performance cluster

2023-01-12 Thread Matthew Stroud
We have a 14 osd node all ssd cluster and for some reason we are continually getting laggy PGs and those seem to correlate to slow requests on Quincy (doesn't seem to happen on our Pacific clusters). These laggy pgs seem to shift between osds. The network seems solid, as in I'm not seeing

[ceph-users] Re: S3 Deletes in Multisite Sometimes Not Syncing

2022-12-23 Thread Matthew Darwin
Hi Alex, We also have a multi-site setup (17.2.5). I just deleted a bunch of files from one side and some files got deleted on the other side but not others. I waited 10 hours to see if the files would delete. I didn't do an exhaustive test like yours, but seems similar issues. In our case,

[ceph-users] Re: Multi site alternative

2022-11-23 Thread Matthew Leonard (BLOOMBERG/ 120 PARK)
Hey Ivan, I think the answer would be multisite. I know there is a lot of effort currently to work out the last few kinks. This tracker might be of interest as it sounds like an already identified issue, https://tracker.ceph.com/issues/57562#change-228263 Matt From: istvan.sz...@agoda.com

[ceph-users] Re: strange OSD status when rebooting one server

2022-10-14 Thread Matthew Darwin
From: Matthew Darwin Sent: 14 October 2022 18:57:37 To:c...@elchaka.de;ceph-users@ceph.io Subject: [ceph-users] Re: strange OSD status when rebooting one server https://gist.githubusercontent.com/matthewdarwin/aec3c2b16ba5e74beb4af1d49e8cfb1a/raw

[ceph-users] Re: strange OSD status when rebooting one server

2022-10-14 Thread Matthew Darwin
be an hint... Hth Am 14. Oktober 2022 18:45:40 MESZ schrieb Matthew Darwin : Hi, I am hoping someone can help explain this strange message.  I took 1 physical server offline which contains 11 OSDs.  "ceph -s" reports 11 osd down.  Great. But on the next line it say

[ceph-users] strange OSD status when rebooting one server

2022-10-14 Thread Matthew Darwin
Hi, I am hoping someone can help explain this strange message.  I took 1 physical server offline which contains 11 OSDs.  "ceph -s" reports 11 osd down.  Great. But on the next line it says "4 hosts" are impacted.  It should only be 1 single host?  When I look the manager dashboard all the

[ceph-users] Re: Ceph iSCSI rbd-target.api Failed to Load

2022-09-09 Thread Matthew J Black
Hi Li, Yeah, that's what I thought (about having the api_secure), so I checked for the iscsi-gateway.cfg file and there's only one on the system, in the /etc/ceph/ folder. Any other ideas? Cheers PEREGRINE IT Signature On 09/09/2022 18:35, Xiubo Li wrote: On 07/09/2022 17:37, duluxoz

[ceph-users] ceph -s command hangs with an authentication timeout - a reply

2022-08-08 Thread Matthew J Black
Hi Eneko, Sorry for the round-about way of getting back to you (I can't seem to work out how to reply/post to my original message - I'm obviously tired/stupid/whatever  :-) Problem solved (about 15 minutes ago) - turns out I had a typo (one of those small, hard to spot ones) - so a PBCAK

[ceph-users] Re: multi-site replication not syncing metadata

2022-07-04 Thread Matthew Darwin
I did manage to get this working. Not sure what exactly fixed it, but creating the pool "default.rgw.otp" helped.  Why are missing pools not automatically created? Also this: radosgw-admin sync status radosgw-admin metadata sync run On 2022-06-20 19:26, Matthew Darwin wrot

[ceph-users] Re: How to remove TELEMETRY_CHANGED( Telemetry requires re-opt-in) message

2022-06-24 Thread Matthew Darwin
Not sure.  Long enough to try the command and write this email, so at least 10 minutes. I expected it to disappear after 30 seconds or so. On 2022-06-24 10:34, Laura Flores wrote: Hi Matthew, About how long did the warning stay up after you ran the `ceph telemetry on` command? - Laura

[ceph-users] Re: How to remove TELEMETRY_CHANGED( Telemetry requires re-opt-in) message

2022-06-24 Thread Matthew Darwin
Thanks Yaarit, The cluster I was using is just a test cluster with a few OSD and almost no data. Not sure why I have to re-opt in upgrading from 17.2.0 to 17.2.1 On 2022-06-24 09:41, Yaarit Hatuka wrote: Hi Matthew, Thanks for your update. How big is the cluster? Thanks for opting

[ceph-users] Re: How to remove TELEMETRY_CHANGED( Telemetry requires re-opt-in) message

2022-06-23 Thread Matthew Darwin
Sorry. Eventually it goes away.  Just slower than I was expecting. On 2022-06-23 23:42, Matthew Darwin wrote: I just updated quincy from 17.2.0 to 17.2.1.  Ceph status reports "Telemetry requires re-opt-in". I then run $ ceph telemetry on $ ceph telemetry on --license sharing-

[ceph-users] How to remove TELEMETRY_CHANGED( Telemetry requires re-opt-in) message

2022-06-23 Thread Matthew Darwin
I just updated quincy from 17.2.0 to 17.2.1.  Ceph status reports "Telemetry requires re-opt-in". I then run $ ceph telemetry on $ ceph telemetry on --license sharing-1-0 Still the message "TELEMETRY_CHANGED( Telemetry requires re-opt-in) message" remains in the log. Any ideas how to get

[ceph-users] multi-site replication not syncing metadata

2022-06-20 Thread Matthew Darwin
Hi all, Running into some trouble. I just setup ceph multi-site replication.  Good news is that it is syncing the data. But the metadata is NOT syncing. I was trying to follow the instructions from here: https://docs.ceph.com/en/quincy/radosgw/multisite/#create-a-secondary-zone I see there

[ceph-users] Re: osd_disk_thread_ioprio_class deprecated?

2022-05-18 Thread Matthew H
See this PR https://github.com/ceph/ceph/pull/19973 From: Josh Baergen Sent: Wednesday, May 18, 2022 10:54 AM To: Richard Bade Cc: Ceph Users Subject: [ceph-users] Re: osd_disk_thread_ioprio_class deprecated? Hi Richard, > Could anyone confirm this? And

[ceph-users] Re: OS suggestion for further ceph installations (centos stream, rocky, ubuntu)?

2022-02-04 Thread Matthew Vernon
the Ubuntu Cloud Archive helpful if you want a more recent Ceph than the version your release shipped with; it can also help you decouple Ceph upgrades from OS upgrades. HTH, Matthew ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send

[ceph-users] Re: [RGW] bi_list(): (5) Input/output error blocking resharding

2022-01-10 Thread Matthew Vernon
Hi, On 07/01/2022 18:39, Gilles Mocellin wrote: Anyone who had that problem find a workaround ? Are you trying to reshard a bucket in a multisite setup? That isn't expected to work (and, IIRC, the changes to support doing so aren't going to make it into quincy). Regards, Matthew

[ceph-users] Re: switching ceph-ansible from /dev/sd to /dev/disk/by-path

2022-01-07 Thread Matthew Vernon
-ansible looks at what ceph-volume lvm list says when working out whether it needs to build new OSDs; I would hope it would correctly follow symlinks back to the correct point when working this out. I'd try it on a handy test cluster and see :) Regards, Matthew

[ceph-users] Re: Why you might want packages not containers for Ceph deployments

2021-11-17 Thread Matthew Vernon
was a real boon. Regards, Matthew ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Stretch cluster experiences in production?

2021-10-19 Thread Matthew Vernon
Hi, On 18/10/2021 23:34, Gregory Farnum wrote: On Fri, Oct 15, 2021 at 8:22 AM Matthew Vernon wrote: Also, if I'm using RGWs, will they do the right thing location-wise? i.e. DC A RGWs will talk to DC A OSDs wherever possible? Stretch clusters are entirely a feature of the RADOS layer

[ceph-users] Stretch cluster experiences in production?

2021-10-15 Thread Matthew Vernon
location-wise? i.e. DC A RGWs will talk to DC A OSDs wherever possible? Thanks, Matthew [0] https://docs.ceph.com/en/latest/rados/operations/stretch-mode/ ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le

[ceph-users] Re: OSD Service Advanced Specification db_slots

2021-09-10 Thread Matthew Vernon
. [not attempted this with cephadm, this was ceph-ansible] Regards, Matthew ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: [Ceph Upgrade] - Rollback Support during Upgrade failure

2021-09-08 Thread Matthew Vernon
tract data from OSDs (e.g. https://docs.ceph.com/en/latest/man/8/ceph-objectstore-tool/ ), you won't get complete objects this way. Instead, the advice would be to try and get enough mons back up to get your cluster at least to a read-only state and then attempt recovery that way. HTH

[ceph-users] Re: [Ceph Upgrade] - Rollback Support during Upgrade failure

2021-09-03 Thread Matthew Vernon
; I'd expect a cluster mid-upgrade to still be operational, so you should still be able to access your OSDs. Regards, Matthew ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: cephadm Pacific bootstrap hangs waiting for mon

2021-09-02 Thread Matthew Pounsett
On Thu, 2 Sept 2021 at 04:47, Sebastian Wagner wrote: > > by chance do you still have the logs of the mon the never went up? > > https://docs.ceph.com/en/latest/cephadm/troubleshooting/#checking-cephadm-logs > Not

[ceph-users] New Pacific deployment, "failed to find osd.# in keyring" errors

2021-09-02 Thread Matthew Pounsett
I'm trying to bring up a new cluster, just installed, and I'm getting errors while trying to deploy OSDs. Of the 85 candidates found, I've got 63 in and 0 up. All of the hosts were successfully added to the cluster using 'ceph orch host add ...' , but I'm seeing things in the logs like the

[ceph-users] Re: cephadm Pacific bootstrap hangs waiting for mon

2021-08-31 Thread Matthew Pounsett
On Tue, 31 Aug 2021 at 03:24, Arnaud MARTEL wrote: > > Hi Matthew, > > I dont' know if it will be helpful but I had the same problem using debian 10 > and the solution was to install docker from docker.io and not from the debian > package (too old). > Ah, that makes sens

[ceph-users] cephadm Pacific bootstrap hangs waiting for mon

2021-08-30 Thread Matthew Pounsett
I'm just getting started with Pacific, and I've run into this problem trying to get bootstrapped. cephadm is waiting for the mon to start, and waiting, and waiting ... checking docker ps it looks like it's running, but I guess it's never finishing its startup tasks? I waited about 30 minutes

[ceph-users] Re: Howto upgrade AND change distro

2021-08-27 Thread Matthew Vernon
of rebalancing you have to do when it rejoins the cluster post upgrade. Regards, Matthew [one good thing about Ubuntu's cloud archive is that e.g. you can get the same version that's default in 20.04 available as packages for 18.04 via UCA meaning you can upgrade Ceph first, and then do

[ceph-users] RGW Swift & multi-site

2021-08-16 Thread Matthew Vernon
Hi, Are there any issues to be aware of when using RGW's newer multi-site features with the Swift front-end? I've, perhaps unfairly, gathered the impression that the Swift support in RGW gets less love than S3... Thanks, Matthew ps: new email address, as I've moved employer

  1   2   >