[ceph-users] Re: Large number of misplaced PGs but little backfill going on

2024-03-24 Thread Tyler Stachecki
On Sat, Mar 23, 2024, 4:26 AM Torkil Svensgaard wrote: > Hi > > ... Using mclock with high_recovery_ops profile. > > What is the bottleneck here? I would have expected a huge number of > simultaneous backfills. Backfill reservation logjam? > mClock is very buggy in my experience and frequently

[ceph-users] Re: ceph-crash NOT reporting crashes due to wrong permissions on /var/lib/ceph/crash/posted (Debian / Ubuntu packages)

2024-03-02 Thread Tyler Stachecki
> On 23.02.24 16:18, Christian Rohmann wrote: > > I just noticed issues with ceph-crash using the Debian /Ubuntu > > packages (package: ceph-base): > > > > While the /var/lib/ceph/crash/posted folder is created by the package > > install, > > it's not properly chowned to ceph:ceph by the postinst

[ceph-users] Re: Ceph 16.2.14: osd crash, bdev() _aio_thread got r=-1 ((1) Operation not permitted)

2023-12-05 Thread Tyler Stachecki
On Tue, Dec 5, 2023 at 10:13 AM Zakhar Kirpichenko wrote: > > Any input from anyone? > > /Z IIt's not clear whether or not these issues are related. I see three things in this e-mail chain: 1) bdev() _aio_thread with EPERM, as in the subject of this e-mail chain 2) bdev() _aio_thread with the

[ceph-users] Re: Many pgs inactive after node failure

2023-11-05 Thread Tyler Stachecki
On Sat, Nov 4, 2023, 6:44 AM Matthew Booth wrote: > I have a 3 node ceph cluster in my home lab. One of the pools spans 3 > hdds, one on each node, and has size 2, min size 1. One of my nodes is > currently down, and I have 160 pgs in 'unknown' state. The other 2 > hosts are up and the cluster

[ceph-users] Re: Problem with upgrade

2023-10-26 Thread Tyler Stachecki
h mgr is active and/or try stopping all but the Octopus mgr when stopping the mon as well? Cheers, Tyler > On Thu, Oct 26, 2023 at 4:57 PM Tyler Stachecki > wrote: > >> On Thu, Oct 26, 2023 at 6:52 PM Jorge Garcia >> wrote: >> > >> > Hi Tyler, >> >

[ceph-users] Re: Problem with upgrade

2023-10-26 Thread Tyler Stachecki
On Thu, Oct 26, 2023 at 6:52 PM Jorge Garcia wrote: > > Hi Tyler, > > Maybe you didn't read the full message, but in the message you will notice > that I'm doing exactly that, and the problem just occurred when I was doing > the upgrade from Octopus to Pacific. I'm nowhere near Quincy yet. The

[ceph-users] Re: Problem with upgrade

2023-10-26 Thread Tyler Stachecki
On Thu, Oct 26, 2023 at 6:36 PM Tyler Stachecki wrote: > > > On Thu, Oct 26, 2023, 6:16 PM Jorge Garcia wrote: > > > > from Centos 7 and Nautilus to > > Rocky 9 and Quincy. > > I hate to be the bearer of bad news here, but: > https://docs.ceph.com/en/latest

[ceph-users] Re: Problem with upgrade

2023-10-26 Thread Tyler Stachecki
On Thu, Oct 26, 2023, 6:16 PM Jorge Garcia wrote: > > from Centos 7 and Nautilus to > Rocky 9 and Quincy. I hate to be the bearer of bad news here, but: https://docs.ceph.com/en/latest/releases/quincy/#upgrading-from-pre-octopus-releases-like-nautilus "You *must* first upgrade to Octopus

[ceph-users] Re: Ceph 16.2.14: OSDs randomly crash in bstore_kv_sync

2023-10-20 Thread Tyler Stachecki
ar between some of the operators and developers. Wallaby is still marked EM upstream, but even so did not get this patch. The story is, unfortunately, the same here: the only way out of some of these holes is to upgrade... Regards, Tyler > > On Fri, 20 Oct 2023 at 15:39, Tyler Stachecki > wr

[ceph-users] Re: Ceph 16.2.14: OSDs randomly crash in bstore_kv_sync

2023-10-20 Thread Tyler Stachecki
On Fri, Oct 20, 2023, 8:11 AM Zakhar Kirpichenko wrote: > Thank you, Igor. > > It is somewhat disappointing that fixing this bug in Pacific has such a low > priority, considering its impact on existing clusters. > Unfortunately, the hard truth here is that Pacific (stable) was released over 30

[ceph-users] Re: Nautilus - Octopus upgrade - more questions

2023-10-17 Thread Tyler Stachecki
On Tue, Oct 17, 2023, 8:19 PM Dave Hall wrote: > Hello, > > I have a Nautilus cluster built using Ceph packages from Debian 10 > Backports, deployed with Ceph-Ansible. > > I see that Debian does not offer Ceph 15/Octopus packages. However, > download.ceph.com does offer such packages. > >

[ceph-users] Re: Time to Upgrade from Nautilus

2023-10-14 Thread Tyler Stachecki
On Sat, Oct 14, 2023 at 5:52 PM Tyler Stachecki wrote: > > On Sat, Oct 14, 2023 at 5:14 PM Dave Hall wrote: > > > > Hello. > > > > It's been a while. For the past couple years I've had a cluster running > > Nautilus on Debian 10 using the Debian Ceph packages

[ceph-users] Re: Time to Upgrade from Nautilus

2023-10-14 Thread Tyler Stachecki
On Sat, Oct 14, 2023 at 5:14 PM Dave Hall wrote: > > Hello. > > It's been a while. For the past couple years I've had a cluster running > Nautilus on Debian 10 using the Debian Ceph packages, and deployed with > Ceph-Ansible. It's not a huge cluster - 10 OSD nodes with 80 x 12TB HDD > OSDs,

[ceph-users] Re: 16.2.14: [progress WARNING root] complete: ev {UUID} does not exist

2023-09-29 Thread Tyler Stachecki
ctions so that it can render progress bars ETAs in e.g. `ceph -s` output. > /Z > > On Fri, 29 Sept 2023 at 14:13, Tyler Stachecki > wrote: > >> On Fri, Sep 29, 2023, 5:55 AM Zakhar Kirpichenko >> wrote: >> >>> Thank you, Eugen. >>> >>> Indeed it l

[ceph-users] Re: 16.2.14: [progress WARNING root] complete: ev {UUID} does not exist

2023-09-29 Thread Tyler Stachecki
On Fri, Sep 29, 2023, 5:55 AM Zakhar Kirpichenko wrote: > Thank you, Eugen. > > Indeed it looks like the progress module had some stale events from the > time when we added new OSDs and set a specific number of PGs for pools, > while the autoscaler tried to scale them down. Somehow the

[ceph-users] Re: Separating Mons and OSDs in Ceph Cluster

2023-09-09 Thread Tyler Stachecki
On Sat, Sep 9, 2023 at 10:48 AM Anthony D'Atri wrote: > There was also at point an issue where clients wouldn’t get a runtime update > of new mons. There's also 8+ year old unresolved bugs like this in OpenStack Cinder that will bite you if the relocated mons have new IP addresses:

[ceph-users] Re: Decrepit ceph cluster performance

2023-08-13 Thread Tyler Stachecki
On Sun, Aug 13, 2023 at 10:43 PM Anthony D'Atri wrote: > Think you meant s/SSD/SAS|SATA/. Indeed, (SATA/SAS) SSD - thanks! > The OP implies that the cluster's performance *degraded* with the Quincy > upgrade.I wonder if there was a kernel change at the same time. Also a good point. OP: do you

[ceph-users] Re: Decrepit ceph cluster performance

2023-08-13 Thread Tyler Stachecki
On Sun, Aug 13, 2023 at 10:19 PM J David wrote: > > I have a cluster that is performing *very* poorly. > > It has 12 physical servers and 72 400GB SATA Intel DC mixed-use SSD > OSDs (S3700's & S3710's). All servers have bonded 10GB NICs for 20Gbps > aggregate networking to a dedicated switch. At

[ceph-users] Re: [External Email] Re: Natuilus: Taking out OSDs that are 'Failure Pending' [EXT]

2023-08-04 Thread Tyler Stachecki
On Fri, Aug 4, 2023 at 11:33 AM Dave Hall wrote: > > Dave, > > Actually, my failure domain is OSD since I so far only have 9 OSD nodes but > EC 8+2. However, the drives are still functioning, except that one has > failed multiple times in the last few days, requiring a node power-cycle to >

[ceph-users] Regressed tail (p99.99+) write latency for RBD workloads in Quincy (vs. pre-Pacific)?

2023-07-24 Thread Tyler Stachecki
Hi all, Has anyone else noticed any p99.99+ tail latency regression for RBD workloads in Quincy vs. pre-Pacific, i.e., before the kv_onode cache existed? Some notes from what I have seen thus far: * Restarting OSDs temporarily resolves the problem... then as activity accrues over time, the

[ceph-users] Re: osd crash randomly

2022-10-23 Thread Tyler Stachecki
On Sun, Oct 23, 2022 at 11:04 PM can zhu wrote: > > crash info: > > { > "backtrace": [ > "/lib64/libpthread.so.0(+0x12ce0) [0x7f82e87cece0]", > "(BlueStore::Onode::put()+0x1a3) [0x55bd21a422e3]", > "(std::_Hashtable boost::intrusive_ptr >, >

[ceph-users] Re: laggy OSDs and staling krbd IO after upgrade from nautilus to octopus

2022-09-26 Thread Tyler Stachecki
Just a datapoint - we upgraded several large Mimic-born clusters straight to 15.2.12 with the quick fsck disabled in ceph.conf, then did require-osd-release, and finally did the omap conversion offline after the cluster was upgraded using the bluestore tool while the OSDs were down (all done in

[ceph-users] Re: 1 PG remains remapped after recovery

2022-08-27 Thread Tyler Stachecki
You seem to have an OSD that's down and out (status says 9 OSDs, 8 up and in). One possibility is that the pg is not able to fully recover because of existing CRUSH rules and the virtue that the only OSD that could store the last replica is down and out. So, what do your CRUSH rules and

[ceph-users] Re: Upgrade from Octopus to Pacific cannot get monitor to join

2022-07-27 Thread Tyler Stachecki
You're supposed to upgrade the mons first... https://docs.ceph.com/en/quincy/releases/pacific/#upgrading-non-cephadm-clusters Maybe try downgrading the mgrs back to Octopus? That's a bit of a scary situation. Tyler On Wed, Jul 27, 2022, 1:24 PM wrote: > Currently running Octopus 15.2.16,

[ceph-users] Re: active+undersized+degraded due to OSD size differences?

2022-06-19 Thread Tyler Stachecki
Your first assumption was correct. You can set the 'size' parameter of the pool to 2 (ceph osd pool set size 2), but you'll also want to either want to drop min_size to 1 or accept the fact that you cannot ever have either metadata OSD go down. It's fine for a toy cluster, but for any production

[ceph-users] Re: Maintenance mode?

2022-05-29 Thread Tyler Stachecki
Ceph always aims to provide high availability. So, if you do not set cluster flags that prevent Ceph from trying to self-heal, it will self-heal. Based on your description, it sounds like you want to consider the 'noout' flag. By default, after 10(?) minutes of an OSD being down, Ceph will begin

[ceph-users] Re: HELP! Upgrading monitors from 14.2.22 to 16.2.7 immediately crashes in FSMap::decode()

2022-03-20 Thread Tyler Stachecki
What does 'ceph mon dump | grep min_mon_release' say? You're running msgrv2 and all Ceph daemons are talking on v2, since you're on Nautilus, right? Was the cluster conceived on Nautilus, or something earlier? Tyler On Sun, Mar 20, 2022 at 10:30 PM Clippinger, Sam wrote: > > Hello! > > I need

[ceph-users] Re: advise to Ceph upgrade from mimic to ***

2022-01-31 Thread Tyler Stachecki
You can upgrade straight from Mimic to Octopus without a full outage window. One thing to keep in mind that wasn't spelled out clearly in the docs (I thought): you will have to bounce all components twice to get the full benefits of MSGRv2. After you bounce everything in the correct order and run

[ceph-users] Re: What exactly does the number of monitors depends on

2022-01-27 Thread Tyler Stachecki
They maintain and distribute the OSD map, i.e. the state of the OSDs. Generally 3 or 5 recommended as there always needs to be a quorum of mons available or the cluster can stall. With 3 mons, you can bring down one for servicing, patching, etc. but if another mon fails while you intentionally

[ceph-users] Re: switch restart facilitating cluster/client network.

2022-01-25 Thread Tyler Stachecki
I would still set noout on relevant parts of the cluster in case something goes south and it does take longer than 2 minutes. Otherwise OSDs will start outing themselves after 10 minutes or so by default and then you have a lot of churn going on. The monitors monitors will be fine unless you lose

[ceph-users] qd=1 bs=4k tuning on a toy cluster

2022-01-16 Thread Tyler Stachecki
I'm curious to hear if anyone has looked into kernel scheduling tweaks or changes in order to improve qd=1/bs=4k performance (while we patiently wait for Seastar!) Using this toy cluster: 3x OSD nodes: Atom C3758 (8 core, 2.2GHz) with 1x Intel S4500 SSD each Debian Bullseye with Linux 5.15.14 and