[ceph-users] Re: Unable to start MDS and access CephFS after upgrade to 17.2.6

2023-06-15 Thread Henning Achterrath
Hello again, we were able to start our MDS-server again. We performed the following steps: ceph fs fail fdi-cephfs cephfs-journal-tool --rank=cephfs:0 journal export backup.bin cephfs-journal-tool --rank=cephfs:0 journal inspect cephfs-journal-tool --rank=cephfs:all event get list cephfs-j

[ceph-users] Re: MDS Upgrade from 17.2.5 to 17.2.6 not possible

2023-05-24 Thread Henning Achterrath
someone can explain to us what these counters mean in detail. The perf schema is not very revealing. Our idea is to add a standbye-replay (hot-standbye mds) temporary, to ensure the journal is replayable before we resume the upgrade. I would be grateful for any advise. best regards Henning On 2

[ceph-users] Re: MDS Upgrade from 17.2.5 to 17.2.6 not possible

2023-05-23 Thread Henning Achterrath
In addition, i would like to mention that the number of "strays_created" also increases after this action, but the number of num_strays is lower now. If desired, we can provide debug logs from mds at the time the mds was in stopping state and we did a systemctl restart mds1. The only active md

[ceph-users] MDS Upgrade from 17.2.5 to 17.2.6 not possible

2023-05-17 Thread Henning Achterrath
Hi all, we did a major update from Pacific to Quincy (17.2.5) a month ago without any problems. Now we have tried a minor update from 17.2.5 to 17.2.6 (ceph orch upgrade). It stucks at mds upgrade phase. At this point the cluster tries to scale down mds (ceph fs set max_mds 1). We waited a f