Hello again,
we were able to start our MDS-server again. We performed the following
steps:
ceph fs fail fdi-cephfs
cephfs-journal-tool --rank=cephfs:0 journal export backup.bin
cephfs-journal-tool --rank=cephfs:0 journal inspect
cephfs-journal-tool --rank=cephfs:all event get list
cephfs-j
someone can explain to us what these counters mean in detail. The
perf schema is not very revealing.
Our idea is to add a standbye-replay (hot-standbye mds) temporary, to
ensure the journal is replayable before we resume the upgrade.
I would be grateful for any advise.
best regards
Henning
On 2
In addition, i would like to mention that the number of "strays_created"
also increases after this action, but the number of num_strays is lower
now. If desired, we can provide debug logs from mds at the time the mds
was in stopping state and we did a systemctl restart mds1.
The only active md
Hi all,
we did a major update from Pacific to Quincy (17.2.5) a month ago
without any problems.
Now we have tried a minor update from 17.2.5 to 17.2.6 (ceph orch
upgrade). It stucks at mds upgrade phase. At this point the cluster
tries to scale down mds (ceph fs set max_mds 1). We waited a f