The fix was actually backported to v18.2.3. The tracker was wrong. On Wed, May 29, 2024 at 3:26 PM <kvesli...@cs.hr> wrote: > > Hi, > > we have a stretched cluster (Reef 18.2.1) with 5 nodes (2 nodes on each side > + witness). You can se our daemon placement below. > > [admin] > ceph-admin01 labels="['_admin', 'mon', 'mgr']" > > [nodes] > [DC1] > ceph-node01 labels="['mon', 'mgr', 'mds', 'osd']" > ceph-node02 labels="['mon', 'rgw', 'mds', 'osd']" > [DC2] > ceph-node03 labels="['mon', 'mgr', 'mds', 'osd']" > ceph-node04 labels="['mon', 'rgw', 'mds', 'osd']" > > We have been testing CephFS HA and noticed when we have active MDS (we have > two active MDS daemons at all times) and active MGR (MGR is either on admin > node or in one of the DC's) in one DC and when we shutdown that site (DC) we > have a problem when one of the MDS metadata can't be retrieved thus showing > in logs as: > > "mgr finish mon failed to return metadata for mds" > > After we turn that site back on the problem persists and metadata of MDS in > question can't be retrieved with "ceph mds metadata" > > After I manually fail MDS daemon in question with "ceph mds fail" the problem > is solved and I can retrieve MDS metadata. > > My question is, would this be related to the following bug > (https://tracker.ceph.com/issues/63166) - I can see that it is showed as > backported to 18.2.1 but I can't find it in release notes for Reef. > > Second question is should this work in current configuration at all as MDS > and MGR are both at the same moment disconnected from the rest of the cluster? > > And final question would be what would be the solution here and is there any > loss of data when this happens? > > Any help is appreciated. > _______________________________________________ > ceph-users mailing list -- ceph-users@ceph.io > To unsubscribe send an email to ceph-users-le...@ceph.io >
-- Patrick Donnelly, Ph.D. He / Him / His Red Hat Partner Engineer IBM, Inc. GPG: 19F28A586F808C2402351B93C3301A3E258DD79D _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io