The fix was actually backported to v18.2.3. The tracker was wrong.

On Wed, May 29, 2024 at 3:26 PM <kvesli...@cs.hr> wrote:
>
> Hi,
>
> we have a stretched cluster (Reef 18.2.1) with 5 nodes (2 nodes on each side 
> + witness). You can se our daemon placement below.
>
> [admin]
> ceph-admin01 labels="['_admin', 'mon', 'mgr']"
>
> [nodes]
> [DC1]
> ceph-node01 labels="['mon', 'mgr', 'mds', 'osd']"
> ceph-node02 labels="['mon', 'rgw', 'mds', 'osd']"
> [DC2]
> ceph-node03 labels="['mon', 'mgr', 'mds', 'osd']"
> ceph-node04 labels="['mon', 'rgw', 'mds', 'osd']"
>
> We have been testing CephFS HA and noticed when we have active MDS (we have 
> two active MDS daemons at all times) and active MGR (MGR is either on admin 
> node or in one of the DC's) in one DC and when we shutdown that site (DC) we 
> have a problem when one of the MDS metadata can't be retrieved thus showing 
> in logs as:
>
> "mgr finish mon failed to return metadata for mds"
>
> After we turn that site back on the problem persists and metadata of MDS in 
> question can't be retrieved with "ceph mds metadata"
>
> After I manually fail MDS daemon in question with "ceph mds fail" the problem 
> is solved and I can retrieve MDS metadata.
>
> My question is, would this be related to the following bug 
> (https://tracker.ceph.com/issues/63166) - I can see that it is showed as 
> backported to 18.2.1 but I can't find it in release notes for Reef.
>
> Second question is should this work in current configuration at all as MDS 
> and MGR are both at the same moment disconnected from the rest of the cluster?
>
> And final question would be what would be the solution here and is there any 
> loss of data when this happens?
>
> Any help is appreciated.
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>


-- 
Patrick Donnelly, Ph.D.
He / Him / His
Red Hat Partner Engineer
IBM, Inc.
GPG: 19F28A586F808C2402351B93C3301A3E258DD79D
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to