[ceph-users] Re: mds stuck in standby, not one active
On Tue, Dec 13, 2022 at 2:02 PM Mevludin Blazevic wrote: > > Hi all, > > in Ceph Pacific 6.2.5, the MDS failover function does not working. The > one host with the active MDS hat to be rebooted and after that, the > standby deamons did not jump in. The fs was not accessible, instead all > mds remain until now to standby. Also the cluster remains in Ceph Error > due to inactive mds so I did the following: > > ceph fs set cephfs false > ceph fs set cephfs max_mds 2 > > We also tried to restart the mds by the given yaml file, nothing works. > > The Ceph FS pool is green and clean. Please share: ceph status ceph fs dump -- Patrick Donnelly, Ph.D. He / Him / His Principal Software Engineer Red Hat, Inc. GPG: 19F28A586F808C2402351B93C3301A3E258DD79D ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io
[ceph-users] Re: mds stuck in standby, not one active
Hi, thanks for the quick response! CEPH STATUS: cluster: id: 8c774934-1535-11ec-973e-525400130e4f health: HEALTH_ERR 7 failed cephadm daemon(s) There are daemons running an older version of ceph 1 filesystem is degraded 1 filesystem has a failed mds daemon 1 filesystem is offline 1 filesystem is online with fewer MDS than max_mds 23 daemons have recently crashed services: mon: 2 daemons, quorum cephadm-vm,store2 (age 12d) mgr: store1.uevcpd(active, since 34m), standbys: cephadm-vm.zwagng mds: 0/1 daemons up (1 failed), 4 standby osd: 324 osds: 318 up (since 3h), 318 in (since 2h) data: volumes: 0/1 healthy, 1 failed pools: 6 pools, 257 pgs objects: 2.61M objects, 9.8 TiB usage: 29 TiB used, 2.0 PiB / 2.0 PiB avail pgs: 257 active+clean io: client: 0 B/s rd, 2.8 MiB/s wr, 435 op/s rd, 496 op/s wr FS DUMP: e60 enable_multiple, ever_enabled_multiple: 1,1 default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2} legacy client fscid: 1 Filesystem 'ceph_fs' (1) fs_name ceph_fs epoch 58 flags 32 created 2022-11-28T12:05:17.203346+ modified 2022-12-13T19:03:46.707236+ tableserver 0 root 0 session_timeout 60 session_autoclose 300 max_file_size 1099511627776 required_client_features {} last_failure 0 last_failure_osd_epoch 196035 compat compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2} max_mds 2 in 0 up {} failed 0 damaged stopped data_pools [4] metadata_pool 5 inline_data disabled balancer standby_count_wanted 1 Standby daemons: [mds.ceph_fs.store5.gnlqqm{-1:152180029} state up:standby seq 1 join_fscid=1 addr [v2:192.168.50.135:6800/3548272808,v1:192.168.50.135:6801/3548272808] compat {c=[1],r=[1],i=[1]}] [mds.ceph_fs.store6.fxgvoj{:915af89} state up:standby seq 1 join_fscid=1 addr [v2:192.168.50.136:1b70/4fde2aa0,v1:192.168.50.136:1b71/4fde2aa0] compat {c=[1],r=[1],i=[1]}] [mds.ceph_fs.store4.mhvpot{:916a09d} state up:standby seq 1 join_fscid=1 addr [v2:192.168.50.134:1a90/b8b1f33c,v1:192.168.50.134:1a91/b8b1f33c] compat {c=[1],r=[1],i=[1]}] [mds.ceph_fs.store3.vcnwzh{:916aff7} state up:standby seq 1 join_fscid=1 addr [v2:192.168.50.133:1a90/49cb4e4,v1:192.168.50.133:1a91/49cb4e4] compat {c=[1],r=[1],i=[1]}] dumped fsmap epoch 60 Am 13.12.2022 um 20:11 schrieb Patrick Donnelly: On Tue, Dec 13, 2022 at 2:02 PM Mevludin Blazevic wrote: Hi all, in Ceph Pacific 6.2.5, the MDS failover function does not working. The one host with the active MDS hat to be rebooted and after that, the standby deamons did not jump in. The fs was not accessible, instead all mds remain until now to standby. Also the cluster remains in Ceph Error due to inactive mds so I did the following: ceph fs set cephfs false ceph fs set cephfs max_mds 2 We also tried to restart the mds by the given yaml file, nothing works. The Ceph FS pool is green and clean. Please share: ceph status ceph fs dump ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io
[ceph-users] Re: mds stuck in standby, not one active
On Tue, Dec 13, 2022 at 2:21 PM Mevludin Blazevic wrote: > > Hi, > > thanks for the quick response! > > CEPH STATUS: > > cluster: > id: 8c774934-1535-11ec-973e-525400130e4f > health: HEALTH_ERR > 7 failed cephadm daemon(s) > There are daemons running an older version of ceph > 1 filesystem is degraded > 1 filesystem has a failed mds daemon > 1 filesystem is offline > 1 filesystem is online with fewer MDS than max_mds > 23 daemons have recently crashed > >services: > mon: 2 daemons, quorum cephadm-vm,store2 (age 12d) > mgr: store1.uevcpd(active, since 34m), standbys: cephadm-vm.zwagng > mds: 0/1 daemons up (1 failed), 4 standby > osd: 324 osds: 318 up (since 3h), 318 in (since 2h) > >data: > volumes: 0/1 healthy, 1 failed > pools: 6 pools, 257 pgs > objects: 2.61M objects, 9.8 TiB > usage: 29 TiB used, 2.0 PiB / 2.0 PiB avail > pgs: 257 active+clean > >io: > client: 0 B/s rd, 2.8 MiB/s wr, 435 op/s rd, 496 op/s wr > > FS DUMP: > > e60 > enable_multiple, ever_enabled_multiple: 1,1 > default compat: compat={},rocompat={},incompat={1=base v0.20,2=client > writeable ranges,3=default file layouts on dirs,4=dir inode in separate > object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no > anchor table,9=file layout v2,10=snaprealm v2} > legacy client fscid: 1 > > Filesystem 'ceph_fs' (1) > fs_name ceph_fs > epoch 58 > flags 32 > created 2022-11-28T12:05:17.203346+ > modified2022-12-13T19:03:46.707236+ > tableserver 0 > root0 > session_timeout 60 > session_autoclose 300 > max_file_size 1099511627776 > required_client_features{} > last_failure0 > last_failure_osd_epoch 196035 > compat compat={},rocompat={},incompat={1=base v0.20,2=client writeable > ranges,3=default file layouts on dirs,4=dir inode in separate > object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no > anchor table,9=file layout v2,10=snaprealm v2} > max_mds 2 > in 0 > up {} > failed 0 > damaged > stopped > data_pools [4] > metadata_pool 5 > inline_data disabled > balancer > standby_count_wanted1 > > > Standby daemons: > > [mds.ceph_fs.store5.gnlqqm{-1:152180029} state up:standby seq 1 > join_fscid=1 addr > [v2:192.168.50.135:6800/3548272808,v1:192.168.50.135:6801/3548272808] > compat {c=[1],r=[1],i=[1]}] > [mds.ceph_fs.store6.fxgvoj{:915af89} state up:standby seq 1 > join_fscid=1 addr > [v2:192.168.50.136:1b70/4fde2aa0,v1:192.168.50.136:1b71/4fde2aa0] compat > {c=[1],r=[1],i=[1]}] > [mds.ceph_fs.store4.mhvpot{:916a09d} state up:standby seq 1 > join_fscid=1 addr > [v2:192.168.50.134:1a90/b8b1f33c,v1:192.168.50.134:1a91/b8b1f33c] compat > {c=[1],r=[1],i=[1]}] > [mds.ceph_fs.store3.vcnwzh{:916aff7} state up:standby seq 1 > join_fscid=1 addr > [v2:192.168.50.133:1a90/49cb4e4,v1:192.168.50.133:1a91/49cb4e4] compat > {c=[1],r=[1],i=[1]}] > dumped fsmap epoch 60 You're encountering a bug fixed in v16.2.7. Please upgrade to the latest version. -- Patrick Donnelly, Ph.D. He / Him / His Principal Software Engineer Red Hat, Inc. GPG: 19F28A586F808C2402351B93C3301A3E258DD79D ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io
[ceph-users] Re: mds stuck in standby, not one active
Hi, while upgrading to ceph pacific 6.2.7, the upgrade process stuck exactly at the mds daemons. Before, I have tried to increase/shrink the placement size of them, but nothing happens. Currently I have 4/3 running daemons. One daemon should be stopped and removed. Do you suggest to force remove these daemons or what could be the preferred workaround? Regards, Mevludin Am 13.12.2022 um 20:32 schrieb Patrick Donnelly: On Tue, Dec 13, 2022 at 2:21 PM Mevludin Blazevic wrote: Hi, thanks for the quick response! CEPH STATUS: cluster: id: 8c774934-1535-11ec-973e-525400130e4f health: HEALTH_ERR 7 failed cephadm daemon(s) There are daemons running an older version of ceph 1 filesystem is degraded 1 filesystem has a failed mds daemon 1 filesystem is offline 1 filesystem is online with fewer MDS than max_mds 23 daemons have recently crashed services: mon: 2 daemons, quorum cephadm-vm,store2 (age 12d) mgr: store1.uevcpd(active, since 34m), standbys: cephadm-vm.zwagng mds: 0/1 daemons up (1 failed), 4 standby osd: 324 osds: 318 up (since 3h), 318 in (since 2h) data: volumes: 0/1 healthy, 1 failed pools: 6 pools, 257 pgs objects: 2.61M objects, 9.8 TiB usage: 29 TiB used, 2.0 PiB / 2.0 PiB avail pgs: 257 active+clean io: client: 0 B/s rd, 2.8 MiB/s wr, 435 op/s rd, 496 op/s wr FS DUMP: e60 enable_multiple, ever_enabled_multiple: 1,1 default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2} legacy client fscid: 1 Filesystem 'ceph_fs' (1) fs_name ceph_fs epoch 58 flags 32 created 2022-11-28T12:05:17.203346+ modified2022-12-13T19:03:46.707236+ tableserver 0 root0 session_timeout 60 session_autoclose 300 max_file_size 1099511627776 required_client_features{} last_failure0 last_failure_osd_epoch 196035 compat compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2} max_mds 2 in 0 up {} failed 0 damaged stopped data_pools [4] metadata_pool 5 inline_data disabled balancer standby_count_wanted1 Standby daemons: [mds.ceph_fs.store5.gnlqqm{-1:152180029} state up:standby seq 1 join_fscid=1 addr [v2:192.168.50.135:6800/3548272808,v1:192.168.50.135:6801/3548272808] compat {c=[1],r=[1],i=[1]}] [mds.ceph_fs.store6.fxgvoj{:915af89} state up:standby seq 1 join_fscid=1 addr [v2:192.168.50.136:1b70/4fde2aa0,v1:192.168.50.136:1b71/4fde2aa0] compat {c=[1],r=[1],i=[1]}] [mds.ceph_fs.store4.mhvpot{:916a09d} state up:standby seq 1 join_fscid=1 addr [v2:192.168.50.134:1a90/b8b1f33c,v1:192.168.50.134:1a91/b8b1f33c] compat {c=[1],r=[1],i=[1]}] [mds.ceph_fs.store3.vcnwzh{:916aff7} state up:standby seq 1 join_fscid=1 addr [v2:192.168.50.133:1a90/49cb4e4,v1:192.168.50.133:1a91/49cb4e4] compat {c=[1],r=[1],i=[1]}] dumped fsmap epoch 60 You're encountering a bug fixed in v16.2.7. Please upgrade to the latest version. -- Mevludin Blazevic, M.Sc. University of Koblenz-Landau Computing Centre (GHRKO) Universitaetsstrasse 1 D-56070 Koblenz, Germany Room A023 Tel: +49 261/287-1326 ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io
[ceph-users] Re: mds stuck in standby, not one active
On Thu, Dec 15, 2022 at 7:24 AM Mevludin Blazevic wrote: > > Hi, > > while upgrading to ceph pacific 6.2.7, the upgrade process stuck exactly > at the mds daemons. Before, I have tried to increase/shrink the > placement size of them, but nothing happens. Currently I have 4/3 > running daemons. One daemon should be stopped and removed. > > Do you suggest to force remove these daemons or what could be the > preferred workaround? Hard to say without more information. Please share: ceph fs dump ceph status ceph health detail -- Patrick Donnelly, Ph.D. He / Him / His Principal Software Engineer Red Hat, Inc. GPG: 19F28A586F808C2402351B93C3301A3E258DD79D ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io
[ceph-users] Re: mds stuck in standby, not one active
Ceph fs dump: e62 enable_multiple, ever_enabled_multiple: 1,1 default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2}legacy client fscid: 1 Filesystem 'ceph_fs' (1) fs_name ceph_fs epoch 62 flags 12 created 2022-11-28T12:05:17.203346+ modified 2022-12-15T12:09:14.091724+ tableserver 0 root 0 session_timeout 60 session_autoclose 300 max_file_size 1099511627776 required_client_features {} last_failure 0 last_failure_osd_epoch 196035 compat compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2} max_mds 1 in 0 up {} failed 0 damaged stopped data_pools [4] metadata_pool 5 inline_data disabled balancer standby_count_wanted 1 f Standby daemons: [mds.ceph_fs.store5.gnlqqm{-1:152180029} state up:standby seq 1 join_fscid=1 addr [v2:192.168.50.135:6800/3548272808,v1:192.168.50.135:6801/3548272808] compat {c=[1],r=[1],i=[1]}] [mds.ceph_fs.store6.fxgvoj{-1:152416137} state up:standby seq 1 join_fscid=1 addr [v2:192.168.50.136:7024/1339959968,v1:192.168.50.136:7025/1339959968] compat {c=[1],r=[1],i=[1]}] [mds.ceph_fs.store4.mhvpot{-1:152477853} state up:standby seq 1 join_fscid=1 addr [v2:192.168.50.134:6800/3098669884,v1:192.168.50.134:6801/3098669884] compat {c=[1],r=[1],i=[1]}] [mds.ceph_fs.store3.vcnwzh{-1:152481783} state up:standby seq 1 join_fscid=1 addr [v2:192.168.50.133:6800/77378788,v1:192.168.50.133:6801/77378788] compat {c=[1],r=[1],i=[1]}] dumped fsmap epoch 62 Ceph Status: cluster: id: 8c774934-1535-11ec-973e-525400130e4f health: HEALTH_ERR 1 filesystem is degraded 1 filesystem has a failed mds daemon 1 filesystem is offline 26 daemons have recently crashed services: mon: 2 daemons, quorum cephadm-vm,store2 (age 2d) mgr: store1.uevcpd(active, since 2d), standbys: cephadm-vm.zwagng mds: 0/1 daemons up (1 failed), 4 standby osd: 312 osds: 312 up (since 8h), 312 in (since 17h) data: volumes: 0/1 healthy, 1 failed pools: 7 pools, 289 pgs objects: 2.62M objects, 9.8 TiB usage: 29 TiB used, 1.9 PiB / 1.9 PiB avail pgs: 286 active+clean 3 active+clean+scrubbing+deep io: client: 945 KiB/s rd, 3.3 MiB/s wr, 516 op/s rd, 562 op/s wr Ceph Health detail: HEALTH_ERR 1 filesystem is degraded; 1 filesystem has a failed mds daemon; 1 filesystem is offline; 26 daemons have recently crashed [WRN] FS_DEGRADED: 1 filesystem is degraded fs ceph_fs is degraded [WRN] FS_WITH_FAILED_MDS: 1 filesystem has a failed mds daemon fs ceph_fs has 1 failed mds [ERR] MDS_ALL_DOWN: 1 filesystem is offline fs ceph_fs is offline because no MDS is active for it. [WRN] RECENT_CRASH: 26 daemons have recently crashed osd.323 crashed on host store7 at 2022-12-12T14:03:23.857874Z osd.323 crashed on host store7 at 2022-12-12T14:03:43.945625Z osd.323 crashed on host store7 at 2022-12-12T14:04:03.282797Z osd.323 crashed on host store7 at 2022-12-12T14:04:22.612037Z osd.323 crashed on host store7 at 2022-12-12T14:04:41.630473Z osd.323 crashed on host store7 at 2022-12-12T14:34:49.237008Z osd.323 crashed on host store7 at 2022-12-12T14:35:09.903922Z osd.323 crashed on host store7 at 2022-12-12T14:35:28.621955Z osd.323 crashed on host store7 at 2022-12-12T14:35:46.985517Z osd.323 crashed on host store7 at 2022-12-12T14:36:05.375758Z osd.323 crashed on host store7 at 2022-12-12T15:01:57.235785Z osd.323 crashed on host store7 at 2022-12-12T15:02:16.581335Z osd.323 crashed on host store7 at 2022-12-12T15:02:33.212653Z osd.323 crashed on host store7 at 2022-12-12T15:02:49.775560Z osd.323 crashed on host store7 at 2022-12-12T15:03:06.303861Z mgr.cephadm-vm.zwagng crashed on host cephadm-vm at 2022-12-13T13:21:41.149773Z mgr.cephadm-vm.zwagng crashed on host cephadm-vm at 2022-12-13T13:22:15.413105Z mgr.cephadm-vm.zwagng crashed on host cephadm-vm at 2022-12-13T13:23:39.888401Z mgr.cephadm-vm.zwagng crashed on host cephadm-vm at 2022-12-13T13:27:56.458529Z mgr.cephadm-vm.zwagng crashed on host cephadm-vm at 2022-12-13T13:31:03.791532Z mgr.cephadm-vm.zwagng crashed on host cephadm-vm at 2022-12-13T13:34:24.023106Z osd.98 crashed on host store3 at 2022-12-13T16:11:38.064735Z mgr.store1.uevcpd crashed on host store1 at 2022-12-13T18:39:33.091261Z osd.322 crashed on host store6 at 2022-12-14T06:06:14.193437Z osd.234 crashed on host store8 at 2022-12-15T02:32:13.009795Z osd.311 crashed on host store8 at 2022-12-15T02:32:18.40
[ceph-users] Re: mds stuck in standby, not one active
On Thu, Dec 15, 2022 at 3:17 PM Mevludin Blazevic wrote: > > Ceph fs dump: > > e62 > enable_multiple, ever_enabled_multiple: 1,1 > default compat: compat={},rocompat={},incompat={1=base v0.20,2=client > writeable ranges,3=default file layouts on dirs,4=dir inode in separate > object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no > anchor table,9=file layout v2,10=snaprealm v2}legacy client fscid: 1 > > Filesystem 'ceph_fs' (1) > fs_name ceph_fs > epoch 62 > flags 12 > created 2022-11-28T12:05:17.203346+ > modified2022-12-15T12:09:14.091724+ > tableserver 0 > root0 > session_timeout 60 > session_autoclose 300 > max_file_size 1099511627776 > required_client_features{} > last_failure0 > last_failure_osd_epoch 196035 > compat compat={},rocompat={},incompat={1=base v0.20,2=client writeable > ranges,3=default file layouts on dirs,4=dir inode in separate > object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no > anchor table,9=file layout v2,10=snaprealm v2} > max_mds 1 > in 0 > up {} > failed 0 > damaged > stopped > data_pools [4] > metadata_pool 5 > inline_data disabled > balancer > standby_count_wanted1 > > f > Standby daemons: > > [mds.ceph_fs.store5.gnlqqm{-1:152180029} state up:standby seq 1 > join_fscid=1 addr > [v2:192.168.50.135:6800/3548272808,v1:192.168.50.135:6801/3548272808] > compat {c=[1],r=[1],i=[1]}] > [mds.ceph_fs.store6.fxgvoj{-1:152416137} state up:standby seq 1 > join_fscid=1 addr > [v2:192.168.50.136:7024/1339959968,v1:192.168.50.136:7025/1339959968] > compat {c=[1],r=[1],i=[1]}] > [mds.ceph_fs.store4.mhvpot{-1:152477853} state up:standby seq 1 > join_fscid=1 addr > [v2:192.168.50.134:6800/3098669884,v1:192.168.50.134:6801/3098669884] > compat {c=[1],r=[1],i=[1]}] > [mds.ceph_fs.store3.vcnwzh{-1:152481783} state up:standby seq 1 > join_fscid=1 addr > [v2:192.168.50.133:6800/77378788,v1:192.168.50.133:6801/77378788] compat > {c=[1],r=[1],i=[1]}] > dumped fsmap epoch 62 > > Ceph Status: > >cluster: > id: 8c774934-1535-11ec-973e-525400130e4f > health: HEALTH_ERR > 1 filesystem is degraded > 1 filesystem has a failed mds daemon > 1 filesystem is offline > 26 daemons have recently crashed > >services: > mon: 2 daemons, quorum cephadm-vm,store2 (age 2d) > mgr: store1.uevcpd(active, since 2d), standbys: cephadm-vm.zwagng > mds: 0/1 daemons up (1 failed), 4 standby > osd: 312 osds: 312 up (since 8h), 312 in (since 17h) > >data: > volumes: 0/1 healthy, 1 failed > pools: 7 pools, 289 pgs > objects: 2.62M objects, 9.8 TiB > usage: 29 TiB used, 1.9 PiB / 1.9 PiB avail > pgs: 286 active+clean > 3 active+clean+scrubbing+deep > >io: > client: 945 KiB/s rd, 3.3 MiB/s wr, 516 op/s rd, 562 op/s wr > > Ceph Health detail: > > HEALTH_ERR 1 filesystem is degraded; 1 filesystem has a failed mds > daemon; 1 filesystem is offline; 26 daemons have recently crashed > [WRN] FS_DEGRADED: 1 filesystem is degraded > fs ceph_fs is degraded > [WRN] FS_WITH_FAILED_MDS: 1 filesystem has a failed mds daemon > fs ceph_fs has 1 failed mds > [ERR] MDS_ALL_DOWN: 1 filesystem is offline > fs ceph_fs is offline because no MDS is active for it. > [WRN] RECENT_CRASH: 26 daemons have recently crashed > osd.323 crashed on host store7 at 2022-12-12T14:03:23.857874Z > osd.323 crashed on host store7 at 2022-12-12T14:03:43.945625Z > osd.323 crashed on host store7 at 2022-12-12T14:04:03.282797Z > osd.323 crashed on host store7 at 2022-12-12T14:04:22.612037Z > osd.323 crashed on host store7 at 2022-12-12T14:04:41.630473Z > osd.323 crashed on host store7 at 2022-12-12T14:34:49.237008Z > osd.323 crashed on host store7 at 2022-12-12T14:35:09.903922Z > osd.323 crashed on host store7 at 2022-12-12T14:35:28.621955Z > osd.323 crashed on host store7 at 2022-12-12T14:35:46.985517Z > osd.323 crashed on host store7 at 2022-12-12T14:36:05.375758Z > osd.323 crashed on host store7 at 2022-12-12T15:01:57.235785Z > osd.323 crashed on host store7 at 2022-12-12T15:02:16.581335Z > osd.323 crashed on host store7 at 2022-12-12T15:02:33.212653Z > osd.323 crashed on host store7 at 2022-12-12T15:02:49.775560Z > osd.323 crashed on host store7 at 2022-12-12T15:03:06.303861Z > mgr.cephadm-vm.zwagng crashed on host cephadm-vm at > 2022-12-13T13:21:41.149773Z > mgr.cephadm-vm.zwagng crashed on host cephadm-vm at > 2022-12-13T13:22:15.413105Z > mgr.cephadm-vm.zwagng crashed on host cephadm-vm at > 2022-12-13T13:23:39.888401Z > mgr.cephadm-vm.zwagng crashed on host cephadm-vm at > 2022-12-13T13:27:56.458529Z > mgr.cephadm-vm.zwagng crashed on host cephadm-vm at > 2022-12-13T13:31:03.791532Z > mgr.cephadm-vm.zwagng crashed on host cephadm-vm at > 2022-12-13T13:34:24.023106Z > osd.98 crashe