The osds are continuously flapping up/down due to the slow MDS metadata IOs
..
what is causing the slow MDS metadata IOs ?
currently, there are 2 mds and 3 monitors deployed ..
would it help to just one mds and one monitor ?

thanks!

On Tue, Oct 5, 2021 at 1:42 PM Eugen Block <ebl...@nde.ag> wrote:

> All your PGs are inactive, if two of four OSDs are down and you
> probably have a pool size of 3 then no IO can be served. You’d need at
> least three up ODSs to resolve that.
>
>
> Zitat von Abdelillah Asraoui <aasra...@gmail.com>:
>
> > Ceph is reporting warning on slow metdataIOs on one of the MDS server,
> > this is
> >
> > a new cluster with no upgrade..
> >
> > Anyone has encountered this and is there a workaround ..
> >
> > ceph -s
> >
> >   cluster:
> >
> >     id:     801691e6xx-x-xx-xx-xx
> >
> >     health: HEALTH_WARN
> >
> >             1 MDSs report slow metadata IOs
> >
> >             noscrub,nodeep-scrub flag(s) set
> >
> >             2 osds down
> >
> >             2 hosts (2 osds) down
> >
> >             Reduced data availability: 97 pgs inactive, 66 pgs peering,
> 53
> > pgs stale
> >
> >             Degraded data redundancy: 31 pgs undersized
> >
> >             2 slow ops, oldest one blocked for 30 sec, osd.0 has slow ops
> >
> >
> >
> >   services:
> >
> >     mon: 3 daemons, quorum a,c,f (age 15h)
> >
> >     mgr: a(active, since 17h)
> >
> >     mds: myfs:1 {0=myfs-a=up:creating} 1 up:standby
> >
> >     osd: 4 osds: 2 up (since 36s), 4 in (since 10h)
> >
> >          flags noscrub,nodeep-scrub
> >
> >
> >
> >   data:
> >
> >     pools:   4 pools, 97 pgs
> >
> >     objects: 0 objects, 0 B
> >
> >     usage:   1.0 GiB used, 1.8 TiB / 1.8 TiB avail
> >
> >     pgs:     100.000% pgs not active
> >
> >              44 creating+peering
> >
> >              31 stale+undersized+peered
> >
> >              22 stale+creating+peering
> >
> >
> >
> >   progress:
> >
> >     Rebalancing after osd.2 marked in (10h)
> >
> >       [............................]
> >
> >     Rebalancing after osd.3 marked in (10h)
> >
> >       [............................]
> >
> >
> > Thanks!
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@ceph.io
> > To unsubscribe send an email to ceph-users-le...@ceph.io
>
>
>
>
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to