[ceph-users] Re: 1 MDS report slow metadata IOs

2021-10-27 Thread Abdelillah Asraoui
OSD flapping was due to ports blocked by firewall.
while mounting the file system, the directory structure shows csi volumes
subfolder folders
as in /tmp/cephFS/csi/csi-vol-/container-name-log
is there a way to not show the csi volumes in the path to the container log
as an example:
/tmp/cephFS/container-name-log
this would make it easier to update the containers configuration without
dealing the volumes sub folders in the path ..

Thanks!

On Wed, Oct 6, 2021 at 2:58 AM Eugen Block  wrote:

> > what is causing the slow MDS metadata IOs ?
>
> Your flapping OSDs.
>
> > currently, there are 2 mds and 3 monitors deployed ..
> > would it help to just one mds and one monitor ?
>
> No, you need to figure out why your OSDs crash. More details about
> your setup (ceph version, deployment method, hardware resources) and
> the logs from a crashing OSD could help identify the issue.
>
>
> Zitat von Abdelillah Asraoui :
>
> > The osds are continuously flapping up/down due to the slow MDS metadata
> IOs
> > ..
> > what is causing the slow MDS metadata IOs ?
> > currently, there are 2 mds and 3 monitors deployed ..
> > would it help to just one mds and one monitor ?
> >
> > thanks!
> >
> > On Tue, Oct 5, 2021 at 1:42 PM Eugen Block  wrote:
> >
> >> All your PGs are inactive, if two of four OSDs are down and you
> >> probably have a pool size of 3 then no IO can be served. You’d need at
> >> least three up ODSs to resolve that.
> >>
> >>
> >> Zitat von Abdelillah Asraoui :
> >>
> >> > Ceph is reporting warning on slow metdataIOs on one of the MDS server,
> >> > this is
> >> >
> >> > a new cluster with no upgrade..
> >> >
> >> > Anyone has encountered this and is there a workaround ..
> >> >
> >> > ceph -s
> >> >
> >> >   cluster:
> >> >
> >> > id: 801691e6xx-x-xx-xx-xx
> >> >
> >> > health: HEALTH_WARN
> >> >
> >> > 1 MDSs report slow metadata IOs
> >> >
> >> > noscrub,nodeep-scrub flag(s) set
> >> >
> >> > 2 osds down
> >> >
> >> > 2 hosts (2 osds) down
> >> >
> >> > Reduced data availability: 97 pgs inactive, 66 pgs
> peering,
> >> 53
> >> > pgs stale
> >> >
> >> > Degraded data redundancy: 31 pgs undersized
> >> >
> >> > 2 slow ops, oldest one blocked for 30 sec, osd.0 has slow
> ops
> >> >
> >> >
> >> >
> >> >   services:
> >> >
> >> > mon: 3 daemons, quorum a,c,f (age 15h)
> >> >
> >> > mgr: a(active, since 17h)
> >> >
> >> > mds: myfs:1 {0=myfs-a=up:creating} 1 up:standby
> >> >
> >> > osd: 4 osds: 2 up (since 36s), 4 in (since 10h)
> >> >
> >> >  flags noscrub,nodeep-scrub
> >> >
> >> >
> >> >
> >> >   data:
> >> >
> >> > pools:   4 pools, 97 pgs
> >> >
> >> > objects: 0 objects, 0 B
> >> >
> >> > usage:   1.0 GiB used, 1.8 TiB / 1.8 TiB avail
> >> >
> >> > pgs: 100.000% pgs not active
> >> >
> >> >  44 creating+peering
> >> >
> >> >  31 stale+undersized+peered
> >> >
> >> >  22 stale+creating+peering
> >> >
> >> >
> >> >
> >> >   progress:
> >> >
> >> > Rebalancing after osd.2 marked in (10h)
> >> >
> >> >   []
> >> >
> >> > Rebalancing after osd.3 marked in (10h)
> >> >
> >> >   []
> >> >
> >> >
> >> > Thanks!
> >> > ___
> >> > ceph-users mailing list -- ceph-users@ceph.io
> >> > To unsubscribe send an email to ceph-users-le...@ceph.io
> >>
> >>
> >>
> >>
>
>
>
>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: 1 MDS report slow metadata IOs

2021-10-06 Thread Eugen Block

what is causing the slow MDS metadata IOs ?


Your flapping OSDs.


currently, there are 2 mds and 3 monitors deployed ..
would it help to just one mds and one monitor ?


No, you need to figure out why your OSDs crash. More details about  
your setup (ceph version, deployment method, hardware resources) and  
the logs from a crashing OSD could help identify the issue.



Zitat von Abdelillah Asraoui :


The osds are continuously flapping up/down due to the slow MDS metadata IOs
..
what is causing the slow MDS metadata IOs ?
currently, there are 2 mds and 3 monitors deployed ..
would it help to just one mds and one monitor ?

thanks!

On Tue, Oct 5, 2021 at 1:42 PM Eugen Block  wrote:


All your PGs are inactive, if two of four OSDs are down and you
probably have a pool size of 3 then no IO can be served. You’d need at
least three up ODSs to resolve that.


Zitat von Abdelillah Asraoui :

> Ceph is reporting warning on slow metdataIOs on one of the MDS server,
> this is
>
> a new cluster with no upgrade..
>
> Anyone has encountered this and is there a workaround ..
>
> ceph -s
>
>   cluster:
>
> id: 801691e6xx-x-xx-xx-xx
>
> health: HEALTH_WARN
>
> 1 MDSs report slow metadata IOs
>
> noscrub,nodeep-scrub flag(s) set
>
> 2 osds down
>
> 2 hosts (2 osds) down
>
> Reduced data availability: 97 pgs inactive, 66 pgs peering,
53
> pgs stale
>
> Degraded data redundancy: 31 pgs undersized
>
> 2 slow ops, oldest one blocked for 30 sec, osd.0 has slow ops
>
>
>
>   services:
>
> mon: 3 daemons, quorum a,c,f (age 15h)
>
> mgr: a(active, since 17h)
>
> mds: myfs:1 {0=myfs-a=up:creating} 1 up:standby
>
> osd: 4 osds: 2 up (since 36s), 4 in (since 10h)
>
>  flags noscrub,nodeep-scrub
>
>
>
>   data:
>
> pools:   4 pools, 97 pgs
>
> objects: 0 objects, 0 B
>
> usage:   1.0 GiB used, 1.8 TiB / 1.8 TiB avail
>
> pgs: 100.000% pgs not active
>
>  44 creating+peering
>
>  31 stale+undersized+peered
>
>  22 stale+creating+peering
>
>
>
>   progress:
>
> Rebalancing after osd.2 marked in (10h)
>
>   []
>
> Rebalancing after osd.3 marked in (10h)
>
>   []
>
>
> Thanks!
> ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io








___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: 1 MDS report slow metadata IOs

2021-10-05 Thread Abdelillah Asraoui
The osds are continuously flapping up/down due to the slow MDS metadata IOs
..
what is causing the slow MDS metadata IOs ?
currently, there are 2 mds and 3 monitors deployed ..
would it help to just one mds and one monitor ?

thanks!

On Tue, Oct 5, 2021 at 1:42 PM Eugen Block  wrote:

> All your PGs are inactive, if two of four OSDs are down and you
> probably have a pool size of 3 then no IO can be served. You’d need at
> least three up ODSs to resolve that.
>
>
> Zitat von Abdelillah Asraoui :
>
> > Ceph is reporting warning on slow metdataIOs on one of the MDS server,
> > this is
> >
> > a new cluster with no upgrade..
> >
> > Anyone has encountered this and is there a workaround ..
> >
> > ceph -s
> >
> >   cluster:
> >
> > id: 801691e6xx-x-xx-xx-xx
> >
> > health: HEALTH_WARN
> >
> > 1 MDSs report slow metadata IOs
> >
> > noscrub,nodeep-scrub flag(s) set
> >
> > 2 osds down
> >
> > 2 hosts (2 osds) down
> >
> > Reduced data availability: 97 pgs inactive, 66 pgs peering,
> 53
> > pgs stale
> >
> > Degraded data redundancy: 31 pgs undersized
> >
> > 2 slow ops, oldest one blocked for 30 sec, osd.0 has slow ops
> >
> >
> >
> >   services:
> >
> > mon: 3 daemons, quorum a,c,f (age 15h)
> >
> > mgr: a(active, since 17h)
> >
> > mds: myfs:1 {0=myfs-a=up:creating} 1 up:standby
> >
> > osd: 4 osds: 2 up (since 36s), 4 in (since 10h)
> >
> >  flags noscrub,nodeep-scrub
> >
> >
> >
> >   data:
> >
> > pools:   4 pools, 97 pgs
> >
> > objects: 0 objects, 0 B
> >
> > usage:   1.0 GiB used, 1.8 TiB / 1.8 TiB avail
> >
> > pgs: 100.000% pgs not active
> >
> >  44 creating+peering
> >
> >  31 stale+undersized+peered
> >
> >  22 stale+creating+peering
> >
> >
> >
> >   progress:
> >
> > Rebalancing after osd.2 marked in (10h)
> >
> >   []
> >
> > Rebalancing after osd.3 marked in (10h)
> >
> >   []
> >
> >
> > Thanks!
> > ___
> > ceph-users mailing list -- ceph-users@ceph.io
> > To unsubscribe send an email to ceph-users-le...@ceph.io
>
>
>
>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: 1 MDS report slow metadata IOs

2021-10-05 Thread Eugen Block
All your PGs are inactive, if two of four OSDs are down and you  
probably have a pool size of 3 then no IO can be served. You’d need at  
least three up ODSs to resolve that.



Zitat von Abdelillah Asraoui :


Ceph is reporting warning on slow metdataIOs on one of the MDS server,
this is

a new cluster with no upgrade..

Anyone has encountered this and is there a workaround ..

ceph -s

  cluster:

id: 801691e6xx-x-xx-xx-xx

health: HEALTH_WARN

1 MDSs report slow metadata IOs

noscrub,nodeep-scrub flag(s) set

2 osds down

2 hosts (2 osds) down

Reduced data availability: 97 pgs inactive, 66 pgs peering, 53
pgs stale

Degraded data redundancy: 31 pgs undersized

2 slow ops, oldest one blocked for 30 sec, osd.0 has slow ops



  services:

mon: 3 daemons, quorum a,c,f (age 15h)

mgr: a(active, since 17h)

mds: myfs:1 {0=myfs-a=up:creating} 1 up:standby

osd: 4 osds: 2 up (since 36s), 4 in (since 10h)

 flags noscrub,nodeep-scrub



  data:

pools:   4 pools, 97 pgs

objects: 0 objects, 0 B

usage:   1.0 GiB used, 1.8 TiB / 1.8 TiB avail

pgs: 100.000% pgs not active

 44 creating+peering

 31 stale+undersized+peered

 22 stale+creating+peering



  progress:

Rebalancing after osd.2 marked in (10h)

  []

Rebalancing after osd.3 marked in (10h)

  []


Thanks!
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io




___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io