What does

‘ceph osd tree’ show?

How many OSD’s should you have 7 or 10?

> On 22 Feb 2022, at 14:40, Michel Niyoyita <mico...@gmail.com> wrote:
> 
> Actually one of my colleagues tried to reboot all nodes and he did not 
> prepare the node like setting noout , norecover ......, once all node are up 
> the cluster is no longer accessible and above are messages we are getting. I 
> did not remove any osd . except are marked down.
> below is my ceph.conf:
> 
> mon initial members = ceph-mon1,ceph-mon2,ceph-mon3
> mon_allow_pool_delete = True
> mon_clock_drift_allowed = 0.5
> mon_max_pg_per_osd = 400
> mon_osd_allow_primary_affinity = 1
> mon_pg_warn_max_object_skew = 0
> mon_pg_warn_max_per_osd = 0
> mon_pg_warn_min_per_osd = 0
> osd pool default crush rule = -1
> osd_pool_default_min_size = 1
> osd_pool_default_size = 2
> public network = 0.0.0.0/0 <http://0.0.0.0/0>
> 
> On Tue, Feb 22, 2022 at 4:32 PM <ash...@amerrick.co.uk 
> <mailto:ash...@amerrick.co.uk>> wrote:
> You have 1 OSD offline, has this disk failed or you aware of what has caused 
> this to go offline?
> Shows you have 10 OSD’s but only 7in, have you removed the other 3? Was the 
> data fully drained off these first?
> 
> I see you have 11 Pool’s what are these setup as, type and min/max size?
> 
> > On 22 Feb 2022, at 14:15, Michel Niyoyita <mico...@gmail.com 
> > <mailto:mico...@gmail.com>> wrote:
> > 
> > Dear Ceph Users,
> > 
> > Kindly help me to repair my cluster is down from yesterday up to now I am
> > not able to make it up and running . below are some findings:
> > 
> >    id:     6ad86187-2738-42d8-8eec-48b2a43c298f
> >    health: HEALTH_ERR
> >            mons are allowing insecure global_id reclaim
> >            1/3 mons down, quorum ceph-mon1,ceph-mon3
> >            10/32332 objects unfound (0.031%)
> >            1 osds down
> >            3 scrub errors
> >            Reduced data availability: 124 pgs inactive, 60 pgs down, 411
> > pgs stale
> >            Possible data damage: 9 pgs recovery_unfound, 1 pg
> > backfill_unfound, 1 pg inconsistent
> >            Degraded data redundancy: 6009/64664 objects degraded (9.293%),
> > 55 pgs degraded, 80 pgs undersized
> >            11 pgs not deep-scrubbed in time
> >            5 slow ops, oldest one blocked for 1638 sec, osd.9 has slow ops
> > 
> >  services:
> >    mon: 3 daemons, quorum ceph-mon1,ceph-mon3 (age 3h), out of quorum:
> > ceph-mon2
> >    mgr: ceph-mon1(active, since 9h), standbys: ceph-mon2
> >    osd: 10 osds: 6 up (since 7h), 7 in (since 9h); 43 remapped pgs
> > 
> >  data:
> >    pools:   11 pools, 560 pgs
> >    objects: 32.33k objects, 159 GiB
> >    usage:   261 GiB used, 939 GiB / 1.2 TiB avail
> >    pgs:     11.429% pgs unknown
> >             10.714% pgs not active
> >             6009/64664 objects degraded (9.293%)
> >             1384/64664 objects misplaced (2.140%)
> >             10/32332 objects unfound (0.031%)
> >             245 stale+active+clean
> >             70  active+clean
> >             64  unknown
> >             48  stale+down
> >             45  stale+active+undersized+degraded
> >             37  stale+active+clean+remapped
> >             28  stale+active+undersized
> >             12  down
> >             2   stale+active+recovery_unfound+degraded
> >             2   stale+active+recovery_unfound+undersized+degraded
> >             2   stale+active+recovery_unfound+undersized+degraded+remapped
> >             2   active+recovery_unfound+undersized+degraded+remapped
> >             1   active+clean+inconsistent
> >             1   stale+active+recovery_unfound+degraded+remapped
> >             1   stale+active+backfill_unfound+undersized+degraded+remapped
> > 
> > If someone faced same issue please help me.
> > 
> > Best Regards.
> > 
> > Michel
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@ceph.io <mailto:ceph-users@ceph.io>
> > To unsubscribe send an email to ceph-users-le...@ceph.io 
> > <mailto:ceph-users-le...@ceph.io>
> 
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io <mailto:ceph-users@ceph.io>
> To unsubscribe send an email to ceph-users-le...@ceph.io 
> <mailto:ceph-users-le...@ceph.io>

_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to