On Fri, May 4, 2018 at 1:59 AM John Spray <jsp...@redhat.com> wrote:

> On Fri, May 4, 2018 at 7:21 AM, Tracy Reed <tr...@ultraviolet.org> wrote:
> > My ceph status says:
> >
> >   cluster:
> >     id:     b2b00aae-f00d-41b4-a29b-58859aa41375
> >     health: HEALTH_OK
> >
> >   services:
> >     mon: 3 daemons, quorum ceph01,ceph03,ceph07
> >     mgr: ceph01(active), standbys: ceph-ceph07, ceph03
> >     osd: 78 osds: 78 up, 78 in
> >
> >   data:
> >     pools:   4 pools, 3240 pgs
> >     objects: 4384k objects, 17533 GB
> >     usage:   53141 GB used, 27311 GB / 80452 GB avail
> >     pgs:     3240 active+clean
> >
> >   io:
> >     client:   4108 kB/s rd, 10071 kB/s wr, 27 op/s rd, 331 op/s wr
> >
> > but my mgr dashboard web interface says:
> >
> >
> > Health
> > Overall status: HEALTH_WARN
> >
> >     PG_AVAILABILITY: Reduced data availability: 2563 pgs inactive
> >
> >
> > Anyone know why the discrepency? Hopefully the dashboard is very
> > mistaken! Everything seems to be operating normally. If I had 2/3 of my
> > pgs inactive I'm sure all of my rbd backing my VMs would be blocked etc.
>
> A situation like this probably indicates that something is going wrong
> with the mon->mgr synchronisation of health state (it's all calculated
> in one place and the mon updates the mgr every few seconds).
>
> 1. Look for errors in your monitor logs
> 2. You'll probably find that everything gets back in sync if you
> restart a mgr daemon
>
> John
>

Isn't that the wrong direction for sync issues, though? I mean, the manager
is where the PG reports actually go. So if the cluster's still running, the
monitor says it's active+clean, and the *dashboard* says the PGs are
inactive, it sounds like the monitor has the correct view and something has
gone wrong between the rest of the manager guts and the dashboard display.
-Greg


>
> > I'm running ceph-12.2.4-0.el7.x86_64 on CentOS 7. Almost all filestore
> > except for one OSD which recently had to be replaced which I made
> > bluestore. I plan to slowly migrate everything over to bluestore over
> > the course of the next month.
> >
> > Thanks!
> >
> > --
> > Tracy Reed
> > http://tracyreed.org
> > Digital signature attached for your safety.
> >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to