Hello,
ceph health detail says my 5-node cluster is healthy, yet when I ran
ceph orch upgrade start --ceph-version 16.2.7 everything seemed to go
fine until we got to the OSD section, now for the past hour, every 15
seconds a new log entry of 'Upgrade: unsafe to stop osd(s) at this time
(1 PGs are or would become offline)' appears in the logs.
ceph pg dump_stuck (unclean, degraded, etc) shows "ok" for everything
too. Yet somehow 1 PG is (apparently) holding up all the OSD upgrades
and not letting the process finish. Should I stop the upgrade and try it
again? (I haven't done that before so was just nervous to try it). Any
other ideas?
cluster:
id: 9aa000e8-b999-11eb-82f2-ecf4bbcc0ac0
health: HEALTH_OK
services:
mon: 4 daemons, quorum ceph05,ceph04,ceph01,ceph03 (age 92m)
mgr: ceph03.futetp(active, since 97m), standbys: ceph01.fblojp
mds: 1/1 daemons up, 1 hot standby
osd: 33 osds: 33 up (since 2h), 33 in (since 4h); 9 remapped pgs
data:
volumes: 1/1 healthy
pools: 7 pools, 193 pgs
objects: 3.72k objects, 14 GiB
usage: 43 GiB used, 64 TiB / 64 TiB avail
pgs: 231/11170 objects misplaced (2.068%)
185 active+clean
8 active+clean+remapped
io:
client: 1.2 KiB/s rd, 2 op/s rd, 0 op/s wr
progress:
Upgrade to 16.2.7 (5m)
[=====.......................] (remaining: 24m)
--
Zach
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io