Hey, I would really appreciate any help I can get on this as googling has
led me to a dead end.

We have 2 data centers each with 4 servers running ceph on kubernetes in
multisite config, everything is working great but recently the master
cluster changed status to HEALTH_WARN and the issues are large omap objects
in the .rgw.log pool. Second cluster is still HEALTH_OK

Viewing the sync error log from master shows a lot of very ancient logs
related to a bucket that has since been deleted.

Is there any way to clear this log?

bash-4.4$ radosgw-admin sync error list | wc -l
352162

I believe, although I'm not sure that this is a massive part of the data
stored in the .rgw.log pool, I haven't been able to find any info on this
except for several other posts about clearing the error log but none of
them had a resolution.

I am tempted to increase the PG's for this pool from 16 to 32 to see if it
helps but holding off because that is not an ideal solution just to get rid
of this warning when all I want is to get rid of the errors related to a
bucket that no longer exists.

Thanks to anyone that can offer advice!

Sarah
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to