I can't directly answer your question, but I'm currently redeploying all my OSDs with `bluestore_elastic_shared_blobs` disabled. I don't know for your situation, but you might want to investigate if that's a good idea before upgrading from 18.2.7 to 19.2.3.
AFAIK the issue has also only be reported on EC pools. Not on replicated pools that I'm aware of. Wannes ________________________________ From: Văn Trần Quốc via ceph-users <[email protected]> Sent: Thursday, December 18, 2025 08:09 To: [email protected] <[email protected]> Subject: [ceph-users] Upgrade Advice: Ceph Reef (18.2.7) to Squid (19.2.3) on Proxmox VMs Hi Ceph Users, I am planning a major upgrade for our production cluster from Reef (18. 2. 7) to Squid (19. 2. 3) and would like to seek advice regarding stability and potential risks. Infrastructure Overview: + Deployment: Cephadm. + Cluster Size: ZjQcmQRYFpfptBannerStart This Message Is From an External Sender This message came from outside your organization. ZjQcmQRYFpfptBannerEnd Hi Ceph Users, I am planning a major upgrade for our production cluster from Reef (18.2.7) to Squid (19.2.3) and would like to seek advice regarding stability and potential risks. Infrastructure Overview: + Deployment: Cephadm. + Cluster Size: 7 Nodes total. + Hardware/Virtualization: Each node is a Virtual Machine hosted on Proxmox. + OSD Layout: 7 OSDs total (7 OSD per node). + Other Daemons: 5 MONs, 7 MDSs, 3 MGRs. Services include RGW, CephFS, and Block Devices. + Pool Type: Replicated. The Concern: We are currently running stable on Reef 18.2.7. However, I have been following recent discussions on the mailing list and tracker regarding critical failures when upgrading to Squid, specifically concerning OSD crashes and data corruption. I am particularly worried about the issues reported in these threads, where users experienced failures during or after the upgrade: + https://urldefense.com/v3/__https://www.mail-archive.com/[email protected]/msg30399.html__;!!FtrhtPsWDhZ6tw!DsPi6pIv4HTyn0JCNNnmIKmLY-ziJgiC9lwTpP92pxuKNqtMOROVIVhYLzFigdllIqGzG2gw495sw7kS$[mail-archive[.]com] + https://urldefense.com/v3/__https://www.mail-archive.com/[email protected]/msg31238.html__;!!FtrhtPsWDhZ6tw!DsPi6pIv4HTyn0JCNNnmIKmLY-ziJgiC9lwTpP92pxuKNqtMOROVIVhYLzFigdllIqGzG2gw4yUJ3xsW$[mail-archive[.]com] + https://urldefense.com/v3/__https://tracker.ceph.com/issues/70390__;!!FtrhtPsWDhZ6tw!DsPi6pIv4HTyn0JCNNnmIKmLY-ziJgiC9lwTpP92pxuKNqtMOROVIVhYLzFigdllIqGzG2gw40vjAxeH$[tracker[.]ceph[.]com] Given our deployment topology and the jump to a new major version (Squid), I have a few questions: + Stability & Success Stories: Is Ceph Squid 19.2.3 considered safe regarding the OSD crash/corruption bugs mentioned in the links above? Has anyone in the community successfully completed the upgrade from 18.2.7 to 19.2.3 without issues? Confirmation of a clean upgrade path would be very reassuring. + Upgrade Path: Are there any known regressions or critical "gotchas" when moving from 18.2.7 directly to 19.2.3 in a virtualized environment? Any experiences or warnings from those running Squid in similar environments would be greatly appreciated. Thank you, Van Tran _______________________________________________ ceph-users mailing list -- [email protected] To unsubscribe send an email to [email protected] _______________________________________________ ceph-users mailing list -- [email protected] To unsubscribe send an email to [email protected]
