As already mentioned in another reply, you will have to make space by either reweighting the OSD or adding more OSDs.

Another option, though quite a bit more radical, would be to reduce the size of the pool.
Right now, you hopefully have a size/min_size of 3/2.

ceph osd pool get {pool} size
ceph osd pool get {pool} min_size

By reducing the size to 2, you will gain about 1/3 of space which can help you get out of the situation. But the pool will become IO blocked once a single OSD is down. So that should only be done as an emergency measure to get operational again. But then you need to address the actual issue ASAP (get more space) so that you can increase the size back to 3 again.

ceph osd pool set {pool} size 2

And please plan an upgrade of the cluster soon! Proxmox VE 6 is EOL since last summer and the intricacies of that version and the Ceph versions along with it are fading away ;)

https://pve.proxmox.com/wiki/Upgrade_from_6.x_to_7.0

Cheers,
Aaron

On 5/4/23 08:39, Joseph John wrote:
Dear All,
Good morning
We have a proxmox setup, with 4 nodes, with 6.3-3
We   have    Node 1 and Node 2 running with 6.3-3
and              Node 3 and Node 4 running with 6.4-15

today we noticed that , we were not able to ssh to the virtual instance ,
or not able to login using the console based option
when I checked the summary, we could see that in "HEALTH_ERR" we are
getting message that 1 full osd(s)



Thanks
Joseph John
00971-50-7451809
_______________________________________________
pve-user mailing list
[email protected]
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user




_______________________________________________
pve-user mailing list
[email protected]
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user

Reply via email to