On 01/06/22 11:29, Aaron Lauterer wrote:
Do you get additional errors if you run the following command? Assuming that the storage is also called pool1.

pvesm list pool1

No additional errors:

root@pmx-14:~# pvesm list pool1
rbd error: rbd: listing images failed: (2) No such file or directory



Do you have VMs with disk images on that storage? If so, do they start normally?

Yes, we have a lot of VMs with disk on that storage and yes they seems to start normally (last start yesterday when we first notice the GUI behaviour)


Can you show the configuration of that storage and the one of the working pool? (/etc/pve/storage.cfg)

Sure (edited the IP addresses and pool names):

[cit /etc/pve/storage.cfg]
...
rbd: pool1
        content images
        monhost 172.16.1.1;1172.16.1.2;172.16.1.3
        pool pool1
        username admin

rbd: pool2
        content images
        monhost 172.16.1.1;172.16.1.2;172.16.1.3
        pool pool2
        username admin
...
[/cit]

Thanks!

Iztok


On 6/1/22 11:13, Iztok Gregori wrote:
Hi to all!

I have a Proxmox cluster (7.1) connected to an external CEPH cluster (octopus).  From the GUI I cannot list the content (disks) of one pool (but I'm able to list all the other pools):

rbd error: rbd: listing images failed: (2) No such file or directory (500)

The pveproxy/access.log shows the error for "pool1":

"GET /api2/json/nodes/pmx-14/storage/pool1/content?content=images HTTP/1.1" 500 13

but when I try another pool ("pool2") it works:

"GET /api2/json/nodes/pmx-14/storage/pool2/content?content=images HTTP/1.1" 200 841

 From the command line "rbd ls pool1" is working fine (because I don't have a ceph.conf I ran it with "rbd -m 172.16.1.1 --keyring /etc/pve/priv/ceph/pool1.keyring ls pool1") and I see the pool contents.

The cluster is running fine, the VMs access the disks on that pool without a problem

What can it be?

The cluster is a mix of freshly installed nodes and upgraded ones, all the 17 nodes (but one which is 6.4 but without any running VMs) are running:

root@pmx-14:~# pveversion -v
proxmox-ve: 7.1-1 (running kernel: 5.13.19-6-pve)
pve-manager: 7.1-12 (running version: 7.1-12/b3c09de3)
pve-kernel-helper: 7.1-14
pve-kernel-5.13: 7.1-9
pve-kernel-5.13.19-6-pve: 5.13.19-15
pve-kernel-5.13.19-2-pve: 5.13.19-4
ceph-fuse: 15.2.15-pve1
corosync: 3.1.5-pve2
criu: 3.15-1+pve-1
glusterfs-client: 9.2-1
ifupdown2: 3.1.0-1+pmx3
ksm-control-daemon: 1.4-1
libjs-extjs: 7.0.0-1
libknet1: 1.22-pve2
libproxmox-acme-perl: 1.4.1
libproxmox-backup-qemu0: 1.2.0-1
libpve-access-control: 7.1-7
libpve-apiclient-perl: 3.2-1
libpve-common-perl: 7.1-5
libpve-guest-common-perl: 4.1-1
libpve-http-server-perl: 4.1-1
libpve-storage-perl: 7.1-1
libspice-server1: 0.14.3-2.1
lvm2: 2.03.11-2.1
lxc-pve: 4.0.11-1
lxcfs: 4.0.11-pve1
novnc-pve: 1.3.0-2
proxmox-backup-client: 2.1.5-1
proxmox-backup-file-restore: 2.1.5-1
proxmox-mini-journalreader: 1.3-1
proxmox-widget-toolkit: 3.4-7
pve-cluster: 7.1-3
pve-container: 4.1-4
pve-docs: 7.1-2
pve-edk2-firmware: 3.20210831-2
pve-firewall: 4.2-5
pve-firmware: 3.3-6
pve-ha-manager: 3.3-3
pve-i18n: 2.6-2
pve-qemu-kvm: 6.1.1-2
pve-xtermjs: 4.16.0-1
qemu-server: 7.1-4
smartmontools: 7.2-1
spiceterm: 3.2-2
swtpm: 0.7.1~bpo11+1
vncterm: 1.7-1
zfsutils-linux: 2.1.4-pve1

I can provide other information if it's needed.

Cheers
Iztok Gregori





_______________________________________________
pve-user mailing list
[email protected]
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user

Reply via email to