Hi all, I am trying to add Ceph RBD (pacific) as a new Primary Storage for my fresh Cloudstack 4.15.1 installation. I have currently an NFS server as Primary storage running and after connecting Cloudstack with Ceph, I would then remove the NFS server. Unfortunately, I am running into the same problem, no matter If I am trying to add the Ceph storage Cluster or Zone-wide. The output of /cloudstack/agent/agent.log is as follows:
2021-08-31 12:43:44,247 INFO [kvm.storage.LibvirtStorageAdaptor] (agentRequest-Handler-4:null) (logid:cb99bb9f) Asking libvirt to refresh storage pool 84aa6a27-0413-39ad-87ca-5e08078b9b84 2021-08-31 12:44:40,699 INFO [kvm.storage.LibvirtStorageAdaptor] (agentRequest-Handler-5:null) (logid:cae1fff8) Attempting to create storage pool fc6d0942-21ac-3cd1-b9f3-9e158cf4d75d (RBD) in libvirt 2021-08-31 12:44:40,701 WARN [kvm.storage.LibvirtStorageAdaptor] (agentRequest-Handler-5:null) (logid:cae1fff8) Storage pool fc6d0942-21ac-3cd1-b9f3-9e158cf4d75d was not found running in libvirt. Need to create it. 2021-08-31 12:44:40,701 INFO [kvm.storage.LibvirtStorageAdaptor] (agentRequest-Handler-5:null) (logid:cae1fff8) Didn't find an existing storage pool fc6d0942-21ac-3cd1-b9f3-9e158cf4d75d by UUID, checking for pools with duplicate paths 2021-08-31 12:44:44,286 INFO [kvm.storage.LibvirtStorageAdaptor] (agentRequest-Handler-3:null) (logid:725f7dcf) Trying to fetch storage pool 84aa6a27-0413-39ad-87ca-5e08078b9b84 from libvirt 2021-08-31 12:44:44,290 INFO [kvm.storage.LibvirtStorageAdaptor] (agentRequest-Handler-3:null) (logid:725f7dcf) Asking libvirt to refresh storage pool 84aa6a27-0413-39ad-87ca-5e08078b9b84 2021-08-31 12:45:10,780 ERROR [kvm.storage.LibvirtStorageAdaptor] (agentRequest-Handler-5:null) (logid:cae1fff8) Failed to create RBD storage pool: org.libvirt.LibvirtException: failed to create the RBD IoCTX. Does the pool 'cloudstack' exist?: No such file or directory 2021-08-31 12:45:10,780 ERROR [kvm.storage.LibvirtStorageAdaptor] (agentRequest-Handler-5:null) (logid:cae1fff8) Failed to create the RBD storage pool, cleaning up the libvirt secret 2021-08-31 12:45:10,781 WARN [cloud.agent.Agent] (agentRequest-Handler-5:null) (logid:cae1fff8) Caught: com.cloud.utils.exception.CloudRuntimeException: Failed to create storage pool: fc6d0942-21ac-3cd1-b9f3-9e158cf4d75d at com.cloud.hypervisor.kvm.storage.LibvirtStorageAdaptor.createStoragePool(LibvirtStorageAdaptor.java:645) at com.cloud.hypervisor.kvm.storage.KVMStoragePoolManager.createStoragePool(KVMStoragePoolManager.java:329) at com.cloud.hypervisor.kvm.storage.KVMStoragePoolManager.createStoragePool(KVMStoragePoolManager.java:323) at com.cloud.hypervisor.kvm.resource.wrapper.LibvirtModifyStoragePoolCommandWrapper.execute(LibvirtModifyStoragePoolCommandWrapper.java:42) at com.cloud.hypervisor.kvm.resource.wrapper.LibvirtModifyStoragePoolCommandWrapper.execute(LibvirtModifyStoragePoolCommandWrapper.java:35) at com.cloud.hypervisor.kvm.resource.wrapper.LibvirtRequestWrapper.execute(LibvirtRequestWrapper.java:78) at com.cloud.hypervisor.kvm.resource.LibvirtComputingResource.executeRequest(LibvirtComputingResource.java:1646) at com.cloud.agent.Agent.processRequest(Agent.java:661) at com.cloud.agent.Agent$AgentRequestHandler.doTask(Agent.java:1079) at com.cloud.utils.nio.Task.call(Task.java:83) at com.cloud.utils.nio.Task.call(Task.java:29) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) Details of my setup: - Ceph pacific is installed and configured on a test environment. Cluster health is ok. - rbd pool and user created as described in the Ceph doc: https://docs.ceph.com/en/pacific/rbd/rbd-cloudstack/?highlight=cloudstack - The IP of my Ceph Mon with the rbd pool is 192.168.1.4, the firewall is disabled there - I have also tried to copy the keyring and ceph.conf from the monitor node to the kvm machine (In the test environment I have only one kvm host), still the same problem Do you have any ideas how to resolve the problem? Cheers, Mevludin