Hi all,

I am trying to add Ceph RBD (pacific) as a new Primary Storage for my fresh 
Cloudstack 4.15.1 installation. I have currently an NFS server as Primary 
storage running and after connecting Cloudstack with Ceph, I would then remove 
the NFS server. Unfortunately, I am running into the same problem, no matter If 
I am trying to add the Ceph storage Cluster or Zone-wide. The output of 
/cloudstack/agent/agent.log is as follows:

2021-08-31 12:43:44,247 INFO  [kvm.storage.LibvirtStorageAdaptor] 
(agentRequest-Handler-4:null) (logid:cb99bb9f) Asking libvirt to refresh 
storage pool 84aa6a27-0413-39ad-87ca-5e08078b9b84
2021-08-31 12:44:40,699 INFO  [kvm.storage.LibvirtStorageAdaptor] 
(agentRequest-Handler-5:null) (logid:cae1fff8) Attempting to create storage 
pool fc6d0942-21ac-3cd1-b9f3-9e158cf4d75d (RBD) in libvirt
2021-08-31 12:44:40,701 WARN  [kvm.storage.LibvirtStorageAdaptor] 
(agentRequest-Handler-5:null) (logid:cae1fff8) Storage pool 
fc6d0942-21ac-3cd1-b9f3-9e158cf4d75d was not found running in libvirt. Need to 
create it.
2021-08-31 12:44:40,701 INFO  [kvm.storage.LibvirtStorageAdaptor] 
(agentRequest-Handler-5:null) (logid:cae1fff8) Didn't find an existing storage 
pool fc6d0942-21ac-3cd1-b9f3-9e158cf4d75d by UUID, checking for pools with 
duplicate paths
2021-08-31 12:44:44,286 INFO  [kvm.storage.LibvirtStorageAdaptor] 
(agentRequest-Handler-3:null) (logid:725f7dcf) Trying to fetch storage pool 
84aa6a27-0413-39ad-87ca-5e08078b9b84 from libvirt
2021-08-31 12:44:44,290 INFO  [kvm.storage.LibvirtStorageAdaptor] 
(agentRequest-Handler-3:null) (logid:725f7dcf) Asking libvirt to refresh 
storage pool 84aa6a27-0413-39ad-87ca-5e08078b9b84
2021-08-31 12:45:10,780 ERROR [kvm.storage.LibvirtStorageAdaptor] 
(agentRequest-Handler-5:null) (logid:cae1fff8) Failed to create RBD storage 
pool: org.libvirt.LibvirtException: failed to create the RBD IoCTX. Does the 
pool 'cloudstack' exist?: No such file or directory
2021-08-31 12:45:10,780 ERROR [kvm.storage.LibvirtStorageAdaptor] 
(agentRequest-Handler-5:null) (logid:cae1fff8) Failed to create the RBD storage 
pool, cleaning up the libvirt secret
2021-08-31 12:45:10,781 WARN  [cloud.agent.Agent] (agentRequest-Handler-5:null) 
(logid:cae1fff8) Caught:
com.cloud.utils.exception.CloudRuntimeException: Failed to create storage pool: 
fc6d0942-21ac-3cd1-b9f3-9e158cf4d75d
        at 
com.cloud.hypervisor.kvm.storage.LibvirtStorageAdaptor.createStoragePool(LibvirtStorageAdaptor.java:645)
        at 
com.cloud.hypervisor.kvm.storage.KVMStoragePoolManager.createStoragePool(KVMStoragePoolManager.java:329)
        at 
com.cloud.hypervisor.kvm.storage.KVMStoragePoolManager.createStoragePool(KVMStoragePoolManager.java:323)
        at 
com.cloud.hypervisor.kvm.resource.wrapper.LibvirtModifyStoragePoolCommandWrapper.execute(LibvirtModifyStoragePoolCommandWrapper.java:42)
        at 
com.cloud.hypervisor.kvm.resource.wrapper.LibvirtModifyStoragePoolCommandWrapper.execute(LibvirtModifyStoragePoolCommandWrapper.java:35)
        at 
com.cloud.hypervisor.kvm.resource.wrapper.LibvirtRequestWrapper.execute(LibvirtRequestWrapper.java:78)
        at 
com.cloud.hypervisor.kvm.resource.LibvirtComputingResource.executeRequest(LibvirtComputingResource.java:1646)
        at com.cloud.agent.Agent.processRequest(Agent.java:661)
        at com.cloud.agent.Agent$AgentRequestHandler.doTask(Agent.java:1079)
        at com.cloud.utils.nio.Task.call(Task.java:83)
        at com.cloud.utils.nio.Task.call(Task.java:29)
        at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
        at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
        at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
        at java.base/java.lang.Thread.run(Thread.java:829)

Details of my setup:
- Ceph pacific is installed and configured on a test environment. Cluster 
health is ok. 
- rbd pool and user created as described in the Ceph doc: 
https://docs.ceph.com/en/pacific/rbd/rbd-cloudstack/?highlight=cloudstack
- The IP of my Ceph Mon with the rbd pool is 192.168.1.4, the firewall is 
disabled there
- I have also tried to copy the keyring and ceph.conf from the monitor node to 
the kvm machine (In the test environment I have only one kvm host), still the 
same problem

Do you have any ideas how to resolve the problem?

Cheers,

Mevludin

Reply via email to