Hi all,
I am trying to add Ceph RBD (pacific) as a new Primary Storage for my fresh
Cloudstack 4.15.1 installation. I have currently an NFS server as Primary
storage running and after connecting Cloudstack with Ceph, I would then remove
the NFS server. Unfortunately, I am running into the same problem, no matter If
I am trying to add the Ceph storage Cluster or Zone-wide. The output of
/cloudstack/agent/agent.log is as follows:
2021-08-31 12:43:44,247 INFO [kvm.storage.LibvirtStorageAdaptor]
(agentRequest-Handler-4:null) (logid:cb99bb9f) Asking libvirt to refresh
storage pool 84aa6a27-0413-39ad-87ca-5e08078b9b84
2021-08-31 12:44:40,699 INFO [kvm.storage.LibvirtStorageAdaptor]
(agentRequest-Handler-5:null) (logid:cae1fff8) Attempting to create storage
pool fc6d0942-21ac-3cd1-b9f3-9e158cf4d75d (RBD) in libvirt
2021-08-31 12:44:40,701 WARN [kvm.storage.LibvirtStorageAdaptor]
(agentRequest-Handler-5:null) (logid:cae1fff8) Storage pool
fc6d0942-21ac-3cd1-b9f3-9e158cf4d75d was not found running in libvirt. Need to
create it.
2021-08-31 12:44:40,701 INFO [kvm.storage.LibvirtStorageAdaptor]
(agentRequest-Handler-5:null) (logid:cae1fff8) Didn't find an existing storage
pool fc6d0942-21ac-3cd1-b9f3-9e158cf4d75d by UUID, checking for pools with
duplicate paths
2021-08-31 12:44:44,286 INFO [kvm.storage.LibvirtStorageAdaptor]
(agentRequest-Handler-3:null) (logid:725f7dcf) Trying to fetch storage pool
84aa6a27-0413-39ad-87ca-5e08078b9b84 from libvirt
2021-08-31 12:44:44,290 INFO [kvm.storage.LibvirtStorageAdaptor]
(agentRequest-Handler-3:null) (logid:725f7dcf) Asking libvirt to refresh
storage pool 84aa6a27-0413-39ad-87ca-5e08078b9b84
2021-08-31 12:45:10,780 ERROR [kvm.storage.LibvirtStorageAdaptor]
(agentRequest-Handler-5:null) (logid:cae1fff8) Failed to create RBD storage
pool: org.libvirt.LibvirtException: failed to create the RBD IoCTX. Does the
pool 'cloudstack' exist?: No such file or directory
2021-08-31 12:45:10,780 ERROR [kvm.storage.LibvirtStorageAdaptor]
(agentRequest-Handler-5:null) (logid:cae1fff8) Failed to create the RBD storage
pool, cleaning up the libvirt secret
2021-08-31 12:45:10,781 WARN [cloud.agent.Agent] (agentRequest-Handler-5:null)
(logid:cae1fff8) Caught:
com.cloud.utils.exception.CloudRuntimeException: Failed to create storage pool:
fc6d0942-21ac-3cd1-b9f3-9e158cf4d75d
at
com.cloud.hypervisor.kvm.storage.LibvirtStorageAdaptor.createStoragePool(LibvirtStorageAdaptor.java:645)
at
com.cloud.hypervisor.kvm.storage.KVMStoragePoolManager.createStoragePool(KVMStoragePoolManager.java:329)
at
com.cloud.hypervisor.kvm.storage.KVMStoragePoolManager.createStoragePool(KVMStoragePoolManager.java:323)
at
com.cloud.hypervisor.kvm.resource.wrapper.LibvirtModifyStoragePoolCommandWrapper.execute(LibvirtModifyStoragePoolCommandWrapper.java:42)
at
com.cloud.hypervisor.kvm.resource.wrapper.LibvirtModifyStoragePoolCommandWrapper.execute(LibvirtModifyStoragePoolCommandWrapper.java:35)
at
com.cloud.hypervisor.kvm.resource.wrapper.LibvirtRequestWrapper.execute(LibvirtRequestWrapper.java:78)
at
com.cloud.hypervisor.kvm.resource.LibvirtComputingResource.executeRequest(LibvirtComputingResource.java:1646)
at com.cloud.agent.Agent.processRequest(Agent.java:661)
at com.cloud.agent.Agent$AgentRequestHandler.doTask(Agent.java:1079)
at com.cloud.utils.nio.Task.call(Task.java:83)
at com.cloud.utils.nio.Task.call(Task.java:29)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:829)
Details of my setup:
- Ceph pacific is installed and configured on a test environment. Cluster
health is ok.
- rbd pool and user created as described in the Ceph doc:
https://docs.ceph.com/en/pacific/rbd/rbd-cloudstack/?highlight=cloudstack
- The IP of my Ceph Mon with the rbd pool is 192.168.1.4, the firewall is
disabled there
- I have also tried to copy the keyring and ceph.conf from the monitor node to
the kvm machine (In the test environment I have only one kvm host), still the
same problem
Do you have any ideas how to resolve the problem?
Cheers,
Mevludin