Hi Jeremy,

The CMK looks fine, you need to review the logs to find what caused the import 
failure. [1]
In the UI , you may want to change the selection to shared storage instead of 
local storage.

[1] 
https://docs.cloudstack.apache.org/en/4.22.0.0/adminguide/troubleshooting.html

-Jithin

From: Jeremy Hansen <[email protected]>
Date: Friday, 6 February 2026 at 8:13 AM
To: Jithin Raju <[email protected]>
Subject: Re: Adding an existing cluster as a new region?

Thank you so much for your help.  I just want to clarify that I’m going 
throught the proper steps.

I placed the qcow here:

nas.sifu.intra:/volume1/cloud/primary                 42T   18T   25T  42% 
/mnt/630a7d01-1f6d-36f1-b587-1f561ccc82dd/8629c278-8eae-4c92-a3b6-831ab20a9c2b

Then I ran this command:

cmk -p la1
Apache CloudStack 🐵 CloudMonkey 6.4.0-rc
Report issues: https://github.com/apache/cloudstack-cloudmonkey/issues

(la1) 🐱 > import vm name=ca-fs-corp displayname=ca-fs-corp importsource=shared 
hypervisor=kvm storageid=630a7d01-1f6d-36f1-b587-1f561ccc82dd 
diskpath=8629c278-8eae-4c92-a3b6-831ab20a9c2b 
networkid=d3619d80-11cd-483b-86f2-3245c688a52c migrateallowed=true 
zoneid=758cd49e-a193-4445-a8a9-576d16f4c24b 
clusterid=38e832ba-8837-4aa6-8b0b-e92a841c0ef2 details[0].UEFI=LEGACY 
serviceofferingid=88de7d7a-02c1-4a51-831f-8b96ce2227ae



Which results in this:

import vm name=ca-fs-corp displayname=ca-fs-corp importsource=shared 
hypervisor=kvm storageid=630a7d01-1f6d-36f1-b587-1f561ccc82dd 
diskpath=8629c278-8eae-4c92-a3b6-831ab20a9c2b 
networkid=d3619d80-11cd-483b-86f2-3245c688a52c migrateallowed=true 
zoneid=758cd49e-a193-4445-a8a9-576d16f4c24b 
clusterid=38e832ba-8837-4aa6-8b0b-e92a841c0ef2 details[0].UEFI=LEGACY 
serviceofferingid=88de7d7a-02c1-4a51-831f-8b96ce2227ae
{
  "account": "admin",
  "accountid": "1c4424d2-f4e8-11f0-888d-df0da2073b14",
  "cmd": "org.apache.cloudstack.api.command.admin.vm.ImportVmCmd",
  "completed": "2026-02-05T18:38:48-0800",
  "created": "2026-02-05T18:38:48-0800",
  "domainid": "deefed4b-f4e7-11f0-888d-df0da2073b14",
  "domainpath": "ROOT",
  "jobid": "3f4f6be4-f67a-4673-9678-e49bec95c701",
  "jobprocstatus": 0,
  "jobresult": {
    "errorcode": 530,
    "errortext": "Import failed for Vm: i-2-54-VM. Suitable deployment 
destination not found"
  },
  "jobresultcode": 530,
  "jobresulttype": "object",
  "jobstatus": 2,
  "userid": "1c4465a9-f4e8-11f0-888d-df0da2073b14"
}
🙈 Error: async API failed for job 3f4f6be4-f67a-4673-9678-e49bec95c701



I also tried to do this through the web interface but when I go to this section:

See attached screenshot.  Storage Pool doesn’t give me any options.




Thank you
-jeremy



On Wednesday, Feb 04, 2026 at 8:37 PM, Jithin Raju 
<[email protected]<mailto:[email protected]>> wrote:
You should be able to import the QCOW2 files from the backup on one region to 
another region using KVM import methods : Import QCOW2 image from local/Shared 
storage [1]

You need to copy the QCOW2 files to the target Primary storage.

Cloudmonkey example :
cmk import vm \
name=<vm-name> \
displayname=<vm-display-name> \
importsource=shared \
hypervisor=kvm \
storageid=<primary-storage-uuid> \
diskpath=<disk-file.qcow2> \
networkid=<guest-network-uuid> \
serviceofferingid=<service-offering-uuid> \
migrateallowed=true \
zoneid=<zone-uuid> \
clusterid=<cluster-uuid> \
details[0].UEFI=LEGACY

Note: optionally pass templateid if required.

cmk import volume \
name=<volume-name> \
zoneid=<zone-uuid> \
diskofferingid=<disk-offering-uuid> \
storageid=<primary-storage-uuid> \
diskpath=<data-disk.qcow2> \
format=QCOW2

cmk attach volume \
id=<volume-uuid> \
virtualmachineid=<vm-uuid>

[1] https://www.shapeblue.com/kvm-import/


-Jithin

From: [email protected] <[email protected]>
Date: Thursday, 5 February 2026 at 7:34 AM
To: [email protected] <[email protected]>
Subject: Re: Adding an existing cluster as a new region?

Am I off base here? Just trying to get some clarification. Can a standalone 
cluster be converted to a region so I would be able to restore from a backup 
made in a different location?

Thanks



On Wednesday, Feb 04, 2026 at 12:18 PM, 
<[email protected]<mailto:[email protected]>> wrote:
Is this just not possible with Cloudstack?

Is there an alternative to backing up a vm and being able to restore it in a 
totally different standalone cluster?

My overall goal is to be able to move these vm to a new cluster that has no 
relation to the other.

Thanks.



On Tuesday, Feb 03, 2026 at 9:57 PM, 
<[email protected]<mailto:[email protected]>> wrote:
I’m trying to test the backup and restore process in 4.22. As of right now I 
have two completely separate 4.22 clusters in two separate locations. To use 
the cross region backup and restore features in 4.22, it seems I need to link 
the cluster or add the second cluster as a region.

Docs say to specify the region during the second cluster’s db creation. What if 
the second cluster was already built and running?

Any advice is greatly appreciated.

-jeremy


Reply via email to