This worked out quite well. Once again, thank you for your help. > On Thursday, Feb 05, 2026 at 9:45 PM, <[email protected] > (mailto:[email protected])> wrote: > When I leave out the “details[0].UEFI=LEGACY”, the job seems to complete. I > haven’t tried to boot the instance yet because inadvertently removed all the > instances. Running the cloudmonkey command created a long list of instances > despite the error. I cleaned up and it removed the qcow2 file, which I guess > is a good thing. It means things were in the right place. > > Recopying the qcow2 file and I will try to start the instance again. > > Thanks for your help. > > > > > On Thursday, Feb 05, 2026 at 8:59 PM, Jithin Raju > > <[email protected] (mailto:[email protected])> wrote: > > Hi Jeremy, > > > > The CMK looks fine, you need to review the logs to find what caused the > > import failure. [1] > > In the UI , you may want to change the selection to shared storage instead > > of local storage. > > > > [1] > > https://docs.cloudstack.apache.org/en/4.22.0.0/adminguide/troubleshooting.html > > > > -Jithin > > > > > > > > From: Jeremy Hansen <[email protected]> > > Date: Friday, 6 February 2026 at 8:13 AM > > To: Jithin Raju <[email protected]> > > Subject: Re: Adding an existing cluster as a new region? > > > > Thank you so much for your help. I just want to clarify that I’m going > > throught the proper steps. > > > > I placed the qcow here: > > > > nas.sifu.intra:/volume1/cloud/primary 42T 18T 25T 42% > > /mnt/630a7d01-1f6d-36f1-b587-1f561ccc82dd/8629c278-8eae-4c92-a3b6-831ab20a9c2b > > > > Then I ran this command: > > > > cmk -p la1 > > Apache CloudStack 🐵 CloudMonkey 6.4.0-rc > > Report issues: https://github.com/apache/cloudstack-cloudmonkey/issues > > > > (la1) 🐱 > import vm name=ca-fs-corp displayname=ca-fs-corp > > importsource=shared hypervisor=kvm > > storageid=630a7d01-1f6d-36f1-b587-1f561ccc82dd > > diskpath=8629c278-8eae-4c92-a3b6-831ab20a9c2b > > networkid=d3619d80-11cd-483b-86f2-3245c688a52c migrateallowed=true > > zoneid=758cd49e-a193-4445-a8a9-576d16f4c24b > > clusterid=38e832ba-8837-4aa6-8b0b-e92a841c0ef2 details[0].UEFI=LEGACY > > serviceofferingid=88de7d7a-02c1-4a51-831f-8b96ce2227ae > > > > > > > > Which results in this: > > > > import vm name=ca-fs-corp displayname=ca-fs-corp importsource=shared > > hypervisor=kvm storageid=630a7d01-1f6d-36f1-b587-1f561ccc82dd > > diskpath=8629c278-8eae-4c92-a3b6-831ab20a9c2b > > networkid=d3619d80-11cd-483b-86f2-3245c688a52c migrateallowed=true > > zoneid=758cd49e-a193-4445-a8a9-576d16f4c24b > > clusterid=38e832ba-8837-4aa6-8b0b-e92a841c0ef2 details[0].UEFI=LEGACY > > serviceofferingid=88de7d7a-02c1-4a51-831f-8b96ce2227ae > > { > > "account": "admin", > > "accountid": "1c4424d2-f4e8-11f0-888d-df0da2073b14", > > "cmd": "org.apache.cloudstack.api.command.admin.vm.ImportVmCmd", > > "completed": "2026-02-05T18:38:48-0800", > > "created": "2026-02-05T18:38:48-0800", > > "domainid": "deefed4b-f4e7-11f0-888d-df0da2073b14", > > "domainpath": "ROOT", > > "jobid": "3f4f6be4-f67a-4673-9678-e49bec95c701", > > "jobprocstatus": 0, > > "jobresult": { > > "errorcode": 530, > > "errortext": "Import failed for Vm: i-2-54-VM. Suitable deployment > > destination not found" > > }, > > "jobresultcode": 530, > > "jobresulttype": "object", > > "jobstatus": 2, > > "userid": "1c4465a9-f4e8-11f0-888d-df0da2073b14" > > } > > 🙈 Error: async API failed for job 3f4f6be4-f67a-4673-9678-e49bec95c701 > > > > > > > > I also tried to do this through the web interface but when I go to this > > section: > > > > See attached screenshot. Storage Pool doesn’t give me any options. > > > > > > > > > > Thank you > > -jeremy > > > > > > > > > On Wednesday, Feb 04, 2026 at 8:37 PM, Jithin Raju > > > <[email protected] (mailto:[email protected])> wrote: > > > You should be able to import the QCOW2 files from the backup on one > > > region to another region using KVM import methods : Import QCOW2 image > > > from local/Shared storage [1] > > > > > > You need to copy the QCOW2 files to the target Primary storage. > > > > > > Cloudmonkey example : > > > cmk import vm \ > > > name=<vm-name> \ > > > displayname=<vm-display-name> \ > > > importsource=shared \ > > > hypervisor=kvm \ > > > storageid=<primary-storage-uuid> \ > > > diskpath=<disk-file.qcow2> \ > > > networkid=<guest-network-uuid> \ > > > serviceofferingid=<service-offering-uuid> \ > > > migrateallowed=true \ > > > zoneid=<zone-uuid> \ > > > clusterid=<cluster-uuid> \ > > > details[0].UEFI=LEGACY > > > > > > Note: optionally pass templateid if required. > > > > > > cmk import volume \ > > > name=<volume-name> \ > > > zoneid=<zone-uuid> \ > > > diskofferingid=<disk-offering-uuid> \ > > > storageid=<primary-storage-uuid> \ > > > diskpath=<data-disk.qcow2> \ > > > format=QCOW2 > > > > > > cmk attach volume \ > > > id=<volume-uuid> \ > > > virtualmachineid=<vm-uuid> > > > > > > [1] https://www.shapeblue.com/kvm-import/ > > > > > > > > > -Jithin > > > > > > From: [email protected] <[email protected]> > > > Date: Thursday, 5 February 2026 at 7:34 AM > > > To: [email protected] <[email protected]> > > > Subject: Re: Adding an existing cluster as a new region? > > > > > > Am I off base here? Just trying to get some clarification. Can a > > > standalone cluster be converted to a region so I would be able to restore > > > from a backup made in a different location? > > > > > > Thanks > > > > > > > > > > > > On Wednesday, Feb 04, 2026 at 12:18 PM, > > > <[email protected]<mailto:[email protected]>> wrote: > > > Is this just not possible with Cloudstack? > > > > > > Is there an alternative to backing up a vm and being able to restore it > > > in a totally different standalone cluster? > > > > > > My overall goal is to be able to move these vm to a new cluster that has > > > no relation to the other. > > > > > > Thanks. > > > > > > > > > > > > On Tuesday, Feb 03, 2026 at 9:57 PM, > > > <[email protected]<mailto:[email protected]>> wrote: > > > I’m trying to test the backup and restore process in 4.22. As of right > > > now I have two completely separate 4.22 clusters in two separate > > > locations. To use the cross region backup and restore features in 4.22, > > > it seems I need to link the cluster or add the second cluster as a region. > > > > > > Docs say to specify the region during the second cluster’s db creation. > > > What if the second cluster was already built and running? > > > > > > Any advice is greatly appreciated. > > > > > > -jeremy > > > > > >
signature.asc
Description: PGP signature
