23 om 16:57
> Aan: users@cloudstack.apache.org
> Onderwerp: Re: Documentation on instances live migration with KVM
> :D
> looking forward to your doc PR ;)
>
> On Mon, Nov 20, 2023 at 2:34 PM Jimmy Huybrechts wrote:
> >
> > I noticed the documentation on instances is
I actually have never created a PR before ;) Where do I start?
--
Jimmy
Van: Daan Hoogland
Datum: maandag, 20 november 2023 om 16:57
Aan: users@cloudstack.apache.org
Onderwerp: Re: Documentation on instances live migration with KVM
:D
looking forward to your doc PR ;)
On Mon, Nov 20, 2023
:D
looking forward to your doc PR ;)
On Mon, Nov 20, 2023 at 2:34 PM Jimmy Huybrechts wrote:
>
> I noticed the documentation on instances is not entirely correct as it it
> says:
>
> (KVM) The Instance must not be using local disk storage. (On XenServer and
> VMware, Insta
I noticed the documentation on instances is not entirely correct as it it says:
(KVM) The Instance must not be using local disk storage. (On XenServer and
VMware, Instance live migration with local disk is enabled by CloudStack
support for XenMotion and vMotion.)
Well, I’m using local storage
:34 AM
To: users@cloudstack.apache.org
Subject: Re: preventing VM Live migration between Pods
Hi Gary
The other 2 settings as mentioned by Simon should do the trick. We are doing
the same as a safeguard albeit to a different problem.
Regards,
Shiv
(Sent from mobile device. Please excuse brevity
,
> distribute or take any action or reliance on its contents. Please destroy
> any hard copies and delete this message.
> -Original Message-
> From: Simon Weller
> Sent: Tuesday, May 16, 2023 4:12 PM
> To: users@cloudstack.apache.org
> Subject: Re: preventing VM Live
s. Please destroy any hard copies and
delete this message.
-Original Message-
From: Simon Weller
Sent: Tuesday, May 16, 2023 4:12 PM
To: users@cloudstack.apache.org
Subject: Re: preventing VM Live migration between Pods
Gary,
There are some global settings you can enable/disable to preve
:*Granwille Strauss
*Sent:* Tuesday, May 16, 2023 3:51 PM
*To:* users@cloudstack.apache.org
*Cc:* gary.di...@quadris.co.uk.INVALID
*Subject:* Re: preventing VM Live migration between Pods
Hi Gary
I am still fairly new to ACS myself, but as far as I can recall, using
the 'host-passthrough' option
across primary
storage
enable.ha.storage.migration - Enable/disable storage migration across
primary storage during HA
-Si
On Tue, May 16, 2023 at 8:13 AM Gary Dixon
wrote:
> Hi everyone
>
>
>
> Other than disabling a Pod – is there a way to prevent live migration of
> VM’s b
...@quadris.co.uk.INVALID
Subject: Re: preventing VM Live migration between Pods
Hi Gary
I am still fairly new to ACS myself, but as far as I can recall, using the
'host-passthrough' option is prone to cause problems during migrations, this is
also mentioned in the documentation:
https
Other than disabling a Pod – is there a way to prevent live migration
of VM’s between Pods in ACS ?
We are on version 4.15.2 with Ubuntu 20.04 KVM hosts. Each Pod
contains a single cluster of Homogenous hosts – however there are only
slight differences between the CPU’s on the physical hosts
Hi everyone
Other than disabling a Pod – is there a way to prevent live migration of VM’s
between Pods in ACS ?
We are on version 4.15.2 with Ubuntu 20.04 KVM hosts. Each Pod contains a
single cluster of Homogenous hosts – however there are only slight differences
between the CPU’s
Hey,
That's right, recreating systemvm fixed the problem.
Regards,
Piotr
-Original Message-
From: Rohit Yadav
Sent: Tuesday, January 17, 2023 8:59 AM
To: users@cloudstack.apache.org
Subject: Re: sysvm live migration 4.17.2
Hi Piotr,
You've hit a bug, wherein somehow the systemvm.iso
.
From: Wei ZHOU
Sent: Sunday, January 15, 2023 04:09
To: users@cloudstack.apache.org
Subject: Re: sysvm live migration 4.17.2
Looks like a bug.
systemvm.iso does not exist on 4.17+, it should be ejected when you
patched the system vms.
Can you retry after ejecting iso
.2 I have a strange symptom, live
> migration system vm is not working. It ends with the message:
>
> Exception during migrate: org.libvirt.LibvirtException: Cannot access
> storage file '/usr/share/cloudstack-common/vms/systemvm.iso': No such
> file or directory
>
> In the director
Hi,
After upgrading CS from 4.17.0 to 4.17.2 I have a strange symptom, live
migration system vm is not working. It ends with the message:
Exception during migrate: org.libvirt.LibvirtException: Cannot access storage
file '/usr/share/cloudstack-common/vms/systemvm.iso': No such file
cs/bonds.
> >
> > Can you deploy VMs to the new nodes?
> >
> > Regards,
> > Alex
> >
> >
> >
> >
> > -----Original Message-
> > From: S.Fuller
> > Sent: 28 October 2022 17:36
> > To: users@cloudstack.apache.org
> >
deploy VMs to the new nodes?
>
> Regards,
> Alex
>
>
>
>
> -Original Message-
> From: S.Fuller
> Sent: 28 October 2022 17:36
> To: users@cloudstack.apache.org
> Subject: Live Migration fails - Cannot get interface MTU - No such device
>
> I'm wor
ards,
Alex
-Original Message-
From: S.Fuller
Sent: 28 October 2022 17:36
To: users@cloudstack.apache.org
Subject: Live Migration fails - Cannot get interface MTU - No such device
I'm working on migrating an existing cluster to new servers, I have two new
servers, which I have tested by a
he bridge networks, plus one for storage. The new servers have two
bonded interfaces (each with two physical nics assigned) which are then
assigned to the bridged networks.
Seeing as how the live migration worked between the new nodes in their own
cluster, as well as from the new nodes to the old nodes
I found the answer here:
https://dacat.cc/1969.html
I don’t understand why this was an issue though. First time I’ve encountered
this and I haven’t specified a UUID on any other nodes.
-jeremy
> On Wednesday, Sep 01, 2021 at 7:54 PM, Jeremy Hansen (mailto:jer...@skidrow.la)> wrote:
> I
I noticed trying to move a VM between specific hosts results in this error:
Exception during migrate: org.libvirt.LibvirtException: internal error: Attempt
to migrate guest to the same host 03000200-0400-0500-0006-000700080009
2021-09-01 19:51:44,166 - Executing command: get_rule_logs_for_vms
local storage from host to host (whole VM with its disks) - @Gabriel
>> Beims Bräscher can confirm this, afaik?
>>
>> If you are using Ubuntu - all fine - qemu-kvm supports live storage
>> migrations from Ubuntu 14.04 at least, an onwards.
>> If you are using CentOS 7, you
supports live
storage migrations from Ubuntu 14.04 at least, an onwards.
If you are using CentOS 7, you have to use qemu-kvm-ev from the
oVirt repo ONLY - all other versions of qemu-kvm do NOT support
storage live migration (Redhat revoked it for $$$ reasons, while
it was working fine
om host to host (whole VM with its disks) - @Gabriel
> Beims Bräscher can confirm this, afaik?
>
> If you are using Ubuntu - all fine - qemu-kvm supports live storage
> migrations from Ubuntu 14.04 at least, an onwards.
> If you are using CentOS 7, you have to use qemu-kvm-ev
fine - qemu-kvm supports live storage
migrations from Ubuntu 14.04 at least, an onwards.
If you are using CentOS 7, you have to use qemu-kvm-ev from the oVirt
repo ONLY - all other versions of qemu-kvm do NOT support storage live
migration (Redhat revoked it for $$$ reasons, while it was working
ds.
If you are using CentOS 7, you have to use qemu-kvm-ev from the oVirt repo
ONLY - all other versions of qemu-kvm do NOT support storage live migration
(Redhat revoked it for $$$ reasons, while it was working fine in CentOS6)
If your tested it and it worked from Gluster to NFS - that's (great) news
(for
storage.
That is indeed not supported with KVM and also not on the roadmap at the
moment.
Wido
Best regards.
Em 26/05/2021 04:08, Wido den Hollander escreveu:
On 25/05/2021 13:32, Kalil de Albuquerque Carvalho wrote:
Hello all.
Reading the manual I discovery that live migration is not support
.
Reading the manual I discovery that live migration is not support for
KVM hypervisor. I was wander if there are studies or predictions for
this features on KVM hosts.
Where did you read this? Live Migration with the KVM hypervisor works
just fine.
Wido
Yet on the manual citation, it said
On 25/05/2021 13:32, Kalil de Albuquerque Carvalho wrote:
Hello all.
Reading the manual I discovery that live migration is not support for
KVM hypervisor. I was wander if there are studies or predictions for
this features on KVM hosts.
Where did you read this? Live Migration
escreveu:
Hello all.
Reading the manual I discovery that live migration is not support for
KVM hypervisor. I was wander if there are studies or predictions for
this features on KVM hosts.
Yet on the manual citation, it said that migration just can occur from
CEPH/NFS to "SolidFire Ma
Hello all.
Reading the manual I discovery that live migration is not support for
KVM hypervisor. I was wander if there are studies or predictions for
this features on KVM hosts.
Yet on the manual citation, it said that migration just can occur from
CEPH/NFS to "SolidFire Managed St
if there is different storage involve, expecting there wont be live
migration , there is time need to move the data from one Storage to another
Storage.
On Sat, Jan 9, 2021 at 4:07 AM Vivek Kumar
wrote:
> U have created the PCS cluster with GFS2 and mounted across the all hosts,
> that’s
>> e-mail and attachments for viruses, it does not guarantee that both are
>>> virus-free and accepts no liability for any damage sustained as a result
>> of
>>> viruses.
>>>
>>>> On 08-Jan-2021, at 7:47 PM, Hean Seng wrote:
>>>&g
achments for viruses, it does not guarantee that both are
> > virus-free and accepts no liability for any damage sustained as a result
> of
> > viruses.
> >
> > > On 08-Jan-2021, at 7:47 PM, Hean Seng wrote:
> > >
> > > are you running share stora
a result of
> viruses.
>
> > On 08-Jan-2021, at 7:47 PM, Hean Seng wrote:
> >
> > are you running share storage or local storage.
> >
> > Live migration work, i tried it on NFS.
> >
> > It not work on LocalStorage, it is understood .
> >
&
that both are virus-free and accepts no
liability for any damage sustained as a result of viruses.
> On 08-Jan-2021, at 7:47 PM, Hean Seng wrote:
>
> are you running share storage or local storage.
>
> Live migration work, i tried it on NFS.
>
> It not work on
are you running share storage or local storage.
Live migration work, i tried it on NFS.
It not work on LocalStorage, it is understood .
On Fri, Jan 8, 2021 at 8:47 PM Vivek Kumar
wrote:
> Hello Guys,
>
> Can we do storage live migration on KVM environment ? In ACS, it’s giving
>
Hello Guys,
Can we do storage live migration on KVM environment ? In ACS, it’s giving the
option only after shutting down the VM.
Vivek Kumar
The only problem is with live migration and when cpu frequency is set to
the highest rate possible, while new cpu has lower rate.
Other problem is when passthrough is set and cpus belong to various
families like x5600 and E5-2600 V4. Then you should avoid passthrough and
set in agent config
you might be good even with slighly older
> > CPU on this new hosts that you want to add to a cluster.
> >
> > Do some reading around libvirt CPU model and simply just add one host and
> > test live migration - as simple as that.
> >
> > Kind regards,
> >
using some level of emulation of CPU model
(versus hos-passthrough), then you might be good even with slighly older
CPU on this new hosts that you want to add to a cluster.
Do some reading around libvirt CPU model and simply just add one host and
test live migration - as simple as that.
Kind regards
on this new hosts that you want to add to a cluster.
Do some reading around libvirt CPU model and simply just add one host and
test live migration - as simple as that.
Kind regards,
Andrija
On Tue, 19 Feb 2019 at 12:39, Piotr Pisz wrote:
> Hi Ugo,
>
> [root@cstack-1 ~]# cat /etc/cloudst
).
Regards,
Piotr
-Original Message-
From: Ugo Vasi
Sent: Tuesday, February 19, 2019 12:27 PM
To: users@cloudstack.apache.org
Subject: CPU live migration question
Hi all,
I would like to know if it is possible to add to an existing cluster a host
that has the cpu different from
Hi all,
I would like to know if it is possible to add to an existing cluster a
host that has the cpu different from that of the pre-existing hosts or
which features must be the same.
I have read around that there may be problems at the libvirt level but
it is not clear to me which parameters
Sent: Wednesday, September 19, 2018 7:46:40 PM
To: users@cloudstack.apache.org
Subject: Live Migration From VMware to KVM?
Hello folks,
Is it possible to add a VMware datacenter and live migrate all VMs from
VMware host to a KVM host?
Thanks in advance.
rohit.ya...@shapeblue.com
www.shapeblue.com
A
Hello,
Live Migratiom across different hypervisors isn't possible, you can use
third party tools.
ср, 19 сент. 2018 г., 17:16 Fariborz Navidan :
> Hello folks,
>
> Is it possible to add a VMware datacenter and live migrate all VMs from
> VMware host to a KVM host?
>
> Thanks in advance.
>
Live migration across different hypervisors is not possible. You could use a
tool such as Veeam agent to backup the running VM and then import it as an
image to KVM and start it.
As the hypervisors use different paravirtualized drivers, you're going to need
to do some work to be able
Hello folks,
Is it possible to add a VMware datacenter and live migrate all VMs from
VMware host to a KVM host?
Thanks in advance.
[mailto:m...@swen.io]
Gesendet: Dienstag, 10. April 2018 09:29
An: users@cloudstack.apache.org
Betreff: XenServer Storage Live Migration between pools
Hi all,
we added a new XenServer 6.5 SP1 cluster (pool) to our CS installation. When
we do a live migration between pools we run into a problem
Hi all,
we added a new XenServer 6.5 SP1 cluster (pool) to our CS installation. When
we do a live migration between pools we run into a problem with the VM. The
migration itself is working fine and the VM is reachable all the time. I can
see in XenCenter that the same network has been created
s on my installation, without recompiling jar
>> files?
>>
>> Regards,
>> IT engineer
>> Farheap, Russia
>> Ivan Derbenev
>>
>> -Original Message-
>> From: Koushik Das [mailto:koushik@accelerite.com]
>> Sent: Tuesday, May 24, 2
this on?
>>>>>>
>>>>>> Thanks,
>>>>>>
>>>>>> - Dave
>>>>>>
>>>>>>> On May 24, 2016, at 10:57 AM, Dave Garbus <davegar...@spiceworks.com>
>>>>
gt;>>>>>
>>>>>> I updated the issue referenced in this email (
>>>> https://issues.apache.org/jira/browse/CLOUDSTACK-9353) to include
>>>> CloudStack 4.8 as an affected version. If we could at least get someone
>>>>
rowse/CLOUDSTACK-9353) to include
> >> CloudStack 4.8 as an affected version. If we could at least get someone
> >> assigned to it, that’d be a good start.
> >>>>
> >>>> Thanks,
> >>>>
> >>>> - Dave
> >>>>
> >>>
t;>> Thanks,
>>>>
>>>> - Dave
>>>>
>>>> On May 24, 2016, at 9:42 AM, Ivan Derbenev <
>> ivan.derbe...@tech-corps.com<mailto:ivan.derbe...@tech-corps.com>> wrote:
>>>>
>>>> Is there a way I ca
iles?
> >>
> >> Regards,
> >> IT engineer
> >> Farheap, Russia
> >> Ivan Derbenev
> >>
> >> -Original Message-
> >> From: Koushik Das [mailto:koushik@accelerite.com]
> >> Sent: Tuesday, May 24, 2
ivan.derbe...@tech-corps.com>> wrote:
>>
>> Is there a way I can fix this on my installation, without recompiling jar
>> files?
>>
>> Regards,
>> IT engineer
>> Farheap, Russia
>> Ivan Derbenev
>>
>> -----Original Message-
>
ineer
> Farheap, Russia
> Ivan Derbenev
>
> -Original Message-
> From: Koushik Das [mailto:koushik@accelerite.com]
> Sent: Tuesday, May 24, 2016 2:06 PM
> To: users@cloudstack.apache.org<mailto:users@cloudstack.apache.org>
> Subject: Re: Xenserver 6.5 live migra
6 2:06 PM
> To: users@cloudstack.apache.org<mailto:users@cloudstack.apache.org>
> Subject: Re: Xenserver 6.5 live migration with local volumes
>
> This is a product bug and needs to be fixed. There is some serialization
> issue that needs fixing.
>
> In the be
Tuesday, May 24, 2016 2:06 PM
To: users@cloudstack.apache.org<mailto:users@cloudstack.apache.org>
Subject: Re: Xenserver 6.5 live migration with local volumes
This is a product bug and needs to be fixed. There is some serialization issue
that needs fixing.
In the below code volumeToFi
: Xenserver 6.5 live migration with local volumes
This is a product bug and needs to be fixed. There is some serialization issue
that needs fixing.
In the below code volumeToFiler is null as Map is not a serializable type. It
needs to be replaced with getVolumeToFilerAsList() and then the data
This is a product bug and needs to be fixed. There is some serialization issue
that needs fixing.
In the below code volumeToFiler is null as Map is not a serializable type. It
needs to be replaced with getVolumeToFilerAsList() and then the data needs to
be read appropriately.
final
Hello!
After the migration onto XS 6.5 (and installing all the updates) and CS 4.8 I
can't migrate VMs with Local Storage
The issue is similar to this:
https://issues.apache.org/jira/browse/CLOUDSTACK-9353, but it doesn't have any
solutions
Any ideas?
Errors:
2016-05-24 12:10:16,440 DEBUG
advice is helpful.
>
> 2016-05-05 21:55 GMT+08:00 Lv Haijiao <luhaij...@gmail.com>:
>
>> Hi, Dag
>>
>> Thanks for your hint !
>>
>> We don't have chance to reproduce this on our production platform, but
>> we notice most of these VMs hitting this issue
Hi Lv,
glad you got it sorted. Apologies - I missed your previous post - but DB
servers are where you typically see these problems first since they tend to be
I/O intensive.
You are right in that shared storage allows for live migration between
hypervisor hosts without the underlying disks
e chance to reproduce this on our production platform, but we
> notice most of these VMs hitting this issue are DB server.
>
> Even that, live migration is supposed to consume less bandwidth as
> majority of VM data is still on the shared storage, need no replication,
> except
imary storage
>>>>> device are marked as “removed”, so if I shutdown the instances, the
>>>>> cloudstack may kick off a storage cleanup job to remove them from new
>>>>> cluster’s primary storage before I can get the fixes in.
>>>>>
>&
Hi, Dag
Thanks for your hint !
We don't have chance to reproduce this on our production platform, but we
notice most of these VMs hitting this issue are DB server.
Even that, live migration is supposed to consume less bandwidth as majority
of VM data is still on the shared storage, need
ot;Lv Haijiao" <luhaij...@gmail.com> wrote:
>Hi,
>
>Here's our environment,
>
>
> - ACS 4.7.1
> - Ubuntu 14.04 KVM
> - NFS as shared, primary storage
>
>
>While we do a couple of VMs live migration in sequence from physical server
>A to B, the file sys
Hi,
Here's our environment,
- ACS 4.7.1
- Ubuntu 14.04 KVM
- NFS as shared, primary storage
While we do a couple of VMs live migration in sequence from physical server
A to B, the file system of some VMs goes into 'ready only' status unless
we reboot the VM again.
No clue why we
>
>>>> Is there a way to temporarily disable storage cleanups ?
>>>>
>>>> Yiping
>>>>
>>>>
>>>>
>>>>
>>>> On 5/4/16, 3:22 PM, "Yiping Zhang" <yzh...@marketo.com> wrote:
>>>
an get the fixes in.
>>>
>>> Is there a way to temporarily disable storage cleanups ?
>>>
>>> Yiping
>>>
>>>
>>>
>>>
>>> On 5/4/16, 3:22 PM, "Yiping Zhang" <yzh...@marketo.com> wrote:
>>>
>>
k may kick off a storage cleanup job to remove them from new
>> cluster’s primary storage before I can get the fixes in.
>>
>> Is there a way to temporarily disable storage cleanups ?
>>
>> Yiping
>>
>>
>>
>>
>> On 5/4/16, 3:22 PM,
> Is there a way to temporarily disable storage cleanups ?
>
> Yiping
>
>
>
>
> On 5/4/16, 3:22 PM, "Yiping Zhang" <yzh...@marketo.com> wrote:
>
>> Hi, all:
>>
>> I am in a situation that I need some help:
>>
>>
ping
>
>
>
>
> On 5/4/16, 3:22 PM, "Yiping Zhang" <yzh...@marketo.com> wrote:
>
> >Hi, all:
> >
> >I am in a situation that I need some help:
> >
> >I did a live migration with storage migration required for a production
> VM instance f
;
>I am in a situation that I need some help:
>
>I did a live migration with storage migration required for a production VM
>instance from one cluster to another. The first migration attempt failed
>after some time, but the second attempt succeeded. During all this time the VM
>
Hi, all:
I am in a situation that I need some help:
I did a live migration with storage migration required for a production VM
instance from one cluster to another. The first migration attempt failed after
some time, but the second attempt succeeded. During all this time the VM
instance
Hi,
we ran into a problem with storage live migration in CLoudstack and need your
help to verify this.
We are using XenServer with local storage.
Steps to reproduce:
1. create an instance from an ISO (not template!) on local storage
2. while instance is running migrate it to another XenServer
elweingart...@gmail.com]
Sent: Monday, October 12, 2015 10:12 PM
To: users@cloudstack.apache.org
Subject: [Questionable] Re: Timeout with live migration
There is your problem, there are currently two distinct values conrolling
those async jobs.
Change that value and everything will work for u.
Can you o
ils once I have it created?
From: Rafael Weingärtner [rafaelweingart...@gmail.com]
Sent: Tuesday, October 13, 2015 7:52 AM
To: users@cloudstack.apache.org
Subject: [Questionable] Re: [Questionable] Re: Timeout with live migration
I guess so, for s
Monday, October 12, 2015 10:12 PM
> To: users@cloudstack.apache.org
> Subject: [Questionable] Re: Timeout with live migration
>
> There is your problem, there are currently two distinct values conrolling
> those async jobs.
> Change that value and everything will work for u.
> Can you open a
;
> From: Rafael Weingärtner [rafaelweingart...@gmail.com]
> Sent: Tuesday, October 13, 2015 7:52 AM
> To: users@cloudstack.apache.org
> Subject: [Questionable] Re: [Questionable] Re: Timeout with live migration
>
> I guess so, for some r
, October 13, 2015 3:43 PM
To: users@cloudstack.apache.org
Subject: [Questionable] Re: Timeout with live migration
That is good. Now I can report to you what is in the code.
Let’s start:
First: when I looked at the problem at first time, I went straight to the
class that sends commands to Xen
agent.api.storage.MigrateVolumeAnswer":{"volumePath":"00db15be-3ccd-4648-8928-35ca90924d7c","result":true,"wait":0}}]
> }
> 2015-10-13 15:27:13,397 DEBUG [c.c.a.m.AgentAttache]
> (DirectAgent-430:ctx-ac6d7aeb) Seq 38-1788936343: No more commands found
> 2015-10-13 15:27:13,397 DEBUG [c.c.a.t.Requ
] Re: Timeout with live migration
Nice, thanks.
Did that solve your problem? Did you migrate the volume?
On Tue, Oct 13, 2015 at 7:00 PM, Ryan Farrington <rfarring...@remitdata.com>
wrote:
> Issue #1) Terrible parameter names
> https://issues.apache.org/jira/browse/CLOUDSTACK-8946
>
&
f volumes without issue. Thanks
> for helping nail this down!
>
>
> From: Rafael Weingärtner [rafaelweingart...@gmail.com]
> Sent: Tuesday, October 13, 2015 6:12 PM
> To: users@cloudstack.apache.org
> Subject: [Questionable] Re: Timeou
We are experiencing a failure in cloudstack waiting for an async job performing
a live migration of a volume to finish. I've copied the relevant log entries
below.We acknowledge that the migration will take a few hours based on the
volume of the data and we are looking for a way to increase
.f.j.i.AsyncJobMonitor]
(Job-Executor-63:ctx-f7b6817d) Remove job-5257 from job monitoring
____
From: Rafael Weingärtner [rafaelweingart...@gmail.com]
Sent: Monday, October 12, 2015 8:24 PM
To: users@cloudstack.apache.org
Subject: [Questionable] Re:
t setting is set
> to 7200 and increase it. If not you could also check
> job.cancel.threshold.minutes and job.expire.minutes.
>
> -Jan-Arve
>
> 2015-10-13 0:46 GMT+02:00 Ryan Farrington <rfarring...@remitdata.com>:
>
> > We are experiencing a failure in cloudstack wait
)
job.cancel.threshold.minutes = 60 (1 hour)
job.expire.minutes = 1440 (24 hours)
From: Jan-Arve Nygård [jan.arve.nyg...@gmail.com]
Sent: Monday, October 12, 2015 6:19 PM
To: users@cloudstack.apache.org
Subject: [Questionable] Re: Timeout with live migration
, October 12, 2015 6:40 PM
To: users@cloudstack.apache.org
Subject: [Questionable] Re: Timeout with live migration
I would first check your NICs' speed and load, the amount of RAM allocated
for the migrating VM and than check the hypervisor log files.
On Mon, Oct 12, 2015 at 8:19 PM, Jan-Arve
elweingart...@gmail.com]
Sent: Monday, October 12, 2015 8:46 PM
To: users@cloudstack.apache.org
Subject: [Questionable] Re: Timeout with live migration
I thought you using the command “migrateVirtualMachineWithVolume” but it
seems that you are using “migrateVolume” command from ACS's API.
For the code
l.com]
> Sent: Monday, October 12, 2015 6:40 PM
> To: users@cloudstack.apache.org
> Subject: [Questionable] Re: Timeout with live migration
>
> I would first check your NICs' speed and load, the amount of RAM allocated
> for the migrating VM and than check the hypervisor log fi
: Timeout with live migration
Are you live migrating a VM, or migrating a volume of a stopped VM to a
different primary storage?
If it is a running VM, is the VM allocated in a shared storage or local
storage?
On Mon, Oct 12, 2015 at 9:17 PM, Ryan Farrington <rfarring...@remitdata.com>
il.com]
Sent: Monday, October 12, 2015 7:53 PM
To: users@cloudstack.apache.org
Subject: [Questionable] Re: Timeout with live migration
what do you mean with livre migrating data volume ?!
I understand a live migration of a VM, but volumes...
do you mean live migrating a VM that has a volume attache
a failure in cloudstack waiting for an async job
> performing a live migration of a volume to finish. I've copied the relevant
> log entries below.We acknowledge that the migration will take a few hours
> based on the volume of the data and we are looking for a way to increase
> the ti
what do you mean with livre migrating data volume ?!
I understand a live migration of a VM, but volumes...
do you mean live migrating a VM that has a volume attached?
are you migrating that volume to a different cluster? or just a different
storage in the same cluster?
What hypervisor are you
Now I understand what you are doing, I am familiar with that concept (live
migration of VM within a cluster, having the VHD being moved from one SR to
another).
I just got confused when I read live migration of volumes (a volume does
not run by itself, so that why I asked a little for some more
onfigured to be 3600
>
>
>
>
> From: Rafael Weingärtner [rafaelweingart...@gmail.com]
> Sent: Monday, October 12, 2015 9:46 PM
> To: users@cloudstack.apache.org
> Subject: [Questionable] Re: Timeout with live migration
>
> I found somet
1 - 100 of 155 matches
Mail list logo