Cloudstack will create the old bridges as it used to, you just need to have
a single 'cloudVirBr' device to trigger the behavior.
On Aug 5, 2013 6:35 PM, "Simon Weller" <swel...@ena.com> wrote:

> Thanks Marcus. We already thought of creating the old bridge, but we were
> hoping we had better options ;-)
>
>
>
> ----- Original Message -----
>
> From: "Marcus Sorensen" <shadow...@gmail.com>
> To: dev@cloudstack.apache.org
> Cc: cloudstack-...@incubator.apache.org
> Sent: Monday, August 5, 2013 5:42:38 PM
> Subject: Re: 4.1.x KVM cloudVirBr<vlan> to br<dev>-<vlan>
>
> Yes, the vm definition already has the bridge name chosen (on initial
> startup), and that definition is migrated between hosts. This is
> further exacerbated by the fact that the bridges are created on the
> destination host prior to migration (so they'll be ready), rather than
> somehow being able to see the existing configuration and prepare for
> the vm based on that. That ultimately probably doesn't matter much
> anyway, since we can't host two different bridges for the same vlan
> (the advantage of detecting what bridge name a migrating VM has would
> be to bring up the required bridge name for migrating VMs, and use the
> new bridge name for freshly started VMs, which isn't possible).
>
> As a workaround, if you want to force any rebooted, upgraded KVM hosts
> to use the old style naming, you can create any random bridge named
> 'cloudVirBr', and the agent will detect it and continue to use the old
> style naming until such point when you can or need to switch to the
> capabilities of the new style naming. At that point you'll need to
> stop any and all VMs that are using the old style name, remove any old
> style bridges (or reboot the host), and then start things back up.
>
> This really should have been documented in the release notes. I think
> it was a misunderstanding that the upgrade process would require
> everything be restarted.
>
> On Mon, Aug 5, 2013 at 4:05 PM, Simon Weller <swel...@ena.com> wrote:
> > All,
> >
> > As most know, the upgrade from 4.0 to 4.1 changed the interface naming
> schema. When a host in a cluster is rebooted, the interface naming changes.
> When this occurs, live migration breaks to that host.
> >
> > Example config:
> > All Management and hosts running CS 4.1.1
> > Hypervisor: KVM on RHEL 6.3
> > Host 1 has older 4.0 interface naming schema
> > Host 2 was rebooted and has newer interface schema
> >
> > Live Migration is looking for older interface schema name (i.e.
> cloudVirBr<vlan>) when attempting a migration from Host 1 to Host 2.
> >
> > Here's a sample log:
> >
> >
> > 2013-08-05 16:45:21,846 DEBUG [agent.transport.Request]
> (Job-Executor-34:job-1660) Seq 19-1921285594: Sending { Cmd , MgmtId:
> 159090354809909, via: 19, Ver: v1, Flags: 100111,
> [{"MigrateCommand":{"vmName":"i-44-255-VM","destIp":"<hostip>","hostGuid":"91e9b633-f46b-31f3-9a4b-92285fd94b62-LibvirtComputingResource","isWindows":false,"wait":0}}]
> }
> > 2013-08-05 16:45:21,926 DEBUG [agent.transport.Request]
> (StatsCollector-1:null) Seq 1-1768126050: Received: { Ans: , MgmtId:
> 159090354809909, via: 1, Ver: v1, Flags: 10, { GetVmStatsAnswer } }
> > 2013-08-05 16:45:21,963 DEBUG [agent.manager.AgentManagerImpl]
> (AgentManager-Handler-7:null) Ping from 5
> > 2013-08-05 16:45:22,012 DEBUG [agent.transport.Request]
> (AgentManager-Handler-9:null) Seq 19-1921285594: Processing: { Ans: ,
> MgmtId: 159090354809909, via: 19, Ver: v1, Flags: 110,
> [{"MigrateAnswer":{"result":false,"details":"Cannot get interface MTU on
> 'cloudVirBr18': No such device","wait":0}}] }
> > 2013-08-05 16:45:22,012 DEBUG [agent.manager.AgentAttache]
> (AgentManager-Handler-9:null) Seq 19-1921285594: No more commands found
> > 2013-08-05 16:45:22,012 DEBUG [agent.transport.Request]
> (Job-Executor-34:job-1660) Seq 19-1921285594: Received: { Ans: , MgmtId:
> 159090354809909, via: 19, Ver: v1, Flags: 110, { MigrateAnswer } }
> > 2013-08-05 16:45:22,012 ERROR [cloud.vm.VirtualMachineManagerImpl]
> (Job-Executor-34:job-1660) Unable to migrate due to Cannot get interface
> MTU on 'cloudVirBr18': No such device
> > 2013-08-05 16:45:22,013 INFO [cloud.vm.VirtualMachineManagerImpl]
> (Job-Executor-34:job-1660) Migration was unsuccessful. Cleaning up:
> VM[User|app01-dev]
> > 2013-08-05 16:45:22,018
> >
> > Is there any current way to change the destination network CS Management
> uses so that a complete VM shutdown and restart isn't required to re-enable
> migration between hosts?
> >
> > Any ideas would be appreciated.
> >
> > - Si
>
>

Reply via email to