Ok it looks like Qemu is unable to access the instance state.
could you perform a $ virsh list --all from the second node and tell me what you see ?
as for the second message, make sure you installed the "dbus" package

Regards,
Razique

Nuage & Co - Razique Mahroua 

Le 10 juil. 2012 à 11:29, Leander Bessa Beernaert a écrit :

Well i've checked the libvirt logs on both nodes and i found these two lines:

2012-07-09 13:58:27.179+0000: 10227: warning : qemuDomainObjTaint:1134 : Domain id=2 name='instance-00000002' uuid=57aca8a6-d062-4a08-8d87-e4d11d259ac7 is tainted: high-privileges
2012-07-09 13:58:27.736+0000: 10226: error : qemuMonitorIORead:513 : Unable to read from monitor: Connection reset by peer

The log is alos filled with the message below, it gets repeated over and over and over.

2012-07-10 09:26:02.244+0000: 10229: error : virExecWithHook:328 : Cannot find 'pm-is-supported' in path: No such file or directory
2012-07-10 09:26:02.244+0000: 10229: warning : qemuCapsInit:856 : Failed to get host power management capabilities

On Tue, Jul 10, 2012 at 8:16 AM, Razique Mahroua <[email protected]> wrote:
Hi Leander,
try to check libvirtd.log files, 
is the instance still running on the first node while you are lauching the migration process ?

Razique
Nuage & Co - Razique Mahroua 
<NUAGECO-LOGO-Fblan_petit.jpg>

Le 9 juil. 2012 à 16:09, Leander Bessa Beernaert a écrit :

Ok, so i've updated to the test packages from 

The migration still fails, but i see no errors in the logs. I'm trying to migrate a VM with the m1.tiny flavor from one machine to another. Their hardware are identical and they have more than enough resources to support the m1.tiny flavor:

cloud35         (total)                         4    3867     186
cloud35         (used_now)                      0     312       5
cloud35         (used_max)                      0       0       0

These are the logs from the origin compute node:  http://paste.openstack.org/show/19319/  and  the destination compute node: http://paste.openstack.org/show/19318/ . The scheduler's log has no visible errors or stack traces.

I'm still using nfsv4. 

Any ideas?


On Fri, Jul 6, 2012 at 7:57 PM, Leander Bessa Beernaert <[email protected]> wrote:
Thanks for the tip, it's a better than nothing :)

Regards,
Leander

On Fri, Jul 6, 2012 at 6:32 PM, Mandar Vaze / मंदार वझे <[email protected]> wrote:
Not sure if you are able to debug this, but a while ago there was a bug where instance.id was passed where instance.uuid was expected. This used to cause some problem.
It looks like you are using distribution package rather than devstack installation, so it is likely that the issue is now fixed. Can you try latest packages (and/or try devstack if you can)

I wish I could help more.

-Mandar


On Fri, Jul 6, 2012 at 3:26 PM, Leander Bessa Beernaert <[email protected]> wrote:
Hello,

I've recently setup a system to test out the live migration feature. So far i've been able to launch the instances with the shared nfs folder. However, when i run the live-migration command i encounter this error in the destination compute node:

2012-07-05 09:33:48 ERROR nova.manager [-] Error during ComputeManager.update_available_resource: Domain not found: no domain with matching id 2
2012-07-05 09:33:48 TRACE nova.manager Traceback (most recent call last):
2012-07-05 09:33:48 TRACE nova.manager   File "/usr/lib/python2.7/dist-packages/nova/manager.py", line 155, in periodic_tasks
2012-07-05 09:33:48 TRACE nova.manager     task(self, context)
2012-07-05 09:33:48 TRACE nova.manager   File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 2409, in update_available_resource
2012-07-05 09:33:48 TRACE nova.manager     self.driver.update_available_resource(context, self.host)
2012-07-05 09:33:48 TRACE nova.manager   File "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/connection.py", line 1936, in update_available_resource
2012-07-05 09:33:48 TRACE nova.manager     'vcpus_used': self.get_vcpu_used(),
2012-07-05 09:33:48 TRACE nova.manager   File "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/connection.py", line 1743, in get_vcpu_used
2012-07-05 09:33:48 TRACE nova.manager     dom = self._conn.lookupByID(dom_id)
2012-07-05 09:33:48 TRACE nova.manager   File "/usr/lib/python2.7/dist-packages/libvirt.py", line 2363, in lookupByID
2012-07-05 09:33:48 TRACE nova.manager     if ret is None:raise libvirtError('virDomainLookupByID() failed', conn=self)
2012-07-05 09:33:48 TRACE nova.manager libvirtError: Domain not found: no domain with matching id 2

Any ideas on how to solve this?

Regards,
Leander

_______________________________________________
Mailing list: https://launchpad.net/~openstack
Post to     : [email protected]
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp




_______________________________________________
Mailing list: https://launchpad.net/~openstack
Post to     : [email protected]
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp



_______________________________________________
Mailing list: https://launchpad.net/~openstack
Post to     : [email protected]
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp

Reply via email to