I forgot to ask, did you enable the vnc console? If so, with which parameters?
On Tue, Jul 10, 2012 at 11:48 AM, Sébastien Han <han.sebast...@gmail.com>wrote: > Hi! > > Usually you get: > > 2012-07-09 13:58:27.179+0000: 10227: warning : qemuDomainObjTaint:1134 : >> Domain id=2 name='instance-00000002' >> uuid=57aca8a6-d062-4a08-8d87-e4d11d259ac7 is tainted: high-privileges > > > when you change permission in libvirt (root I presumed) which is not > necessary. > > 2012-07-10 09:26:02.244+0000: 10229: error : virExecWithHook:328 : Cannot >> find 'pm-is-supported' in path: No such file or directory > > > This error is harmless and can be easily solved by installing the > following package: > > sudo apt-get install pm-utils -y > > > Do you have something in the nova-scheduler logs? > > Cheers! > > On Tue, Jul 10, 2012 at 11:29 AM, Leander Bessa Beernaert < > leande...@gmail.com> wrote: > >> Well i've checked the libvirt logs on both nodes and i found these two >> lines: >> >> 2012-07-09 13:58:27.179+0000: 10227: warning : qemuDomainObjTaint:1134 : >>> Domain id=2 name='instance-00000002' >>> uuid=57aca8a6-d062-4a08-8d87-e4d11d259ac7 is tainted: high-privileges >>> 2012-07-09 13:58:27.736+0000: 10226: error : qemuMonitorIORead:513 : >>> Unable to read from monitor: Connection reset by peer >> >> >> The log is alos filled with the message below, it gets repeated over and >> over and over. >> >> 2012-07-10 09:26:02.244+0000: 10229: error : virExecWithHook:328 : Cannot >>> find 'pm-is-supported' in path: No such file or directory >>> 2012-07-10 09:26:02.244+0000: 10229: warning : qemuCapsInit:856 : Failed >>> to get host power management capabilities >> >> >> On Tue, Jul 10, 2012 at 8:16 AM, Razique Mahroua < >> razique.mahr...@gmail.com> wrote: >> >>> Hi Leander, >>> try to check libvirtd.log files, >>> is the instance still running on the first node while you are lauching >>> the migration process ? >>> >>> Razique >>> *Nuage & Co - Razique Mahroua** * >>> razique.mahr...@gmail.com >>> >>> >>> Le 9 juil. 2012 à 16:09, Leander Bessa Beernaert a écrit : >>> >>> Ok, so i've updated to the test packages from >>> >>> The migration still fails, but i see no errors in the logs. I'm trying >>> to migrate a VM with the m1.tiny flavor from one machine to another. Their >>> hardware are identical and they have more than enough resources to support >>> the m1.tiny flavor: >>> >>> cloud35 (total) 4 3867 186 >>>> cloud35 (used_now) 0 312 5 >>>> cloud35 (used_max) 0 0 0 >>> >>> >>> These are the logs from the origin compute node: >>> http://paste.openstack.org/show/19319/ and the destination compute >>> node: http://paste.openstack.org/show/19318/ . The scheduler's log has >>> no visible errors or stack traces. >>> >>> I'm still using nfsv4. >>> >>> Any ideas? >>> >>> >>> On Fri, Jul 6, 2012 at 7:57 PM, Leander Bessa Beernaert < >>> leande...@gmail.com> wrote: >>> >>>> Thanks for the tip, it's a better than nothing :) >>>> >>>> Regards, >>>> Leander >>>> >>>> On Fri, Jul 6, 2012 at 6:32 PM, Mandar Vaze / मंदार वझे < >>>> mandarv...@gmail.com> wrote: >>>> >>>>> Not sure if you are able to debug this, but a while ago there was a >>>>> bug where instance.id was passed where instance.uuid was expected. >>>>> This used to cause some problem. >>>>> It looks like you are using distribution package rather than devstack >>>>> installation, so it is likely that the issue is now fixed. Can you try >>>>> latest packages (and/or try devstack if you can) >>>>> >>>>> I wish I could help more. >>>>> >>>>> -Mandar >>>>> >>>>> >>>>> On Fri, Jul 6, 2012 at 3:26 PM, Leander Bessa Beernaert < >>>>> leande...@gmail.com> wrote: >>>>> >>>>>> Hello, >>>>>> >>>>>> I've recently setup a system to test out the live migration feature. >>>>>> So far i've been able to launch the instances with the shared nfs folder. >>>>>> However, when i run the live-migration command i encounter this error in >>>>>> the destination compute node: >>>>>> >>>>>> 2012-07-05 09:33:48 ERROR nova.manager [-] Error during >>>>>>> ComputeManager.update_available_resource: Domain not found: no domain >>>>>>> with >>>>>>> matching id 2 >>>>>>> 2012-07-05 09:33:48 TRACE nova.manager Traceback (most recent call >>>>>>> last): >>>>>>> 2012-07-05 09:33:48 TRACE nova.manager File >>>>>>> "/usr/lib/python2.7/dist-packages/nova/manager.py", line 155, in >>>>>>> periodic_tasks >>>>>>> 2012-07-05 09:33:48 TRACE nova.manager task(self, context) >>>>>>> 2012-07-05 09:33:48 TRACE nova.manager File >>>>>>> "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 2409, >>>>>>> in >>>>>>> update_available_resource >>>>>>> 2012-07-05 09:33:48 TRACE nova.manager >>>>>>> self.driver.update_available_resource(context, self.host) >>>>>>> 2012-07-05 09:33:48 TRACE nova.manager File >>>>>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/connection.py", line >>>>>>> 1936, in update_available_resource >>>>>>> 2012-07-05 09:33:48 TRACE nova.manager 'vcpus_used': >>>>>>> self.get_vcpu_used(), >>>>>>> 2012-07-05 09:33:48 TRACE nova.manager File >>>>>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/connection.py", line >>>>>>> 1743, in get_vcpu_used >>>>>>> 2012-07-05 09:33:48 TRACE nova.manager dom = >>>>>>> self._conn.lookupByID(dom_id) >>>>>>> 2012-07-05 09:33:48 TRACE nova.manager File >>>>>>> "/usr/lib/python2.7/dist-packages/libvirt.py", line 2363, in lookupByID >>>>>>> 2012-07-05 09:33:48 TRACE nova.manager if ret is None:raise >>>>>>> libvirtError('virDomainLookupByID() failed', conn=self) >>>>>>> 2012-07-05 09:33:48 TRACE nova.manager libvirtError: Domain not >>>>>>> found: no domain with matching id 2 >>>>>> >>>>>> >>>>>> Any ideas on how to solve this? >>>>>> >>>>>> Regards, >>>>>> Leander >>>>>> >>>>>> _______________________________________________ >>>>>> Mailing list: https://launchpad.net/~openstack >>>>>> Post to : openstack@lists.launchpad.net >>>>>> Unsubscribe : https://launchpad.net/~openstack >>>>>> More help : https://help.launchpad.net/ListHelp >>>>>> >>>>>> >>>>> >>>> >>> _______________________________________________ >>> Mailing list: https://launchpad.net/~openstack >>> Post to : openstack@lists.launchpad.net >>> Unsubscribe : https://launchpad.net/~openstack >>> More help : https://help.launchpad.net/ListHelp >>> >>> >>> >> >> _______________________________________________ >> Mailing list: https://launchpad.net/~openstack >> Post to : openstack@lists.launchpad.net >> Unsubscribe : https://launchpad.net/~openstack >> More help : https://help.launchpad.net/ListHelp >> >> >
_______________________________________________ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp