That did! Thanks :) Do you by change have any pointer on getting the live-migration to work without running libvirt under root?
On Tue, Jul 10, 2012 at 10:55 AM, Sébastien Han <[email protected]>wrote: > Change the vncserver_listen to 0.0.0.0 and re-try the live-migration, you > should get better results :) > > > > On Tue, Jul 10, 2012 at 11:52 AM, Leander Bessa Beernaert < > [email protected]> wrote: > >> If i don't run libvirt with root, it can't write to the shared folder. >> It's the only way i've been able to get this to work. :S >> >> Below follows the configuration of one of the compute nodes. 10.0.1.1 is >> the controller and 10.0.1.2 is the compute node. >> >> novncproxy_base_url=http://10.0.1.1:6080/vnc_auto.html >>> xvpvncproxy_base_url=http://10.0.1.1:6081/console >>> vncserver_proxyclient_address=10.0.1.2 >>> vncserver_listen=10.0.1.2 >> >> >> On Tue, Jul 10, 2012 at 10:49 AM, Sébastien Han >> <[email protected]>wrote: >> >>> I forgot to ask, did you enable the vnc console? >>> >>> If so, with which parameters? >>> >>> >>> On Tue, Jul 10, 2012 at 11:48 AM, Sébastien Han <[email protected] >>> > wrote: >>> >>>> Hi! >>>> >>>> Usually you get: >>>> >>>> 2012-07-09 13:58:27.179+0000: 10227: warning : qemuDomainObjTaint:1134 >>>>> : Domain id=2 name='instance-00000002' >>>>> uuid=57aca8a6-d062-4a08-8d87-e4d11d259ac7 is tainted: high-privileges >>>> >>>> >>>> when you change permission in libvirt (root I presumed) which is not >>>> necessary. >>>> >>>> 2012-07-10 09:26:02.244+0000: 10229: error : virExecWithHook:328 : >>>>> Cannot find 'pm-is-supported' in path: No such file or directory >>>> >>>> >>>> This error is harmless and can be easily solved by installing the >>>> following package: >>>> >>>> sudo apt-get install pm-utils -y >>>> >>>> >>>> Do you have something in the nova-scheduler logs? >>>> >>>> Cheers! >>>> >>>> On Tue, Jul 10, 2012 at 11:29 AM, Leander Bessa Beernaert < >>>> [email protected]> wrote: >>>> >>>>> Well i've checked the libvirt logs on both nodes and i found these two >>>>> lines: >>>>> >>>>> 2012-07-09 13:58:27.179+0000: 10227: warning : qemuDomainObjTaint:1134 >>>>>> : Domain id=2 name='instance-00000002' >>>>>> uuid=57aca8a6-d062-4a08-8d87-e4d11d259ac7 is tainted: high-privileges >>>>>> 2012-07-09 13:58:27.736+0000: 10226: error : qemuMonitorIORead:513 : >>>>>> Unable to read from monitor: Connection reset by peer >>>>> >>>>> >>>>> The log is alos filled with the message below, it gets repeated over >>>>> and over and over. >>>>> >>>>> 2012-07-10 09:26:02.244+0000: 10229: error : virExecWithHook:328 : >>>>>> Cannot find 'pm-is-supported' in path: No such file or directory >>>>>> 2012-07-10 09:26:02.244+0000: 10229: warning : qemuCapsInit:856 : >>>>>> Failed to get host power management capabilities >>>>> >>>>> >>>>> On Tue, Jul 10, 2012 at 8:16 AM, Razique Mahroua < >>>>> [email protected]> wrote: >>>>> >>>>>> Hi Leander, >>>>>> try to check libvirtd.log files, >>>>>> is the instance still running on the first node while you are >>>>>> lauching the migration process ? >>>>>> >>>>>> Razique >>>>>> *Nuage & Co - Razique Mahroua** * >>>>>> [email protected] >>>>>> >>>>>> >>>>>> Le 9 juil. 2012 à 16:09, Leander Bessa Beernaert a écrit : >>>>>> >>>>>> Ok, so i've updated to the test packages from >>>>>> >>>>>> The migration still fails, but i see no errors in the logs. I'm >>>>>> trying to migrate a VM with the m1.tiny flavor from one machine to >>>>>> another. >>>>>> Their hardware are identical and they have more than enough resources to >>>>>> support the m1.tiny flavor: >>>>>> >>>>>> cloud35 (total) 4 3867 186 >>>>>>> cloud35 (used_now) 0 312 5 >>>>>>> cloud35 (used_max) 0 0 0 >>>>>> >>>>>> >>>>>> These are the logs from the origin compute node: >>>>>> http://paste.openstack.org/show/19319/ and the destination compute >>>>>> node: http://paste.openstack.org/show/19318/ . The scheduler's log >>>>>> has no visible errors or stack traces. >>>>>> >>>>>> I'm still using nfsv4. >>>>>> >>>>>> Any ideas? >>>>>> >>>>>> >>>>>> On Fri, Jul 6, 2012 at 7:57 PM, Leander Bessa Beernaert < >>>>>> [email protected]> wrote: >>>>>> >>>>>>> Thanks for the tip, it's a better than nothing :) >>>>>>> >>>>>>> Regards, >>>>>>> Leander >>>>>>> >>>>>>> On Fri, Jul 6, 2012 at 6:32 PM, Mandar Vaze / मंदार वझे < >>>>>>> [email protected]> wrote: >>>>>>> >>>>>>>> Not sure if you are able to debug this, but a while ago there was a >>>>>>>> bug where instance.id was passed where instance.uuid was expected. >>>>>>>> This used to cause some problem. >>>>>>>> It looks like you are using distribution package rather than >>>>>>>> devstack installation, so it is likely that the issue is now fixed. >>>>>>>> Can you >>>>>>>> try latest packages (and/or try devstack if you can) >>>>>>>> >>>>>>>> I wish I could help more. >>>>>>>> >>>>>>>> -Mandar >>>>>>>> >>>>>>>> >>>>>>>> On Fri, Jul 6, 2012 at 3:26 PM, Leander Bessa Beernaert < >>>>>>>> [email protected]> wrote: >>>>>>>> >>>>>>>>> Hello, >>>>>>>>> >>>>>>>>> I've recently setup a system to test out the live migration >>>>>>>>> feature. So far i've been able to launch the instances with the >>>>>>>>> shared nfs >>>>>>>>> folder. However, when i run the live-migration command i encounter >>>>>>>>> this >>>>>>>>> error in the destination compute node: >>>>>>>>> >>>>>>>>> 2012-07-05 09:33:48 ERROR nova.manager [-] Error during >>>>>>>>>> ComputeManager.update_available_resource: Domain not found: no >>>>>>>>>> domain with >>>>>>>>>> matching id 2 >>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager Traceback (most recent >>>>>>>>>> call last): >>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager File >>>>>>>>>> "/usr/lib/python2.7/dist-packages/nova/manager.py", line 155, in >>>>>>>>>> periodic_tasks >>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager task(self, context) >>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager File >>>>>>>>>> "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line >>>>>>>>>> 2409, in >>>>>>>>>> update_available_resource >>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager >>>>>>>>>> self.driver.update_available_resource(context, self.host) >>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager File >>>>>>>>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/connection.py", >>>>>>>>>> line >>>>>>>>>> 1936, in update_available_resource >>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager 'vcpus_used': >>>>>>>>>> self.get_vcpu_used(), >>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager File >>>>>>>>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/connection.py", >>>>>>>>>> line >>>>>>>>>> 1743, in get_vcpu_used >>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager dom = >>>>>>>>>> self._conn.lookupByID(dom_id) >>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager File >>>>>>>>>> "/usr/lib/python2.7/dist-packages/libvirt.py", line 2363, in >>>>>>>>>> lookupByID >>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager if ret is None:raise >>>>>>>>>> libvirtError('virDomainLookupByID() failed', conn=self) >>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager libvirtError: Domain not >>>>>>>>>> found: no domain with matching id 2 >>>>>>>>> >>>>>>>>> >>>>>>>>> Any ideas on how to solve this? >>>>>>>>> >>>>>>>>> Regards, >>>>>>>>> Leander >>>>>>>>> >>>>>>>>> _______________________________________________ >>>>>>>>> Mailing list: https://launchpad.net/~openstack >>>>>>>>> Post to : [email protected] >>>>>>>>> Unsubscribe : https://launchpad.net/~openstack >>>>>>>>> More help : https://help.launchpad.net/ListHelp >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> _______________________________________________ >>>>>> Mailing list: https://launchpad.net/~openstack >>>>>> Post to : [email protected] >>>>>> Unsubscribe : https://launchpad.net/~openstack >>>>>> More help : https://help.launchpad.net/ListHelp >>>>>> >>>>>> >>>>>> >>>>> >>>>> _______________________________________________ >>>>> Mailing list: https://launchpad.net/~openstack >>>>> Post to : [email protected] >>>>> Unsubscribe : https://launchpad.net/~openstack >>>>> More help : https://help.launchpad.net/ListHelp >>>>> >>>>> >>>> >>> >> >
_______________________________________________ Mailing list: https://launchpad.net/~openstack Post to : [email protected] Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp

