Np ;) On Tue, Jul 10, 2012 at 12:33 PM, Leander Bessa Beernaert < [email protected]> wrote:
> Ok. Thx for the help :) > > > On Tue, Jul 10, 2012 at 11:30 AM, Sébastien Han > <[email protected]>wrote: > >> It's production ready, RedHat offers a commercial support on it. >> Just keep in mind that it's owned by Redhat ;) >> >> >> >> On Tue, Jul 10, 2012 at 12:24 PM, Leander Bessa Beernaert < >> [email protected]> wrote: >> >>> Is GlusterFS be more viable for a production environment? >>> >>> >>> On Tue, Jul 10, 2012 at 11:17 AM, Sébastien Han <[email protected] >>> > wrote: >>> >>>> Great! >>>> >>>> The last time I ran the live-migration, it was with GlusterFS and >>>> CephFS and I didn't changed any permissions in libvirt. I did the >>>> live-migration with NFS once but it was in Diablo (horrible), I don't >>>> really remember my setup. Maybe you should consider to try GlusterFS. >>>> >>>> >>>> On Tue, Jul 10, 2012 at 12:07 PM, Leander Bessa Beernaert < >>>> [email protected]> wrote: >>>> >>>>> That did! Thanks :) >>>>> >>>>> Do you by change have any pointer on getting the live-migration to >>>>> work without running libvirt under root? >>>>> >>>>> >>>>> On Tue, Jul 10, 2012 at 10:55 AM, Sébastien Han < >>>>> [email protected]> wrote: >>>>> >>>>>> Change the vncserver_listen to 0.0.0.0 and re-try the live-migration, >>>>>> you should get better results :) >>>>>> >>>>>> >>>>>> >>>>>> On Tue, Jul 10, 2012 at 11:52 AM, Leander Bessa Beernaert < >>>>>> [email protected]> wrote: >>>>>> >>>>>>> If i don't run libvirt with root, it can't write to the shared >>>>>>> folder. It's the only way i've been able to get this to work. :S >>>>>>> >>>>>>> Below follows the configuration of one of the compute nodes. >>>>>>> 10.0.1.1 is the controller and 10.0.1.2 is the compute node. >>>>>>> >>>>>>> novncproxy_base_url=http://10.0.1.1:6080/vnc_auto.html >>>>>>>> xvpvncproxy_base_url=http://10.0.1.1:6081/console >>>>>>>> vncserver_proxyclient_address=10.0.1.2 >>>>>>>> vncserver_listen=10.0.1.2 >>>>>>> >>>>>>> >>>>>>> On Tue, Jul 10, 2012 at 10:49 AM, Sébastien Han < >>>>>>> [email protected]> wrote: >>>>>>> >>>>>>>> I forgot to ask, did you enable the vnc console? >>>>>>>> >>>>>>>> If so, with which parameters? >>>>>>>> >>>>>>>> >>>>>>>> On Tue, Jul 10, 2012 at 11:48 AM, Sébastien Han < >>>>>>>> [email protected]> wrote: >>>>>>>> >>>>>>>>> Hi! >>>>>>>>> >>>>>>>>> Usually you get: >>>>>>>>> >>>>>>>>> 2012-07-09 13:58:27.179+0000: 10227: warning : >>>>>>>>>> qemuDomainObjTaint:1134 : Domain id=2 name='instance-00000002' >>>>>>>>>> uuid=57aca8a6-d062-4a08-8d87-e4d11d259ac7 is tainted: high-privileges >>>>>>>>> >>>>>>>>> >>>>>>>>> when you change permission in libvirt (root I presumed) which is >>>>>>>>> not necessary. >>>>>>>>> >>>>>>>>> 2012-07-10 09:26:02.244+0000: 10229: error : virExecWithHook:328 : >>>>>>>>>> Cannot find 'pm-is-supported' in path: No such file or directory >>>>>>>>> >>>>>>>>> >>>>>>>>> This error is harmless and can be easily solved by installing the >>>>>>>>> following package: >>>>>>>>> >>>>>>>>> sudo apt-get install pm-utils -y >>>>>>>>> >>>>>>>>> >>>>>>>>> Do you have something in the nova-scheduler logs? >>>>>>>>> >>>>>>>>> Cheers! >>>>>>>>> >>>>>>>>> On Tue, Jul 10, 2012 at 11:29 AM, Leander Bessa Beernaert < >>>>>>>>> [email protected]> wrote: >>>>>>>>> >>>>>>>>>> Well i've checked the libvirt logs on both nodes and i found >>>>>>>>>> these two lines: >>>>>>>>>> >>>>>>>>>> 2012-07-09 13:58:27.179+0000: 10227: warning : >>>>>>>>>>> qemuDomainObjTaint:1134 : Domain id=2 name='instance-00000002' >>>>>>>>>>> uuid=57aca8a6-d062-4a08-8d87-e4d11d259ac7 is tainted: >>>>>>>>>>> high-privileges >>>>>>>>>>> 2012-07-09 13:58:27.736+0000: 10226: error : >>>>>>>>>>> qemuMonitorIORead:513 : Unable to read from monitor: Connection >>>>>>>>>>> reset by >>>>>>>>>>> peer >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> The log is alos filled with the message below, it gets repeated >>>>>>>>>> over and over and over. >>>>>>>>>> >>>>>>>>>> 2012-07-10 09:26:02.244+0000: 10229: error : virExecWithHook:328 >>>>>>>>>>> : Cannot find 'pm-is-supported' in path: No such file or directory >>>>>>>>>>> 2012-07-10 09:26:02.244+0000: 10229: warning : qemuCapsInit:856 >>>>>>>>>>> : Failed to get host power management capabilities >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Tue, Jul 10, 2012 at 8:16 AM, Razique Mahroua < >>>>>>>>>> [email protected]> wrote: >>>>>>>>>> >>>>>>>>>>> Hi Leander, >>>>>>>>>>> try to check libvirtd.log files, >>>>>>>>>>> is the instance still running on the first node while you are >>>>>>>>>>> lauching the migration process ? >>>>>>>>>>> >>>>>>>>>>> Razique >>>>>>>>>>> *Nuage & Co - Razique Mahroua** * >>>>>>>>>>> [email protected] >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Le 9 juil. 2012 à 16:09, Leander Bessa Beernaert a écrit : >>>>>>>>>>> >>>>>>>>>>> Ok, so i've updated to the test packages from >>>>>>>>>>> >>>>>>>>>>> The migration still fails, but i see no errors in the logs. I'm >>>>>>>>>>> trying to migrate a VM with the m1.tiny flavor from one machine to >>>>>>>>>>> another. >>>>>>>>>>> Their hardware are identical and they have more than enough >>>>>>>>>>> resources to >>>>>>>>>>> support the m1.tiny flavor: >>>>>>>>>>> >>>>>>>>>>> cloud35 (total) 4 3867 186 >>>>>>>>>>>> cloud35 (used_now) 0 312 >>>>>>>>>>>> 5 >>>>>>>>>>>> cloud35 (used_max) 0 0 >>>>>>>>>>>> 0 >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> These are the logs from the origin compute node: >>>>>>>>>>> http://paste.openstack.org/show/19319/ and the destination >>>>>>>>>>> compute node: http://paste.openstack.org/show/19318/ . The >>>>>>>>>>> scheduler's log has no visible errors or stack traces. >>>>>>>>>>> >>>>>>>>>>> I'm still using nfsv4. >>>>>>>>>>> >>>>>>>>>>> Any ideas? >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Fri, Jul 6, 2012 at 7:57 PM, Leander Bessa Beernaert < >>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>> >>>>>>>>>>>> Thanks for the tip, it's a better than nothing :) >>>>>>>>>>>> >>>>>>>>>>>> Regards, >>>>>>>>>>>> Leander >>>>>>>>>>>> >>>>>>>>>>>> On Fri, Jul 6, 2012 at 6:32 PM, Mandar Vaze / मंदार वझे < >>>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> Not sure if you are able to debug this, but a while ago there >>>>>>>>>>>>> was a bug where instance.id was passed where instance.uuid >>>>>>>>>>>>> was expected. This used to cause some problem. >>>>>>>>>>>>> It looks like you are using distribution package rather than >>>>>>>>>>>>> devstack installation, so it is likely that the issue is now >>>>>>>>>>>>> fixed. Can you >>>>>>>>>>>>> try latest packages (and/or try devstack if you can) >>>>>>>>>>>>> >>>>>>>>>>>>> I wish I could help more. >>>>>>>>>>>>> >>>>>>>>>>>>> -Mandar >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On Fri, Jul 6, 2012 at 3:26 PM, Leander Bessa Beernaert < >>>>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> Hello, >>>>>>>>>>>>>> >>>>>>>>>>>>>> I've recently setup a system to test out the live migration >>>>>>>>>>>>>> feature. So far i've been able to launch the instances with the >>>>>>>>>>>>>> shared nfs >>>>>>>>>>>>>> folder. However, when i run the live-migration command i >>>>>>>>>>>>>> encounter this >>>>>>>>>>>>>> error in the destination compute node: >>>>>>>>>>>>>> >>>>>>>>>>>>>> 2012-07-05 09:33:48 ERROR nova.manager [-] Error during >>>>>>>>>>>>>>> ComputeManager.update_available_resource: Domain not found: no >>>>>>>>>>>>>>> domain with >>>>>>>>>>>>>>> matching id 2 >>>>>>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager Traceback (most >>>>>>>>>>>>>>> recent call last): >>>>>>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager File >>>>>>>>>>>>>>> "/usr/lib/python2.7/dist-packages/nova/manager.py", line 155, in >>>>>>>>>>>>>>> periodic_tasks >>>>>>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager task(self, >>>>>>>>>>>>>>> context) >>>>>>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager File >>>>>>>>>>>>>>> "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", >>>>>>>>>>>>>>> line 2409, in >>>>>>>>>>>>>>> update_available_resource >>>>>>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager >>>>>>>>>>>>>>> self.driver.update_available_resource(context, self.host) >>>>>>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager File >>>>>>>>>>>>>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/connection.py", >>>>>>>>>>>>>>> line >>>>>>>>>>>>>>> 1936, in update_available_resource >>>>>>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager 'vcpus_used': >>>>>>>>>>>>>>> self.get_vcpu_used(), >>>>>>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager File >>>>>>>>>>>>>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/connection.py", >>>>>>>>>>>>>>> line >>>>>>>>>>>>>>> 1743, in get_vcpu_used >>>>>>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager dom = >>>>>>>>>>>>>>> self._conn.lookupByID(dom_id) >>>>>>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager File >>>>>>>>>>>>>>> "/usr/lib/python2.7/dist-packages/libvirt.py", line 2363, in >>>>>>>>>>>>>>> lookupByID >>>>>>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager if ret is >>>>>>>>>>>>>>> None:raise libvirtError('virDomainLookupByID() failed', >>>>>>>>>>>>>>> conn=self) >>>>>>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager libvirtError: Domain >>>>>>>>>>>>>>> not found: no domain with matching id 2 >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> Any ideas on how to solve this? >>>>>>>>>>>>>> >>>>>>>>>>>>>> Regards, >>>>>>>>>>>>>> Leander >>>>>>>>>>>>>> >>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>> Mailing list: https://launchpad.net/~openstack >>>>>>>>>>>>>> Post to : [email protected] >>>>>>>>>>>>>> Unsubscribe : https://launchpad.net/~openstack >>>>>>>>>>>>>> More help : https://help.launchpad.net/ListHelp >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> Mailing list: https://launchpad.net/~openstack >>>>>>>>>>> Post to : [email protected] >>>>>>>>>>> Unsubscribe : https://launchpad.net/~openstack >>>>>>>>>>> More help : https://help.launchpad.net/ListHelp >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>>> Mailing list: https://launchpad.net/~openstack >>>>>>>>>> Post to : [email protected] >>>>>>>>>> Unsubscribe : https://launchpad.net/~openstack >>>>>>>>>> More help : https://help.launchpad.net/ListHelp >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >
_______________________________________________ Mailing list: https://launchpad.net/~openstack Post to : [email protected] Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp

