What Stephan said, appart that, yes, it's production ready :-)

Nuage & Co - Razique Mahroua 

Le 10 juil. 2012 à 11:52, Leander Bessa Beernaert a écrit :

If i don't run libvirt with root, it can't write to the shared folder. It's the only way i've been able to get this to work. :S

Below follows the configuration of one of the compute nodes. 10.0.1.1 is the controller and 10.0.1.2 is the compute node.

novncproxy_base_url=http://10.0.1.1:6080/vnc_auto.html
xvpvncproxy_base_url=http://10.0.1.1:6081/console
vncserver_proxyclient_address=10.0.1.2
vncserver_listen=10.0.1.2

On Tue, Jul 10, 2012 at 10:49 AM, Sébastien Han <han.sebast...@gmail.com> wrote:
I forgot to ask, did you enable the vnc console?

If so, with which parameters?


On Tue, Jul 10, 2012 at 11:48 AM, Sébastien Han <han.sebast...@gmail.com> wrote:
Hi!

Usually you get:

2012-07-09 13:58:27.179+0000: 10227: warning : qemuDomainObjTaint:1134 : Domain id=2 name='instance-00000002' uuid=57aca8a6-d062-4a08-8d87-e4d11d259ac7 is tainted: high-privileges

when you change permission in libvirt (root I presumed) which is not necessary.

2012-07-10 09:26:02.244+0000: 10229: error : virExecWithHook:328 : Cannot find 'pm-is-supported' in path: No such file or directory

This error is harmless and can be easily solved by installing the following package:

sudo apt-get install pm-utils -y

Do you have something in the nova-scheduler logs?

Cheers!

On Tue, Jul 10, 2012 at 11:29 AM, Leander Bessa Beernaert <leande...@gmail.com> wrote:
Well i've checked the libvirt logs on both nodes and i found these two lines:

2012-07-09 13:58:27.179+0000: 10227: warning : qemuDomainObjTaint:1134 : Domain id=2 name='instance-00000002' uuid=57aca8a6-d062-4a08-8d87-e4d11d259ac7 is tainted: high-privileges
2012-07-09 13:58:27.736+0000: 10226: error : qemuMonitorIORead:513 : Unable to read from monitor: Connection reset by peer

The log is alos filled with the message below, it gets repeated over and over and over.

2012-07-10 09:26:02.244+0000: 10229: error : virExecWithHook:328 : Cannot find 'pm-is-supported' in path: No such file or directory
2012-07-10 09:26:02.244+0000: 10229: warning : qemuCapsInit:856 : Failed to get host power management capabilities

On Tue, Jul 10, 2012 at 8:16 AM, Razique Mahroua <razique.mahr...@gmail.com> wrote:
Hi Leander,
try to check libvirtd.log files, 
is the instance still running on the first node while you are lauching the migration process ?

Razique
Nuage & Co - Razique Mahroua 

Le 9 juil. 2012 à 16:09, Leander Bessa Beernaert a écrit :

Ok, so i've updated to the test packages from 

The migration still fails, but i see no errors in the logs. I'm trying to migrate a VM with the m1.tiny flavor from one machine to another. Their hardware are identical and they have more than enough resources to support the m1.tiny flavor:

cloud35         (total)                         4    3867     186
cloud35         (used_now)                      0     312       5
cloud35         (used_max)                      0       0       0

These are the logs from the origin compute node:  http://paste.openstack.org/show/19319/  and  the destination compute node: http://paste.openstack.org/show/19318/ . The scheduler's log has no visible errors or stack traces.

I'm still using nfsv4. 

Any ideas?


On Fri, Jul 6, 2012 at 7:57 PM, Leander Bessa Beernaert <leande...@gmail.com> wrote:
Thanks for the tip, it's a better than nothing :)

Regards,
Leander

On Fri, Jul 6, 2012 at 6:32 PM, Mandar Vaze / मंदार वझे <mandarv...@gmail.com> wrote:
Not sure if you are able to debug this, but a while ago there was a bug where instance.id was passed where instance.uuid was expected. This used to cause some problem.
It looks like you are using distribution package rather than devstack installation, so it is likely that the issue is now fixed. Can you try latest packages (and/or try devstack if you can)

I wish I could help more.

-Mandar


On Fri, Jul 6, 2012 at 3:26 PM, Leander Bessa Beernaert <leande...@gmail.com> wrote:
Hello,

I've recently setup a system to test out the live migration feature. So far i've been able to launch the instances with the shared nfs folder. However, when i run the live-migration command i encounter this error in the destination compute node:

2012-07-05 09:33:48 ERROR nova.manager [-] Error during ComputeManager.update_available_resource: Domain not found: no domain with matching id 2
2012-07-05 09:33:48 TRACE nova.manager Traceback (most recent call last):
2012-07-05 09:33:48 TRACE nova.manager   File "/usr/lib/python2.7/dist-packages/nova/manager.py", line 155, in periodic_tasks
2012-07-05 09:33:48 TRACE nova.manager     task(self, context)
2012-07-05 09:33:48 TRACE nova.manager   File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 2409, in update_available_resource
2012-07-05 09:33:48 TRACE nova.manager     self.driver.update_available_resource(context, self.host)
2012-07-05 09:33:48 TRACE nova.manager   File "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/connection.py", line 1936, in update_available_resource
2012-07-05 09:33:48 TRACE nova.manager     'vcpus_used': self.get_vcpu_used(),
2012-07-05 09:33:48 TRACE nova.manager   File "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/connection.py", line 1743, in get_vcpu_used
2012-07-05 09:33:48 TRACE nova.manager     dom = self._conn.lookupByID(dom_id)
2012-07-05 09:33:48 TRACE nova.manager   File "/usr/lib/python2.7/dist-packages/libvirt.py", line 2363, in lookupByID
2012-07-05 09:33:48 TRACE nova.manager     if ret is None:raise libvirtError('virDomainLookupByID() failed', conn=self)
2012-07-05 09:33:48 TRACE nova.manager libvirtError: Domain not found: no domain with matching id 2

Any ideas on how to solve this?

Regards,
Leander

_______________________________________________
Mailing list: https://launchpad.net/~openstack
Post to     : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp




_______________________________________________
Mailing list: https://launchpad.net/~openstack
Post to     : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp



_______________________________________________
Mailing list: https://launchpad.net/~openstack
Post to     : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp




_______________________________________________
Mailing list: https://launchpad.net/~openstack
Post to     : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp

_______________________________________________
Mailing list: https://launchpad.net/~openstack
Post to     : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp

Reply via email to