I got hit up with the same issue - I think it had something to do with latency or timeout to my NAS/SAN, despite an isolated & prioritised storage network. I'm guessing the VDSM puts VM's in a paused state whenever it hits a slight panic or speed-wobble.

I've, for the time-being, moved to a NFS/POSIX share & the issues has not presented again.

- J

On 22/03/13 22:59, Nicolas Ecarnot wrote:
Hi,

After two months of stable usage of this 3.1 oVirt setup, here comes the first blocking issue for which I've no other mean to ask some hint.

When I'm starting a VM, the start process is running fine. Being fast enough, we can ssh-connect to it, but 5 seconds later, the VM is paused.

In the manager, I see that :

2013-03-22 09:42:57,435 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.VdsBrokerObjectsBuilder] (QuartzScheduler_Worker-40) Error in parsing vm pause status. Setting value to NONE 2013-03-22 09:42:57,436 INFO [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] (QuartzScheduler_Worker-40) VM serv-chk-adm3 3e17586d-bf8f-465b-8075-defaac90bc95 moved from PoweringUp --> Paused

And on the host, I see one warning message, no error msg, and many looping repeated messages :

* Warning :

Thread-1968::WARNING::2013-03-22 09:19:18,536::libvirtvm::1547::vm.Vm::(_readPauseCode) vmId=`3e17586d-bf8f-465b-8075-defaac90bc95`::_readPauseCode unsupported by libvirt vm

* Repeated msgs, amongst other repeated ones :
Thread-1973::DEBUG::2013-03-22 09:19:20,247::libvirtvm::220::vm.Vm::(_getNetworkStats) vmId=`3e17586d-bf8f-465b-8075-defaac90bc95`::Network stats not available Thread-1973::DEBUG::2013-03-22 09:19:20,247::libvirtvm::240::vm.Vm::(_getDiskStats) vmId=`3e17586d-bf8f-465b-8075-defaac90bc95`::Disk hdc stats not available


I made my homework and found some bugs that could be similar :
https://bugzilla.redhat.com/show_bug.cgi?id=660598
https://bugzilla.redhat.com/show_bug.cgi?id=672208
and moreover :
https://bugzilla.redhat.com/show_bug.cgi?id=695393

- I tried to restart the node's vds daemon : same behavior
- I tried to reboot the node : same behavior
- I tried to to restart the manager's engine : same behavior
- I tried to run this VM on another node : same behavior
- I tried to run another VM on the node I saw the issue : the other VM is running fine.

I don't know if I have to conclude that this issue is specific to this VM, but I sounds like yes.
Things to say about this VM :
- it it a RH6 IIRC. It has already been successfully started, migrated, stopped and rebooted many times in the past.
- it has 3 disks : one for the system and two for datas.
- it has no snapshots
- it has no different or complicated network setup

My storage domain is a SAN, iSCSI linked, and doing good job since months.

I must admit I'm a bit stuck. Last thing I haven't tried is to reboot the manager, though I'm not sure that would help.


_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to