On 02/23/2017 10:28 AM, Francesco Romani wrote:
> On 02/23/2017 04:20 PM, Michael Watters wrote:
>> I have an ovirt cluster running ovirt 4.0 and I am seeing several errors
>> when I attempt to put one of our nodes into maintenance mode.  The logs
>> on the source server show errors as follows.
>>
>> Feb 23 10:15:08 ovirt-node-production3.example.com libvirtd[18800]: 
>> operation aborted: migration job: canceled by client
>> Feb 23 10:15:08 ovirt-node-production3.example.com libvirtd[18800]: internal 
>> error: qemu unexpectedly closed the monitor: 2017-02-23T15:12:58.289459Z 
>> qemu-kvm: warning: CPU(s) not present in any NUMA nodes: 2 3 4 5 6 7 8 9 10 
>> 11 12 13 14 15                                                               
>>                 
>>                                                                              
>>   2017-02-23T15:12:58.289684Z qemu-kvm: warning: All CPU(s) up to maxcpus 
>> should be described in NUMA config                                           
>>                                                                              
>>                          
>>                                                                              
>>   2017-02-23T15:15:07.889891Z qemu-kvm: Unknown combination of migration 
>> flags: 0
>>                                                                              
>>   2017-02-23T15:15:07.890821Z qemu-kvm: error while loading state section id 
>> 2(ram)
>>                                                                              
>>   2017-02-23T15:15:07.892357Z qemu-kvm: load of migration failed: Invalid 
>> argument
>>
>> This cluster does *not* have NUMA enabled so I am not sure why this
>> error is happening. 
> It's one implementation detail. NUMA is enabled transparently because it
> is required for memory hotplug support.
> It should be fully transparent.
>
>
>> Some migrations did succeed after being restarted
>> on a different host however I have two VMs that appear to be stuck.  Is
>> there a way to resolve this?
> The load/save state errors are most often found when the two sides of
> the migration have different and incompatible version of QEMU.
> In turn, this is quite often a bug, because forward migrations (e.g.
> from 2.3.0 to 2.4.0) are always supported, for obvious upgrade needs.
>
> So, which version of libvirt and qemu do you have on the sides of
> failing migration paths?
>
> Bests,
>

_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to