Thank you all!

Francesco,

[root@alpha ~]# journalctl -u libvirtd
-- Logs begin at Sun 2017-10-29 09:39:58 EET, end at Mon 2017-10-30 16:35:36 
EET. --
Oct 29 09:41:24 alpha.datamax.bg systemd[1]: Starting Virtualization daemon...
Oct 29 09:41:26 alpha.datamax.bg systemd[1]: Started Virtualization daemon.
Oct 29 22:30:26 alpha.datamax.bg libvirtd[4207]: libvirt version: 2.0.0, 
package: 10.el7_3.9 (CentOS BuildSystem < http://bugs.centos.org >, 
2017-05-25-20:52:28, c1bm.rdu2.centos.org)
Oct 29 22:30:26 alpha.datamax.bg libvirtd[4207]: hostname: alpha.datamax.bg
Oct 29 22:30:26 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double 
use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on 
function 0)
Oct 29 22:41:30 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double 
use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on 
function 0)
Oct 29 23:17:08 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double 
use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on 
function 0)
Oct 29 23:22:47 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double 
use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on 
function 0)
Oct 29 23:26:11 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double 
use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on 
function 0)
Oct 29 23:42:42 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double 
use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on 
function 0)
Oct 29 23:45:09 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double 
use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on 
function 0)
Oct 29 23:45:43 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double 
use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on 
function 0)
Oct 29 23:56:52 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double 
use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on 
function 0)
Oct 29 23:56:58 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double 
use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on 
function 0)
Oct 29 23:59:13 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double 
use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on 
function 0)
Oct 30 16:34:48 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double 
use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on 
function 0)

the first mistake for XML error in red matches exactly the time when we 
restarted it.
Тhese errors have at all three nodes and appears when trying to start engine 
virtual machine.

Also attaching vdsm.log.

Have a nice day!
>Понедельник, 30 октября 2017, 16:32 +02:00 от Francesco Romani < 
>from...@redhat.com >:
>
>Hi,
>
>anything in the journal about libvirt? (journalctl -u libvirtd)
>
>could you share a bigger chunk of the vdsm log, demonstrating the
>      failed VM start?
>
>Bests,
>
>On 10/30/2017 03:28 PM, Hristo Pavlov
>      wrote:
>>I tried
>>          it already,  doesn't start on any of the nodes. 
>>
>>
>>In a log /var/log/libvirt/qemu/HostedEngine.log at all three nodes has 
>>nothing to do with starting, as if
>>                  it had not started.
>>
>>
>>>Понедельник, 30 октября 2017, 16:11 +02:00
>>>        от Nathanaël Blanchet  <blanc...@abes.fr> :
>>>
>>>Hi,
>>>It happened to me too, after a live migration of
>>>                    it, I shut down the hosted engine on the targeted
>>>                    host, and I couldn't restart it anymore on this
>>>                    specific host. But I was able to start it on the
>>>                    initial one, where I initially where I deployed the
>>>                    HE. It was like the lease and the libvirt host
>>>                    definition staid on the first host after migration.
>>>So try you may try  hosted-engine --vm-start on one of
>>>                        other hosts...
>>>
>>>Le
>>>                    30/10/2017 à 12:28, Hristo Pavlov a écrit :
>>>>Hi
>>>>                    All,
>>>>
>>>>Our oVirt cluster is with 3 nodes with shared fibre
>>>>                    channel storage, the engine virtual machine is self
>>>>                    hosted.
>>>> 
>>>>Hypervisors OS: CentOS Linux release 7.3 / x86_64,
>>>>                    oVirt version is  4.1.2.2. The environment has been
>>>>                            working for about a year without any
>>>>                            problems .
>>>>
>>>>After shutdown of the hosted engine
>>>>                        virtual machine, it doesn't start. 
>>>>
>>>>Тhese commands that were
>>>>                            executed:
>>>>
>>>>hosted-engine --set-maintenance
>>>>                            --mode=global
>>>>hosted-engine --vm-shutdown
>>>>
>>>>after the status of engine vm was down, we
>>>>                        executed start.
>>>>
>>>>[root@alpha] hosted-engine --vm-start
>>>>VM exists and is down, destroying it
>>>>Exception in thread Client localhost:54321 (most
>>>>                        likely raised during interpreter shutdown):
>>>>
>>>>we noticied that at vdsm.log 
>>>>
>>>>017-10-30 13:11:04,863+0200 INFO  (jsonrpc/1)
>>>>                      [jsonrpc.JsonRpcServer] RPC call
>>>>                      StorageDomain.getStats succeeded in 0.26 seconds
>>>>                      (__init__:533)
>>>>2017-10-30 13:11:05,802+0200 INFO  (jsonrpc/6)
>>>>                      [jsonrpc.JsonRpcServer] RPC call
>>>>                      Host.getAllVmStats succeeded in 0.01 seconds
>>>>                      (__init__:533)
>>>>2017-10-30 13:11:05,825+0200 WARN  (jsonrpc/2)
>>>>                      [virt.vm]
>>>>                      (vmId='da98112d-b9fb-4098-93fa-1f1374b41e46')
>>>>                      Failed to get metadata, domain not connected.
>>>>                      (vm:2765)
>>>>2017-10-30 13:11:05,825+0200 ERROR (jsonrpc/2)
>>>>                      [jsonrpc.JsonRpcServer] Internal server error
>>>>                      (__init__:570)
>>>>Traceback (most recent call last):
>>>>  File
>>>>                      
>>>> "/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py",
>>>>                      line 565, in _handle_request
>>>>    res = method(**params)
>>>>  File
>>>>                      "/usr/lib/python2.7/site-packages/vdsm/rpc/Bridge.py",
>>>>                      line 202, in _dynamicMethod
>>>>    result = fn(*methodArgs)
>>>>  File "/usr/share/vdsm/API.py", line 1454, in
>>>>                      getAllVmIoTunePolicies
>>>>    io_tune_policies_dict =
>>>>                      self._cif.getAllVmIoTunePolicies()
>>>>  File "/usr/share/vdsm/clientIF.py", line 448, in
>>>>                      getAllVmIoTunePolicies
>>>>    'current_values': v.getIoTune()}
>>>>  File "/usr/share/vdsm/virt/vm.py", line 2803, in
>>>>                      getIoTune
>>>>    result = self.getIoTuneResponse()
>>>>  File "/usr/share/vdsm/virt/vm.py", line 2816, in
>>>>                      getIoTuneResponse
>>>>    res = self._dom.blockIoTune(
>>>>  File
>>>>                      
>>>> "/usr/lib/python2.7/site-packages/vdsm/virt/virdomain.py",
>>>>                      line 47, in __getattr__
>>>>    % self.vmid)
>>>>NotConnectedError: VM
>>>>                      u'da98112d-b9fb-4098-93fa-1f1374b41e46' was not
>>>>                      started yet or was shut down
>>>>
>>>>
>>>>The storage of self hosted engine multipath, pvs,
>>>>                      lvs, seems ok... 
>>>>
>>>>At the moment of the three nodes
>>>>                          there is a working about 100 virtual machines
>>>>                          and we can't manage them.
>>>>
>>>>Does
>>>>                              anyone have any ideas, what can be done то 
>>>> recover self hosted engine virtual machine?
>>>>
>>>>Thahk You!
>>>>Have a nice day!
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>_______________________________________________
>>>>Users mailing list
>>>>Users@ovirt.org
>>>>http://lists.ovirt.org/mailman/listinfo/users
>>>>
>>>-- 
>>>Nathanaël Blanchet
>>>
>>>Supervision réseau
>>>Pôle Infrastrutures Informatiques
>>>227 avenue Professeur-Jean-Louis-Viala
>>>34193 MONTPELLIER CEDEX 5    
>>>Tél. 33 (0)4 67 54 84 55
>>>Fax  33 (0)4 67 54 84 14
>>>blanc...@abes.fr 
>>
>>
>>
>>
>>_______________________________________________
>>Users mailing list
>>Users@ovirt.org
>>http://lists.ovirt.org/mailman/listinfo/users
>>
>-- 
>Francesco Romani
>Senior SW Eng., Virtualization R&D
>Red Hat
>IRC: fromani github: @fromanirh
>_______________________________________________
>Users mailing list
>Users@ovirt.org
>http://lists.ovirt.org/mailman/listinfo/users



_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to