[ovirt-users] self hosted engine vm won't start

Hristo Pavlov ichkata_66 at mail.ru
Mon Oct 30 14:56:55 UTC 2017


Thank you all!

Francesco,

[root at alpha ~]# journalctl -u libvirtd
-- Logs begin at Sun 2017-10-29 09:39:58 EET, end at Mon 2017-10-30 16:35:36 EET. --
Oct 29 09:41:24 alpha.datamax.bg systemd[1]: Starting Virtualization daemon...
Oct 29 09:41:26 alpha.datamax.bg systemd[1]: Started Virtualization daemon.
Oct 29 22:30:26 alpha.datamax.bg libvirtd[4207]: libvirt version: 2.0.0, package: 10.el7_3.9 (CentOS BuildSystem < http://bugs.centos.org >, 2017-05-25-20:52:28, c1bm.rdu2.centos.org)
Oct 29 22:30:26 alpha.datamax.bg libvirtd[4207]: hostname: alpha.datamax.bg
Oct 29 22:30:26 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on function 0)
Oct 29 22:41:30 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on function 0)
Oct 29 23:17:08 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on function 0)
Oct 29 23:22:47 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on function 0)
Oct 29 23:26:11 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on function 0)
Oct 29 23:42:42 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on function 0)
Oct 29 23:45:09 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on function 0)
Oct 29 23:45:43 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on function 0)
Oct 29 23:56:52 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on function 0)
Oct 29 23:56:58 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on function 0)
Oct 29 23:59:13 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on function 0)
Oct 30 16:34:48 alpha.datamax.bg libvirtd[4207]: XML error: Attempted double use of PCI slot 0000:00:06.0 (may need "multifunction='on'" for device on function 0)

the first mistake for XML error in red matches exactly the time when we restarted it.
Тhese errors have at all three nodes and appears when trying to start engine virtual machine.

Also attaching vdsm.log.

Have a nice day!
>Понедельник, 30 октября 2017, 16:32 +02:00 от Francesco Romani < fromani at redhat.com >:
>
>Hi,
>
>anything in the journal about libvirt? (journalctl -u libvirtd)
>
>could you share a bigger chunk of the vdsm log, demonstrating the
>      failed VM start?
>
>Bests,
>
>On 10/30/2017 03:28 PM, Hristo Pavlov
>      wrote:
>>I tried
>>          it already,  doesn't start on any of the nodes. 
>>
>>
>>In a log /var/log/libvirt/qemu/HostedEngine.log at all three nodes has nothing to do with starting, as if
>>                  it had not started.
>>
>>
>>>Понедельник, 30 октября 2017, 16:11 +02:00
>>>        от Nathanaël Blanchet  <blanchet at abes.fr> :
>>>
>>>Hi,
>>>It happened to me too, after a live migration of
>>>                    it, I shut down the hosted engine on the targeted
>>>                    host, and I couldn't restart it anymore on this
>>>                    specific host. But I was able to start it on the
>>>                    initial one, where I initially where I deployed the
>>>                    HE. It was like the lease and the libvirt host
>>>                    definition staid on the first host after migration.
>>>So try you may try  hosted-engine --vm-start on one of
>>>                        other hosts...
>>>
>>>Le
>>>                    30/10/2017 à 12:28, Hristo Pavlov a écrit :
>>>>Hi
>>>>                    All,
>>>>
>>>>Our oVirt cluster is with 3 nodes with shared fibre
>>>>                    channel storage, the engine virtual machine is self
>>>>                    hosted.
>>>> 
>>>>Hypervisors OS: CentOS Linux release 7.3 / x86_64,
>>>>                    oVirt version is  4.1.2.2. The environment has been
>>>>                            working for about a year without any
>>>>                            problems .
>>>>
>>>>After shutdown of the hosted engine
>>>>                        virtual machine, it doesn't start. 
>>>>
>>>>Тhese commands that were
>>>>                            executed:
>>>>
>>>>hosted-engine --set-maintenance
>>>>                            --mode=global
>>>>hosted-engine --vm-shutdown
>>>>
>>>>after the status of engine vm was down, we
>>>>                        executed start.
>>>>
>>>>[root at alpha] hosted-engine --vm-start
>>>>VM exists and is down, destroying it
>>>>Exception in thread Client localhost:54321 (most
>>>>                        likely raised during interpreter shutdown):
>>>>
>>>>we noticied that at vdsm.log 
>>>>
>>>>017-10-30 13:11:04,863+0200 INFO  (jsonrpc/1)
>>>>                      [jsonrpc.JsonRpcServer] RPC call
>>>>                      StorageDomain.getStats succeeded in 0.26 seconds
>>>>                      (__init__:533)
>>>>2017-10-30 13:11:05,802+0200 INFO  (jsonrpc/6)
>>>>                      [jsonrpc.JsonRpcServer] RPC call
>>>>                      Host.getAllVmStats succeeded in 0.01 seconds
>>>>                      (__init__:533)
>>>>2017-10-30 13:11:05,825+0200 WARN  (jsonrpc/2)
>>>>                      [virt.vm]
>>>>                      (vmId='da98112d-b9fb-4098-93fa-1f1374b41e46')
>>>>                      Failed to get metadata, domain not connected.
>>>>                      (vm:2765)
>>>>2017-10-30 13:11:05,825+0200 ERROR (jsonrpc/2)
>>>>                      [jsonrpc.JsonRpcServer] Internal server error
>>>>                      (__init__:570)
>>>>Traceback (most recent call last):
>>>>  File
>>>>                      "/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py",
>>>>                      line 565, in _handle_request
>>>>    res = method(**params)
>>>>  File
>>>>                      "/usr/lib/python2.7/site-packages/vdsm/rpc/Bridge.py",
>>>>                      line 202, in _dynamicMethod
>>>>    result = fn(*methodArgs)
>>>>  File "/usr/share/vdsm/API.py", line 1454, in
>>>>                      getAllVmIoTunePolicies
>>>>    io_tune_policies_dict =
>>>>                      self._cif.getAllVmIoTunePolicies()
>>>>  File "/usr/share/vdsm/clientIF.py", line 448, in
>>>>                      getAllVmIoTunePolicies
>>>>    'current_values': v.getIoTune()}
>>>>  File "/usr/share/vdsm/virt/vm.py", line 2803, in
>>>>                      getIoTune
>>>>    result = self.getIoTuneResponse()
>>>>  File "/usr/share/vdsm/virt/vm.py", line 2816, in
>>>>                      getIoTuneResponse
>>>>    res = self._dom.blockIoTune(
>>>>  File
>>>>                      "/usr/lib/python2.7/site-packages/vdsm/virt/virdomain.py",
>>>>                      line 47, in __getattr__
>>>>    % self.vmid)
>>>>NotConnectedError: VM
>>>>                      u'da98112d-b9fb-4098-93fa-1f1374b41e46' was not
>>>>                      started yet or was shut down
>>>>
>>>>
>>>>The storage of self hosted engine multipath, pvs,
>>>>                      lvs, seems ok... 
>>>>
>>>>At the moment of the three nodes
>>>>                          there is a working about 100 virtual machines
>>>>                          and we can't manage them.
>>>>
>>>>Does
>>>>                              anyone have any ideas, what can be done то recover self hosted engine virtual machine?
>>>>
>>>>Thahk You!
>>>>Have a nice day!
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>_______________________________________________
>>>>Users mailing list
>>>>Users at ovirt.org
>>>>http://lists.ovirt.org/mailman/listinfo/users
>>>>
>>>-- 
>>>Nathanaël Blanchet
>>>
>>>Supervision réseau
>>>Pôle Infrastrutures Informatiques
>>>227 avenue Professeur-Jean-Louis-Viala
>>>34193 MONTPELLIER CEDEX 5 	
>>>Tél. 33 (0)4 67 54 84 55
>>>Fax  33 (0)4 67 54 84 14
>>>blanchet at abes.fr 
>>
>>
>>
>>
>>_______________________________________________
>>Users mailing list
>>Users at ovirt.org
>>http://lists.ovirt.org/mailman/listinfo/users
>>
>-- 
>Francesco Romani
>Senior SW Eng., Virtualization R&D
>Red Hat
>IRC: fromani github: @fromanirh
>_______________________________________________
>Users mailing list
>Users at ovirt.org
>http://lists.ovirt.org/mailman/listinfo/users



-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20171030/2ce113dc/attachment.html>


More information about the Users mailing list