Process to remove ovirt metrics store
by Jayme
I'm just wondering if there is documentation available for complete removal
of ovirt metrics store. I was testing it out but no longer want to
continue using it at this time. I'm mostly interested in reverting the
changes it makes to hosts/engine for logging.
Thanks!
5 years, 1 month
multiple vm got stack after gluster server hardware failure
by g.vasilopoulos@uoc.gr
Hello
We have a cluster with 9 ovirt host and a storage cluster with 3 storage servers, we run ovirt 4.3.6.6-1.
Storage is setup is replica 3 with arbiter.
There is another unmanaged gluster volume which is replica 3 based on local disks of ovirt hosts and the engine and a few small pfsense vm reside on it.
Yesterday we experienced a hardware failure on one raid controler on one of the storage servers, so the server stopped working.
On that particular server on it's local disks was a nfs storage domain used as ISO.
After the storage server failure multiple machines got stuck and in not responding state, others kept working but we could not manage them from ovirt. many machines had iso mounted so it may have to do with this. https://bugzilla.redhat.com/show_bug.cgi?id=1337073
But machines with ISO mounted were not the only ones affected. Allmost every host on ovirt clusterd switched from non responding to ok status, many machines were on unknown status
even if they did not have mounted iso or even resided in any storage domain of the gluster cluster. I.E. pfsense machines residing on local drives as the engine. These machines could not be Powered off from inside the engine. In any case I could not put the iso domain on maintenance because isos were still mounted on varius virtual machines.
I'm wondering what caused the allmost total failure, was it the nfs domain that could not be mounted ? And of course how could I prevent such a situation.
I'm sending you the logs from the critical time period from one of the hosts. Please let me know if you need more debug info.
2019-10-17 06:01:43,039+0300 INFO (jsonrpc/2) [jsonrpc.JsonRpcServer] RPC call Host.getStats succeeded in 0.25 seconds (__init__:312)
2019-10-17 06:01:46,638+0300 WARN (check/loop) [storage.check] Checker u'/rhev/data-center/mnt/g2-car0133.gfs-int.uoc.gr:_nfs-exports_isos/ec67be6b-367f-4e41-a686-695ab50c0034/dom_md/metadata' is blocked for 39
0.00 seconds (check:282)
2019-10-17 06:01:49,318+0300 WARN (itmap/0) [storage.scanDomains] Could not collect metadata file for domain path /rhev/data-center/mnt/g2-car0133.gfs-int.uoc.gr:_nfs-exports_isos (fileSD:913)
Traceback (most recent call last):
File "/usr/lib/python2.7/site-packages/vdsm/storage/fileSD.py", line 902, in collectMetaFiles
metaFiles = oop.getProcessPool(client_name).glob.glob(mdPattern)
File "/usr/lib/python2.7/site-packages/vdsm/storage/outOfProcess.py", line 110, in glob
return self._iop.glob(pattern)
File "/usr/lib/python2.7/site-packages/ioprocess/__init__.py", line 561, in glob
return self._sendCommand("glob", {"pattern": pattern}, self.timeout)
File "/usr/lib/python2.7/site-packages/ioprocess/__init__.py", line 442, in _sendCommand
raise Timeout(os.strerror(errno.ETIMEDOUT))
Timeout: Connection timed out
2019-10-17 06:01:49,319+0300 ERROR (monitor/ec67be6) [storage.Monitor] Error checking domain ec67be6b-367f-4e41-a686-695ab50c0034 (monitor:425)
Traceback (most recent call last):
File "/usr/lib/python2.7/site-packages/vdsm/storage/monitor.py", line 406, in _checkDomainStatus
self.domain.selftest()
File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 48, in __getattr__
return getattr(self.getRealDomain(), attrName)
File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 51, in getRealDomain
return self._cache._realProduce(self._sdUUID)
File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 134, in _realProduce
domain = self._findDomain(sdUUID)
File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 151, in _findDomain
return findMethod(sdUUID)
File "/usr/lib/python2.7/site-packages/vdsm/storage/nfsSD.py", line 145, in findDomain
return NfsStorageDomain(NfsStorageDomain.findDomainPath(sdUUID))
File "/usr/lib/python2.7/site-packages/vdsm/storage/nfsSD.py", line 135, in findDomainPath
raise se.StorageDomainDoesNotExist(sdUUID)
StorageDomainDoesNotExist: Storage domain does not exist: (u'ec67be6b-367f-4e41-a686-695ab50c0034',)
2019-10-17 06:01:50,253+0300 INFO (periodic/42) [vdsm.api] START repoStats(domains=()) from=internal, task_id=25ab44e7-26f7-4494-a970-f73f01892156 (api:48)
2019-10-17 06:01:50,253+0300 INFO (periodic/42) [vdsm.api] FINISH repoStats return={u'3c973b94-a57d-4462-bcb3-3cef13290327': {'code': 0, 'actual': True, 'version': 5, 'acquired': True, 'delay': '0.000932761', '
lastCheck': '3.6', 'valid': True}, u'ec67be6b-367f-4e41-a686-695ab50c0034': {'code': 2001, 'actual': True, 'version': -1, 'acquired': False, 'delay': '0', 'lastCheck': '0.9', 'valid': False}, u'a1ecca6f-1487-492
d-b9bf-d3fb60f1bc99': {'code': 0, 'actual': True, 'version': 5, 'acquired': True, 'delay': '0.00142361', 'lastCheck': '3.7', 'valid': True}, u'd5c52bc8-1a83-4915-b82f-d98e33db7c99': {'code': 0, 'actual': True, '
version': 5, 'acquired': True, 'delay': '0.00124982', 'lastCheck': '3.6', 'valid': True}, u'453eb3c2-c3a5-4d5f-8c1d-b16588fc5eff': {'code': 0, 'actual': True, 'version': 5, 'acquired': True, 'delay': '0.001343',
'lastCheck': '3.6', 'valid': True}} from=internal, task_id=25ab44e7-26f7-4494-a970-f73f01892156 (api:54)
2019-10-17 06:01:50,403+0300 WARN (vdsm.Scheduler) [Executor] executor state: count=8 workers=set([<Worker name=qgapoller/58 waiting task#=1 at 0x7fcd61483310>, <Worker name=qgapoller/53 running <Task discardab
le <Operation action=<VmDispatcher operation=<function <lambda> at 0x7fcd80036668> at 0x7fcd80492850> at 0x7fcd80492890> timeout=10, duration=30.00 at 0x7fcd615adcd0> discarded task#=11 at 0x7fcd615ad2d0>, <Work
er name=qgapoller/56 running <Task discardable <Operation action=<VmDispatcher operation=<function <lambda> at 0x7fcd80036668> at 0x7fcd80492850> at 0x7fcd80492890> timeout=10, duration=10.00 at 0x7fcd614cd8d0>
discarded task#=11 at 0x7fcd610013d0>, <Worker name=qgapoller/60 waiting task#=0 at 0x7fcd612cd910>, <Worker name=qgapoller/57 running <Task discardable <Operation action=<VmDispatcher operation=<function <lambd
a> at 0x7fcd80036668> at 0x7fcd80492850> at 0x7fcd80492890> timeout=10, duration=20.00 at 0x7fcd8003f190> discarded task#=0 at 0x7fcd3c355310>, <Worker name=qgapoller/55 running <Task discardable <Operation acti
on=<VmDispatcher operation=<function <lambda> at 0x7fcd80036668> at 0x7fcd80492850> at 0x7fcd80492890> timeout=10, duration=40.00 at 0x7fcd61483bd0> discarded task#=57 at 0x7fcd611dd150>, <Worker name=qgapoller/
59 waiting task#=0 at 0x7fcd614318d0>, <Worker name=qgapoller/54 waiting task#=14 at 0x7fcd3c3f5fd0>]) (executor:213)
2019-10-17 06:01:50,403+0300 INFO (vdsm.Scheduler) [Executor] Worker discarded: <Worker name=qgapoller/56 running <Task discardable <Operation action=<VmDispatcher operation=<function <lambda> at 0x7fcd80036668
> at 0x7fcd80492850> at 0x7fcd80492890> timeout=10, duration=10.00 at 0x7fcd614cd8d0> discarded task#=11 at 0x7fcd610013d0> (executor:355)
2019-10-17 06:01:52,094+0300 WARN (vdsm.Scheduler) [Executor] executor state: count=10 workers=set([<Worker name=periodic/42 waiting task#=4 at 0x7fcd3c592850>, <Worker name=periodic/43 waiting task#=0 at 0x7fc
d6124a090>, <Worker name=periodic/39 running <Task discardable <Operation action=<vdsm.virt.sampling.VMBulkstatsMonitor object at 0x7fcd3c5a5090> at 0x7fcd3c5a52d0> timeout=7.5, duration=37.50 at 0x7fcd61483610>
discarded task#=10 at 0x7fcd8003fed0>, <Worker name=periodic/2 running <Task discardable <Operation action=<vdsm.virt.sampling.VMBulkstatsMonitor object at 0x7fcd3c5a5090> at 0x7fcd3c5a52d0> timeout=7.5, durati
on=832.50 at 0x7fcd3c505f10> discarded task#=287729 at 0x7fcd806020d0>, <Worker name=periodic/41 waiting task#=7 at 0x7fcd3c5927d0>, <Worker name=periodic/3 running <Task discardable <Operation action=<vdsm.virt
.sampling.VMBulkstatsMonitor object at 0x7fcd3c5a5090> at 0x7fcd3c5a52d0> timeout=7.5, duration=802.50 at 0x7fcd6162d550> discarded task#=287772 at 0x7fcd3c592910>, <Worker name=periodic/34 running <Task discard
able <Operation action=<vdsm.virt.sampling.VMBulkstatsMonitor object at 0x7fcd3c5a5090> at 0x7fcd3c5a52d0> timeout=7.5, duration=22.50 at 0x7fcd3c529d10> discarded task#=43 at 0x7fcd3c4add50>, <Worker name=perio
dic/40 running <Task discardable <Operation action=<VmDispatcher operation=<class 'vdsm.virt.periodic.UpdateVolumes'> at 0x7fcd3c5924d0> at 0x7fcd3c592e10> timeout=30.0, duration=2.89 at 0x7fcd3c433710> task#=20
at 0x7fcd612cd890>, <Worker name=periodic/38 running <Task discardable <Operation action=<vdsm.virt.sampling.VMBulkstatsMonitor object at 0x7fcd3c5a5090> at 0x7fcd3c5a52d0> timeout=7.5, duration=7.50 at 0x7fcd3
c315fd0> discarded task#=18 at 0x7fcd6124a1d0>, <Worker name=periodic/4 running <Task discardable <Operation action=<vdsm.virt.sampling.VMBulkstatsMonitor object at 0x7fcd3c5a5090> at 0x7fcd3c5a52d0> timeout=7.5
, duration=817.50 at 0x7fcd615d62d0> discarded task#=284811 at 0x7fcd614495d0>]) (executor:213)
2019-10-17 06:01:52,094+0300 INFO (vdsm.Scheduler) [Executor] Worker discarded: <Worker name=periodic/38 running <Task discardable <Operation action=<vdsm.virt.sampling.VMBulkstatsMonitor object at 0x7fcd3c5a5090> at 0x7fcd3c5a52d0> timeout=7.5, duration=7.50 at 0x7fcd3c315fd0> discarded task#=18 at 0x7fcd6124a1d0> (executor:355)
2019-10-17 06:01:52,495+0300 INFO (monitor/453eb3c) [IOProcessClient] (/g2-car0133.gfs-int.uoc.gr:_nfs-exports_isos) Closing client (__init__:610)
2019-10-17 06:01:52,495+0300 INFO (monitor/453eb3c) [IOProcessClient] (/glusterSD/10.252.166.128:_volume1) Closing client (__init__:610)
2019-10-17 06:01:52,496+0300 INFO (monitor/453eb3c) [IOProcessClient] (/glusterSD/10.252.166.130:_volume3) Closing client (__init__:610)
2019-10-17 06:01:52,496+0300 INFO (monitor/453eb3c) [IOProcessClient] (/glusterSD/10.252.166.129:_volume2) Closing client (__init__:610)
2019-10-17 06:01:52,496+0300 INFO (monitor/453eb3c) [IOProcessClient] (/glusterSD/o5-car0118.gfs-int.uoc.gr:_engine) Closing client (__init__:610)
2019-10-17 06:01:53,123+0300 INFO (itmap/0) [IOProcessClient] (/glusterSD/o5-car0118.gfs-int.uoc.gr:_engine) Starting client (__init__:308)
2019-10-17 06:01:53,133+0300 INFO (itmap/1) [IOProcessClient] (/glusterSD/10.252.166.130:_volume3) Starting client (__init__:308)
2019-10-17 06:01:53,136+0300 INFO (ioprocess/28525) [IOProcess] (/glusterSD/o5-car0118.gfs-int.uoc.gr:_engine) Starting ioprocess (__init__:434)
2019-10-17 06:01:53,149+0300 INFO (itmap/2) [IOProcessClient] (/glusterSD/10.252.166.128:_volume1) Starting client (__init__:308)
2019-10-17 06:01:53,152+0300 INFO (ioprocess/28534) [IOProcess] (/glusterSD/10.252.166.130:_volume3) Starting ioprocess (__init__:434)
2019-10-17 06:01:53,161+0300 INFO (itmap/3) [IOProcessClient] (/glusterSD/10.252.166.129:_volume2) Starting client (__init__:308)
2019-10-17 06:01:53,165+0300 WARN (monitor/a1ecca6) [storage.PersistentDict] Could not parse line `^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@
^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^
so on
2019-10-17 06:01:53,166+0300 INFO (ioprocess/28540) [IOProcess] (/glusterSD/10.252.166.128:_volume1) Starting ioprocess (__init__:434)
2019-10-17 06:01:53,175+0300 INFO (ioprocess/28547) [IOProcess] (/glusterSD/10.252.166.129:_volume2) Starting ioprocess (__init__:434)
2019-10-17 06:01:54,763+0300 INFO (monitor/a1ecca6) [storage.StorageDomain] Removing remnants of deleted images [] (fileSD:726)
2019-10-17 06:01:55,271+0300 INFO (jsonrpc/4) [api.host] START getAllVmStats() from=::1,32814 (api:48)
2019-10-17 06:01:55,274+0300 WARN (jsonrpc/4) [virt.vm] (vmId='38586008-37a0-4400-83b2-771090c50a9a') monitor became unresponsive (command timeout, age=850.680000001) (vm:6061)
2019-10-17 06:01:55,277+0300 WARN (vdsm.Scheduler) [Executor] Worker blocked: <Worker name=jsonrpc/3 running <Task <JsonRpcTask {'params': {}, 'jsonrpc': '2.0', 'method': u'Host.getAllVmStats', 'id': u'33b72799-68cd-419c-bb6c-d9c9f61664b5'} at 0x7fcd3c329d10> timeout=60, duration=60.01 at 0x7fcd3c3298d0> task#=39322 at 0x7fcd3c581ed0>, traceback:
File: "/usr/lib64/python2.7/threading.py", line 785, in __bootstrap
self.__bootstrap_inner()
File: "/usr/lib64/python2.7/threading.py", line 812, in __bootstrap_inner
self.run()
File: "/usr/lib64/python2.7/threading.py", line 765, in run
self.__target(*self.__args, **self.__kwargs)
File: "/usr/lib/python2.7/site-packages/vdsm/common/concurrent.py", line 195, in run
ret = func(*args, **kwargs)
File: "/usr/lib/python2.7/site-packages/vdsm/executor.py", line 301, in _run
self._execute_task()
File: "/usr/lib/python2.7/site-packages/vdsm/executor.py", line 315, in _execute_task
task()
File: "/usr/lib/python2.7/site-packages/vdsm/executor.py", line 391, in __call__
self._callable()
File: "/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py", line 262, in __call__
self._handler(self._ctx, self._req)
File: "/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py", line 305, in _serveRequest
response = self._handle_request(req, ctx)
File: "/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py", line 345, in _handle_request
res = method(**params)
File: "/usr/lib/python2.7/site-packages/vdsm/rpc/Bridge.py", line 198, in _dynamicMethod
result = fn(*methodArgs)
File: "<string>", line 2, in getAllVmStats
File: "/usr/lib/python2.7/site-packages/vdsm/common/api.py", line 50, in method
ret = func(*args, **kwargs)
File: "/usr/lib/python2.7/site-packages/vdsm/API.py", line 1395, in getAllVmStats
statsList = self._cif.getAllVmStats()
File: "/usr/lib/python2.7/site-packages/vdsm/clientIF.py", line 569, in getAllVmStats
return [v.getStats() for v in self.getVMs().values()]
File: "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 1781, in getStats
oga_stats = self._getGuestStats()
File: "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 2000, in _getGuestStats
self._update_guest_disk_mapping()
File: "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 2015, in _update_guest_disk_mapping
self._sync_metadata()
File: "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 5142, in _sync_metadata
self._md_desc.dump(self._dom)
File: "/usr/lib/python2.7/site-packages/vdsm/virt/metadata.py", line 515, in dump
0)
File: "/usr/lib/python2.7/site-packages/vdsm/virt/virdomain.py", line 100, in f
ret = attr(*args, **kwargs)
File: "/usr/lib/python2.7/site-packages/vdsm/common/libvirtconnection.py", line 131, in wrapper
ret = f(*args, **kwargs)
File: "/usr/lib/python2.7/site-packages/vdsm/common/function.py", line 94, in wrapper
return func(inst, *args, **kwargs)
File: "/usr/lib64/python2.7/site-packages/libvirt.py", line 2364, in setMetadata
ret = libvirtmod.virDomainSetMetadata(self._o, type, metadata, key, uri, flags) (executor:363)
2019-10-17 06:01:55,279+0300 WARN (jsonrpc/3) [virt.vm] (vmId='b14cbec9-7acb-4f80-ad5e-3169e17d5328') Couldn't update metadata: Timed out during operation: cannot acquire state change lock (held by monitor=remoteDispatchConnectGetAllDomainStats) (vm:2018)
2019-10-17 06:01:55,285+0300 INFO (jsonrpc/3) [api.host] FINISH getAllVmStats return={'status': {'message': 'Done', 'code': 0}, 'statsList': (suppressed)} from=::1,32814 (api:54)
2019-10-17 06:01:55,290+0300 INFO (jsonrpc/3) [jsonrpc.JsonRpcServer] RPC call Host.getAllVmStats succeeded in 60.02 seconds (__init__:312)
5 years, 1 month
Ovirt Node 4.3.6 Logical Network Issue - Network Down
by jeremy_tourville@hotmail.com
I have built a new host and my setup is a single hyperconverged node.. I followed the directions to create a new logical network. I see that the engine has marked the network as being down. (Hosts>Network Interfaces Tab>Setup Host Networks)
Here is my network config on the host-
[root@vmh ~]# ip a
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
inet 127.0.0.1/8 scope host lo
valid_lft forever preferred_lft forever
inet6 ::1/128 scope host
valid_lft forever preferred_lft forever
2: enp96s0f0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master ovirtmgmt state UP group default qlen 1000
link/ether 0c:c4:7a:f9:b9:88 brd ff:ff:ff:ff:ff:ff
3: enp96s0f1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP group default qlen 1000
link/ether 0c:c4:7a:f9:b9:89 brd ff:ff:ff:ff:ff:ff
inet 172.30.51.2/30 brd 172.30.51.3 scope global noprefixroute dynamic enp96s0f1
valid_lft 82411sec preferred_lft 82411sec
inet6 fe80::d899:439c:5ee8:e292/64 scope link noprefixroute
valid_lft forever preferred_lft forever
19: ;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group default qlen 1000
link/ether 96:e4:00:aa:71:f7 brd ff:ff:ff:ff:ff:ff
23: ovs-system: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group default qlen 1000
link/ether 36:0f:60:69:e1:2b brd ff:ff:ff:ff:ff:ff
24: br-int: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group default qlen 1000
link/ether f6:3b:14:e1:15:48 brd ff:ff:ff:ff:ff:ff
25: ovirtmgmt: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP group default qlen 1000
link/ether 0c:c4:7a:f9:b9:88 brd ff:ff:ff:ff:ff:ff
inet 172.30.50.3/24 brd 172.30.50.255 scope global dynamic ovirtmgmt
valid_lft 84156sec preferred_lft 84156sec
inet6 fe80::ec4:7aff:fef9:b988/64 scope link
valid_lft forever preferred_lft forever
26: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master ovirtmgmt state UNKNOWN group default qlen 1000
link/ether fe:16:3e:50:53:cd brd ff:ff:ff:ff:ff:ff
inet6 fe80::fc16:3eff:fe50:53cd/64 scope link
valid_lft forever preferred_lft forever
VLAN 101 is attached to enp96s0f0 which is my ovirtmgmt interface. -IP range 172.30.50.x
VLAN 102 is attached to enp96s0f1 which is my storage NIC for gluster. -IP range 172.30.51.x
VLAN 103 is attached to enp96s0f1 is intended for most of my VMs that are not infrastructure related. -IP range 192.168.2.x
I am pretty confident my router/switch is setup correctly.. As a test, I can go to localhost>networking>add VLAN and assign enp96s0f1 to VLAN 103 and it does get an IP address in the 192.168.2.x range. The host can also ping the 192.168.2.1 gateway.
Why doesn't the engine think the VLAN is up? Which logs do I need to review?
5 years, 1 month
Disk encryption in oVirt
by MIMMIK _
Is there a way to get a full disk encryption on virtual disks used by VMs in oVirt?
Regards
5 years, 1 month
oVirt Metrics Store Installation - Error during SSO authentication
by Markus Schaufler
Hi,
oVirt 4.3.5
changed certificate to one from an official CA
configured active directory auth
no kerberos / no ldapS
as per: https://access.redhat.com/documentation/en-us/red_hat_virtualization/4.3/...
ANSIBLE_JINJA2_EXTENSIONS="jinja2.ext.do" ./configure_ovirt_machines_for_metrics.sh --playbook=ovirt-metrics-store-installation.yml --ask-vault-pass -vvvv
During installation of the metrics store following error appears:
TASK [oVirt.image-template : Login to oVirt] *********************************************************************************************************
task path: /usr/share/ansible/roles/ovirt.image-template/tasks/qcow2_image.yml:41
<localhost> ESTABLISH LOCAL CONNECTION FOR USER: root
<localhost> EXEC /bin/sh -c 'echo ~root && sleep 0'
<localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp/ansible-tmp-1571121133.7-63276692983944 `" && echo ansible-tmp-1571121133.7-63276692983944="` echo /root/.ansible/tmp/ansible-tmp-1571121133.7-63276692983944 `" ) && sleep 0'
Using module file /usr/lib/python2.7/site-packages/ansible/modules/cloud/ovirt/ovirt_auth.py
<localhost> PUT /root/.ansible/tmp/ansible-local-32046wbKds4/tmpyM5Ro2 TO /root/.ansible/tmp/ansible-tmp-1571121133.7-63276692983944/AnsiballZ_ovirt_auth.py
<localhost> EXEC /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1571121133.7-63276692983944/ /root/.ansible/tmp/ansible-tmp-1571121133.7-63276692983944/AnsiballZ_ovirt_auth.py && sleep 0'
<localhost> EXEC /bin/sh -c '/usr/bin/python /root/.ansible/tmp/ansible-tmp-1571121133.7-63276692983944/AnsiballZ_ovirt_auth.py && sleep 0'
<localhost> EXEC /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1571121133.7-63276692983944/ > /dev/null 2>&1 && sleep 0'
The full traceback is:
Traceback (most recent call last):
File "/tmp/ansible_ovirt_auth_payload_2cmBY9/__main__.py", line 276, in main
token = connection.authenticate()
File "/usr/lib64/python2.7/site-packages/ovirtsdk4/__init__.py", line 382, in authenticate
self._sso_token = self._get_access_token()
File "/usr/lib64/python2.7/site-packages/ovirtsdk4/__init__.py", line 628, in _get_access_token
sso_error[1]
AuthError: Error during SSO authentication access_denied : Cannot authenticate user 'xxx(a)xxxx.LOCAL': No valid profile found in credentials..
fatal: [localhost]: FAILED! => {
"changed": false,
"invocation": {
"module_args": {
"ca_file": "/etc/pki/ovirt-engine/apache-ca.pem",
"compress": true,
"headers": null,
"hostname": "ovirt-poc.xxxx.at",
"insecure": false,
"kerberos": false,
"ovirt_auth": null,
"password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
"state": "present",
"timeout": 0,
"token": null,
"url": "https://ovirt-poc.xxxx.at/ovirt-engine/api",
"username": "xxx(a)xxx.LOCAL"
}
},
"msg": "Error during SSO authentication access_denied : Cannot authenticate user 'xxx(a)xxx.LOCAL': No valid profile found in credentials.."
}
TASK [oVirt.image-template : Remove downloaded image] ************************************************************************************************
task path: /usr/share/ansible/roles/ovirt.image-template/tasks/qcow2_image.yml:210
skipping: [localhost] => {
"changed": false,
"skip_reason": "Conditional result was False"
}
TASK [oVirt.image-template : Remove vm] **************************************************************************************************************
task path: /usr/share/ansible/roles/ovirt.image-template/tasks/qcow2_image.yml:216
fatal: [localhost]: FAILED! => {
"msg": "The conditional check 'ovirt_templates | length == 0' failed. The error was: error while evaluating conditional (ovirt_templates | length == 0): 'ovirt_templates' is undefined\n\nThe error appears to be in '/usr/share/ansible/roles/ovirt.image-template/tasks/qcow2_image.yml': line 216, column 7, but may\nbe elsewhere in the file depending on the exact syntax problem.\n\nThe offending line appears to be:\n\n\n - name: Remove vm\n ^ here\n"
}
###############################################
Double-checked the credentials in "metrics-store-config.yml" and "secure_vars.yaml" and tried with admin@internal with same error.
Any hints on this?
5 years, 1 month
Help/guidelines/ideas on how to create new hosted-engine on final server with old config
by Lars Bækmark
I'm looking for guidelines/ideas for moving/migrating/Hosted-engine from the initial install with local storage, to the "Final" server.
Initial it was a "Test" to see if this will work or not, but this has now become the production, where the initial host named "ovirt" that is hosting the hosted-engine on local storage is "just" doing that. My hosted-engine is called "ovirtmgr" and I created a new host to run Virtual machines called "Hermod" - getting better in the naming thing, right? This test was so successful that I converted the old Xen/Oracle based host to CentOS 7 - the same as host Ovirt and Hermod:
Linux hermod 3.10.0-1062.1.1.el7.x86_64 #1 SMP Fri Sep 13 22:55:44 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux <https://www.facebook.com/linux.lanka?fref=gs&__tn__=%2CdK-R-R&eid=ARDadyp...>
Linux fernis 3.10.0-1062.el7.x86_64 #1 SMP Wed Aug 7 18:08:02 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux
oVirt version is: 4.3.5.4-1.el7
I want to move the hosted-engine to Fenris and at the same time use a NFS share for that. I can shut the whole setup down, as long as it runs again in daytime.
I do know that I should have started with the hosted-engine in the correct place, but that's not the case anymore. There are eleven VM’s in total
Can anyone suggest a plan on how to move/create a new hosted-engine on my server Fenris? The history of the old server is not important, I just want the hosted-engine on fenris using NFS share
I did find some sugesting on the web, but they all look quite complicated: https://www.ovirt.org/…/engine/migrate-to-hosted-engine.html <https://l.facebook.com/l.php?u=https%3A%2F%2Fwww.ovirt.org%2Fdevelop%2Fde...>
I just love the product! I'm now running VM's 2-3 times faster by converting from Xen to KVM, but still use the same storage and server! :-D
Appreciate any help! 😀
/Lars
Lars Bækmark
lars(a)baekmark.dk <mailto:lars@baekmark.dk>
+45 2241 0500
Mølleengen 6
3550 Slangerup
Danmark
-------------------------------------
Oversteer is when you hit the wall with the rear of the car.
Horsepower is how fast you hit the wall.
Torque is how far you take the wall with you
5 years, 1 month
Fencing Agent using SuperMicro IPMI Fails
by jeremy_tourville@hotmail.com
I am trying to setup my power management fencing agent on the host.
Address: I put in the IP address of the web page for IPMI. (Which does work via my web browser without issue)
Username: ADMIN
PW: my_password
Type: ipmilan
Options: lanplus=1
I click on the test button and it fails.
Board Manufacturer: Supermicro
Board Product Name: X11DPi-N
Any suggestions for troubleshooting?
5 years, 1 month