Manual VM Migration fails
by Anton Louw
Hello Everybody,
I am having a strange issue. When I try and manually migrate a VM from one host to another, I get an error stating:
"Migration failed (VM: VM1, Source: node6.example.com, Destination: node3.example.com)"
I have tried with a few different machines, and it pops up with the same error. I have attached the VDSM logs of both source and destination nodes. The time frame is 16:01
I see the below in that time frame, but not quite sure what I need to change:
2020-10-23 16:01:14,419+0200 ERROR (migsrc/42186f82) [virt.vm] (vmId='42186f82-b84c-7e65-e736-e6331acd04ed') Failed to migrate (migration:450)
Traceback (most recent call last):
File "/usr/lib/python2.7/site-packages/vdsm/virt/migration.py", line 431, in _regular_run
time.time(), migrationParams, machineParams
File "/usr/lib/python2.7/site-packages/vdsm/virt/migration.py", line 505, in _startUnderlyingMigration
self._perform_with_conv_schedule(duri, muri)
File "/usr/lib/python2.7/site-packages/vdsm/virt/migration.py", line 591, in _perform_with_conv_schedule
self._perform_migration(duri, muri)
File "/usr/lib/python2.7/site-packages/vdsm/virt/migration.py", line 525, in _perform_migration
self._migration_flags)
File "/usr/lib/python2.7/site-packages/vdsm/virt/virdomain.py", line 100, in f
ret = attr(*args, **kwargs)
File "/usr/lib/python2.7/site-packages/vdsm/common/libvirtconnection.py", line 131, in wrapper
ret = f(*args, **kwargs)
File "/usr/lib/python2.7/site-packages/vdsm/common/function.py", line 94, in wrapper
return func(inst, *args, **kwargs)
File "/usr/lib64/python2.7/site-packages/libvirt.py", line 1781, in migrateToURI3
if ret == -1: raise libvirtError ('virDomainMigrateToURI3() failed', dom=self)
libvirtError: operation failed: guest CPU doesn't match specification: missing features: spec-ctrl,ssbd
Anton Louw
Cloud Engineer: Storage and Virtualization
______________________________________
D: 087 805 1572 | M: N/A
A: Rutherford Estate, 1 Scott Street, Waverley, Johannesburg
anton.louw(a)voxtelecom.co.za
www.vox.co.za
4 years, 5 months
oVirt test clustered DOA after upgrade to latest 4.4.2 (AFAIK). Network issue?
by Gilboa Davara
Hello all,
I just upgraded one of my test oVirt setups to latest.
Post reboot, the machine lost network and hosted engine didn't start.
When I connected to the machine, I noticed all the
/etc/sysconfig/network-scripts/ifcfg-* files disappeared. No idea why /
how. (Possibly I did something wrong and forgot about it.)
Long story story short, I copied the two missing files (ifcfg-onb0 ethernet
device configuration and the ifcfg-ovirtmgmt bridge configuration) from
another ovirt host, changed the ovirtmgmt IP address and UUID to match the
UUID reported in logs as missing and restarted NetworkManager and restarted
all the oVirt related services (vdsmd, supervdsmd, ovirt-*, etc).
Sadly enough, even with both onb0 and ovirtmgmt up, vdsm still complains
about the missing network (ovirtmgmt) and refuses to start the hosted
engine.
Reboot doesn't seem to change anything.
In the main log I see the following errors:
Oct 22 16:33:31 office-wx-otest vdsm[2634]: WARN Attempting to remove a non
existing network: ovirtmgmt/1da8c5b7-999c-4ada-8287-1f35de6ce21d
Oct 22 16:33:31 office-wx-otest vdsm[2634]: WARN Attempting to remove a non
existing net user: ovirtmgmt/1da8c5b7-999c-4ada-8287-1f35de6ce21d
Oct 22 16:33:31 office-wx-otest vdsm[2634]: WARN Attempting to remove a non
existing network: ovirtmgmt/1da8c5b7-999c-4ada-8287-1f35de6ce21d
Oct 22 16:33:31 office-wx-otest vdsm[2634]: WARN Attempting to remove a non
existing net user: ovirtmgmt/1da8c5b7-999c-4ada-8287-1f35de6ce21d
As it is one of my oVirt test setup, I can simply redeploy the host and
continue from there, but I rather use this experience to learn how to fix
oVirt such issues in the future.
Logs attached.
https://drive.google.com/file/d/12ugy6CuaFaMvXYt6uGT4D_EHIW6nXttb/view?us...
$ PAGER= nmcli connection show
NAME UUID TYPE DEVICE
ovirtmgmt 1da8c5b7-999c-4ada-8287-1f35de6ce21d bridge ovirtmgmt
onb0 48332db3-8939-bff3-6b71-772a28c9e7b8 ethernet onb0
$ PAGER= nmcli device show
GENERAL.DEVICE: ovirtmgmt
GENERAL.TYPE: bridge
GENERAL.HWADDR: FC:AA:14:6B:A8:E0
GENERAL.MTU: 1500
GENERAL.STATE: 100 (connected)
GENERAL.CONNECTION: ovirtmgmt
GENERAL.CON-PATH:
/org/freedesktop/NetworkManager/ActiveConnection/2
IP4.ADDRESS[1]: 192.168.2.117/24
IP4.GATEWAY: 192.168.2.100
IP4.ROUTE[1]: dst = 192.168.2.0/24, nh = 0.0.0.0,
mt = 425
IP4.ROUTE[2]: dst = 0.0.0.0/0, nh =
192.168.2.100, mt = 425
IP4.DNS[1]: 192.168.2.100
IP4.DNS[2]: 8.8.8.8
IP6.GATEWAY: --
GENERAL.DEVICE: onb0
GENERAL.TYPE: ethernet
GENERAL.HWADDR: FC:AA:14:6B:A8:E0
GENERAL.MTU: 1500
GENERAL.STATE: 100 (connected)
GENERAL.CONNECTION: onb0
GENERAL.CON-PATH:
/org/freedesktop/NetworkManager/ActiveConnection/3
WIRED-PROPERTIES.CARRIER: on
IP4.GATEWAY: --
GENERAL.DEVICE: ;vdsmdummy;
GENERAL.TYPE: bridge
GENERAL.HWADDR: 92:8B:9A:5E:C1:3E
GENERAL.MTU: 1500
GENERAL.STATE: 10 (unmanaged)
GENERAL.CONNECTION: --
GENERAL.CON-PATH: --
IP4.GATEWAY: --
IP6.GATEWAY: --
GENERAL.DEVICE: lo
GENERAL.TYPE: loopback
GENERAL.HWADDR: 00:00:00:00:00:00
GENERAL.MTU: 65536
GENERAL.STATE: 10 (unmanaged)
GENERAL.CONNECTION: --
GENERAL.CON-PATH: --
IP4.ADDRESS[1]: 127.0.0.1/8
IP4.GATEWAY: --
IP6.GATEWAY: --
GENERAL.DEVICE: br-int
GENERAL.TYPE: openvswitch
GENERAL.HWADDR: 8E:15:6A:F8:3C:45
GENERAL.MTU: 1500
GENERAL.STATE: 10 (unmanaged)
GENERAL.CONNECTION: --
GENERAL.CON-PATH: --
IP4.GATEWAY: --
IP6.GATEWAY: --
GENERAL.DEVICE: ovs-system
GENERAL.TYPE: openvswitch
GENERAL.HWADDR: E2:09:EA:A2:BD:70
GENERAL.MTU: 1500
GENERAL.STATE: 10 (unmanaged)
GENERAL.CONNECTION: --
GENERAL.CON-PATH: --
IP4.GATEWAY: --
IP6.GATEWAY:
- Gilboa
office-wx-otest-vdsm.bz2
<https://drive.google.com/file/d/12ugy6CuaFaMvXYt6uGT4D_EHIW6nXttb/view?us...>
4 years, 5 months
sr-iov network config assistance IOMMU group error
by Edward Berger
I'm' trying to bring up a test ovirt 4.4 cluster with sr-iov virtual
infiniband interfaces on Mellanox ConnectX-6 cards in some older Opteron
servers while we await delivery of new Epyc servers.
I have the hypervisors configured for sr-iov and showing the interfaces,
but when I try to bring up a VM with the host device selected it fails to
boot. "Failed Launching ..." shows up on the engine UI.
[image: sriov-hostdev.JPG]
The /var/log/libvirt/qemu log for the VM seems to think something is wrong
with the IOMMU group. If I scroll over to the right it says N/A under the
IOMMU column. Is there some step I need to 'create IOMMU groups' or any
other suggestions?
[root@n001 qemu]# more C*log
2020-10-21 20:26:39.266+0000: starting up libvirt version: 6.0.0, package:
25.2.
el8 (CBS <cbs(a)centos.org>, 2020-09-10-13:20:10, ), qemu version:
4.2.0qemu-kvm-4
.2.0-29.el8.3, kernel: 4.18.0-193.19.1.el8_2.x86_64, hostname:
n001.bil.psc.edu
LC_ALL=C \
PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin \
HOME=/var/lib/libvirt/qemu/domain-1-C8.2test \
XDG_DATA_HOME=/var/lib/libvirt/qemu/domain-1-C8.2test/.local/share \
XDG_CACHE_HOME=/var/lib/libvirt/qemu/domain-1-C8.2test/.cache \
XDG_CONFIG_HOME=/var/lib/libvirt/qemu/domain-1-C8.2test/.config \
QEMU_AUDIO_DRV=spice \
/usr/libexec/qemu-kvm \
-name guest=C8.2test,debug-threads=on \
-S \
-object
secret,id=masterKey0,format=raw,file=/var/lib/libvirt/qemu/domain-1-C8.2
test/master-key.aes \
-machine pc-q35-rhel8.1.0,accel=kvm,usb=off,dump-guest-core=off \
-cpu Opteron_G5 \
-m size=16777216k,slots=16,maxmem=67108864k \
-overcommit mem-lock=off \
-smp 4,maxcpus=16,sockets=16,dies=1,cores=1,threads=1 \
-numa node,nodeid=0,cpus=0-15,mem=16384 \
-uuid 07eb98af-0d9f-4f9c-967a-8af302c68703 \
-smbios
type=1,manufacturer=oVirt,product=RHEL,version=8.2-2.2004.0.2.el8,serial
=47513848-0036-2500-905b-0025905bff22,uuid=07eb98af-0d9f-4f9c-967a-8af302c68703,
family=oVirt \
-no-user-config \
-nodefaults \
-chardev socket,id=charmonitor,fd=39,server,nowait \
-mon chardev=charmonitor,id=monitor,mode=control \
-rtc base=2020-10-21T20:26:38,driftfix=slew \
-global kvm-pit.lost_tick_policy=delay \
-no-hpet \
-no-shutdown \
-global ICH9-LPC.disable_s3=1 \
-global ICH9-LPC.disable_s4=1 \
-boot strict=on \
-device
pcie-root-port,port=0x10,chassis=1,id=pci.1,bus=pcie.0,multifunction=on,
addr=0x2 \
-device pcie-root-port,port=0x11,chassis=2,id=pci.2,bus=pcie.0,addr=0x2.0x1
\
-device pcie-root-port,port=0x12,chassis=3,id=pci.3,bus=pcie.0,addr=0x2.0x2
\
-device pcie-root-port,port=0x13,chassis=4,id=pci.4,bus=pcie.0,addr=0x2.0x3
\
-device pcie-root-port,port=0x14,chassis=5,id=pci.5,bus=pcie.0,addr=0x2.0x4
\
-device pcie-root-port,port=0x15,chassis=6,id=pci.6,bus=pcie.0,addr=0x2.0x5
\
-device pcie-root-port,port=0x16,chassis=7,id=pci.7,bus=pcie.0,addr=0x2.0x6
\
-device pcie-root-port,port=0x17,chassis=8,id=pci.8,bus=pcie.0,addr=0x2.0x7
\
-device
pcie-root-port,port=0x18,chassis=9,id=pci.9,bus=pcie.0,multifunction=on,
addr=0x3 \
-device
pcie-root-port,port=0x19,chassis=10,id=pci.10,bus=pcie.0,addr=0x3.0x1 \
-device
pcie-root-port,port=0x1a,chassis=11,id=pci.11,bus=pcie.0,addr=0x3.0x2 \
-device
pcie-root-port,port=0x1b,chassis=12,id=pci.12,bus=pcie.0,addr=0x3.0x3 \
-device
pcie-root-port,port=0x1c,chassis=13,id=pci.13,bus=pcie.0,addr=0x3.0x4 \
-device
pcie-root-port,port=0x1d,chassis=14,id=pci.14,bus=pcie.0,addr=0x3.0x5 \
-device
pcie-root-port,port=0x1e,chassis=15,id=pci.15,bus=pcie.0,addr=0x3.0x6 \
-device
pcie-root-port,port=0x1f,chassis=16,id=pci.16,bus=pcie.0,addr=0x3.0x7 \
-device
qemu-xhci,p2=8,p3=8,id=ua-cd3a4abd-57e9-4f2e-a61f-01128ead8702,bus=pci.2
,addr=0x0 \
-device
virtio-scsi-pci,id=ua-64b19c77-98d3-4946-a0f8-c43c1347a5dc,bus=pci.1,add
r=0x0 \
-device virtio-serial-pci,id=virtio-serial0,max_ports=16,bus=pci.4,addr=0x0
\
-device
ide-cd,bus=ide.2,id=ua-271c6d06-1f3c-4e45-99cc-a39f39a6dd1b,werror=repor
t,rerror=report \
-blockdev '{"driver":"file","filename":"/rhev/data-center/mnt/10.200.16.152:
_var
_nfs__shares_data/662d0028-ceaf-4641-87fa-839254e15209/images/4c332ece-f6db-4b22
-879c-b6912c5b27c0/2918a5f1-c801-4a34-8cf7-89045265f6f3","aio":"threads","node-n
ame":"libvirt-1-storage","cache":{"direct":true,"no-flush":false},"auto-read-onl
y":true,"discard":"unmap"}' \
-blockdev
'{"node-name":"libvirt-1-format","read-only":false,"cache":{"direct":t
rue,"no-flush":false},"driver":"raw","file":"libvirt-1-storage"}' \
-device
scsi-hd,bus=ua-64b19c77-98d3-4946-a0f8-c43c1347a5dc.0,channel=0,scsi-id=
0,lun=0,device_id=4c332ece-f6db-4b22-879c-b6912c5b27c0,drive=libvirt-1-format,id
=ua-4c332ece-f6db-4b22-879c-b6912c5b27c0,bootindex=1,write-cache=on,serial=4c332
ece-f6db-4b22-879c-b6912c5b27c0,werror=stop,rerror=stop \
-chardev socket,id=charchannel0,fd=41,server,nowait \
-device
virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel
0,name=ovirt-guest-agent.0 \
-chardev socket,id=charchannel1,fd=42,server,nowait \
-device
virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel1,id=channel
1,name=org.qemu.guest_agent.0 \
-chardev spicevmc,id=charchannel2,name=vdagent \
-device
virtserialport,bus=virtio-serial0.0,nr=3,chardev=charchannel2,id=channel
2,name=com.redhat.spice.0 \
-device
usb-tablet,id=input0,bus=ua-cd3a4abd-57e9-4f2e-a61f-01128ead8702.0,port=
1 \
-vnc 128.182.82.72:0,password \
-k en-us \
-spice
port=5901,tls-port=5902,addr=128.182.82.72,x509-dir=/etc/pki/vdsm/libvirt
-spice,tls-channel=main,tls-channel=display,tls-channel=inputs,tls-channel=curso
r,tls-channel=playback,tls-channel=record,tls-channel=smartcard,tls-channel=usbr
edir,seamless-migration=on \
-device
qxl-vga,id=ua-bd6ad9d1-2d02-40fa-b087-4fc4786a2f44,ram_size=67108864,vra
m_size=33554432,vram64_size_mb=0,vgamem_mb=16,max_outputs=1,bus=pcie.0,addr=0x1
\
-device
vfio-pci,host=0000:04:00.1,id=ua-7fce29ba-fcf7-4bd3-a680-21b200d44e42,bu
s=pci.3,addr=0x0 \
-object
rng-random,id=objua-e1fb82d8-2455-4763-a5d9-22d7a9903806,filename=/dev/u
random \
-device
virtio-rng-pci,rng=objua-e1fb82d8-2455-4763-a5d9-22d7a9903806,id=ua-e1fb
82d8-2455-4763-a5d9-22d7a9903806,bus=pci.5,addr=0x0 \
-sandbox
on,obsolete=deny,elevateprivileges=deny,spawn=deny,resourcecontrol=deny
\
-msg timestamp=on
2020-10-21 20:26:39.267+0000: Domain id=1 is tainted:
custom-hypervisor-feature
libvirt: error : internal error: Invalid device 0000:04:00.1 iommu_group
file /
sys/bus/pci/devices/0000:04:00.1/iommu_group is not a symlink
2020-10-21 20:26:39.298+0000: shutting down, reason=failed
2020-10-21 20:27:41.683+0000: starting up libvirt version: 6.0.0, package:
25.2.
el8 (CBS <cbs(a)centos.org>, 2020-09-10-13:20:10, ), qemu version:
4.2.0qemu-kvm-4
.2.0-29.el8.3, kernel: 4.18.0-193.19.1.el8_2.x86_64, hostname:
n001.bil.psc.edu
LC_ALL=C \
PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin \
HOME=/var/lib/libvirt/qemu/domain-2-C8.2test \
XDG_DATA_HOME=/var/lib/libvirt/qemu/domain-2-C8.2test/.local/share \
XDG_CACHE_HOME=/var/lib/libvirt/qemu/domain-2-C8.2test/.cache \
XDG_CONFIG_HOME=/var/lib/libvirt/qemu/domain-2-C8.2test/.config \
QEMU_AUDIO_DRV=spice \
/usr/libexec/qemu-kvm \
-name guest=C8.2test,debug-threads=on \
-S \
-object
secret,id=masterKey0,format=raw,file=/var/lib/libvirt/qemu/domain-2-C8.2
test/master-key.aes \
-machine pc-q35-rhel8.1.0,accel=kvm,usb=off,dump-guest-core=off \
-cpu Opteron_G5 \
-m size=16777216k,slots=16,maxmem=67108864k \
-overcommit mem-lock=off \
-smp 4,maxcpus=16,sockets=16,dies=1,cores=1,threads=1 \
-numa node,nodeid=0,cpus=0-15,mem=16384 \
-uuid 07eb98af-0d9f-4f9c-967a-8af302c68703 \
-smbios
type=1,manufacturer=oVirt,product=RHEL,version=8.2-2.2004.0.2.el8,serial
=47513848-0036-2500-905b-0025905bff22,uuid=07eb98af-0d9f-4f9c-967a-8af302c68703,
family=oVirt \
-no-user-config \
-nodefaults \
-chardev socket,id=charmonitor,fd=38,server,nowait \
-mon chardev=charmonitor,id=monitor,mode=control \
-rtc base=2020-10-21T20:27:40,driftfix=slew \
-global kvm-pit.lost_tick_policy=delay \
-no-hpet \
-no-shutdown \
-global ICH9-LPC.disable_s3=1 \
-global ICH9-LPC.disable_s4=1 \
-boot strict=on \
-device
pcie-root-port,port=0x10,chassis=1,id=pci.1,bus=pcie.0,multifunction=on,
addr=0x2 \
-device pcie-root-port,port=0x11,chassis=2,id=pci.2,bus=pcie.0,addr=0x2.0x1
\
-device pcie-root-port,port=0x12,chassis=3,id=pci.3,bus=pcie.0,addr=0x2.0x2
\
-device pcie-root-port,port=0x13,chassis=4,id=pci.4,bus=pcie.0,addr=0x2.0x3
\
-device pcie-root-port,port=0x14,chassis=5,id=pci.5,bus=pcie.0,addr=0x2.0x4
\
-device pcie-root-port,port=0x15,chassis=6,id=pci.6,bus=pcie.0,addr=0x2.0x5
\
-device pcie-root-port,port=0x16,chassis=7,id=pci.7,bus=pcie.0,addr=0x2.0x6
\
-device pcie-root-port,port=0x17,chassis=8,id=pci.8,bus=pcie.0,addr=0x2.0x7
\
-device
pcie-root-port,port=0x18,chassis=9,id=pci.9,bus=pcie.0,multifunction=on,
addr=0x3 \
-device
pcie-root-port,port=0x19,chassis=10,id=pci.10,bus=pcie.0,addr=0x3.0x1 \
-device
pcie-root-port,port=0x1a,chassis=11,id=pci.11,bus=pcie.0,addr=0x3.0x2 \
-device
pcie-root-port,port=0x1b,chassis=12,id=pci.12,bus=pcie.0,addr=0x3.0x3 \
-device
pcie-root-port,port=0x1c,chassis=13,id=pci.13,bus=pcie.0,addr=0x3.0x4 \
-device
pcie-root-port,port=0x1d,chassis=14,id=pci.14,bus=pcie.0,addr=0x3.0x5 \
-device
pcie-root-port,port=0x1e,chassis=15,id=pci.15,bus=pcie.0,addr=0x3.0x6 \
-device
pcie-root-port,port=0x1f,chassis=16,id=pci.16,bus=pcie.0,addr=0x3.0x7 \
-device
qemu-xhci,p2=8,p3=8,id=ua-cd3a4abd-57e9-4f2e-a61f-01128ead8702,bus=pci.2
,addr=0x0 \
-device
virtio-scsi-pci,id=ua-64b19c77-98d3-4946-a0f8-c43c1347a5dc,bus=pci.1,add
r=0x0 \
-device virtio-serial-pci,id=virtio-serial0,max_ports=16,bus=pci.4,addr=0x0
\
-device
ide-cd,bus=ide.2,id=ua-271c6d06-1f3c-4e45-99cc-a39f39a6dd1b,werror=repor
t,rerror=report \
-blockdev '{"driver":"file","filename":"/rhev/data-center/mnt/10.200.16.152:
_var
_nfs__shares_data/662d0028-ceaf-4641-87fa-839254e15209/images/4c332ece-f6db-4b22
-879c-b6912c5b27c0/2918a5f1-c801-4a34-8cf7-89045265f6f3","aio":"threads","node-n
ame":"libvirt-1-storage","cache":{"direct":true,"no-flush":false},"auto-read-onl
y":true,"discard":"unmap"}' \
-blockdev
'{"node-name":"libvirt-1-format","read-only":false,"cache":{"direct":t
rue,"no-flush":false},"driver":"raw","file":"libvirt-1-storage"}' \
-device
scsi-hd,bus=ua-64b19c77-98d3-4946-a0f8-c43c1347a5dc.0,channel=0,scsi-id=
0,lun=0,device_id=4c332ece-f6db-4b22-879c-b6912c5b27c0,drive=libvirt-1-format,id
=ua-4c332ece-f6db-4b22-879c-b6912c5b27c0,bootindex=1,write-cache=on,serial=4c332
ece-f6db-4b22-879c-b6912c5b27c0,werror=stop,rerror=stop \
-chardev socket,id=charchannel0,fd=40,server,nowait \
-device
virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel
0,name=ovirt-guest-agent.0 \
-chardev socket,id=charchannel1,fd=41,server,nowait \
-device
virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel1,id=channel
1,name=org.qemu.guest_agent.0 \
-chardev spicevmc,id=charchannel2,name=vdagent \
-device
virtserialport,bus=virtio-serial0.0,nr=3,chardev=charchannel2,id=channel
2,name=com.redhat.spice.0 \
-device
usb-tablet,id=input0,bus=ua-cd3a4abd-57e9-4f2e-a61f-01128ead8702.0,port=
1 \
-vnc 128.182.82.72:0,password \
-k en-us \
-spice
port=5901,tls-port=5902,addr=128.182.82.72,x509-dir=/etc/pki/vdsm/libvirt
-spice,tls-channel=main,tls-channel=display,tls-channel=inputs,tls-channel=curso
r,tls-channel=playback,tls-channel=record,tls-channel=smartcard,tls-channel=usbr
edir,seamless-migration=on \
-device
qxl-vga,id=ua-bd6ad9d1-2d02-40fa-b087-4fc4786a2f44,ram_size=67108864,vra
m_size=33554432,vram64_size_mb=0,vgamem_mb=16,max_outputs=1,bus=pcie.0,addr=0x1
\
-device
vfio-pci,host=0000:04:00.1,id=ua-7fce29ba-fcf7-4bd3-a680-21b200d44e42,bu
s=pci.3,addr=0x0 \
-object
rng-random,id=objua-e1fb82d8-2455-4763-a5d9-22d7a9903806,filename=/dev/u
random \
-device
virtio-rng-pci,rng=objua-e1fb82d8-2455-4763-a5d9-22d7a9903806,id=ua-e1fb
82d8-2455-4763-a5d9-22d7a9903806,bus=pci.5,addr=0x0 \
-sandbox
on,obsolete=deny,elevateprivileges=deny,spawn=deny,resourcecontrol=deny
\
-msg timestamp=on
2020-10-21 20:27:41.684+0000: Domain id=2 is tainted:
custom-hypervisor-feature
libvirt: error : internal error: Invalid device 0000:04:00.1 iommu_group
file /
sys/bus/pci/devices/0000:04:00.1/iommu_group is not a symlink
2020-10-21 20:27:41.715+0000: shutting down, reason=failed
4 years, 5 months
Unable to connect to postgres database on oVirt engine
by kushagra2agarwal@gmail.com
I am unable to connect to postgres database running on oVirt engine and while troubleshooting found below possible issue.
semanage fcontext -a -t postgresql_exec_t /opt/rh/rh-postgresql10/root/usr/bin/psql
ValueError: File spec /opt/rh/rh-postgresql10/root/usr/bin/psql conflicts with equivalency rule '/opt/rh/rh-postgresql10/root /'; Try adding '//usr/bin/psql' instead
Can someone help to resolve this issue.
4 years, 5 months
VM has snapshot with disk in illegal status
by Massimo Mad
Hi,
I try to delete a snapshot that i have created in the past, now in the
events i have the following error:
Failed to delete snapshot
And the VM now has the exclamation point with the following message:
VM has snapshot with disk in illegal status
How i can fix this problem ?
Regards
Massimo
4 years, 5 months
ovirt Storage Domain Cluster Filesystem?
by lsc.oraes@gmail.com
Hi all
I come from Oracle VM x86 world and we are planning from moving Oracle VM to oVirt.
I am having hard time understanding Storage Domains in oVirt. All our storage are SAN and I wonder how can we manage SAN LUN in oVirt to create a storage domain such that the VM guests can run in any host in the oVirt Cluster?
For example in Oracle VM the Storage Repository (it is the Storage Domain in OVM words) are based on SAN LUNs and on top of that a cluster filesystem is created so all hosts in the cluster have concurrent access to the storage repository and the VM guest can be started in any of the hosts in the cluster.
How do we accomplish the same in oVirt with SAN Storage? Which Cluster Filesystem is supported in Storage Domain?
Or perhaps in oVirt the mechanism is totally different?
Thank you
4 years, 5 months
Re: Gluster volume not responding
by Strahil Nikolov
Hi Simon,
Usually it is the network, but you need real-world data. I would open screen sessions and run ping continiously . Something like this:
while true; do echo -n "$(date) "; timeout -s 9 1 ping -c 1 ovirt2 | grep icmp_seq; sleep 1; done | tee -a /tmp/icmp_log
Are all systems in the same network ?
What about dns resolution - do you have entries in /etc/hosts ?
Best Regards,
Strahil Nikolov
В неделя, 11 октомври 2020 г., 11:54:47 Гринуич+3, Simon Scott <simon(a)justconnect.ie> написа:
Thanks Strahil.
I have found between 1 & 4 Gluster peer rpc-clnt-ping timer expired messages in the rhev-data-center-mnt-glusterSD-hostname-strg:_pltfm_data01.log on the storage network IP. Of the 6 Hosts only 1 does not have these timeouts.
Fencing has been disabled but can you identify which logs are key to identifying the cause please.
It's a bonded (bond1) 10GB ovirt-mgmt logical network and Prod VM VLAN interface AND a bonded (bond2) 10GB Gluster storage network.
Dropped packets are seen incrementing in the vdsm.log but neither ethtool -S or kernel logs are showing dropped packets. I am wondering if they are being dropped due to the ring buffers being small.
Kind Regards
Shimme
________________________________
From: Strahil Nikolov <hunter86_bg(a)yahoo.com>
Sent: Thursday 8 October 2020 20:40
To: users(a)ovirt.org <users(a)ovirt.org>; Simon Scott <simon(a)justconnect.ie>
Subject: Re: [ovirt-users] Gluster volume not responding
>Every Monday and Wednesday morning there are gluster connectivity timeouts >but all checks of the network and network configs are ok.
Based on this one I make the following conclusions:
1. Issue is reoccuring
2. You most probably have a network issue
Have you checked the following:
- are there any ping timeouts between fuse clients and gluster nodes
- Have you tried to disable fencing and check the logs after the issue reoccurs
- Are you sharing Blackup and Prod networks ? Is it possible some backup/other production load in your environment to "black-out" your oVirt ?
- Have you check the gluster cluster's logs for anything meaningful ?
Best Regards,
Strahil Nikolov
_______________________________________________
Users mailing list -- users(a)ovirt.org
To unsubscribe send an email to users-leave(a)ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/U527TGUQR6R...
4 years, 5 months
New host: no network interfaces visible
by Richard Chan
What could cause no Network Interfaces visible in installing a new host?
I have added a new host to oVirt 4.3.10, the initial SSH installs all the
packages (all with failed=0 in host-deploy/*log). The installation is shown
as failed without any meaningful message.
However the Network Interfaces page is blank, what could cause this?
Host in CentOS 7.8 tried with/without biosdevname.
--
Richard Chan
4 years, 5 months