oVirt test clustered DOA after upgrade to latest 4.4.2 (AFAIK). Network issue?
by Gilboa Davara
Hello all,
I just upgraded one of my test oVirt setups to latest.
Post reboot, the machine lost network and hosted engine didn't start.
When I connected to the machine, I noticed all the
/etc/sysconfig/network-scripts/ifcfg-* files disappeared. No idea why /
how. (Possibly I did something wrong and forgot about it.)
Long story story short, I copied the two missing files (ifcfg-onb0 ethernet
device configuration and the ifcfg-ovirtmgmt bridge configuration) from
another ovirt host, changed the ovirtmgmt IP address and UUID to match the
UUID reported in logs as missing and restarted NetworkManager and restarted
all the oVirt related services (vdsmd, supervdsmd, ovirt-*, etc).
Sadly enough, even with both onb0 and ovirtmgmt up, vdsm still complains
about the missing network (ovirtmgmt) and refuses to start the hosted
engine.
Reboot doesn't seem to change anything.
In the main log I see the following errors:
Oct 22 16:33:31 office-wx-otest vdsm[2634]: WARN Attempting to remove a non
existing network: ovirtmgmt/1da8c5b7-999c-4ada-8287-1f35de6ce21d
Oct 22 16:33:31 office-wx-otest vdsm[2634]: WARN Attempting to remove a non
existing net user: ovirtmgmt/1da8c5b7-999c-4ada-8287-1f35de6ce21d
Oct 22 16:33:31 office-wx-otest vdsm[2634]: WARN Attempting to remove a non
existing network: ovirtmgmt/1da8c5b7-999c-4ada-8287-1f35de6ce21d
Oct 22 16:33:31 office-wx-otest vdsm[2634]: WARN Attempting to remove a non
existing net user: ovirtmgmt/1da8c5b7-999c-4ada-8287-1f35de6ce21d
As it is one of my oVirt test setup, I can simply redeploy the host and
continue from there, but I rather use this experience to learn how to fix
oVirt such issues in the future.
Logs attached.
https://drive.google.com/file/d/12ugy6CuaFaMvXYt6uGT4D_EHIW6nXttb/view?us...
$ PAGER= nmcli connection show
NAME UUID TYPE DEVICE
ovirtmgmt 1da8c5b7-999c-4ada-8287-1f35de6ce21d bridge ovirtmgmt
onb0 48332db3-8939-bff3-6b71-772a28c9e7b8 ethernet onb0
$ PAGER= nmcli device show
GENERAL.DEVICE: ovirtmgmt
GENERAL.TYPE: bridge
GENERAL.HWADDR: FC:AA:14:6B:A8:E0
GENERAL.MTU: 1500
GENERAL.STATE: 100 (connected)
GENERAL.CONNECTION: ovirtmgmt
GENERAL.CON-PATH:
/org/freedesktop/NetworkManager/ActiveConnection/2
IP4.ADDRESS[1]: 192.168.2.117/24
IP4.GATEWAY: 192.168.2.100
IP4.ROUTE[1]: dst = 192.168.2.0/24, nh = 0.0.0.0,
mt = 425
IP4.ROUTE[2]: dst = 0.0.0.0/0, nh =
192.168.2.100, mt = 425
IP4.DNS[1]: 192.168.2.100
IP4.DNS[2]: 8.8.8.8
IP6.GATEWAY: --
GENERAL.DEVICE: onb0
GENERAL.TYPE: ethernet
GENERAL.HWADDR: FC:AA:14:6B:A8:E0
GENERAL.MTU: 1500
GENERAL.STATE: 100 (connected)
GENERAL.CONNECTION: onb0
GENERAL.CON-PATH:
/org/freedesktop/NetworkManager/ActiveConnection/3
WIRED-PROPERTIES.CARRIER: on
IP4.GATEWAY: --
GENERAL.DEVICE: ;vdsmdummy;
GENERAL.TYPE: bridge
GENERAL.HWADDR: 92:8B:9A:5E:C1:3E
GENERAL.MTU: 1500
GENERAL.STATE: 10 (unmanaged)
GENERAL.CONNECTION: --
GENERAL.CON-PATH: --
IP4.GATEWAY: --
IP6.GATEWAY: --
GENERAL.DEVICE: lo
GENERAL.TYPE: loopback
GENERAL.HWADDR: 00:00:00:00:00:00
GENERAL.MTU: 65536
GENERAL.STATE: 10 (unmanaged)
GENERAL.CONNECTION: --
GENERAL.CON-PATH: --
IP4.ADDRESS[1]: 127.0.0.1/8
IP4.GATEWAY: --
IP6.GATEWAY: --
GENERAL.DEVICE: br-int
GENERAL.TYPE: openvswitch
GENERAL.HWADDR: 8E:15:6A:F8:3C:45
GENERAL.MTU: 1500
GENERAL.STATE: 10 (unmanaged)
GENERAL.CONNECTION: --
GENERAL.CON-PATH: --
IP4.GATEWAY: --
IP6.GATEWAY: --
GENERAL.DEVICE: ovs-system
GENERAL.TYPE: openvswitch
GENERAL.HWADDR: E2:09:EA:A2:BD:70
GENERAL.MTU: 1500
GENERAL.STATE: 10 (unmanaged)
GENERAL.CONNECTION: --
GENERAL.CON-PATH: --
IP4.GATEWAY: --
IP6.GATEWAY:
- Gilboa
office-wx-otest-vdsm.bz2
<https://drive.google.com/file/d/12ugy6CuaFaMvXYt6uGT4D_EHIW6nXttb/view?us...>
4 years, 2 months
sr-iov network config assistance IOMMU group error
by Edward Berger
I'm' trying to bring up a test ovirt 4.4 cluster with sr-iov virtual
infiniband interfaces on Mellanox ConnectX-6 cards in some older Opteron
servers while we await delivery of new Epyc servers.
I have the hypervisors configured for sr-iov and showing the interfaces,
but when I try to bring up a VM with the host device selected it fails to
boot. "Failed Launching ..." shows up on the engine UI.
[image: sriov-hostdev.JPG]
The /var/log/libvirt/qemu log for the VM seems to think something is wrong
with the IOMMU group. If I scroll over to the right it says N/A under the
IOMMU column. Is there some step I need to 'create IOMMU groups' or any
other suggestions?
[root@n001 qemu]# more C*log
2020-10-21 20:26:39.266+0000: starting up libvirt version: 6.0.0, package:
25.2.
el8 (CBS <cbs(a)centos.org>, 2020-09-10-13:20:10, ), qemu version:
4.2.0qemu-kvm-4
.2.0-29.el8.3, kernel: 4.18.0-193.19.1.el8_2.x86_64, hostname:
n001.bil.psc.edu
LC_ALL=C \
PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin \
HOME=/var/lib/libvirt/qemu/domain-1-C8.2test \
XDG_DATA_HOME=/var/lib/libvirt/qemu/domain-1-C8.2test/.local/share \
XDG_CACHE_HOME=/var/lib/libvirt/qemu/domain-1-C8.2test/.cache \
XDG_CONFIG_HOME=/var/lib/libvirt/qemu/domain-1-C8.2test/.config \
QEMU_AUDIO_DRV=spice \
/usr/libexec/qemu-kvm \
-name guest=C8.2test,debug-threads=on \
-S \
-object
secret,id=masterKey0,format=raw,file=/var/lib/libvirt/qemu/domain-1-C8.2
test/master-key.aes \
-machine pc-q35-rhel8.1.0,accel=kvm,usb=off,dump-guest-core=off \
-cpu Opteron_G5 \
-m size=16777216k,slots=16,maxmem=67108864k \
-overcommit mem-lock=off \
-smp 4,maxcpus=16,sockets=16,dies=1,cores=1,threads=1 \
-numa node,nodeid=0,cpus=0-15,mem=16384 \
-uuid 07eb98af-0d9f-4f9c-967a-8af302c68703 \
-smbios
type=1,manufacturer=oVirt,product=RHEL,version=8.2-2.2004.0.2.el8,serial
=47513848-0036-2500-905b-0025905bff22,uuid=07eb98af-0d9f-4f9c-967a-8af302c68703,
family=oVirt \
-no-user-config \
-nodefaults \
-chardev socket,id=charmonitor,fd=39,server,nowait \
-mon chardev=charmonitor,id=monitor,mode=control \
-rtc base=2020-10-21T20:26:38,driftfix=slew \
-global kvm-pit.lost_tick_policy=delay \
-no-hpet \
-no-shutdown \
-global ICH9-LPC.disable_s3=1 \
-global ICH9-LPC.disable_s4=1 \
-boot strict=on \
-device
pcie-root-port,port=0x10,chassis=1,id=pci.1,bus=pcie.0,multifunction=on,
addr=0x2 \
-device pcie-root-port,port=0x11,chassis=2,id=pci.2,bus=pcie.0,addr=0x2.0x1
\
-device pcie-root-port,port=0x12,chassis=3,id=pci.3,bus=pcie.0,addr=0x2.0x2
\
-device pcie-root-port,port=0x13,chassis=4,id=pci.4,bus=pcie.0,addr=0x2.0x3
\
-device pcie-root-port,port=0x14,chassis=5,id=pci.5,bus=pcie.0,addr=0x2.0x4
\
-device pcie-root-port,port=0x15,chassis=6,id=pci.6,bus=pcie.0,addr=0x2.0x5
\
-device pcie-root-port,port=0x16,chassis=7,id=pci.7,bus=pcie.0,addr=0x2.0x6
\
-device pcie-root-port,port=0x17,chassis=8,id=pci.8,bus=pcie.0,addr=0x2.0x7
\
-device
pcie-root-port,port=0x18,chassis=9,id=pci.9,bus=pcie.0,multifunction=on,
addr=0x3 \
-device
pcie-root-port,port=0x19,chassis=10,id=pci.10,bus=pcie.0,addr=0x3.0x1 \
-device
pcie-root-port,port=0x1a,chassis=11,id=pci.11,bus=pcie.0,addr=0x3.0x2 \
-device
pcie-root-port,port=0x1b,chassis=12,id=pci.12,bus=pcie.0,addr=0x3.0x3 \
-device
pcie-root-port,port=0x1c,chassis=13,id=pci.13,bus=pcie.0,addr=0x3.0x4 \
-device
pcie-root-port,port=0x1d,chassis=14,id=pci.14,bus=pcie.0,addr=0x3.0x5 \
-device
pcie-root-port,port=0x1e,chassis=15,id=pci.15,bus=pcie.0,addr=0x3.0x6 \
-device
pcie-root-port,port=0x1f,chassis=16,id=pci.16,bus=pcie.0,addr=0x3.0x7 \
-device
qemu-xhci,p2=8,p3=8,id=ua-cd3a4abd-57e9-4f2e-a61f-01128ead8702,bus=pci.2
,addr=0x0 \
-device
virtio-scsi-pci,id=ua-64b19c77-98d3-4946-a0f8-c43c1347a5dc,bus=pci.1,add
r=0x0 \
-device virtio-serial-pci,id=virtio-serial0,max_ports=16,bus=pci.4,addr=0x0
\
-device
ide-cd,bus=ide.2,id=ua-271c6d06-1f3c-4e45-99cc-a39f39a6dd1b,werror=repor
t,rerror=report \
-blockdev '{"driver":"file","filename":"/rhev/data-center/mnt/10.200.16.152:
_var
_nfs__shares_data/662d0028-ceaf-4641-87fa-839254e15209/images/4c332ece-f6db-4b22
-879c-b6912c5b27c0/2918a5f1-c801-4a34-8cf7-89045265f6f3","aio":"threads","node-n
ame":"libvirt-1-storage","cache":{"direct":true,"no-flush":false},"auto-read-onl
y":true,"discard":"unmap"}' \
-blockdev
'{"node-name":"libvirt-1-format","read-only":false,"cache":{"direct":t
rue,"no-flush":false},"driver":"raw","file":"libvirt-1-storage"}' \
-device
scsi-hd,bus=ua-64b19c77-98d3-4946-a0f8-c43c1347a5dc.0,channel=0,scsi-id=
0,lun=0,device_id=4c332ece-f6db-4b22-879c-b6912c5b27c0,drive=libvirt-1-format,id
=ua-4c332ece-f6db-4b22-879c-b6912c5b27c0,bootindex=1,write-cache=on,serial=4c332
ece-f6db-4b22-879c-b6912c5b27c0,werror=stop,rerror=stop \
-chardev socket,id=charchannel0,fd=41,server,nowait \
-device
virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel
0,name=ovirt-guest-agent.0 \
-chardev socket,id=charchannel1,fd=42,server,nowait \
-device
virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel1,id=channel
1,name=org.qemu.guest_agent.0 \
-chardev spicevmc,id=charchannel2,name=vdagent \
-device
virtserialport,bus=virtio-serial0.0,nr=3,chardev=charchannel2,id=channel
2,name=com.redhat.spice.0 \
-device
usb-tablet,id=input0,bus=ua-cd3a4abd-57e9-4f2e-a61f-01128ead8702.0,port=
1 \
-vnc 128.182.82.72:0,password \
-k en-us \
-spice
port=5901,tls-port=5902,addr=128.182.82.72,x509-dir=/etc/pki/vdsm/libvirt
-spice,tls-channel=main,tls-channel=display,tls-channel=inputs,tls-channel=curso
r,tls-channel=playback,tls-channel=record,tls-channel=smartcard,tls-channel=usbr
edir,seamless-migration=on \
-device
qxl-vga,id=ua-bd6ad9d1-2d02-40fa-b087-4fc4786a2f44,ram_size=67108864,vra
m_size=33554432,vram64_size_mb=0,vgamem_mb=16,max_outputs=1,bus=pcie.0,addr=0x1
\
-device
vfio-pci,host=0000:04:00.1,id=ua-7fce29ba-fcf7-4bd3-a680-21b200d44e42,bu
s=pci.3,addr=0x0 \
-object
rng-random,id=objua-e1fb82d8-2455-4763-a5d9-22d7a9903806,filename=/dev/u
random \
-device
virtio-rng-pci,rng=objua-e1fb82d8-2455-4763-a5d9-22d7a9903806,id=ua-e1fb
82d8-2455-4763-a5d9-22d7a9903806,bus=pci.5,addr=0x0 \
-sandbox
on,obsolete=deny,elevateprivileges=deny,spawn=deny,resourcecontrol=deny
\
-msg timestamp=on
2020-10-21 20:26:39.267+0000: Domain id=1 is tainted:
custom-hypervisor-feature
libvirt: error : internal error: Invalid device 0000:04:00.1 iommu_group
file /
sys/bus/pci/devices/0000:04:00.1/iommu_group is not a symlink
2020-10-21 20:26:39.298+0000: shutting down, reason=failed
2020-10-21 20:27:41.683+0000: starting up libvirt version: 6.0.0, package:
25.2.
el8 (CBS <cbs(a)centos.org>, 2020-09-10-13:20:10, ), qemu version:
4.2.0qemu-kvm-4
.2.0-29.el8.3, kernel: 4.18.0-193.19.1.el8_2.x86_64, hostname:
n001.bil.psc.edu
LC_ALL=C \
PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin \
HOME=/var/lib/libvirt/qemu/domain-2-C8.2test \
XDG_DATA_HOME=/var/lib/libvirt/qemu/domain-2-C8.2test/.local/share \
XDG_CACHE_HOME=/var/lib/libvirt/qemu/domain-2-C8.2test/.cache \
XDG_CONFIG_HOME=/var/lib/libvirt/qemu/domain-2-C8.2test/.config \
QEMU_AUDIO_DRV=spice \
/usr/libexec/qemu-kvm \
-name guest=C8.2test,debug-threads=on \
-S \
-object
secret,id=masterKey0,format=raw,file=/var/lib/libvirt/qemu/domain-2-C8.2
test/master-key.aes \
-machine pc-q35-rhel8.1.0,accel=kvm,usb=off,dump-guest-core=off \
-cpu Opteron_G5 \
-m size=16777216k,slots=16,maxmem=67108864k \
-overcommit mem-lock=off \
-smp 4,maxcpus=16,sockets=16,dies=1,cores=1,threads=1 \
-numa node,nodeid=0,cpus=0-15,mem=16384 \
-uuid 07eb98af-0d9f-4f9c-967a-8af302c68703 \
-smbios
type=1,manufacturer=oVirt,product=RHEL,version=8.2-2.2004.0.2.el8,serial
=47513848-0036-2500-905b-0025905bff22,uuid=07eb98af-0d9f-4f9c-967a-8af302c68703,
family=oVirt \
-no-user-config \
-nodefaults \
-chardev socket,id=charmonitor,fd=38,server,nowait \
-mon chardev=charmonitor,id=monitor,mode=control \
-rtc base=2020-10-21T20:27:40,driftfix=slew \
-global kvm-pit.lost_tick_policy=delay \
-no-hpet \
-no-shutdown \
-global ICH9-LPC.disable_s3=1 \
-global ICH9-LPC.disable_s4=1 \
-boot strict=on \
-device
pcie-root-port,port=0x10,chassis=1,id=pci.1,bus=pcie.0,multifunction=on,
addr=0x2 \
-device pcie-root-port,port=0x11,chassis=2,id=pci.2,bus=pcie.0,addr=0x2.0x1
\
-device pcie-root-port,port=0x12,chassis=3,id=pci.3,bus=pcie.0,addr=0x2.0x2
\
-device pcie-root-port,port=0x13,chassis=4,id=pci.4,bus=pcie.0,addr=0x2.0x3
\
-device pcie-root-port,port=0x14,chassis=5,id=pci.5,bus=pcie.0,addr=0x2.0x4
\
-device pcie-root-port,port=0x15,chassis=6,id=pci.6,bus=pcie.0,addr=0x2.0x5
\
-device pcie-root-port,port=0x16,chassis=7,id=pci.7,bus=pcie.0,addr=0x2.0x6
\
-device pcie-root-port,port=0x17,chassis=8,id=pci.8,bus=pcie.0,addr=0x2.0x7
\
-device
pcie-root-port,port=0x18,chassis=9,id=pci.9,bus=pcie.0,multifunction=on,
addr=0x3 \
-device
pcie-root-port,port=0x19,chassis=10,id=pci.10,bus=pcie.0,addr=0x3.0x1 \
-device
pcie-root-port,port=0x1a,chassis=11,id=pci.11,bus=pcie.0,addr=0x3.0x2 \
-device
pcie-root-port,port=0x1b,chassis=12,id=pci.12,bus=pcie.0,addr=0x3.0x3 \
-device
pcie-root-port,port=0x1c,chassis=13,id=pci.13,bus=pcie.0,addr=0x3.0x4 \
-device
pcie-root-port,port=0x1d,chassis=14,id=pci.14,bus=pcie.0,addr=0x3.0x5 \
-device
pcie-root-port,port=0x1e,chassis=15,id=pci.15,bus=pcie.0,addr=0x3.0x6 \
-device
pcie-root-port,port=0x1f,chassis=16,id=pci.16,bus=pcie.0,addr=0x3.0x7 \
-device
qemu-xhci,p2=8,p3=8,id=ua-cd3a4abd-57e9-4f2e-a61f-01128ead8702,bus=pci.2
,addr=0x0 \
-device
virtio-scsi-pci,id=ua-64b19c77-98d3-4946-a0f8-c43c1347a5dc,bus=pci.1,add
r=0x0 \
-device virtio-serial-pci,id=virtio-serial0,max_ports=16,bus=pci.4,addr=0x0
\
-device
ide-cd,bus=ide.2,id=ua-271c6d06-1f3c-4e45-99cc-a39f39a6dd1b,werror=repor
t,rerror=report \
-blockdev '{"driver":"file","filename":"/rhev/data-center/mnt/10.200.16.152:
_var
_nfs__shares_data/662d0028-ceaf-4641-87fa-839254e15209/images/4c332ece-f6db-4b22
-879c-b6912c5b27c0/2918a5f1-c801-4a34-8cf7-89045265f6f3","aio":"threads","node-n
ame":"libvirt-1-storage","cache":{"direct":true,"no-flush":false},"auto-read-onl
y":true,"discard":"unmap"}' \
-blockdev
'{"node-name":"libvirt-1-format","read-only":false,"cache":{"direct":t
rue,"no-flush":false},"driver":"raw","file":"libvirt-1-storage"}' \
-device
scsi-hd,bus=ua-64b19c77-98d3-4946-a0f8-c43c1347a5dc.0,channel=0,scsi-id=
0,lun=0,device_id=4c332ece-f6db-4b22-879c-b6912c5b27c0,drive=libvirt-1-format,id
=ua-4c332ece-f6db-4b22-879c-b6912c5b27c0,bootindex=1,write-cache=on,serial=4c332
ece-f6db-4b22-879c-b6912c5b27c0,werror=stop,rerror=stop \
-chardev socket,id=charchannel0,fd=40,server,nowait \
-device
virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel
0,name=ovirt-guest-agent.0 \
-chardev socket,id=charchannel1,fd=41,server,nowait \
-device
virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel1,id=channel
1,name=org.qemu.guest_agent.0 \
-chardev spicevmc,id=charchannel2,name=vdagent \
-device
virtserialport,bus=virtio-serial0.0,nr=3,chardev=charchannel2,id=channel
2,name=com.redhat.spice.0 \
-device
usb-tablet,id=input0,bus=ua-cd3a4abd-57e9-4f2e-a61f-01128ead8702.0,port=
1 \
-vnc 128.182.82.72:0,password \
-k en-us \
-spice
port=5901,tls-port=5902,addr=128.182.82.72,x509-dir=/etc/pki/vdsm/libvirt
-spice,tls-channel=main,tls-channel=display,tls-channel=inputs,tls-channel=curso
r,tls-channel=playback,tls-channel=record,tls-channel=smartcard,tls-channel=usbr
edir,seamless-migration=on \
-device
qxl-vga,id=ua-bd6ad9d1-2d02-40fa-b087-4fc4786a2f44,ram_size=67108864,vra
m_size=33554432,vram64_size_mb=0,vgamem_mb=16,max_outputs=1,bus=pcie.0,addr=0x1
\
-device
vfio-pci,host=0000:04:00.1,id=ua-7fce29ba-fcf7-4bd3-a680-21b200d44e42,bu
s=pci.3,addr=0x0 \
-object
rng-random,id=objua-e1fb82d8-2455-4763-a5d9-22d7a9903806,filename=/dev/u
random \
-device
virtio-rng-pci,rng=objua-e1fb82d8-2455-4763-a5d9-22d7a9903806,id=ua-e1fb
82d8-2455-4763-a5d9-22d7a9903806,bus=pci.5,addr=0x0 \
-sandbox
on,obsolete=deny,elevateprivileges=deny,spawn=deny,resourcecontrol=deny
\
-msg timestamp=on
2020-10-21 20:27:41.684+0000: Domain id=2 is tainted:
custom-hypervisor-feature
libvirt: error : internal error: Invalid device 0000:04:00.1 iommu_group
file /
sys/bus/pci/devices/0000:04:00.1/iommu_group is not a symlink
2020-10-21 20:27:41.715+0000: shutting down, reason=failed
4 years, 2 months
Unable to connect to postgres database on oVirt engine
by kushagra2agarwal@gmail.com
I am unable to connect to postgres database running on oVirt engine and while troubleshooting found below possible issue.
semanage fcontext -a -t postgresql_exec_t /opt/rh/rh-postgresql10/root/usr/bin/psql
ValueError: File spec /opt/rh/rh-postgresql10/root/usr/bin/psql conflicts with equivalency rule '/opt/rh/rh-postgresql10/root /'; Try adding '//usr/bin/psql' instead
Can someone help to resolve this issue.
4 years, 2 months
VM has snapshot with disk in illegal status
by Massimo Mad
Hi,
I try to delete a snapshot that i have created in the past, now in the
events i have the following error:
Failed to delete snapshot
And the VM now has the exclamation point with the following message:
VM has snapshot with disk in illegal status
How i can fix this problem ?
Regards
Massimo
4 years, 2 months
ovirt Storage Domain Cluster Filesystem?
by lsc.oraes@gmail.com
Hi all
I come from Oracle VM x86 world and we are planning from moving Oracle VM to oVirt.
I am having hard time understanding Storage Domains in oVirt. All our storage are SAN and I wonder how can we manage SAN LUN in oVirt to create a storage domain such that the VM guests can run in any host in the oVirt Cluster?
For example in Oracle VM the Storage Repository (it is the Storage Domain in OVM words) are based on SAN LUNs and on top of that a cluster filesystem is created so all hosts in the cluster have concurrent access to the storage repository and the VM guest can be started in any of the hosts in the cluster.
How do we accomplish the same in oVirt with SAN Storage? Which Cluster Filesystem is supported in Storage Domain?
Or perhaps in oVirt the mechanism is totally different?
Thank you
4 years, 2 months
Re: Gluster volume not responding
by Strahil Nikolov
Hi Simon,
Usually it is the network, but you need real-world data. I would open screen sessions and run ping continiously . Something like this:
while true; do echo -n "$(date) "; timeout -s 9 1 ping -c 1 ovirt2 | grep icmp_seq; sleep 1; done | tee -a /tmp/icmp_log
Are all systems in the same network ?
What about dns resolution - do you have entries in /etc/hosts ?
Best Regards,
Strahil Nikolov
В неделя, 11 октомври 2020 г., 11:54:47 Гринуич+3, Simon Scott <simon(a)justconnect.ie> написа:
Thanks Strahil.
I have found between 1 & 4 Gluster peer rpc-clnt-ping timer expired messages in the rhev-data-center-mnt-glusterSD-hostname-strg:_pltfm_data01.log on the storage network IP. Of the 6 Hosts only 1 does not have these timeouts.
Fencing has been disabled but can you identify which logs are key to identifying the cause please.
It's a bonded (bond1) 10GB ovirt-mgmt logical network and Prod VM VLAN interface AND a bonded (bond2) 10GB Gluster storage network.
Dropped packets are seen incrementing in the vdsm.log but neither ethtool -S or kernel logs are showing dropped packets. I am wondering if they are being dropped due to the ring buffers being small.
Kind Regards
Shimme
________________________________
From: Strahil Nikolov <hunter86_bg(a)yahoo.com>
Sent: Thursday 8 October 2020 20:40
To: users(a)ovirt.org <users(a)ovirt.org>; Simon Scott <simon(a)justconnect.ie>
Subject: Re: [ovirt-users] Gluster volume not responding
>Every Monday and Wednesday morning there are gluster connectivity timeouts >but all checks of the network and network configs are ok.
Based on this one I make the following conclusions:
1. Issue is reoccuring
2. You most probably have a network issue
Have you checked the following:
- are there any ping timeouts between fuse clients and gluster nodes
- Have you tried to disable fencing and check the logs after the issue reoccurs
- Are you sharing Blackup and Prod networks ? Is it possible some backup/other production load in your environment to "black-out" your oVirt ?
- Have you check the gluster cluster's logs for anything meaningful ?
Best Regards,
Strahil Nikolov
_______________________________________________
Users mailing list -- users(a)ovirt.org
To unsubscribe send an email to users-leave(a)ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/U527TGUQR6R...
4 years, 2 months
New host: no network interfaces visible
by Richard Chan
What could cause no Network Interfaces visible in installing a new host?
I have added a new host to oVirt 4.3.10, the initial SSH installs all the
packages (all with failed=0 in host-deploy/*log). The installation is shown
as failed without any meaningful message.
However the Network Interfaces page is blank, what could cause this?
Host in CentOS 7.8 tried with/without biosdevname.
--
Richard Chan
4 years, 2 months
[ANN] oVirt 4.4.3 Sixth Release Candidate is now available for testing
by Sandro Bonazzola
oVirt 4.4.3 Sixth Release Candidate is now available for testing
The oVirt Project is pleased to announce the availability of oVirt 4.4.3
Sixth Release Candidate for testing, as of October 22th, 2020.
This update is the third in a series of stabilization updates to the 4.4
series.
How to prevent hosts entering emergency mode after upgrade from oVirt 4.4.1
Note: Upgrading from 4.4.2 GA should not require re-doing these steps, if
already performed while upgrading from 4.4.1 to 4.4.2 GA. These are only
required to be done once.
Due to Bug 1837864 <https://bugzilla.redhat.com/show_bug.cgi?id=1837864> -
Host enter emergency mode after upgrading to latest build
If you have your root file system on a multipath device on your hosts you
should be aware that after upgrading from 4.4.1 to 4.4.3 you may get your
host entering emergency mode.
In order to prevent this be sure to upgrade oVirt Engine first, then on
your hosts:
1.
Remove the current lvm filter while still on 4.4.1, or in emergency mode
(if rebooted).
2.
Reboot.
3.
Upgrade to 4.4.3 (redeploy in case of already being on 4.4.3).
4.
Run vdsm-tool config-lvm-filter to confirm there is a new filter in
place.
5.
Only if not using oVirt Node:
- run "dracut --force --add multipath” to rebuild initramfs with the
correct filter configuration
6.
Reboot.
Documentation
-
If you want to try oVirt as quickly as possible, follow the instructions
on the Download <https://ovirt.org/download/> page.
-
For complete installation, administration, and usage instructions, see
the oVirt Documentation <https://ovirt.org/documentation/>.
-
For upgrading from a previous version, see the oVirt Upgrade Guide
<https://ovirt.org/documentation/upgrade_guide/>.
-
For a general overview of oVirt, see About oVirt
<https://ovirt.org/community/about.html>.
Important notes before you try it
Please note this is a pre-release build.
The oVirt Project makes no guarantees as to its suitability or usefulness.
This pre-release must not be used in production.
While upgrading from 4.4.2 to 4.4.3 RC6 You can hit:
Problem: cannot install the best update candidate for package
ovirt-engine-metrics-1.4.1.1-1.el8.noarch
- nothing provides rhel-system-roles >= 1.0-19 needed by
ovirt-engine-metrics-1.4.2-1.el8.noarch
in order to get rhel-system-roles >= 1.0-19 you need
https://buildlogs.centos.org/centos/8/virt/x86_64/ovirt-44/ repo since that
package can be promoted to release only at 4.4.3 GA.
Installation instructions
For installation instructions and additional information please refer to:
https://ovirt.org/documentation/
This release is available now on x86_64 architecture for:
* Red Hat Enterprise Linux 8.2 or newer
* CentOS Linux (or similar) 8.2 or newer
This release supports Hypervisor Hosts on x86_64 and ppc64le architectures
for:
* Red Hat Enterprise Linux 8.2 or newer
* CentOS Linux (or similar) 8.2 or newer
* oVirt Node 4.4 based on CentOS Linux 8.2 (available for x86_64 only)
See the release notes [1] for installation instructions and a list of new
features and bugs fixed.
Notes:
- oVirt Appliance is already available for CentOS Linux 8
- oVirt Node NG is already available for CentOS Linux 8
Additional Resources:
* Read more about the oVirt 4.4.3 release highlights:
http://www.ovirt.org/release/4.4.3/
* Get more oVirt project updates on Twitter: https://twitter.com/ovirt
* Check out the latest project news on the oVirt blog:
http://www.ovirt.org/blog/
[1] http://www.ovirt.org/release/4.4.3/
[2] http://resources.ovirt.org/pub/ovirt-4.4-pre/iso/
--
Sandro Bonazzola
MANAGER, SOFTWARE ENGINEERING, EMEA R&D RHV
Red Hat EMEA <https://www.redhat.com/>
sbonazzo(a)redhat.com
<https://www.redhat.com/>
*Red Hat respects your work life balance. Therefore there is no need to
answer this email out of your office hours.*
* <https://www.redhat.com/it/forums/emea/italy-track>*
4 years, 2 months