Migrating VMs with templates from 4.2.8 to new 4.4.6 instance
by Pavel Strzinek
Hello,
I am having hard time migrating vms based on templates from existing 4.2.8 instance to newly installed 4.4.6 HCI with different storage. When I export VMs and corresponding templates to export NFS storage, detach it from source DC and attach the export storage to a new oVirt environment, the import of templates is failing with error "Failed to import Template XXX to Data Center YYY, Cluster ZZZ". I cannot find anything more specific about the error in logs. Am I missing something?
Exporting to OVA and importing back does work, but I want to make use of template thin provisioning.
2 years, 11 months
oVirt 2021 Spring survey
by Sandro Bonazzola
As we continue to develop oVirt 4.4, the Development and Integration teams
at Red Hat would value insights on how you are deploying the oVirt
environment.
Please help us to hit the mark by completing this short survey. Survey will
close on *May 30th 2021*.
If you're managing multiple oVirt deployments with very different use cases
or very different deployments you can consider answering this survey
multiple times.
*Please note the answers to this survey will be publicly accessible*.
This survey is under oVirt Privacy Policy available at
https://ovirt.org/privacy-policy.html .
The survey form is available at
https://docs.google.com/forms/d/e/1FAIpQLScdJGoBYxuW-4IsIvZGVpbiEWhmt4O-o...
--
Sandro Bonazzola
MANAGER, SOFTWARE ENGINEERING, EMEA R&D RHV
Red Hat EMEA <https://www.redhat.com/>
sbonazzo(a)redhat.com
<https://www.redhat.com/>
*Red Hat respects your work life balance. Therefore there is no need to
answer this email out of your office hours.*
2 years, 11 months
Create Brick from Engine host view
by Harry O
When I try to create a single disk brick via host view "storage devices" on engine, I get the following error.
Error while executing action Create Brick: Internal Engine Error
Failed to create brick lalaf on host hej1.5ervers.lan of cluster Clu1.
I want the brick to be single disk no raid, no cache. Is there a way to create it via CLI? Do I need to pull some logs?
2 years, 11 months
Power management on Dell PowerEdge R320 and R520
by Pavel Strzinek
Hello,
I am having trouble configuring fencing on these two servers with iDrac7 module from freshly installed oVirt 4.4.6. I tried drac5, drac7 and ipmilan modules and neither passes the test. I used ipmilan with option "lanplus=1", as noted in previous threads about idrac usage with oVirt/RHEV, but with no success. Also, I am successfully using ipmilan fencing from older oVirt 4.2 on several SuperMicro server nodes.
This is the error message in engine.log:
2021-05-14 09:28:41,185+02 ERROR [org.ovirt.engine.core.bll.pm.FenceProxyLocator] (default task-6) [12faf352-da4a-4b24-9ab9-af54559cecd1] Can not run fence action on host 'onode1', no suitable proxy host was found.
I can successfully query the idrac module with ipmitool from command line on the node, using the same credentials.
2 years, 11 months
[ANN] Async release for oVirt 4.4.6
by Sandro Bonazzola
On May 14th 2021 the oVirt project released an async update to the
following packages:
- ovirt-hosted-engine-ha-2.4.7
- ovirt-release44-4.4.6.2
- ovirt-engine-4.4.6.8
- oVirt Node 4.4.6.2
Fixing the following bugs:
- Bug 1909888 <https://bugzilla.redhat.com/show_bug.cgi?id=1909888> - [RFE]
Support multiple IQN in hosted-engine.conf for Active-Active DR setup
- Bug 1957253 <https://bugzilla.redhat.com/show_bug.cgi?id=1957253>
- [cinderlib]
Enable using Managed Block Storage on 4.6 cluster by default
- Bug 1958869 <https://bugzilla.redhat.com/show_bug.cgi?id=1958869> - Import
VM from export domain fails - the imported VM remains in 'image locked'
state
oVirt Node Changes:
- Consume above oVirt updates
- Updated hivex (CVE-2021-3504
<https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-3504>)
Full diff list:
*--- ovirt-node-ng-image-4.4.6.1.manifest-rpm* *2021-05-11
08:39:44.714649170 +0200**+++
ovirt-node-ng-image-4.4.6.2.manifest-rpm* *2021-05-14
08:58:12.581488678 +0200*@@ -253 +253
@@-hivex-1.3.18-20.module_el8.5.0+746+bbd5d70c.x86_64+hivex-1.3.18-21.el8s.x86_64(a)@
-638 +638 @@-ovirt-hosted-engine-ha-2.4.6-1.el8.noarch+ovirt-hosted-engine-ha-2.4.7-1.el8.noarch(a)@
-643 +643 @@-ovirt-node-ng-image-update-placeholder-4.4.6.1-1.el8.noarch+ovirt-node-ng-image-update-placeholder-4.4.6.2-1.el8.noarch(a)@
-651,2 +651,2 @(a)-ovirt-release-host-node-4.4.6.1-1.el8.noarch-ovirt-release44-4.4.6.1-1.el8.noarch+ovirt-release-host-node-4.4.6.2-1.el8.noarch+ovirt-release44-4.4.6.2-1.el8.noarch
--
Sandro Bonazzola
MANAGER, SOFTWARE ENGINEERING, EMEA R&D RHV
Red Hat EMEA <https://www.redhat.com/>
sbonazzo(a)redhat.com
<https://www.redhat.com/>
*Red Hat respects your work life balance. Therefore there is no need to
answer this email out of your office hours.*
2 years, 11 months
[OLVM] Host non responsive after installation
by alan@softdrive.co
I am using Oracle Linux Virtualization Manager, following this guide: https://docs.oracle.com/en/virtualization/oracle-linux-virtualization-man...
After adding a host to the engine, the host becomes non responsive due to network errors:
engine.log
2021-04-27 14:53:02,255Z ERROR [org.ovirt.engine.core.bll.hostdeploy.InstallVdsInternalCommand] (EE-ManagedThreadFactory-engine-Thread-32356) [38586e0e] Host installation failed for host 'c97604b3-5774-4260-92fd-633257aa7498', 'GPU2-2': Network error during communication with the host
Help resolving this would be much appreciated!
2 years, 11 months
Moving interfaces for switch maintenance
by David White
So I have two switches.
All 3 of my HCI oVirt servers are connected to both switches.
1 switch serves the ovirtmgmt network (internal, gluster communication and everything else on that subnet)
The other switch serves the "main" front-end network (Private).
It turns out that my datacenter plugged the switches into the wrong power supplies, and now needs to move them.
My switches have single PSUs, so when we move them, the network will go down.
Obviously, I'm just going to move 1 switch at a time.
What I'm wondering, though, is there's an "easy" way to force all the traffic for both networks out one of the single interfaces, so that I don't experience any more downtime during this switch maintenance.
Should I assign another temporary IP address on each of the interfaces for each server that is in the same subnet as the switch we're going to take down?
Sent with ProtonMail Secure Email.
2 years, 11 months
Issue upgrading from 4.3 (Centos 7) to 4.4 (Centos 8)
by ling@aliko.com
Hello,
I have been trying to upgrade my self-hosted engine from 4.3 to 4.4 but running into issue while performing hosted-engine deploy.
Old hypervisor hosts are all running Centos 8 and old ovirt-engine is also running Centos 7.
I created a brand new baremental node running Centos 8, Kernel 4.18.0-240.15.1.el8_3.x86_64 and the following engine versions:
ovirt-hosted-engine-setup-2.4.9-1.el8.noarch
ovirt-hosted-engine-ha-2.4.6-1.el8.noarch
ovirt-engine-appliance-4.4-20210323171213.1.el8.x86_64
python3-ovirt-engine-sdk4-4.4.10-1.el8.x86_64
I have many VLANs in my environment. But on this host, I only have these network devices set up (eth0 is the main network, eth1 for storage):
# nmcli con
NAME UUID TYPE DEVICE
ovirtmgmt 02f64861-d992-4e56-8cec-da1906bac09f bridge ovirtmgmt
System eth1 bd9e565f-bdc3-4e43-bbd3-5875b9d7fed7 ethernet eth1
virbr0 78e6875d-70f6-4c89-89dd-180dbb9250b1 bridge virbr0
eth0 743b0e26-aae7-44b8-9215-3754a537e90b ethernet eth0
vnet0 bcfead6d-c5b6-4428-9f89-41589735be02 tun vnet0
When I run hosted-engine --deploy --restore-from-file=backup_050321.bck, it hangs after showing:
[ INFO ] TASK [ovirt.ovirt.engine_setup : Copy yum configuration file]
[ INFO ] changed: [localhost -> ovirt.safari.apple.com]
[ INFO ] TASK [ovirt.ovirt.engine_setup : Set 'best' to false]
[ INFO ] changed: [localhost -> ovirt.safari.apple.com]
[ INFO ] TASK [ovirt.ovirt.engine_setup : Update all packages]
virsh shows the VM is in paused state:
# virsh list
Id Name State
----------------------------------
1 HostedEngineLocal paused
I was able to ssh onto the VM until that point.
Do I need to set up all the network connection for all the VLANs before running the deploy script?
And how about engine storage domain? I have a new NFS mount ready but it did not ask me about which storage domain to use. Will it ask in later stage?
Thanks.
2 years, 11 months
Something broke & took down multiple VMs for ~20 minutes
by David White
As the subject suggestions, something in oVirt HCI broke. I have no idea what, and it recovered on its own after about 20 minutes or so.
I believe that the issue was limited to a single host (although I don't know that for sure), as we had two VMs go completely unresponsive, but a 3rd VM remained operational. For a while during the outage, I was able to log into the oVirt admin web portal, and I noticed at least 1-2 of my hosts (I have 3 hosts) showed the problematic VMs as being problematic inside of oVirt.
Reviewing the oVirt Events, I see that this basically started right when the ETL Service Started. There were no events before that point since yesterday, but right when the ETL Service started, it seems like all hell broke loose.
oVirt detected "No faulty multipaths" on any of the hosts, but then very quickly started indicating that hosts, vms, and storage targets were unavailable. See my screenshot below.
Around 30 - 35 minutes later, it appears that the Hosted Engine terminated due to a storage issue, and auto recovered on a different host. There's a 2nd screenshot beneath the first.
Everything came back up shortly before 9am, and has been stable since.
In fact, the Volume replication issues that I saw in my environment after I performed maintenance on 1 of my hosts on Friday are no longer present. It appears that the Hosted Engine sees the storage as being perfectly healthy.
How do I even begin to figure out what happened, and try to prevent it from happening again?
[Screenshot from 2021-04-26 16-36-47.png]
[Screenshot from 2021-04-26 16-44-08.png]
Sent with ProtonMail Secure Email.
2 years, 11 months