Reinstall without dataloss
by hans@jiflin.nl
Hi all,
I am new here, been searching the mailing list on a regular base when I encountered problems and till now I always was able to keep the system up & running.
As said... till now...
I have oVirt at home, and I have about 7 vm's running on it. Lately I have had some troubles with my electricity, which results in a complete power outage on a irregular base.
Yesterday I had a power outage which left the system unbootable with the following errors:
error: ../../grub-core/loader/i386/pc/linux.c:170:invalid magic number.
error: ../../grub-core/loader/i386/pc/linux.c:1418:you need to load the kernel first.
Press any key to continue
Normally this can be solved following https://access.redhat.com/solutions/5829141, but his time also the files in /boot had a size of 0 bytes. So basically i did not have a working kernel on the system anymore.
https://www.thegeekdiary.com/centos-rhel-7-how-to-install-kernel-from-res... does work for CentOS 8 also, but the ovirt 4.8 iso does not have the same directory structure. Using CentOS 8 installs a kernel, but not a bootable system.
Is there a way i can start the installer in troubleshooting mode so I can reinstall just the kernel on the system?
3 years, 2 months
about the expiration time of the oVirt certs
by Tommy Sway
As you know, there are many kinds of certificates in Ovirt, used for
communication, authentication and so on.
However, in practice, there is a security risk related to the above
certificates.
That is, you need to generate a new certificate after the certificate
expires. Otherwise, a problem will occur.
In addition, different certificates expire at different times, which brings
a lot of management trouble to users.
Especially in the production system, a huge virtualization cluster may run
thousands of VMS. If a cluster certificate has a problem, the impact is very
serious.
So I felt there was an urgent need for a technical tool that could help
users quickly locate certificates, identify their expiration dates, and
rebuild them.
Even if there is no tool, there should be a way to solve the problems caused
by partial certificate expiration. I think it should include the following
points:
First, how to list the certificate in detail
Second, how to check the certificate expiration time
Third, how to rebuild the certificate
Does anyone else have this kind of confusion? What's a good solution?
Thanks.
3 years, 2 months
Using third-party certificate: PKIX path building failed: sun.security.provider.certpath.SunCertPathBuilderException: unable to find valid certification path to requested target
by nicolas@devels.es
Hi,
I'm making a bare metal oVirt installation, version 4.4.8.
'ovirt-engine' command ends well, however, we're using a third-party
certificate (from LetsEncrypt) both for the apache server and the
ovirt-websocket-proxy. So we changed configuration files regarding httpd
and ovirt-websocket-proxy.
Once changed the configurations, if I try to log in to the oVirt engine,
I get a "PKIX path building failed:
sun.security.provider.certpath.SunCertPathBuilderException: unable to
find valid certification path to requested target" error.
In prior versions we used to add the chain to the
/etc/pki/ovirt-engine/.truststore file, however, simply listing the
current certificates seems not to be working on 4.4.8.
# LANG=C keytool -list -keystore /etc/pki/ovirt-engine/.truststore
-alias intermedia_le -storepass mypass
keytool error: java.io.IOException: Invalid keystore format
Is there something I'm missing here?
Thank
3 years, 2 months
SSl vdmsd failed ssl
by Ingeniero especialista Redhat / Suse
Hello, good evening, I want to consult the following case, we have two
ovirt 3.6 servers with hosted-engine, yesterday some multipath servers were
presented and two machines were blocked, which was not possible to start,
checking we found that the ssl of the nodes They are expired and we change
them only in the nodes. We could start the hosted-engine but the two nodes
are not responsive
comes out in vdsm daemon
vdsm [43067]: vdsm ProtocolDetector.SSLHandshakeDispatcher ERROR Error
during handshake: unexpected eof
I appreciate any ideas or suggestions to be able to recover normal
operation.
Thanks
3 years, 2 months
Intermittent failure to upload ISOs
by aclysma@gmail.com
This may be the same issue as described here:
https://lists.ovirt.org/archives/list/users@ovirt.org/thread/CJISJIDQKSIN...
https://bugzilla.redhat.com/show_bug.cgi?id=1977276
I am on 4.4.8.6-1.el8, installed a couple days ago from the ovirt node ISO. In particular, I noticed if I SSH into the hosted engine and tail -f /var/log/ovirt-imageio/daemon.log, in the failure case I get something like:
2021-09-30 08:15:52,330 INFO (Thread-8) [http] OPEN connection=8 client=::ffff:192.168.1.53
2021-09-30 08:16:23,315 INFO (Thread-8) [http] CLOSE connection=8 client=::ffff:192.168.1.53 [connection 1 ops, 30.984947 s] [dispatch 1 ops, 0.000097 s]
No activity in tail -f /var/log/ovirt-imageio/daemon.log on the host (I only have one host) in the failure case, just the engine. In the success case, there is activity in both logs.
It is very intermittent. Sometimes uploads work most of the time (maybe 4 out of 5), and I've had other times that uploads do not work at all (0 out of 5).
I think when it's behaving particularly badly, restarting the engine (hosted-engine --vm-shutdown, then hosted-engine --vm-start) helps, but I haven't figured out a reliable pattern. (I am logged in as admin.) I've tried several browsers, closing/reopening the browser, etc.
Hoping this info will help in tracking it down.
3 years, 2 months
oVirt - No supported package manager found in your system
by German Sandoval
Probably this isn't the place to ask, but I'm doing a test with an Almalinux Physical host and trying to install a standalone instance and I get this error when I use the Engine-Setup, I'm using a Centos stream guide.
[ INFO ] Stage: Initializing
[ INFO ] Stage: Environment setup
Configuration files: /etc/ovirt-engine-setup.conf.d/10-packaging-jboss.conf, /etc/ovirt-engine-setup.conf.d/10-packaging.conf
Log file: /var/log/ovirt-engine/setup/ovirt-engine-setup-20210915140413-hsjs2f.log
Version: otopi-1.9.5 (otopi-1.9.5-1.el8)
[ ERROR ] Failed to execute stage 'Environment setup': No supported package manager found in your system
[ INFO ] Stage: Clean up
Log file is located at /var/log/ovirt-engine/setup/ovirt-engine-setup-20210915140413-hsjs2f.log
[ INFO ] Generating answer file '/var/lib/ovirt-engine/setup/answers/20210915140414-setup.conf'
[ INFO ] Stage: Pre-termination
[ INFO ] Stage: Termination
[ ERROR ] Execution of setup failed
2021-09-15 14:12:16,421-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_SYSTEM/groupKvm=str:'kvm'
2021-09-15 14:12:16,421-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_SYSTEM/groupVmConsole=str:'ovirt-vmconsole'
2021-09-15 14:12:16,421-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_SYSTEM/hostileServices=str:'ovirt-engine-dwhd,ovirt-engine-notifier'
2021-09-15 14:12:16,422-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_SYSTEM/memCheckEnabled=bool:'True'
2021-09-15 14:12:16,422-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_SYSTEM/memCheckMinimumMB=int:'4096'
2021-09-15 14:12:16,422-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_SYSTEM/memCheckRecommendedMB=int:'16384'
2021-09-15 14:12:16,422-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_SYSTEM/memCheckThreshold=int:'90'
2021-09-15 14:12:16,422-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_SYSTEM/nfsConfigEnabled=NoneType:'None'
2021-09-15 14:12:16,423-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_SYSTEM/nfsConfigEnabled_legacyInPostInstall=bool:'False'
2021-09-15 14:12:16,423-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_SYSTEM/nfsServiceName=NoneType:'None'
2021-09-15 14:12:16,423-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_SYSTEM/reservedPorts=set:'set()'
2021-09-15 14:12:16,423-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_SYSTEM/selinuxBooleans=list:'[]'
2021-09-15 14:12:16,423-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_SYSTEM/selinuxContexts=list:'[]'
2021-09-15 14:12:16,424-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_SYSTEM/selinuxPorts=list:'[]'
2021-09-15 14:12:16,424-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_SYSTEM/selinuxRestorePaths=list:'[]'
2021-09-15 14:12:16,424-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_SYSTEM/shmmax=int:'68719476736'
2021-09-15 14:12:16,424-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_SYSTEM/userApache=str:'apache'
2021-09-15 14:12:16,424-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_SYSTEM/userEngine=str:'ovirt'
2021-09-15 14:12:16,425-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_SYSTEM/userPostgres=str:'postgres'
2021-09-15 14:12:16,425-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_SYSTEM/userRoot=str:'root'
2021-09-15 14:12:16,425-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_SYSTEM/userVdsm=str:'vdsm'
2021-09-15 14:12:16,425-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_SYSTEM/userVmConsole=str:'ovirt-vmconsole'
2021-09-15 14:12:16,426-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_VMCONSOLE_PROXY_CONFIG/vmconsoleProxyConfig=NoneType:'None'
2021-09-15 14:12:16,426-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_VMCONSOLE_PROXY_CONFIG/vmconsoleProxyPort=int:'2222'
2021-09-15 14:12:16,426-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_WSP_RPMDISTRO_PACKAGES=str:'ovirt-engine-websocket-proxy'
2021-09-15 14:12:16,426-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV OVESETUP_WSP_RPMDISTRO_PACKAGES_SETUP=str:'ovirt-engine-setup-plugin-websocket-proxy'
2021-09-15 14:12:16,426-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV PACKAGER/dnfDisabledPlugins=list:'[]'
2021-09-15 14:12:16,427-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV PACKAGER/dnfExpireCache=bool:'True'
2021-09-15 14:12:16,427-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV PACKAGER/dnfRollback=bool:'True'
2021-09-15 14:12:16,427-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV PACKAGER/dnfpackagerEnabled=bool:'False'
2021-09-15 14:12:16,427-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV PACKAGER/keepAliveInterval=int:'30'
2021-09-15 14:12:16,427-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV PACKAGER/yumDisabledPlugins=list:'[]'
2021-09-15 14:12:16,428-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV PACKAGER/yumEnabledPlugins=list:'[]'
2021-09-15 14:12:16,428-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV PACKAGER/yumExpireCache=bool:'True'
2021-09-15 14:12:16,428-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV PACKAGER/yumRollback=bool:'True'
2021-09-15 14:12:16,428-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV PACKAGER/yumpackagerEnabled=bool:'True'
2021-09-15 14:12:16,428-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV SYSTEM/clockMaxGap=int:'5'
2021-09-15 14:12:16,429-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV SYSTEM/clockSet=bool:'False'
2021-09-15 14:12:16,429-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV SYSTEM/commandPath=str:'/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/root/bin:/root/bin'
2021-09-15 14:12:16,429-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV SYSTEM/reboot=bool:'False'
2021-09-15 14:12:16,429-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV SYSTEM/rebootAllow=bool:'True'
2021-09-15 14:12:16,430-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV SYSTEM/rebootDeferTime=int:'10'
2021-09-15 14:12:16,430-0400 DEBUG otopi.context context.dumpEnvironment:779 ENVIRONMENT DUMP - END
2021-09-15 14:12:16,433-0400 DEBUG otopi.context context._executeMethod:127 Stage pre-terminate METHOD otopi.plugins.otopi.dialog.cli.Plugin._pre_terminate
2021-09-15 14:12:16,433-0400 DEBUG otopi.context context._executeMethod:136 otopi.plugins.otopi.dialog.cli.Plugin._pre_terminate condition False
2021-09-15 14:12:16,435-0400 INFO otopi.context context.runSequence:616 Stage: Termination
2021-09-15 14:12:16,435-0400 DEBUG otopi.context context.runSequence:620 STAGE terminate
2021-09-15 14:12:16,437-0400 DEBUG otopi.context context._executeMethod:127 Stage terminate METHOD otopi.plugins.ovirt_engine_common.base.core.misc.Plugin._terminate
2021-09-15 14:12:16,437-0400 ERROR otopi.plugins.ovirt_engine_common.base.core.misc misc._terminate:153 Execution of setup failed
2021-09-15 14:12:16,441-0400 DEBUG otopi.context context._executeMethod:127 Stage terminate METHOD otopi.plugins.otopi.dialog.human.Plugin._terminate
2021-09-15 14:12:16,460-0400 DEBUG otopi.context context._executeMethod:127 Stage terminate METHOD otopi.plugins.otopi.dialog.machine.Plugin._terminate
2021-09-15 14:12:16,460-0400 DEBUG otopi.context context._executeMethod:136 otopi.plugins.otopi.dialog.machine.Plugin._terminate condition False
2021-09-15 14:12:16,463-0400 DEBUG otopi.context context._executeMethod:127 Stage terminate METHOD otopi.plugins.otopi.core.log.Plugin._terminate
I haven't found I guide for Alma Linux, So I can assume maybe oVirt still not supported on this OS, I couldn't find much information regarding this error.
https://bugzilla.redhat.com/show_bug.cgi?id=1908602
https://bugzilla.redhat.com/show_bug.cgi?format=multiple&id=1909965
Thanks for your help.
3 years, 2 months
oVirt/Hyperconverged issue
by topoigerm@gmail.com
I have 4 servers of identical hardware. The documentation says "you need 3", not "you need 3 or more"; is it possible to run hyperconverged with 4 servers. Currently all the 4 nodes server has been crashed n after the 4th node try joining the hyperconverged 3nodes cluster. Kindly advise.
FYI currently i'm trying to reinstall back all the OS back due mentioned incident happen.
BR
Faizal
3 years, 2 months
Help ovirt 3.6
by Ingeniero especialista Redhat / Suse
Hello, good evening, I want to consult the following case, we have two
ovirt 3.6 servers with hosted-engine, yesterday some multipath servers were
presented and two machines were blocked, which was not possible to start,
checking we found that the ssl of the nodes They are expired and we change
them only in the nodes. We could start the hosted-engine but the two nodes
are not responsive
comes out in vdsm daemon
vdsm [43067]: vdsm ProtocolDetector.SSLHandshakeDispatcher ERROR Error
during handshake: unexpected eof
I appreciate any ideas or suggestions to be able to recover normal
operation.
Thanks
3 years, 2 months
Host reboots when network switch goes down
by cen
Hi,
we are experiencing a weird issue with our Ovirt setup. We have two
physical hosts (DC1 and DC2) and mounted Lenovo NAS storage for all VM data.
They are connected via a managed network switch.
What happens is that if switch goes down for whatever reason (firmware
update etc), physical host reboots. Not sure if this is an action
performed by Ovirt but I suspect it is because connection to mounted
storage is lost and it performs some kind of an emergency action. I
would need to get some direction pointers to find out
a) who triggers the reboot and why
c) a way to prevent reboots by increasing storage? timeouts
Switch reboot takes 2-3 minutes.
These are the host /var/log/messages just before reboot occurs:
Sep 28 16:20:00 ovirtnode02 sanlock[10993]: 2021-09-28 16:20:00 7690984
[10993]: s11 check_our_lease warning 72 last_success 7690912
Sep 28 16:20:00 ovirtnode02 sanlock[10993]: 2021-09-28 16:20:00 7690984
[10993]: s3 check_our_lease warning 76 last_success 7690908
Sep 28 16:20:00 ovirtnode02 sanlock[10993]: 2021-09-28 16:20:00 7690984
[10993]: s1 check_our_lease warning 68 last_success 7690916
Sep 28 16:20:00 ovirtnode02 sanlock[10993]: 2021-09-28 16:20:00 7690984
[27983]: s11 delta_renew read timeout 10 sec offset 0
/var/run/vdsm/storage/15514c65-5d45-4ba7-bcd4-cc772351c940/fce598a8-11c3-44f9-8aaf-8712c96e00ce/65413499-6970-4a4c-af04-609ef78891a2
Sep 28 16:20:00 ovirtnode02 sanlock[10993]: 2021-09-28 16:20:00 7690984
[27983]: s11 renewal error -202 delta_length 20 last_success 7690912
Sep 28 16:20:00 ovirtnode02 wdmd[11102]: test warning now 7690984 ping
7690970 close 7690980 renewal 7690912 expire 7690992 client 10993
sanlock_hosted-engine:2
Sep 28 16:20:00 ovirtnode02 wdmd[11102]: test warning now 7690984 ping
7690970 close 7690980 renewal 7690908 expire 7690988 client 10993
sanlock_3cb12f04-5d68-4d79-8663-f33c0655baa6:2
Sep 28 16:20:01 ovirtnode02 systemd: Created slice User Slice of root.
Sep 28 16:20:01 ovirtnode02 systemd: Started Session 15148 of user root.
Sep 28 16:20:01 ovirtnode02 systemd: Removed slice User Slice of root.
Sep 28 16:20:01 ovirtnode02 sanlock[10993]: 2021-09-28 16:20:01 7690985
[10993]: s11 check_our_lease warning 73 last_success 7690912
Sep 28 16:20:01 ovirtnode02 sanlock[10993]: 2021-09-28 16:20:01 7690985
[10993]: s3 check_our_lease warning 77 last_success 7690908
Sep 28 16:20:01 ovirtnode02 sanlock[10993]: 2021-09-28 16:20:01 7690985
[10993]: s1 check_our_lease warning 69 last_success 7690916
Sep 28 16:20:01 ovirtnode02 wdmd[11102]: test warning now 7690985 ping
7690970 close 7690980 renewal 7690912 expire 7690992 client 10993
sanlock_hosted-engine:2
Sep 28 16:20:01 ovirtnode02 wdmd[11102]: test warning now 7690985 ping
7690970 close 7690980 renewal 7690908 expire 7690988 client 10993
sanlock_3cb12f04-5d68-4d79-8663-f33c0655baa6:2
Sep 28 16:20:02 ovirtnode02 sanlock[10993]: 2021-09-28 16:20:02 7690986
[10993]: s11 check_our_lease warning 74 last_success 7690912
Sep 28 16:20:02 ovirtnode02 sanlock[10993]: 2021-09-28 16:20:02 7690986
[10993]: s3 check_our_lease warning 78 last_success 7690908
Sep 28 16:20:02 ovirtnode02 sanlock[10993]: 2021-09-28 16:20:02 7690986
[10993]: s1 check_our_lease warning 70 last_success 7690916
Sep 28 16:20:02 ovirtnode02 wdmd[11102]: test warning now 7690986 ping
7690970 close 7690980 renewal 7690916 expire 7690996 client 10993
sanlock_15514c65-5d45-4ba7-bcd4-cc772351c940:2
Sep 28 16:20:02 ovirtnode02 wdmd[11102]: test warning now 7690986 ping
7690970 close 7690980 renewal 7690912 expire 7690992 client 10993
sanlock_hosted-engine:2
Sep 28 16:20:02 ovirtnode02 wdmd[11102]: test warning now 7690986 ping
7690970 close 7690980 renewal 7690908 expire 7690988 client 10993
sanlock_3cb12f04-5d68-4d79-8663-f33c0655baa6:2
Sep 28 16:20:03 ovirtnode02 sanlock[10993]: 2021-09-28 16:20:03 7690987
[10993]: s11 check_our_lease warning 75 last_success 7690912
Sep 28 16:20:03 ovirtnode02 sanlock[10993]: 2021-09-28 16:20:03 7690987
[10993]: s3 check_our_lease warning 79 last_success 7690908
Sep 28 16:20:03 ovirtnode02 sanlock[10993]: 2021-09-28 16:20:03 7690987
[10993]: s1 check_our_lease warning 71 last_success 7690916
3 years, 2 months