centos stream 8 ovirt 4.4.6 with intel silver 4216 deployment problem
by Anatoliy Radchenko
Hi,
i cannot deploy hyperconverged 4.4.6 replica 3 on "centos stream 8" with
4216 processor.
Demployment with intel 4210 is ok.
Demployment with centos 8.4 is ok.
The differences of cat /proc/cpuinfo are only microcode:
centos 8: 0x5003006
centos 8 stream: 0x5003102
I try to install on centos 8 and successfully upgrade to centos stream 8
but engine cannot change version of cluster from 4.5 to 4.6 with error:
"Cannot change Cluster compatibility version where there are no hosts in
the Cluster which support that version"
and log says:
"Host moved to Non-Operational state as host CPU type is not supported in
this cluster compatibility version or is not supported at all"
Any ideas?
Thanks in advance.
Best regards.
ps: if is need..
dell poweredge R440
cat /proc/cpuinfo:
processor : 31
vendor_id : GenuineIntel
cpu family : 6
model : 85
model name : Intel(R) Xeon(R) Silver 4216 CPU @ 2.10GHz
stepping : 7
microcode : 0x5003102
cpu MHz : 2445.408
cache size : 22528 KB
physical id : 0
siblings : 32
core id : 12
cpu cores : 16
apicid : 25
initial apicid : 25
fpu : yes
fpu_exception : yes
cpuid level : 22
wp : yes
flags : fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat
pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb
rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology
nonstop_tsc cpuid aperfmperf pni pclmulqdq dtes64 monitor ds_cpl vmx smx
est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe
popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm
3dnowprefetch cpuid_fault epb cat_l3 cdp_l3 invpcid_single intel_ppin ssbd
mba ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid
ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid cqm mpx rdt_a
avx512f avx512dq rdseed adx smap clflushopt clwb intel_pt avx512cd avx512bw
avx512vl xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total
cqm_mbm_local dtherm ida arat pln pts pku ospke avx512_vnni md_clear
flush_l1d arch_capabilities
bugs : spectre_v1 spectre_v2 spec_store_bypass swapgs taa itlb_multihit
bogomips : 4200.00
clflush size : 64
cache_alignment : 64
address sizes : 46 bits physical, 48 bits virtual
rpm -qa|grep ovirt:
ovirt-ansible-collection-1.4.2-1.el8.noarch
ovirt-hosted-engine-ha-2.4.7-1.el8.noarch
ovirt-openvswitch-ovn-2.11-0.2020061801.el8.noarch
ovirt-host-dependencies-4.4.6-1.el8.x86_64
cockpit-ovirt-dashboard-0.15.0-1.el8.noarch
ovirt-openvswitch-2.11-0.2020061801.el8.noarch
ovirt-openvswitch-ovn-host-2.11-0.2020061801.el8.noarch
ovirt-imageio-common-2.1.1-1.el8.x86_64
python3-ovirt-engine-sdk4-4.4.12-1.el8.x86_64
ovirt-vmconsole-1.0.9-1.el8.noarch
ovirt-imageio-client-2.1.1-1.el8.x86_64
python3-ovirt-setup-lib-1.3.2-1.el8.noarch
ovirt-release44-4.4.6.3-1.el8.noarch
ovirt-provider-ovn-driver-1.2.33-1.el8.noarch
ovirt-host-4.4.6-1.el8.x86_64
ovirt-vmconsole-host-1.0.9-1.el8.noarch
ovirt-python-openvswitch-2.11-0.2020061801.el8.noarch
ovirt-openvswitch-ovn-common-2.11-0.2020061801.el8.noarch
ovirt-imageio-daemon-2.1.1-1.el8.x86_64
ovirt-hosted-engine-setup-2.5.0-1.el8.noarch
--
_____________________________________
Radchenko Anatolii
via Manoppello, 83 - 00132 Roma
tel. 06 96044328
cel. 329 6030076
Nota di riservatezza : ai sensi e per gli effetti della Legge sulla Tutela
della Riservatezza Personale (Legge 196/03) si precisa che il presente
messaggio, corredato dei relativi allegati, contiene informazioni da
considerarsi strettamente riservate, ed è destinato esclusivamente al
destinatario sopra indicato, il quale è l'unico autorizzato ad usarlo,
copiarlo e, sotto la propria responsabilità, diffonderlo. Chiunque
ricevesse questo messaggio per errore o comunque lo leggesse senza esserne
legittimato è avvertito che trattenerlo, copiarlo, divulgarlo, distribuirlo
a persone diverse dal destinatario è severamente proibito,ed è pregato di
rinviarlo al mittente.
3 years, 7 months
Data Centers status Not Operational
by nexpron@gmail.com
Hi everyone,
I found 3 servers (HV1, HV2, HV3) that contains VMs at another servers room. These servers are hosts based on KVM.
I logged into oVirt Engine Web Administration. Data Centers tab show me only one entry
Name: RCV
Storage: Shared
Status: Not Operational
Compability Version: 3.4
Description: [Empty]
Webbrowser oVirt Engine Web Administrator show me
oVirt Engine Version: 3.4.0-1.el6
One VM on HV3 has stopped, Hosts tab show Non Responsive Status for every host (HV1-3); every VM in Virtual Machines tab show Unknown status.
What do I'll do to change status in Data Center? How to start debug the reason? Last administrator leave documentation in vestigial form :)
DataCenter: RCV
-->Cluster: RCV_Cluster
---->Host: HV1 node
---->Host: Hv2 node
---->Host: HV3 engine
Best regards,
nexpron
3 years, 7 months
Re: [oVirt 4.4.6] Offline deployment of hosted engine
by Yedidyah Bar David
On Sun, Jun 6, 2021 at 9:03 AM sharma.ishan(a)fujitsu.com
<sharma.ishan(a)fujitsu.com> wrote:
>
> Hello Strahil,
>
>
>
> Thank you for your suggestion. Yes, I downloaded ovirt-engine-appliance rpm and installed manually, but it seems that it still tries to access to internet repository as below. Is it possible to do the whole deployment offline?
>
>
>
> [ ERROR ] fatal: [localhost -> ovirt-engine.macc.local]: FAILED! => {"changed": false, "msg": "Failed to download metadata for repo 'appstream': Cannot prepare internal mirrorlist: Curl
>
> error (7): Couldn't connect to server for http://mirrorlist.centos.org/?release=8-stream&arch=x86_64&repo=AppStream... [Failed to connect to jp.fujitsu.com port 123456789: No
>
> route to host]", "rc": 1, "results": []}
Please try with:
- appliance installed
- '--ansible-extra-vars=he_offline_deployment=true'
- latest released code (with your custom patch removed)
If it still fails, please open a bug and attach all relevant logs. Thanks!
Best regards,
--
Didi
3 years, 7 months
How to handle broken NFS storage?
by David White
I'm trying to figure out how to keep a "broken" NFS mount point from causing the entire HCI cluster to crash.
HCI is working beautifully.
Last night, I finished adding some NFS storage to the cluster - this is storage that I don't necessarily need to be HA, and I was hoping to store some backups and less-important VMs on, since my Gluster (sssd) storage availability is pretty limited.
But as a test, after I got everything setup, I stopped the nfs-server.
This caused the entire cluster to go down, and several VMs - that are not stored on the NFS storage - went belly up.
Once I started the NFS server process again, HCI did what it was supposed to do, and was able to automatically recover.
My concern is that NFS is a single point of failure, and if VMs that don't even rely on that storage are affected if the NFS storage goes away, then I don't want anything to do with it.
On the other hand, I'm still struggling to come up with a good way to run on-site backups and snapshots without using up more gluster space on my (more expensive) sssd storage.
Is there any way to setup NFS storage for a Backup Domain - as well as a Data domain (for lesser important VMs) - such that, if the NFS server crashed, all of my non-NFS stuff would be unaffected?
Sent with ProtonMail Secure Email.
3 years, 7 months
deployment fails - debug: var=server_cpu_dict
by lejeczek
Hi guys,
my:
-> $ ovirt-hosted-engine-setup
fails every time with:
...
[ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Convert
CPU model name]
[ ERROR ] fatal: [localhost]: FAILED! => {"msg": "The task
includes an option with an undefined variable. The error
was: 'dict object' has no attribute ''\n\nThe error appears
to be in
'/usr/share/ansible/collections/ansible_collections/ovirt/ovirt/roles/hosted_engine_setup/tasks/create_target_vm/01_create_target_hosted_engine_vm.yml':
line 69, column 15, but may\nbe elsewhere in the file
depending on the exact syntax problem.\n\nThe offending line
appears to be:\n\n - debug:
var=server_cpu_dict\n ^ here\n\nThere appears
to be both 'k=v' shorthand syntax and YAML in this task.
Only one syntax may be used.\n"}
[ ERROR ] Failed to execute stage 'Closing up': Failed
executing ansible-playbook
...
An expert would care to comment?
many thanks, L
3 years, 7 months
Hosted-engine fail and host reboot
by Dominique D
I tried with a 1 baremetal host, 3 baremetals or virtual hosts and I still have the same problem for installing the hosted-engine. Hyperconverge installs well.
I have tried with multiple version of ovirt ISO file 4.4.1 and 4.4.4 and 4.4.6.
when I run hosted-engine --deploy or with the cockpit, it creates a temporary VM in the 192.168.222.x subnet and I am able to connect in ssh on this temporary ip. When the script displays "TASK [ovirt.ovirt.hoted_engine_setup]: Wait for the host to be up" the server reboots and I have nothing left.
problem when creation of the ovirtmgmt?
here all my log https://drive.google.com/drive/folders/1kFFSlIqbjVwSN8t88aQZZR45DYEHPUvt?...
021-05-26 10:15:09,135-0400 INFO otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:111 TASK [ovirt.ovirt.hosted_engine_setup : Always revoke the SSO token]
2021-05-26 10:15:10,439-0400 DEBUG otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:105 {'msg': "You must specify either 'url' or 'hostname'.", 'invocation': {'module_args': {'state': 'absent', 'ovirt_auth': {'changed': False, 'ansible_facts': {'ovirt_auth': {'token': 'Mz2onwB7qWX2x8HnJVgetQIQ9U4eVziRt8TEabfoizI2B98d0PDp-yxTU92a9lbun2vcr_i5yOXRsJKJKhqkVw', 'url': 'https://oe.telecom.lan/ovirt-engine/api', 'ca_file': None, 'insecure': True, 'timeout': 0, 'compress': True, 'kerberos': False, 'headers': None}}, 'failed': False, 'attempts': 1}, 'timeout': 0, 'compress': True, 'kerberos': False, 'url': None, 'hostname': None, 'username': None, 'password': None, 'ca_file': None, 'insecure': None, 'headers': None, 'token': None}}, '_ansible_no_log': False, 'changed': False}
2021-05-26 10:15:10,540-0400 DEBUG otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:105 ignored: [localhost]: FAILED! => {"changed": false, "msg": "You must specify either 'url' or 'hostname'."}
2021-05-26 10:15:11,643-0400 INFO otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:111 TASK [ovirt.ovirt.hosted_engine_setup : include_tasks]
2021-05-26 10:15:12,647-0400 INFO otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:111 ok: [localhost]
2021-05-26 10:15:13,851-0400 INFO otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:111 TASK [ovirt.ovirt.hosted_engine_setup : Obtain SSO token using username/password credentials]
2021-05-26 10:15:15,261-0400 INFO otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:111 ok: [localhost]
2021-05-26 10:15:17,275-0400 INFO otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:111 TASK [ovirt.ovirt.hosted_engine_setup : Wait for the host to be up]
2021-05-26 10:22:54,758-0400 ERROR otopi.context context._executeMethod:154 Failed to execute stage 'Closing up': SIG1
2021-05-26 10:22:54,762-0400 DEBUG otopi.context context.dumpEnvironment:765 ENVIRONMENT DUMP - BEGIN
2021-05-26 10:22:54,763-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV BASE/error=bool:'True'
2021-05-26 10:22:54,763-0400 DEBUG otopi.context context.dumpEnvironment:775 ENV BASE/exceptionInfo=list:'[(<class 'RuntimeError'>, RuntimeError('SIG1',), <traceback object at 0x7fa8fc1ca1c8>)]'
thank you
3 years, 7 months
Adding a Ubuntu Host's NFS share to oVirt
by David White
Hello,
Is it possible to use Ubuntu to share an NFS export with oVirt?I'm trying to setup a Backup Domain for my environment.
I got to the point of actually adding the new Storage Domain.
When I click OK, I see the storage domain appear momentarily before disappearing, at which point I get a message about oVirt not being able to obtain a lock.
It appears I'm running into the issue described in this thread: https://lists.ovirt.org/archives/list/users@ovirt.org/thread/BNVX.... Although the actual export is ext4, not xfs.
From what I'm reading on that thread and elsewhere, it sounds like this problem is a result of SELinux not being present, is that correct?
Is my only option here to install an OS that supports SELinux?
Sent with ProtonMail Secure Email.
3 years, 7 months