On Tue, Oct 13, 2020 at 12:06 PM Gianluca Cecchi <gianluca.cecchi@gmail.com> wrote:
On Sat, Oct 10, 2020 at 10:13 AM Martin Perina <mperina@redhat.com> wrote:

[snip]


Can I replicate the command that the engine would run on host through ssh?

I don't think so there is an easy way to do it 
Let's see what else we can get from the logs... 

Martin 


Hi,
I've run on engine the command
ovirt-log-collector --no-hypervisors
but potentially there is much sensitive information (like the dump of the database).

Is there any particular file you are more interested in that archive I can share?

BTW: can I put engine in debug for the time I'm trying to add the host so that we can see if more messages are shown?
In that case how can I do?

Another information I have noticed is that when the new host command from web admin GUI suddenly fails, anyway the ov200 host is now present in the host list, with the down icon and "Install failed" info.
If I click on it and go in General subtab, in the section "Action Items" I see 3 items with exclamation mark in front of them:

1) Power Management is not configured for this Host.
Enable Power Management
--> OK, I skipped it

2) Host has no default route.
---> I don't know why it says this.

[root@ov200 log]# ip route show
default via 10.4.192.254 dev bond0.68 proto static metric 400
10.4.192.0/24 dev bond0.68 proto kernel scope link src 10.4.192.32 metric 400
192.168.122.0/24 dev virbr0 proto kernel scope link src 192.168.122.1 linkdown
[root@ov200 log]#

On the still in CentOS 7 active host I have:

[root@ov300 ~]# ip route show
default via 10.4.192.254 dev ovirtmgmntZ2Z3
10.4.187.0/24 dev p1p2.187 proto kernel scope link src 10.4.187.100
10.4.192.0/24 dev ovirtmgmntZ2Z3 proto kernel scope link src 10.4.192.33
10.10.100.0/24 dev p1p2 proto kernel scope link src 10.10.100.88
10.10.100.0/24 dev p1p1.100 proto kernel scope link src 10.10.100.87
[root@ov300 ~]#

[root@ov300 ~]# brctl show ovirtmgmntZ2Z3
bridge name bridge id STP enabled interfaces
ovirtmgmntZ2Z3 8000.1803730ba369 no bond0.68
[root@ov300 ~]#

Could it be the fact that for historical reasons my mgmt network has not the name ovirtmgmt but ovirtmgmntZ2Z3 that confuses the installer that expects ovirtmgmt to setup? And erroneously reports the no default route message?

3) The host CPU does not match the Cluster CPU Type and is running in a degraded mode. It is missing the following CPU flags: vmx, ssbd, nx, model_Westmere, aes, spec_ctrl. Please update the host CPU microcode or change the Cluster CPU Type.

The cluster is set as "Intel Westmere IBRS SSBD Family".
all the hosts are the same hw Dell PE M610, with same processor

Host installed in CentOS 8:
[root@ov200 log]# cat /proc/cpuinfo | grep "model name" | sort -u
model name : Intel(R) Xeon(R) CPU           X5690  @ 3.47GHz
[root@ov200 log]#

Host still in CentOS 7:
[root@ov300 ~]# cat /proc/cpuinfo | grep "model name" | sort -u
model name : Intel(R) Xeon(R) CPU           X5690  @ 3.47GHz
[root@ov300 ~]#

If I compare the cpu flags inside the OS I see:

CentOS 8:
[root@ov200 log]# cat /proc/cpuinfo | grep flags | sort -u
flags : fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 cx16 xtpr pdcm pcid dca sse4_1 sse4_2 popcnt aes lahf_lm pti ssbd ibrs ibpb stibp tpr_shadow vnmi flexpriority ept vpid dtherm ida arat flush_l1d
[root@ov200 log]#

CentOS 7:
[root@ov300 ~]# cat /proc/cpuinfo | grep flags | sort -u
flags : fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc aperfmperf eagerfpu pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 cx16 xtpr pdcm pcid dca sse4_1 sse4_2 popcnt aes lahf_lm ssbd ibrs ibpb stibp tpr_shadow vnmi flexpriority ept vpid dtherm ida arat spec_ctrl intel_stibp flush_l1d
[root@ov300 ~]#

When still in CentOS 7, ov200 had the same flags as ov300
ov200 has this more now:
cpuid pti

ov200 has these less now:
eagerfpu spec_ctrl intel_stibp

Gianluca

Any feedback on my latest comments?
In the meantime here:
https://drive.google.com/file/d/1iN37znRtCo2vgyGTH_ymLhBJfs-2pWDr/view?usp=sharing
you can find inside the sosreport in tar.gz format, where I have modified some file names and context in respect of hostnames.
The only file I have not put inside is the dump of the database, but I can run any query you like in case.

Gianluca