So, let's summarize:
- Cannot migrate the HE due to "CPU policy".
- HE's CPU is westmere - just like hosts
- You have enough resources on the second HE host (both CPU + MEMORY)
What is the Cluster's CPU type (you can check in UI) ?
Maybe you should enable debugging on various locations to identify the issue.
Anything interesting in the libvirt's log for the HostedEngine.xml on the destination
host ?
Best Regards,
Strahil Nikolov
В вторник, 22 септември 2020 г., 05:37:18 Гринуич+3, ddqlo <ddqlo(a)126.com> написа:
Yes. I can. The host which does not host the HE could be reinstalled sucessfully in web
UI. After this is done nothing has changed.
在 2020-09-22 03:08:18,"Strahil Nikolov" <hunter86_bg(a)yahoo.com> 写道:
Can you put 1 host in maintenance and use the "Installation"
-> "Reinstall" and enable the HE deployment from one of the tabs ?
Best Regards,
Strahil Nikolov
В понеделник, 21 септември 2020 г., 06:38:06 Гринуич+3, ddqlo <ddqlo(a)126.com>
написа:
so strange! After I set global maintenance, powered off and started H The cpu of HE became
'Westmere'(did not change anything). But HE still could not be migrated.
HE xml:
<cpu mode='custom' match='exact' check='full'>
<model fallback='forbid'>Westmere</model>
<topology sockets='16' cores='4' threads='1'/>
<feature policy='require' name='vme'/>
<feature policy='require' name='pclmuldq'/>
<feature policy='require' name='x2apic'/>
<feature policy='require' name='hypervisor'/>
<feature policy='require' name='arat'/>
<numa>
<cell id='0' cpus='0-3' memory='16777216'
unit='KiB'/>
</numa>
</cpu>
host capabilities:
<model>Westmere</model>
cluster cpu type (UI):
host cpu type (UI):
HE cpu type (UI):
在 2020-09-19 13:27:35,"Strahil Nikolov" <hunter86_bg(a)yahoo.com> 写道:
>Hm... interesting.
>
>The VM is using 'Haswell-noTSX' while the host is 'Westmere'.
>
>In my case I got no difference:
>
>[root@ovirt1 ~]# virsh dumpxml HostedEngine | grep Opteron
> <model fallback='forbid'>Opteron_G5</model>
>[root@ovirt1 ~]# virsh capabilities | grep Opteron
> <model>Opteron_G5</model>
>
>Did you update the cluster holding the Hosted Engine ?
>
>
>I guess you can try to:
>
>- Set global maintenance
>- Power off the HostedEngine VM
>- virsh dumpxml HostedEngine > /root/HE.xml
>- use virsh edit to change the cpu of the HE (non-permanent) change
>- try to power on the modified HE
>
>If it powers on , you can try to migrate it and if it succeeds - then you should make
it permanent.
>
>
>
>
>
>Best Regards,
>Strahil Nikolov
>
>В петък, 18 септември 2020 г., 04:40:39 Гринуич+3, ddqlo <ddqlo(a)126.com> написа:
>
>
>
>
>
>HE:
>
><domain type='kvm' id='1'>
> <name>HostedEngine</name>
> <uuid>b4e805ff-556d-42bd-a6df-02f5902fd01c</uuid>
> <metadata
xmlns:ns0="http://ovirt.org/vm/tune/1.0"
xmlns:ovirt-vm="http://ovirt.org/vm/1.0">
> <ns0:qos/>
> <ovirt-vm:vm
xmlns:ovirt-vm="http://ovirt.org/vm/1.0">
> <ovirt-vm:clusterVersion>4.3</ovirt-vm:clusterVersion>
> <ovirt-vm:destroy_on_reboot
type="bool">False</ovirt-vm:destroy_on_reboot>
> <ovirt-vm:launchPaused>false</ovirt-vm:launchPaused>
> <ovirt-vm:memGuaranteedSize
type="int">1024</ovirt-vm:memGuaranteedSize>
> <ovirt-vm:minGuaranteedMemoryMb
type="int">1024</ovirt-vm:minGuaranteedMemoryMb>
> <ovirt-vm:resumeBehavior>auto_resume</ovirt-vm:resumeBehavior>
> <ovirt-vm:startTime
type="float">1600307555.19</ovirt-vm:startTime>
> <ovirt-vm:device mac_address="56:6f:9b:b0:00:01">
> <ovirt-vm:network>external</ovirt-vm:network>
> <ovirt-vm:custom>
> <ovirt-vm:queues>4</ovirt-vm:queues>
> </ovirt-vm:custom>
> </ovirt-vm:device>
> <ovirt-vm:device mac_address="00:16:3e:50:c1:97">
> <ovirt-vm:network>ovirtmgmt</ovirt-vm:network>
> <ovirt-vm:custom>
> <ovirt-vm:queues>4</ovirt-vm:queues>
> </ovirt-vm:custom>
> </ovirt-vm:device>
> <ovirt-vm:device devtype="disk" name="vda">
>
<ovirt-vm:domainID>c17c1934-332f-464c-8f89-ad72463c00b3</ovirt-vm:domainID>
> <ovirt-vm:guestName>/dev/vda2</ovirt-vm:guestName>
>
<ovirt-vm:imageID>8eca143a-4535-4421-bd35-9f5764d67d70</ovirt-vm:imageID>
>
<ovirt-vm:poolID>00000000-0000-0000-0000-000000000000</ovirt-vm:poolID>
> <ovirt-vm:shared>exclusive</ovirt-vm:shared>
>
<ovirt-vm:volumeID>ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</ovirt-vm:volumeID>
> <ovirt-vm:specParams>
> <ovirt-vm:pinToIoThread>1</ovirt-vm:pinToIoThread>
> </ovirt-vm:specParams>
> <ovirt-vm:volumeChain>
> <ovirt-vm:volumeChainNode>
>
<ovirt-vm:domainID>c17c1934-332f-464c-8f89-ad72463c00b3</ovirt-vm:domainID>
>
<ovirt-vm:imageID>8eca143a-4535-4421-bd35-9f5764d67d70</ovirt-vm:imageID>
> <ovirt-vm:leaseOffset
type="int">108003328</ovirt-vm:leaseOffset>
>
<ovirt-vm:leasePath>/dev/c17c1934-332f-464c-8f89-ad72463c00b3/leases</ovirt-vm:leasePath>
>
<ovirt-vm:path>/rhev/data-center/mnt/blockSD/c17c1934-332f-464c-8f89-ad72463c00b3/images/8eca143a-4535-4421-bd35-9f5764d67d70/ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</ovirt-vm:path>
>
<ovirt-vm:volumeID>ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</ovirt-vm:volumeID>
> </ovirt-vm:volumeChainNode>
> </ovirt-vm:volumeChain>
> </ovirt-vm:device>
> <ovirt-vm:device devtype="disk" name="hdc"/>
></ovirt-vm:vm>
> </metadata>
> <maxMemory slots='16' unit='KiB'>67108864</maxMemory>
> <memory unit='KiB'>16777216</memory>
> <currentMemory unit='KiB'>16777216</currentMemory>
> <vcpu placement='static' current='4'>64</vcpu>
> <iothreads>1</iothreads>
> <resource>
> <partition>/machine</partition>
> </resource>
> <sysinfo type='smbios'>
> <system>
> <entry name='manufacturer'>oVirt</entry>
> <entry name='product'>oVirt Node</entry>
> <entry name='version'>7-5.1804.el7.centos</entry>
> <entry
name='serial'>00000000-0000-0000-0000-0CC47A6B3160</entry>
> <entry
name='uuid'>b4e805ff-556d-42bd-a6df-02f5902fd01c</entry>
> </system>
> </sysinfo>
> <os>
> <type arch='x86_64'
machine='pc-i440fx-rhel7.6.0'>hvm</type>
> <boot dev='hd'/>
> <bios useserial='yes'/>
> <smbios mode='sysinfo'/>
> </os>
> <features>
> <acpi/>
> </features>
> <cpu mode='custom' match='exact' check='full'>
> <model fallback='forbid'>Haswell-noTSX</model>
> <topology sockets='16' cores='4' threads='1'/>
> <feature policy='require' name='vme'/>
> <feature policy='require' name='f16c'/>
> <feature policy='require' name='rdrand'/>
> <feature policy='require' name='hypervisor'/>
> <feature policy='require' name='arat'/>
> <feature policy='require' name='xsaveopt'/>
> <feature policy='require' name='abm'/>
> <numa>
> <cell id='0' cpus='0-3' memory='16777216'
unit='KiB'/>
> </numa>
> </cpu>
> <clock offset='variable' adjustment='0' basis='utc'>
> <timer name='rtc' tickpolicy='catchup'/>
> <timer name='pit' tickpolicy='delay'/>
> <timer name='hpet' present='no'/>
> </clock>
> <on_poweroff>destroy</on_poweroff>
> <on_reboot>destroy</on_reboot>
> <on_crash>destroy</on_crash>
> <pm>
> <suspend-to-mem enabled='no'/>
> <suspend-to-disk enabled='no'/>
> </pm>
> <devices>
> <emulator>/usr/libexec/qemu-kvm</emulator>
> <disk type='file' device='cdrom'>
> <driver name='qemu' error_policy='report'/>
> <source startupPolicy='optional'/>
> <target dev='hdc' bus='ide'/>
> <readonly/>
> <alias name='ua-80fde7d5-ee7f-4201-9118-11bc6c3b8530'/>
> <address type='drive' controller='0' bus='1'
target='0' unit='0'/>
> </disk>
> <disk type='block' device='disk' snapshot='no'>
> <driver name='qemu' type='raw' cache='none'
error_policy='stop' io='native' iothread='1'/>
> <source
dev='/var/run/vdsm/storage/c17c1934-332f-464c-8f89-ad72463c00b3/8eca143a-4535-4421-bd35-9f5764d67d70/ae961104-c3b3-4a43-9f46-7fa6bdc2ac33'>
> <seclabel model='dac' relabel='no'/>
> </source>
> <backingStore/>
> <target dev='vda' bus='virtio'/>
> <serial>8eca143a-4535-4421-bd35-9f5764d67d70</serial>
> <alias name='ua-8eca143a-4535-4421-bd35-9f5764d67d70'/>
> <address type='pci' domain='0x0000' bus='0x00'
slot='0x07' function='0x0'/>
> </disk>
> <controller type='scsi' index='0'
model='virtio-scsi'>
> <driver iothread='1'/>
> <alias name='ua-27331e83-03f4-42a3-9554-c41649c02ba4'/>
> <address type='pci' domain='0x0000' bus='0x00'
slot='0x05' function='0x0'/>
> </controller>
> <controller type='ide' index='0'>
> <alias name='ide'/>
> <address type='pci' domain='0x0000' bus='0x00'
slot='0x01' function='0x1'/>
> </controller>
> <controller type='virtio-serial' index='0'
ports='16'>
> <alias name='ua-8fe74299-b60f-4778-8e80-db05393a9489'/>
> <address type='pci' domain='0x0000' bus='0x00'
slot='0x06' function='0x0'/>
> </controller>
> <controller type='usb' index='0'
model='piix3-uhci'>
> <alias name='usb'/>
> <address type='pci' domain='0x0000' bus='0x00'
slot='0x01' function='0x2'/>
> </controller>
> <controller type='pci' index='0' model='pci-root'>
> <alias name='pci.0'/>
> </controller>
> <lease>
> <lockspace>c17c1934-332f-464c-8f89-ad72463c00b3</lockspace>
> <key>ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</key>
> <target path='/dev/c17c1934-332f-464c-8f89-ad72463c00b3/leases'
offset='108003328'/>
> </lease>
> <interface type='bridge'>
> <mac address='00:16:3e:50:c1:97'/>
> <source bridge='ovirtmgmt'/>
> <target dev='vnet0'/>
> <model type='virtio'/>
> <driver name='vhost' queues='4'/>
> <filterref filter='vdsm-no-mac-spoofing'/>
> <link state='up'/>
> <mtu size='1500'/>
> <alias name='ua-fada74ee-2338-4cde-a7ba-43a9a636ad6e'/>
> <address type='pci' domain='0x0000' bus='0x00'
slot='0x03' function='0x0'/>
> </interface>
> <interface type='bridge'>
> <mac address='56:6f:9b:b0:00:01'/>
> <source bridge='external'/>
> <target dev='vnet1'/>
> <model type='virtio'/>
> <driver name='vhost' queues='4'/>
> <filterref filter='vdsm-no-mac-spoofing'/>
> <link state='up'/>
> <mtu size='1500'/>
> <alias name='ua-f7b4c949-1f9f-4355-811d-88428c88ce4e'/>
> <address type='pci' domain='0x0000' bus='0x00'
slot='0x0a' function='0x0'/>
> </interface>
> <serial type='unix'>
> <source mode='bind'
path='/var/run/ovirt-vmconsole-console/b4e805ff-556d-42bd-a6df-02f5902fd01c.sock'/>
> <target type='isa-serial' port='0'>
> <model name='isa-serial'/>
> </target>
> <alias name='serial0'/>
> </serial>
> <console type='unix'>
> <source mode='bind'
path='/var/run/ovirt-vmconsole-console/b4e805ff-556d-42bd-a6df-02f5902fd01c.sock'/>
> <target type='serial' port='0'/>
> <alias name='serial0'/>
> </console>
> <channel type='unix'>
> <source mode='bind'
path='/var/lib/libvirt/qemu/channels/b4e805ff-556d-42bd-a6df-02f5902fd01c.ovirt-guest-agent.0'/>
> <target type='virtio' name='ovirt-guest-agent.0'
state='connected'/>
> <alias name='channel0'/>
> <address type='virtio-serial' controller='0' bus='0'
port='1'/>
> </channel>
> <channel type='unix'>
> <source mode='bind'
path='/var/lib/libvirt/qemu/channels/b4e805ff-556d-42bd-a6df-02f5902fd01c.org.qemu.guest_agent.0'/>
> <target type='virtio' name='org.qemu.guest_agent.0'
state='connected'/>
> <alias name='channel1'/>
> <address type='virtio-serial' controller='0' bus='0'
port='2'/>
> </channel>
> <channel type='spicevmc'>
> <target type='virtio' name='com.redhat.spice.0'
state='disconnected'/>
> <alias name='channel2'/>
> <address type='virtio-serial' controller='0' bus='0'
port='3'/>
> </channel>
> <channel type='unix'>
> <source mode='bind'
path='/var/lib/libvirt/qemu/channels/b4e805ff-556d-42bd-a6df-02f5902fd01c.org.ovirt.hosted-engine-setup.0'/>
> <target type='virtio' name='org.ovirt.hosted-engine-setup.0'
state='disconnected'/>
> <alias name='channel3'/>
> <address type='virtio-serial' controller='0' bus='0'
port='4'/>
> </channel>
> <input type='tablet' bus='usb'>
> <alias name='input0'/>
> <address type='usb' bus='0' port='1'/>
> </input>
> <input type='mouse' bus='ps2'>
> <alias name='input1'/>
> </input>
> <input type='keyboard' bus='ps2'>
> <alias name='input2'/>
> </input>
> <graphics type='vnc' port='5900' autoport='yes'
listen='192.168.1.22' keymap='en-us'
passwdValidTo='1970-01-01T00:00:01'>
> <listen type='network' address='192.168.1.22'
network='vdsm-external'/>
> </graphics>
> <graphics type='spice' port='5901' tlsPort='5902'
autoport='yes' listen='192.168.1.22'
passwdValidTo='1970-01-01T00:00:01'>
> <listen type='network' address='192.168.1.22'
network='vdsm-external'/>
> <channel name='main' mode='secure'/>
> <channel name='display' mode='secure'/>
> <channel name='inputs' mode='secure'/>
> <channel name='cursor' mode='secure'/>
> <channel name='playback' mode='secure'/>
> <channel name='record' mode='secure'/>
> <channel name='smartcard' mode='secure'/>
> <channel name='usbredir' mode='secure'/>
> </graphics>
> <sound model='ich6'>
> <alias name='ua-bd287767-9b83-4e44-ac6f-8b527f9632b8'/>
> <address type='pci' domain='0x0000' bus='0x00'
slot='0x04' function='0x0'/>
> </sound>
> <video>
> <model type='qxl' ram='65536' vram='32768'
vgamem='16384' heads='1' primary='yes'/>
> <alias name='ua-bcfb6b4b-0b3c-4d5b-ba2d-8ce40a65facd'/>
> <address type='pci' domain='0x0000' bus='0x00'
slot='0x02' function='0x0'/>
> </video>
> <memballoon model='virtio'>
> <stats period='5'/>
> <alias name='ua-39d36063-8808-47db-9fef-a0baad9f9661'/>
> <address type='pci' domain='0x0000' bus='0x00'
slot='0x08' function='0x0'/>
> </memballoon>
> <rng model='virtio'>
> <backend model='random'>/dev/urandom</backend>
> <alias name='ua-75516d34-dd8f-4f0f-8496-e1f222a359a8'/>
> <address type='pci' domain='0x0000' bus='0x00'
slot='0x09' function='0x0'/>
> </rng>
> </devices>
> <seclabel type='dynamic' model='selinux'
relabel='yes'>
> <label>system_u:system_r:svirt_t:s0:c162,c716</label>
> <imagelabel>system_u:object_r:svirt_image_t:s0:c162,c716</imagelabel>
> </seclabel>
> <seclabel type='dynamic' model='dac' relabel='yes'>
> <label>+107:+107</label>
> <imagelabel>+107:+107</imagelabel>
> </seclabel>
></domain>
>
>hosts:
>
><capabilities>
> <host>
> <uuid>b25adcec-eef2-49a5-8663-7cdcfc50891b</uuid>
> <cpu>
> <arch>x86_64</arch>
> <model>Westmere</model>
> <vendor>Intel</vendor>
> <microcode version='34'/>
> <counter name='tsc' frequency='3699996000'
scaling='no'/>
> <topology sockets='1' cores='2' threads='2'/>
> <feature name='vme'/>
> <feature name='ds'/>
> <feature name='acpi'/>
> <feature name='ss'/>
> <feature name='ht'/>
> <feature name='tm'/>
> <feature name='pbe'/>
> <feature name='pclmuldq'/>
> <feature name='dtes64'/>
> <feature name='monitor'/>
> <feature name='ds_cpl'/>
> <feature name='vmx'/>
> <feature name='est'/>
> <feature name='tm2'/>
> <feature name='fma'/>
> <feature name='xtpr'/>
> <feature name='pdcm'/>
> <feature name='pcid'/>
> <feature name='movbe'/>
> <feature name='tsc-deadline'/>
> <feature name='xsave'/>
> <feature name='osxsave'/>
> <feature name='avx'/>
> <feature name='f16c'/>
> <feature name='rdrand'/>
> <feature name='arat'/>
> <feature name='fsgsbase'/>
> <feature name='tsc_adjust'/>
> <feature name='bmi1'/>
> <feature name='avx2'/>
> <feature name='smep'/>
> <feature name='bmi2'/>
> <feature name='erms'/>
> <feature name='invpcid'/>
> <feature name='xsaveopt'/>
> <feature name='pdpe1gb'/>
> <feature name='rdtscp'/>
> <feature name='abm'/>
> <feature name='invtsc'/>
> <pages unit='KiB' size='4'/>
> <pages unit='KiB' size='2048'/>
> <pages unit='KiB' size='1048576'/>
> </cpu>
> <power_management>
> <suspend_mem/>
> <suspend_disk/>
> <suspend_hybrid/>
> </power_management>
> <iommu support='no'/>
> <migration_features>
> <live/>
> <uri_transports>
> <uri_transport>tcp</uri_transport>
> <uri_transport>rdma</uri_transport>
> </uri_transports>
> </migration_features>
> <topology>
> <cells num='1'>
> <cell id='0'>
> <memory unit='KiB'>32903488</memory>
> <pages unit='KiB' size='4'>8225872</pages>
> <pages unit='KiB' size='2048'>0</pages>
> <pages unit='KiB' size='1048576'>0</pages>
> <distances>
> <sibling id='0' value='10'/>
> </distances>
> <cpus num='4'>
> <cpu id='0' socket_id='0' core_id='0'
siblings='0,2'/>
> <cpu id='1' socket_id='0' core_id='1'
siblings='1,3'/>
> <cpu id='2' socket_id='0' core_id='0'
siblings='0,2'/>
> <cpu id='3' socket_id='0' core_id='1'
siblings='1,3'/>
> </cpus>
> </cell>
> </cells>
> </topology>
> <cache>
> <bank id='0' level='3' type='both' size='3'
unit='MiB' cpus='0-3'/>
> </cache>
> <secmodel>
> <model>selinux</model>
> <doi>0</doi>
> <baselabel
type='kvm'>system_u:system_r:svirt_t:s0</baselabel>
> <baselabel
type='qemu'>system_u:system_r:svirt_tcg_t:s0</baselabel>
> </secmodel>
> <secmodel>
> <model>dac</model>
> <doi>0</doi>
> <baselabel type='kvm'>+107:+107</baselabel>
> <baselabel type='qemu'>+107:+107</baselabel>
> </secmodel>
> </host>
> <guest>
> <os_type>hvm</os_type>
> <arch name='i686'>
> <wordsize>32</wordsize>
> <emulator>/usr/libexec/qemu-kvm</emulator>
> <machine maxCpus='240'>pc-i440fx-rhel7.6.0</machine>
> <machine canonical='pc-i440fx-rhel7.6.0'
maxCpus='240'>pc</machine>
> <machine maxCpus='240'>pc-i440fx-rhel7.0.0</machine>
> <machine maxCpus='384'>pc-q35-rhel7.6.0</machine>
> <machine canonical='pc-q35-rhel7.6.0'
maxCpus='384'>q35</machine>
> <machine maxCpus='240'>rhel6.3.0</machine>
> <machine maxCpus='240'>rhel6.4.0</machine>
> <machine maxCpus='240'>rhel6.0.0</machine>
> <machine maxCpus='240'>pc-i440fx-rhel7.5.0</machine>
> <machine maxCpus='240'>pc-i440fx-rhel7.1.0</machine>
> <machine maxCpus='240'>pc-i440fx-rhel7.2.0</machine>
> <machine maxCpus='255'>pc-q35-rhel7.3.0</machine>
> <machine maxCpus='240'>rhel6.5.0</machine>
> <machine maxCpus='384'>pc-q35-rhel7.4.0</machine>
> <machine maxCpus='240'>rhel6.6.0</machine>
> <machine maxCpus='240'>rhel6.1.0</machine>
> <machine maxCpus='240'>rhel6.2.0</machine>
> <machine maxCpus='240'>pc-i440fx-rhel7.3.0</machine>
> <machine maxCpus='240'>pc-i440fx-rhel7.4.0</machine>
> <machine maxCpus='384'>pc-q35-rhel7.5.0</machine>
> <domain type='qemu'/>
> <domain type='kvm'>
> <emulator>/usr/libexec/qemu-kvm</emulator>
> </domain>
> </arch>
> <features>
> <cpuselection/>
> <deviceboot/>
> <disksnapshot default='on' toggle='no'/>
> <acpi default='on' toggle='yes'/>
> <apic default='on' toggle='no'/>
> <pae/>
> <nonpae/>
> </features>
> </guest>
> <guest>
> <os_type>hvm</os_type>
> <arch name='x86_64'>
> <wordsize>64</wordsize>
> <emulator>/usr/libexec/qemu-kvm</emulator>
> <machine maxCpus='240'>pc-i440fx-rhel7.6.0</machine>
> <machine canonical='pc-i440fx-rhel7.6.0'
maxCpus='240'>pc</machine>
> <machine maxCpus='240'>pc-i440fx-rhel7.0.0</machine>
> <machine maxCpus='384'>pc-q35-rhel7.6.0</machine>
> <machine canonical='pc-q35-rhel7.6.0'
maxCpus='384'>q35</machine>
> <machine maxCpus='240'>rhel6.3.0</machine>
> <machine maxCpus='240'>rhel6.4.0</machine>
> <machine maxCpus='240'>rhel6.0.0</machine>
> <machine maxCpus='240'>pc-i440fx-rhel7.5.0</machine>
> <machine maxCpus='240'>pc-i440fx-rhel7.1.0</machine>
> <machine maxCpus='240'>pc-i440fx-rhel7.2.0</machine>
> <machine maxCpus='255'>pc-q35-rhel7.3.0</machine>
> <machine maxCpus='240'>rhel6.5.0</machine>
> <machine maxCpus='384'>pc-q35-rhel7.4.0</machine>
> <machine maxCpus='240'>rhel6.6.0</machine>
> <machine maxCpus='240'>rhel6.1.0</machine>
> <machine maxCpus='240'>rhel6.2.0</machine>
> <machine maxCpus='240'>pc-i440fx-rhel7.3.0</machine>
> <machine maxCpus='240'>pc-i440fx-rhel7.4.0</machine>
> <machine maxCpus='384'>pc-q35-rhel7.5.0</machine>
> <domain type='qemu'/>
> <domain type='kvm'>
> <emulator>/usr/libexec/qemu-kvm</emulator>
> </domain>
> </arch>
> <features>
> <cpuselection/>
> <deviceboot/>
> <disksnapshot default='on' toggle='no'/>
> <acpi default='on' toggle='yes'/>
> <apic default='on' toggle='no'/>
> </features>
> </guest>
></capabilities>
>
>
>在 2020-09-17 12:00:19,"Strahil Nikolov" <hunter86_bg(a)yahoo.com> 写道:
>>It would be easier if you posted the whole xml.
>>
>>What about the sections (in HE xml) starting with:
>>feature policy=
>>
>>Also the hosts have a section which contains:
>>
>><feature name=
>>
>>If you can share a VM's xml sections for a good VM.
>>
>>Best Regards,
>>Strahil Nikolov
>>
>>
>>
>>
>>
>>
>>В четвъртък, 17 септември 2020 г., 05:54:12 Гринуич+3, ddqlo <ddqlo(a)126.com>
написа:
>>
>>
>>
>>
>>
>>HostedEngine:
>>......
>><model fallback='forbid'>Haswell-noTSX</model>
>>......
>>
>>both of the hosts:
>>......
>><model>Westmere</model>
>>......
>>
>>others vms which can be migrated:
>>......
>><model fallback='forbid'>Haswell-noTSX</model>
>>......
>>
>>
>>
>>在 2020-09-17 03:03:24,"Strahil Nikolov" <hunter86_bg(a)yahoo.com>
写道:
>>>Can you verify the HostedEngine's CPU ?
>>>
>>>1. ssh to the host hosting the HE
>>>2. alias virsh='virsh -c
qemu:///system?authfile=/etc/ovirt-hosted-engine/virsh_auth.conf'
>>>3. virsh dumpxml HostedEngine
>>>
>>>
>>>Then set the alias for virsh on all Hosts and 'virsh capabilites'
should show the Hosts' <cpu><model> .
>>>
>>>Best Regards,
>>>Strahil Nikolov
>>>
>>>
>>>
>>>
>>>
>>>
>>>В сряда, 16 септември 2020 г., 10:16:08 Гринуич+3, ddqlo <ddqlo(a)126.com>
написа:
>>>
>>>
>>>
>>>
>>>
>>>My gateway was not pingable. I have fixed this problem and now both nodes have
a score(3400).
>>>Yet, hosted engine could not be migrated. Same log in engine.log:
>>>host filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'CPU'
>>>
>>>
>>>在 2020-09-16 02:11:09,"Strahil Nikolov"
<hunter86_bg(a)yahoo.com> 写道:
>>>>Both nodes have a lower than the usual score (should be 3400 ).
>>>>Based on the score you are probably suffering from gateway-score-penalty
[1][2].
>>>>Check if your gateway is pingable.
>>>>
>>>>Best Regards,
>>>>Strahil Nikolov
>>>>
>>>>1 - https://www.ovirt.org/images/Hosted-Engine-4.3-deep-dive.pdf(page 8)
>>>>2 - /etc/ovirt-hosted-engine-ha/agent.conf
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>В вторник, 15 септември 2020 г., 04:49:48 Гринуич+3, ddqlo
<ddqlo(a)126.com> написа:
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>--== Host node28 (id: 1) status ==--
>>>>
>>>>conf_on_shared_storage : True
>>>>Status up-to-date : True
>>>>Hostname : node28
>>>>Host ID : 1
>>>>Engine status : {"reason": "vm not
running on this host", "health": "bad", "vm":
"down_unexpected", "detail": "unknown"}
>>>>Score : 1800
>>>>stopped : False
>>>>Local maintenance : False
>>>>crc32 : 4ac6105b
>>>>local_conf_timestamp : 1794597
>>>>Host timestamp : 1794597
>>>>Extra metadata (valid at timestamp):
>>>> metadata_parse_version=1
>>>> metadata_feature_version=1
>>>> timestamp=1794597 (Tue Sep 15 09:47:17 2020)
>>>> host-id=1
>>>> score=1800
>>>> vm_conf_refresh_time=1794597 (Tue Sep 15 09:47:17 2020)
>>>> conf_on_shared_storage=True
>>>> maintenance=False
>>>> state=EngineDown
>>>> stopped=False
>>>>
>>>>
>>>>--== Host node22 (id: 2) status ==--
>>>>
>>>>conf_on_shared_storage : True
>>>>Status up-to-date : True
>>>>Hostname : node22
>>>>Host ID : 2
>>>>Engine status : {"health":
"good", "vm": "up", "detail": "Up"}
>>>>Score : 1800
>>>>stopped : False
>>>>Local maintenance : False
>>>>crc32 : ffc41893
>>>>local_conf_timestamp : 1877876
>>>>Host timestamp : 1877876
>>>>Extra metadata (valid at timestamp):
>>>> metadata_parse_version=1
>>>> metadata_feature_version=1
>>>> timestamp=1877876 (Tue Sep 15 09:47:13 2020)
>>>> host-id=2
>>>> score=1800
>>>> vm_conf_refresh_time=1877876 (Tue Sep 15 09:47:13 2020)
>>>> conf_on_shared_storage=True
>>>> maintenance=False
>>>> state=EngineUp
>>>> stopped=False
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>在 2020-09-09 01:32:55,"Strahil Nikolov"
<hunter86_bg(a)yahoo.com> 写道:
>>>>>What is the output of 'hosted-engine --vm-status' on the node
where the HostedEngine is running ?
>>>>>
>>>>>
>>>>>Best Regards,
>>>>>Strahil Nikolov
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>В понеделник, 7 септември 2020 г., 03:53:13 Гринуич+3, ddqlo
<ddqlo(a)126.com> написа:
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>I could not find any logs because the migration button is disabled in
the web UI. It seems that the engine migration operation is prevented at first. Any other
ideas? Thanks!
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>在 2020-09-01 00:06:19,"Strahil Nikolov"
<hunter86_bg(a)yahoo.com> 写道:
>>>>>>I'm running oVirt 4.3.10 and I can migrate my Engine from node
to node.
>>>>>>I had one similar issue , but powering off and on the HE has fixed
it.
>>>>>>
>>>>>>You have to check the vdsm log on the source and on destination in
order to figure out what is going on.
>>>>>>Also you might consider checking the libvirt logs on the
destination.
>>>>>>
>>>>>>Best Regards,
>>>>>>Strahil Nikolov
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>В понеделник, 31 август 2020 г., 10:47:22 Гринуич+3, ddqlo
<ddqlo(a)126.com> написа:
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>Thanks! The scores of all nodes are not '0'. I find that
someone has already asked a question like this. It seems that this feature has been
disabled in 4.3. I am not sure if it is enabled in 4.4.
>>>>>>
>>>>>>
>>>>>>在 2020-08-29 02:27:03,"Strahil Nikolov"
<hunter86_bg(a)yahoo.com> :
>>>>>>>Have you checked under a shell the output of
'hosted-engine --vm-status' . Check the Score of the hosts. Maybe there is a node
with score of '0' ?
>>>>>>>
>>>>>>>Best Regards,
>>>>>>>Strahil Nikolov
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>В вторник, 25 август 2020 г., 13:46:18 Гринуич+3, 董青龙
<ddqlo(a)126.com> написа:
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>Hi all,
>>>>>>> I have an ovirt4.3.10.4 environment of 2 hosts. Normal
vms in this environment could be migrated, but the hosted engine vm could not be migrated.
Anyone can help? Thanks a lot!
>>>>>>>
>>>>>>>hosts status:
>>>>>>>
>>>>>>>normal vm migration:
>>>>>>>
>>>>>>>hosted engine vm migration:
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>_______________________________________________
>>>>>>>Users mailing list -- users(a)ovirt.org
>>>>>>>To unsubscribe send an email to users-leave(a)ovirt.org
>>>>>>>Privacy Statement:
https://www.ovirt.org/privacy-policy.html
>>>>>>>oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
>>>>>>>List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZXHE2AJX42H...
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>_______________________________________________
>>>>>>Users mailing list -- users(a)ovirt.org
>>>>>>To unsubscribe send an email to users-leave(a)ovirt.org
>>>>>>Privacy Statement:
https://www.ovirt.org/privacy-policy.html
>>>>>>oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
>>>>>>List Archives:
>>>>>>https://lists.ovirt.org/archives/list/users@ovirt.org/message/IAYLFLC6K42OUPZSZU3P3ZYAU66LGSCD/
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>_______________________________________________
>>>>>Users mailing list -- users(a)ovirt.org
>>>>>To unsubscribe send an email to users-leave(a)ovirt.org
>>>>>Privacy Statement:
https://www.ovirt.org/privacy-policy.html
>>>>>oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
>>>>>List Archives:
>>>>>https://lists.ovirt.org/archives/list/users@ovirt.org/message/23ZMAP5K7N5KKX3HOKUEJNVCBH7CY4ZL/
>>>>
>>>>
>>>>
>>>>
>>>>
>>>
>>>
>>>
>>>
>>>
>>
>>
>>
>>
>>
>
>
>
>
>
_______________________________________________
Users mailing list -- users(a)ovirt.org
To unsubscribe send an email to users-leave(a)ovirt.org
Privacy Statement:
https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/BM3QAMWVBKU...
_______________________________________________
Users mailing list -- users(a)ovirt.org
To unsubscribe send an email to users-leave(a)ovirt.org
Privacy Statement: