That's quite strange.
Any errors/clues in the Engine's logs ?
Best Regards,
Strahil Nikolov
В понеделник, 21 септември 2020 г., 05:58:35 Гринуич+3, ddqlo <ddqlo(a)126.com>
написа:
so strange! After I set global maintenance, powered off and started H The cpu of HE became
'Westmere'(did not change anything). But HE still could not be migrated.
HE xml:
<cpu mode='custom' match='exact' check='full'>
<model fallback='forbid'>Westmere</model>
<topology sockets='16' cores='4' threads='1'/>
<feature policy='require' name='vme'/>
<feature policy='require' name='pclmuldq'/>
<feature policy='require' name='x2apic'/>
<feature policy='require' name='hypervisor'/>
<feature policy='require' name='arat'/>
<numa>
<cell id='0' cpus='0-3' memory='16777216'
unit='KiB'/>
</numa>
</cpu>
host capabilities:
<model>Westmere</model>
cluster cpu type (UI):
host cpu type (UI):
HE cpu type (UI):
在 2020-09-19 13:27:35,"Strahil Nikolov" <hunter86_bg(a)yahoo.com> 写道:
Hm... interesting.
The VM is using 'Haswell-noTSX' while the host is 'Westmere'.
In my case I got no difference:
[root@ovirt1 ~]# virsh dumpxml HostedEngine | grep Opteron
<model fallback='forbid'>Opteron_G5</model>
[root@ovirt1 ~]# virsh capabilities | grep Opteron
<model>Opteron_G5</model>
Did you update the cluster holding the Hosted Engine ?
I guess you can try to:
- Set global maintenance
- Power off the HostedEngine VM
- virsh dumpxml HostedEngine > /root/HE.xml
- use virsh edit to change the cpu of the HE (non-permanent) change
- try to power on the modified HE
If it powers on , you can try to migrate it and if it succeeds - then you should make it
permanent.
Best Regards,
Strahil Nikolov
В петък, 18 септември 2020 г., 04:40:39 Гринуич+3, ddqlo <ddqlo(a)126.com> написа:
HE:
<domain type='kvm' id='1'>
<name>HostedEngine</name>
<uuid>b4e805ff-556d-42bd-a6df-02f5902fd01c</uuid>
<metadata
xmlns:ns0="http://ovirt.org/vm/tune/1.0"
xmlns:ovirt-vm="http://ovirt.org/vm/1.0">
<ns0:qos/>
<ovirt-vm:vm
xmlns:ovirt-vm="http://ovirt.org/vm/1.0">
<ovirt-vm:clusterVersion>4.3</ovirt-vm:clusterVersion>
<ovirt-vm:destroy_on_reboot
type="bool">False</ovirt-vm:destroy_on_reboot>
<ovirt-vm:launchPaused>false</ovirt-vm:launchPaused>
<ovirt-vm:memGuaranteedSize
type="int">1024</ovirt-vm:memGuaranteedSize>
<ovirt-vm:minGuaranteedMemoryMb
type="int">1024</ovirt-vm:minGuaranteedMemoryMb>
<ovirt-vm:resumeBehavior>auto_resume</ovirt-vm:resumeBehavior>
<ovirt-vm:startTime
type="float">1600307555.19</ovirt-vm:startTime>
<ovirt-vm:device mac_address="56:6f:9b:b0:00:01">
<ovirt-vm:network>external</ovirt-vm:network>
<ovirt-vm:custom>
<ovirt-vm:queues>4</ovirt-vm:queues>
</ovirt-vm:custom>
</ovirt-vm:device>
<ovirt-vm:device mac_address="00:16:3e:50:c1:97">
<ovirt-vm:network>ovirtmgmt</ovirt-vm:network>
<ovirt-vm:custom>
<ovirt-vm:queues>4</ovirt-vm:queues>
</ovirt-vm:custom>
</ovirt-vm:device>
<ovirt-vm:device devtype="disk" name="vda">
<ovirt-vm:domainID>c17c1934-332f-464c-8f89-ad72463c00b3</ovirt-vm:domainID>
<ovirt-vm:guestName>/dev/vda2</ovirt-vm:guestName>
<ovirt-vm:imageID>8eca143a-4535-4421-bd35-9f5764d67d70</ovirt-vm:imageID>
<ovirt-vm:poolID>00000000-0000-0000-0000-000000000000</ovirt-vm:poolID>
<ovirt-vm:shared>exclusive</ovirt-vm:shared>
<ovirt-vm:volumeID>ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</ovirt-vm:volumeID>
<ovirt-vm:specParams>
<ovirt-vm:pinToIoThread>1</ovirt-vm:pinToIoThread>
</ovirt-vm:specParams>
<ovirt-vm:volumeChain>
<ovirt-vm:volumeChainNode>
<ovirt-vm:domainID>c17c1934-332f-464c-8f89-ad72463c00b3</ovirt-vm:domainID>
<ovirt-vm:imageID>8eca143a-4535-4421-bd35-9f5764d67d70</ovirt-vm:imageID>
<ovirt-vm:leaseOffset
type="int">108003328</ovirt-vm:leaseOffset>
<ovirt-vm:leasePath>/dev/c17c1934-332f-464c-8f89-ad72463c00b3/leases</ovirt-vm:leasePath>
<ovirt-vm:path>/rhev/data-center/mnt/blockSD/c17c1934-332f-464c-8f89-ad72463c00b3/images/8eca143a-4535-4421-bd35-9f5764d67d70/ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</ovirt-vm:path>
<ovirt-vm:volumeID>ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</ovirt-vm:volumeID>
</ovirt-vm:volumeChainNode>
</ovirt-vm:volumeChain>
</ovirt-vm:device>
<ovirt-vm:device devtype="disk" name="hdc"/>
</ovirt-vm:vm>
</metadata>
<maxMemory slots='16' unit='KiB'>67108864</maxMemory>
<memory unit='KiB'>16777216</memory>
<currentMemory unit='KiB'>16777216</currentMemory>
<vcpu placement='static' current='4'>64</vcpu>
<iothreads>1</iothreads>
<resource>
<partition>/machine</partition>
</resource>
<sysinfo type='smbios'>
<system>
<entry name='manufacturer'>oVirt</entry>
<entry name='product'>oVirt Node</entry>
<entry name='version'>7-5.1804.el7.centos</entry>
<entry
name='serial'>00000000-0000-0000-0000-0CC47A6B3160</entry>
<entry name='uuid'>b4e805ff-556d-42bd-a6df-02f5902fd01c</entry>
</system>
</sysinfo>
<os>
<type arch='x86_64'
machine='pc-i440fx-rhel7.6.0'>hvm</type>
<boot dev='hd'/>
<bios useserial='yes'/>
<smbios mode='sysinfo'/>
</os>
<features>
<acpi/>
</features>
<cpu mode='custom' match='exact' check='full'>
<model fallback='forbid'>Haswell-noTSX</model>
<topology sockets='16' cores='4' threads='1'/>
<feature policy='require' name='vme'/>
<feature policy='require' name='f16c'/>
<feature policy='require' name='rdrand'/>
<feature policy='require' name='hypervisor'/>
<feature policy='require' name='arat'/>
<feature policy='require' name='xsaveopt'/>
<feature policy='require' name='abm'/>
<numa>
<cell id='0' cpus='0-3' memory='16777216'
unit='KiB'/>
</numa>
</cpu>
<clock offset='variable' adjustment='0' basis='utc'>
<timer name='rtc' tickpolicy='catchup'/>
<timer name='pit' tickpolicy='delay'/>
<timer name='hpet' present='no'/>
</clock>
<on_poweroff>destroy</on_poweroff>
<on_reboot>destroy</on_reboot>
<on_crash>destroy</on_crash>
<pm>
<suspend-to-mem enabled='no'/>
<suspend-to-disk enabled='no'/>
</pm>
<devices>
<emulator>/usr/libexec/qemu-kvm</emulator>
<disk type='file' device='cdrom'>
<driver name='qemu' error_policy='report'/>
<source startupPolicy='optional'/>
<target dev='hdc' bus='ide'/>
<readonly/>
<alias name='ua-80fde7d5-ee7f-4201-9118-11bc6c3b8530'/>
<address type='drive' controller='0' bus='1'
target='0' unit='0'/>
</disk>
<disk type='block' device='disk' snapshot='no'>
<driver name='qemu' type='raw' cache='none'
error_policy='stop' io='native' iothread='1'/>
<source
dev='/var/run/vdsm/storage/c17c1934-332f-464c-8f89-ad72463c00b3/8eca143a-4535-4421-bd35-9f5764d67d70/ae961104-c3b3-4a43-9f46-7fa6bdc2ac33'>
<seclabel model='dac' relabel='no'/>
</source>
<backingStore/>
<target dev='vda' bus='virtio'/>
<serial>8eca143a-4535-4421-bd35-9f5764d67d70</serial>
<alias name='ua-8eca143a-4535-4421-bd35-9f5764d67d70'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x07' function='0x0'/>
</disk>
<controller type='scsi' index='0' model='virtio-scsi'>
<driver iothread='1'/>
<alias name='ua-27331e83-03f4-42a3-9554-c41649c02ba4'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x05' function='0x0'/>
</controller>
<controller type='ide' index='0'>
<alias name='ide'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x01' function='0x1'/>
</controller>
<controller type='virtio-serial' index='0' ports='16'>
<alias name='ua-8fe74299-b60f-4778-8e80-db05393a9489'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x06' function='0x0'/>
</controller>
<controller type='usb' index='0' model='piix3-uhci'>
<alias name='usb'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x01' function='0x2'/>
</controller>
<controller type='pci' index='0' model='pci-root'>
<alias name='pci.0'/>
</controller>
<lease>
<lockspace>c17c1934-332f-464c-8f89-ad72463c00b3</lockspace>
<key>ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</key>
<target path='/dev/c17c1934-332f-464c-8f89-ad72463c00b3/leases'
offset='108003328'/>
</lease>
<interface type='bridge'>
<mac address='00:16:3e:50:c1:97'/>
<source bridge='ovirtmgmt'/>
<target dev='vnet0'/>
<model type='virtio'/>
<driver name='vhost' queues='4'/>
<filterref filter='vdsm-no-mac-spoofing'/>
<link state='up'/>
<mtu size='1500'/>
<alias name='ua-fada74ee-2338-4cde-a7ba-43a9a636ad6e'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x03' function='0x0'/>
</interface>
<interface type='bridge'>
<mac address='56:6f:9b:b0:00:01'/>
<source bridge='external'/>
<target dev='vnet1'/>
<model type='virtio'/>
<driver name='vhost' queues='4'/>
<filterref filter='vdsm-no-mac-spoofing'/>
<link state='up'/>
<mtu size='1500'/>
<alias name='ua-f7b4c949-1f9f-4355-811d-88428c88ce4e'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x0a' function='0x0'/>
</interface>
<serial type='unix'>
<source mode='bind'
path='/var/run/ovirt-vmconsole-console/b4e805ff-556d-42bd-a6df-02f5902fd01c.sock'/>
<target type='isa-serial' port='0'>
<model name='isa-serial'/>
</target>
<alias name='serial0'/>
</serial>
<console type='unix'>
<source mode='bind'
path='/var/run/ovirt-vmconsole-console/b4e805ff-556d-42bd-a6df-02f5902fd01c.sock'/>
<target type='serial' port='0'/>
<alias name='serial0'/>
</console>
<channel type='unix'>
<source mode='bind'
path='/var/lib/libvirt/qemu/channels/b4e805ff-556d-42bd-a6df-02f5902fd01c.ovirt-guest-agent.0'/>
<target type='virtio' name='ovirt-guest-agent.0'
state='connected'/>
<alias name='channel0'/>
<address type='virtio-serial' controller='0' bus='0'
port='1'/>
</channel>
<channel type='unix'>
<source mode='bind'
path='/var/lib/libvirt/qemu/channels/b4e805ff-556d-42bd-a6df-02f5902fd01c.org.qemu.guest_agent.0'/>
<target type='virtio' name='org.qemu.guest_agent.0'
state='connected'/>
<alias name='channel1'/>
<address type='virtio-serial' controller='0' bus='0'
port='2'/>
</channel>
<channel type='spicevmc'>
<target type='virtio' name='com.redhat.spice.0'
state='disconnected'/>
<alias name='channel2'/>
<address type='virtio-serial' controller='0' bus='0'
port='3'/>
</channel>
<channel type='unix'>
<source mode='bind'
path='/var/lib/libvirt/qemu/channels/b4e805ff-556d-42bd-a6df-02f5902fd01c.org.ovirt.hosted-engine-setup.0'/>
<target type='virtio' name='org.ovirt.hosted-engine-setup.0'
state='disconnected'/>
<alias name='channel3'/>
<address type='virtio-serial' controller='0' bus='0'
port='4'/>
</channel>
<input type='tablet' bus='usb'>
<alias name='input0'/>
<address type='usb' bus='0' port='1'/>
</input>
<input type='mouse' bus='ps2'>
<alias name='input1'/>
</input>
<input type='keyboard' bus='ps2'>
<alias name='input2'/>
</input>
<graphics type='vnc' port='5900' autoport='yes'
listen='192.168.1.22' keymap='en-us'
passwdValidTo='1970-01-01T00:00:01'>
<listen type='network' address='192.168.1.22'
network='vdsm-external'/>
</graphics>
<graphics type='spice' port='5901' tlsPort='5902'
autoport='yes' listen='192.168.1.22'
passwdValidTo='1970-01-01T00:00:01'>
<listen type='network' address='192.168.1.22'
network='vdsm-external'/>
<channel name='main' mode='secure'/>
<channel name='display' mode='secure'/>
<channel name='inputs' mode='secure'/>
<channel name='cursor' mode='secure'/>
<channel name='playback' mode='secure'/>
<channel name='record' mode='secure'/>
<channel name='smartcard' mode='secure'/>
<channel name='usbredir' mode='secure'/>
</graphics>
<sound model='ich6'>
<alias name='ua-bd287767-9b83-4e44-ac6f-8b527f9632b8'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x04' function='0x0'/>
</sound>
<video>
<model type='qxl' ram='65536' vram='32768'
vgamem='16384' heads='1' primary='yes'/>
<alias name='ua-bcfb6b4b-0b3c-4d5b-ba2d-8ce40a65facd'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x02' function='0x0'/>
</video>
<memballoon model='virtio'>
<stats period='5'/>
<alias name='ua-39d36063-8808-47db-9fef-a0baad9f9661'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x08' function='0x0'/>
</memballoon>
<rng model='virtio'>
<backend model='random'>/dev/urandom</backend>
<alias name='ua-75516d34-dd8f-4f0f-8496-e1f222a359a8'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x09' function='0x0'/>
</rng>
</devices>
<seclabel type='dynamic' model='selinux' relabel='yes'>
<label>system_u:system_r:svirt_t:s0:c162,c716</label>
<imagelabel>system_u:object_r:svirt_image_t:s0:c162,c716</imagelabel>
</seclabel>
<seclabel type='dynamic' model='dac' relabel='yes'>
<label>+107:+107</label>
<imagelabel>+107:+107</imagelabel>
</seclabel>
</domain>
hosts:
<capabilities>
<host>
<uuid>b25adcec-eef2-49a5-8663-7cdcfc50891b</uuid>
<cpu>
<arch>x86_64</arch>
<model>Westmere</model>
<vendor>Intel</vendor>
<microcode version='34'/>
<counter name='tsc' frequency='3699996000'
scaling='no'/>
<topology sockets='1' cores='2' threads='2'/>
<feature name='vme'/>
<feature name='ds'/>
<feature name='acpi'/>
<feature name='ss'/>
<feature name='ht'/>
<feature name='tm'/>
<feature name='pbe'/>
<feature name='pclmuldq'/>
<feature name='dtes64'/>
<feature name='monitor'/>
<feature name='ds_cpl'/>
<feature name='vmx'/>
<feature name='est'/>
<feature name='tm2'/>
<feature name='fma'/>
<feature name='xtpr'/>
<feature name='pdcm'/>
<feature name='pcid'/>
<feature name='movbe'/>
<feature name='tsc-deadline'/>
<feature name='xsave'/>
<feature name='osxsave'/>
<feature name='avx'/>
<feature name='f16c'/>
<feature name='rdrand'/>
<feature name='arat'/>
<feature name='fsgsbase'/>
<feature name='tsc_adjust'/>
<feature name='bmi1'/>
<feature name='avx2'/>
<feature name='smep'/>
<feature name='bmi2'/>
<feature name='erms'/>
<feature name='invpcid'/>
<feature name='xsaveopt'/>
<feature name='pdpe1gb'/>
<feature name='rdtscp'/>
<feature name='abm'/>
<feature name='invtsc'/>
<pages unit='KiB' size='4'/>
<pages unit='KiB' size='2048'/>
<pages unit='KiB' size='1048576'/>
</cpu>
<power_management>
<suspend_mem/>
<suspend_disk/>
<suspend_hybrid/>
</power_management>
<iommu support='no'/>
<migration_features>
<live/>
<uri_transports>
<uri_transport>tcp</uri_transport>
<uri_transport>rdma</uri_transport>
</uri_transports>
</migration_features>
<topology>
<cells num='1'>
<cell id='0'>
<memory unit='KiB'>32903488</memory>
<pages unit='KiB' size='4'>8225872</pages>
<pages unit='KiB' size='2048'>0</pages>
<pages unit='KiB' size='1048576'>0</pages>
<distances>
<sibling id='0' value='10'/>
</distances>
<cpus num='4'>
<cpu id='0' socket_id='0' core_id='0'
siblings='0,2'/>
<cpu id='1' socket_id='0' core_id='1'
siblings='1,3'/>
<cpu id='2' socket_id='0' core_id='0'
siblings='0,2'/>
<cpu id='3' socket_id='0' core_id='1'
siblings='1,3'/>
</cpus>
</cell>
</cells>
</topology>
<cache>
<bank id='0' level='3' type='both' size='3'
unit='MiB' cpus='0-3'/>
</cache>
<secmodel>
<model>selinux</model>
<doi>0</doi>
<baselabel type='kvm'>system_u:system_r:svirt_t:s0</baselabel>
<baselabel
type='qemu'>system_u:system_r:svirt_tcg_t:s0</baselabel>
</secmodel>
<secmodel>
<model>dac</model>
<doi>0</doi>
<baselabel type='kvm'>+107:+107</baselabel>
<baselabel type='qemu'>+107:+107</baselabel>
</secmodel>
</host>
<guest>
<os_type>hvm</os_type>
<arch name='i686'>
<wordsize>32</wordsize>
<emulator>/usr/libexec/qemu-kvm</emulator>
<machine maxCpus='240'>pc-i440fx-rhel7.6.0</machine>
<machine canonical='pc-i440fx-rhel7.6.0'
maxCpus='240'>pc</machine>
<machine maxCpus='240'>pc-i440fx-rhel7.0.0</machine>
<machine maxCpus='384'>pc-q35-rhel7.6.0</machine>
<machine canonical='pc-q35-rhel7.6.0'
maxCpus='384'>q35</machine>
<machine maxCpus='240'>rhel6.3.0</machine>
<machine maxCpus='240'>rhel6.4.0</machine>
<machine maxCpus='240'>rhel6.0.0</machine>
<machine maxCpus='240'>pc-i440fx-rhel7.5.0</machine>
<machine maxCpus='240'>pc-i440fx-rhel7.1.0</machine>
<machine maxCpus='240'>pc-i440fx-rhel7.2.0</machine>
<machine maxCpus='255'>pc-q35-rhel7.3.0</machine>
<machine maxCpus='240'>rhel6.5.0</machine>
<machine maxCpus='384'>pc-q35-rhel7.4.0</machine>
<machine maxCpus='240'>rhel6.6.0</machine>
<machine maxCpus='240'>rhel6.1.0</machine>
<machine maxCpus='240'>rhel6.2.0</machine>
<machine maxCpus='240'>pc-i440fx-rhel7.3.0</machine>
<machine maxCpus='240'>pc-i440fx-rhel7.4.0</machine>
<machine maxCpus='384'>pc-q35-rhel7.5.0</machine>
<domain type='qemu'/>
<domain type='kvm'>
<emulator>/usr/libexec/qemu-kvm</emulator>
</domain>
</arch>
<features>
<cpuselection/>
<deviceboot/>
<disksnapshot default='on' toggle='no'/>
<acpi default='on' toggle='yes'/>
<apic default='on' toggle='no'/>
<pae/>
<nonpae/>
</features>
</guest>
<guest>
<os_type>hvm</os_type>
<arch name='x86_64'>
<wordsize>64</wordsize>
<emulator>/usr/libexec/qemu-kvm</emulator>
<machine maxCpus='240'>pc-i440fx-rhel7.6.0</machine>
<machine canonical='pc-i440fx-rhel7.6.0'
maxCpus='240'>pc</machine>
<machine maxCpus='240'>pc-i440fx-rhel7.0.0</machine>
<machine maxCpus='384'>pc-q35-rhel7.6.0</machine>
<machine canonical='pc-q35-rhel7.6.0'
maxCpus='384'>q35</machine>
<machine maxCpus='240'>rhel6.3.0</machine>
<machine maxCpus='240'>rhel6.4.0</machine>
<machine maxCpus='240'>rhel6.0.0</machine>
<machine maxCpus='240'>pc-i440fx-rhel7.5.0</machine>
<machine maxCpus='240'>pc-i440fx-rhel7.1.0</machine>
<machine maxCpus='240'>pc-i440fx-rhel7.2.0</machine>
<machine maxCpus='255'>pc-q35-rhel7.3.0</machine>
<machine maxCpus='240'>rhel6.5.0</machine>
<machine maxCpus='384'>pc-q35-rhel7.4.0</machine>
<machine maxCpus='240'>rhel6.6.0</machine>
<machine maxCpus='240'>rhel6.1.0</machine>
<machine maxCpus='240'>rhel6.2.0</machine>
<machine maxCpus='240'>pc-i440fx-rhel7.3.0</machine>
<machine maxCpus='240'>pc-i440fx-rhel7.4.0</machine>
<machine maxCpus='384'>pc-q35-rhel7.5.0</machine>
<domain type='qemu'/>
<domain type='kvm'>
<emulator>/usr/libexec/qemu-kvm</emulator>
</domain>
</arch>
<features>
<cpuselection/>
<deviceboot/>
<disksnapshot default='on' toggle='no'/>
<acpi default='on' toggle='yes'/>
<apic default='on' toggle='no'/>
</features>
</guest>
</capabilities>
在 2020-09-17 12:00:19,"Strahil Nikolov" <hunter86_bg(a)yahoo.com> 写道:
>It would be easier if you posted the whole xml.
>
>What about the sections (in HE xml) starting with:
>feature policy=
>
>Also the hosts have a section which contains:
>
><feature name=
>
>If you can share a VM's xml sections for a good VM.
>
>Best Regards,
>Strahil Nikolov
>
>
>
>
>
>
>В четвъртък, 17 септември 2020 г., 05:54:12 Гринуич+3, ddqlo <ddqlo(a)126.com>
написа:
>
>
>
>
>
>HostedEngine:
>......
><model fallback='forbid'>Haswell-noTSX</model>
>......
>
>both of the hosts:
>......
><model>Westmere</model>
>......
>
>others vms which can be migrated:
>......
><model fallback='forbid'>Haswell-noTSX</model>
>......
>
>
>
>在 2020-09-17 03:03:24,"Strahil Nikolov" <hunter86_bg(a)yahoo.com> 写道:
>>Can you verify the HostedEngine's CPU ?
>>
>>1. ssh to the host hosting the HE
>>2. alias virsh='virsh -c
qemu:///system?authfile=/etc/ovirt-hosted-engine/virsh_auth.conf'
>>3. virsh dumpxml HostedEngine
>>
>>
>>Then set the alias for virsh on all Hosts and 'virsh capabilites' should
show the Hosts' <cpu><model> .
>>
>>Best Regards,
>>Strahil Nikolov
>>
>>
>>
>>
>>
>>
>>В сряда, 16 септември 2020 г., 10:16:08 Гринуич+3, ddqlo <ddqlo(a)126.com>
написа:
>>
>>
>>
>>
>>
>>My gateway was not pingable. I have fixed this problem and now both nodes have a
score(3400).
>>Yet, hosted engine could not be migrated. Same log in engine.log:
>>host filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'CPU'
>>
>>
>>在 2020-09-16 02:11:09,"Strahil Nikolov" <hunter86_bg(a)yahoo.com>
写道:
>>>Both nodes have a lower than the usual score (should be 3400 ).
>>>Based on the score you are probably suffering from gateway-score-penalty
[1][2].
>>>Check if your gateway is pingable.
>>>
>>>Best Regards,
>>>Strahil Nikolov
>>>
>>>1 - https://www.ovirt.org/images/Hosted-Engine-4.3-deep-dive.pdf(page 8)
>>>2 - /etc/ovirt-hosted-engine-ha/agent.conf
>>>
>>>
>>>
>>>
>>>
>>>
>>>В вторник, 15 септември 2020 г., 04:49:48 Гринуич+3, ddqlo
<ddqlo(a)126.com> написа:
>>>
>>>
>>>
>>>
>>>
>>>--== Host node28 (id: 1) status ==--
>>>
>>>conf_on_shared_storage : True
>>>Status up-to-date : True
>>>Hostname : node28
>>>Host ID : 1
>>>Engine status : {"reason": "vm not running
on this host", "health": "bad", "vm":
"down_unexpected", "detail": "unknown"}
>>>Score : 1800
>>>stopped : False
>>>Local maintenance : False
>>>crc32 : 4ac6105b
>>>local_conf_timestamp : 1794597
>>>Host timestamp : 1794597
>>>Extra metadata (valid at timestamp):
>>> metadata_parse_version=1
>>> metadata_feature_version=1
>>> timestamp=1794597 (Tue Sep 15 09:47:17 2020)
>>> host-id=1
>>> score=1800
>>> vm_conf_refresh_time=1794597 (Tue Sep 15 09:47:17 2020)
>>> conf_on_shared_storage=True
>>> maintenance=False
>>> state=EngineDown
>>> stopped=False
>>>
>>>
>>>--== Host node22 (id: 2) status ==--
>>>
>>>conf_on_shared_storage : True
>>>Status up-to-date : True
>>>Hostname : node22
>>>Host ID : 2
>>>Engine status : {"health": "good",
"vm": "up", "detail": "Up"}
>>>Score : 1800
>>>stopped : False
>>>Local maintenance : False
>>>crc32 : ffc41893
>>>local_conf_timestamp : 1877876
>>>Host timestamp : 1877876
>>>Extra metadata (valid at timestamp):
>>> metadata_parse_version=1
>>> metadata_feature_version=1
>>> timestamp=1877876 (Tue Sep 15 09:47:13 2020)
>>> host-id=2
>>> score=1800
>>> vm_conf_refresh_time=1877876 (Tue Sep 15 09:47:13 2020)
>>> conf_on_shared_storage=True
>>> maintenance=False
>>> state=EngineUp
>>> stopped=False
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>在 2020-09-09 01:32:55,"Strahil Nikolov"
<hunter86_bg(a)yahoo.com> 写道:
>>>>What is the output of 'hosted-engine --vm-status' on the node
where the HostedEngine is running ?
>>>>
>>>>
>>>>Best Regards,
>>>>Strahil Nikolov
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>В понеделник, 7 септември 2020 г., 03:53:13 Гринуич+3, ddqlo
<ddqlo(a)126.com> написа:
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>I could not find any logs because the migration button is disabled in the
web UI. It seems that the engine migration operation is prevented at first. Any other
ideas? Thanks!
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>在 2020-09-01 00:06:19,"Strahil Nikolov"
<hunter86_bg(a)yahoo.com> 写道:
>>>>>I'm running oVirt 4.3.10 and I can migrate my Engine from node to
node.
>>>>>I had one similar issue , but powering off and on the HE has fixed
it.
>>>>>
>>>>>You have to check the vdsm log on the source and on destination in
order to figure out what is going on.
>>>>>Also you might consider checking the libvirt logs on the destination.
>>>>>
>>>>>Best Regards,
>>>>>Strahil Nikolov
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>В понеделник, 31 август 2020 г., 10:47:22 Гринуич+3, ddqlo
<ddqlo(a)126.com> написа:
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>Thanks! The scores of all nodes are not '0'. I find that
someone has already asked a question like this. It seems that this feature has been
disabled in 4.3. I am not sure if it is enabled in 4.4.
>>>>>
>>>>>
>>>>>在 2020-08-29 02:27:03,"Strahil Nikolov"
<hunter86_bg(a)yahoo.com> :
>>>>>>Have you checked under a shell the output of 'hosted-engine
--vm-status' . Check the Score of the hosts. Maybe there is a node with score of
'0' ?
>>>>>>
>>>>>>Best Regards,
>>>>>>Strahil Nikolov
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>В вторник, 25 август 2020 г., 13:46:18 Гринуич+3, 董青龙
<ddqlo(a)126.com> написа:
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>Hi all,
>>>>>> I have an ovirt4.3.10.4 environment of 2 hosts. Normal vms
in this environment could be migrated, but the hosted engine vm could not be migrated.
Anyone can help? Thanks a lot!
>>>>>>
>>>>>>hosts status:
>>>>>>
>>>>>>normal vm migration:
>>>>>>
>>>>>>hosted engine vm migration:
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>_______________________________________________
>>>>>>Users mailing list -- users(a)ovirt.org
>>>>>>To unsubscribe send an email to users-leave(a)ovirt.org
>>>>>>Privacy Statement:
https://www.ovirt.org/privacy-policy.html
>>>>>>oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
>>>>>>List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZXHE2AJX42H...
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>_______________________________________________
>>>>>Users mailing list -- users(a)ovirt.org
>>>>>To unsubscribe send an email to users-leave(a)ovirt.org
>>>>>Privacy Statement:
https://www.ovirt.org/privacy-policy.html
>>>>>oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
>>>>>List Archives:
>>>>>https://lists.ovirt.org/archives/list/users@ovirt.org/message/IAYLFLC6K42OUPZSZU3P3ZYAU66LGSCD/
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>_______________________________________________
>>>>Users mailing list -- users(a)ovirt.org
>>>>To unsubscribe send an email to users-leave(a)ovirt.org
>>>>Privacy Statement:
https://www.ovirt.org/privacy-policy.html
>>>>oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
>>>>List Archives:
>>>>https://lists.ovirt.org/archives/list/users@ovirt.org/message/23ZMAP5K7N5KKX3HOKUEJNVCBH7CY4ZL/
>>>
>>>
>>>
>>>
>>>
>>
>>
>>
>>
>>
>
>
>
>
>