Yes. I can. The host which does not host the HE could be reinstalled sucessfully in web UI. After this is done nothing has changed.






在 2020-09-22 03:08:18,"Strahil Nikolov" <hunter86_bg@yahoo.com> 写道: >Can you put 1 host in maintenance and use the "Installation" -> "Reinstall" and enable the HE deployment from one of the tabs ? > >Best Regards, >Strahil Nikolov > > > > > > >В понеделник, 21 септември 2020 г., 06:38:06 Гринуич+3, ddqlo <ddqlo@126.com> написа: > > > > > >so strange! After I set global maintenance, powered off and started H The cpu of HE became 'Westmere'(did not change anything). But HE still could not be migrated. > >HE xml: >  <cpu mode='custom' match='exact' check='full'> >    <model fallback='forbid'>Westmere</model> >    <topology sockets='16' cores='4' threads='1'/> >    <feature policy='require' name='vme'/> >    <feature policy='require' name='pclmuldq'/> >    <feature policy='require' name='x2apic'/> >    <feature policy='require' name='hypervisor'/> >    <feature policy='require' name='arat'/> >    <numa> >      <cell id='0' cpus='0-3' memory='16777216' unit='KiB'/> >    </numa> >  </cpu> > >host capabilities:  ><model>Westmere</model> > >cluster cpu type (UI):  > > >host cpu type (UI): > > >HE cpu type (UI): > > > > > > > >在 2020-09-19 13:27:35,"Strahil Nikolov" <hunter86_bg@yahoo.com> 写道: >>Hm... interesting. >> >>The VM is using 'Haswell-noTSX'  while the host is 'Westmere'. >> >>In my case I got no difference: >> >>[root@ovirt1 ~]# virsh  dumpxml HostedEngine | grep Opteron >>   <model fallback='forbid'>Opteron_G5</model> >>[root@ovirt1 ~]# virsh capabilities | grep Opteron >>     <model>Opteron_G5</model> >> >>Did you update the cluster holding the Hosted Engine ? >> >> >>I guess you can try to: >> >>- Set global maintenance >>- Power off the HostedEngine VM >>- virsh dumpxml HostedEngine > /root/HE.xml >>- use virsh edit to change the cpu of the HE (non-permanent) change >>- try to power on the modified HE >> >>If it powers on , you can try to migrate it and if it succeeds - then you should make it permanent. >> >> >> >> >> >>Best Regards, >>Strahil Nikolov >> >>В петък, 18 септември 2020 г., 04:40:39 Гринуич+3, ddqlo <ddqlo@126.com> написа: >> >> >> >> >> >>HE: >> >><domain type='kvm' id='1'> >>  <name>HostedEngine</name> >>  <uuid>b4e805ff-556d-42bd-a6df-02f5902fd01c</uuid> >>  <metadata xmlns:ns0="http://ovirt.org/vm/tune/1.0" xmlns:ovirt-vm="http://ovirt.org/vm/1.0"> >>    <ns0:qos/> >>    <ovirt-vm:vm xmlns:ovirt-vm="http://ovirt.org/vm/1.0"> >>    <ovirt-vm:clusterVersion>4.3</ovirt-vm:clusterVersion> >>    <ovirt-vm:destroy_on_reboot type="bool">False</ovirt-vm:destroy_on_reboot> >>    <ovirt-vm:launchPaused>false</ovirt-vm:launchPaused> >>    <ovirt-vm:memGuaranteedSize type="int">1024</ovirt-vm:memGuaranteedSize> >>    <ovirt-vm:minGuaranteedMemoryMb type="int">1024</ovirt-vm:minGuaranteedMemoryMb> >>    <ovirt-vm:resumeBehavior>auto_resume</ovirt-vm:resumeBehavior> >>    <ovirt-vm:startTime type="float">1600307555.19</ovirt-vm:startTime> >>    <ovirt-vm:device mac_address="56:6f:9b:b0:00:01"> >>        <ovirt-vm:network>external</ovirt-vm:network> >>        <ovirt-vm:custom> >>            <ovirt-vm:queues>4</ovirt-vm:queues> >>        </ovirt-vm:custom> >>    </ovirt-vm:device> >>    <ovirt-vm:device mac_address="00:16:3e:50:c1:97"> >>        <ovirt-vm:network>ovirtmgmt</ovirt-vm:network> >>        <ovirt-vm:custom> >>            <ovirt-vm:queues>4</ovirt-vm:queues> >>        </ovirt-vm:custom> >>    </ovirt-vm:device> >>    <ovirt-vm:device devtype="disk" name="vda"> >>        <ovirt-vm:domainID>c17c1934-332f-464c-8f89-ad72463c00b3</ovirt-vm:domainID> >>        <ovirt-vm:guestName>/dev/vda2</ovirt-vm:guestName> >>        <ovirt-vm:imageID>8eca143a-4535-4421-bd35-9f5764d67d70</ovirt-vm:imageID> >>        <ovirt-vm:poolID>00000000-0000-0000-0000-000000000000</ovirt-vm:poolID> >>        <ovirt-vm:shared>exclusive</ovirt-vm:shared> >>        <ovirt-vm:volumeID>ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</ovirt-vm:volumeID> >>        <ovirt-vm:specParams> >>            <ovirt-vm:pinToIoThread>1</ovirt-vm:pinToIoThread> >>        </ovirt-vm:specParams> >>        <ovirt-vm:volumeChain> >>            <ovirt-vm:volumeChainNode> >>                <ovirt-vm:domainID>c17c1934-332f-464c-8f89-ad72463c00b3</ovirt-vm:domainID> >>                <ovirt-vm:imageID>8eca143a-4535-4421-bd35-9f5764d67d70</ovirt-vm:imageID> >>                <ovirt-vm:leaseOffset type="int">108003328</ovirt-vm:leaseOffset> >>                <ovirt-vm:leasePath>/dev/c17c1934-332f-464c-8f89-ad72463c00b3/leases</ovirt-vm:leasePath> >>                <ovirt-vm:path>/rhev/data-center/mnt/blockSD/c17c1934-332f-464c-8f89-ad72463c00b3/images/8eca143a-4535-4421-bd35-9f5764d67d70/ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</ovirt-vm:path> >>                <ovirt-vm:volumeID>ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</ovirt-vm:volumeID> >>            </ovirt-vm:volumeChainNode> >>        </ovirt-vm:volumeChain> >>    </ovirt-vm:device> >>    <ovirt-vm:device devtype="disk" name="hdc"/> >></ovirt-vm:vm> >>  </metadata> >>  <maxMemory slots='16' unit='KiB'>67108864</maxMemory> >>  <memory unit='KiB'>16777216</memory> >>  <currentMemory unit='KiB'>16777216</currentMemory> >>  <vcpu placement='static' current='4'>64</vcpu> >>  <iothreads>1</iothreads> >>  <resource> >>    <partition>/machine</partition> >>  </resource> >>  <sysinfo type='smbios'> >>    <system> >>      <entry name='manufacturer'>oVirt</entry> >>      <entry name='product'>oVirt Node</entry> >>      <entry name='version'>7-5.1804.el7.centos</entry> >>      <entry name='serial'>00000000-0000-0000-0000-0CC47A6B3160</entry> >>      <entry name='uuid'>b4e805ff-556d-42bd-a6df-02f5902fd01c</entry> >>    </system> >>  </sysinfo> >>  <os> >>    <type arch='x86_64' machine='pc-i440fx-rhel7.6.0'>hvm</type> >>    <boot dev='hd'/> >>    <bios useserial='yes'/> >>    <smbios mode='sysinfo'/> >>  </os> >>  <features> >>    <acpi/> >>  </features> >>  <cpu mode='custom' match='exact' check='full'> >>    <model fallback='forbid'>Haswell-noTSX</model> >>    <topology sockets='16' cores='4' threads='1'/> >>    <feature policy='require' name='vme'/> >>    <feature policy='require' name='f16c'/> >>    <feature policy='require' name='rdrand'/> >>    <feature policy='require' name='hypervisor'/> >>    <feature policy='require' name='arat'/> >>    <feature policy='require' name='xsaveopt'/> >>    <feature policy='require' name='abm'/> >>    <numa> >>      <cell id='0' cpus='0-3' memory='16777216' unit='KiB'/> >>    </numa> >>  </cpu> >>  <clock offset='variable' adjustment='0' basis='utc'> >>    <timer name='rtc' tickpolicy='catchup'/> >>    <timer name='pit' tickpolicy='delay'/> >>    <timer name='hpet' present='no'/> >>  </clock> >>  <on_poweroff>destroy</on_poweroff> >>  <on_reboot>destroy</on_reboot> >>  <on_crash>destroy</on_crash> >>  <pm> >>    <suspend-to-mem enabled='no'/> >>    <suspend-to-disk enabled='no'/> >>  </pm> >>  <devices> >>    <emulator>/usr/libexec/qemu-kvm</emulator> >>    <disk type='file' device='cdrom'> >>      <driver name='qemu' error_policy='report'/> >>      <source startupPolicy='optional'/> >>      <target dev='hdc' bus='ide'/> >>      <readonly/> >>      <alias name='ua-80fde7d5-ee7f-4201-9118-11bc6c3b8530'/> >>      <address type='drive' controller='0' bus='1' target='0' unit='0'/> >>    </disk> >>    <disk type='block' device='disk' snapshot='no'> >>      <driver name='qemu' type='raw' cache='none' error_policy='stop' io='native' iothread='1'/> >>      <source dev='/var/run/vdsm/storage/c17c1934-332f-464c-8f89-ad72463c00b3/8eca143a-4535-4421-bd35-9f5764d67d70/ae961104-c3b3-4a43-9f46-7fa6bdc2ac33'> >>        <seclabel model='dac' relabel='no'/> >>      </source> >>      <backingStore/> >>      <target dev='vda' bus='virtio'/> >>      <serial>8eca143a-4535-4421-bd35-9f5764d67d70</serial> >>      <alias name='ua-8eca143a-4535-4421-bd35-9f5764d67d70'/> >>      <address type='pci' domain='0x0000' bus='0x00' slot='0x07' function='0x0'/> >>    </disk> >>    <controller type='scsi' index='0' model='virtio-scsi'> >>      <driver iothread='1'/> >>      <alias name='ua-27331e83-03f4-42a3-9554-c41649c02ba4'/> >>      <address type='pci' domain='0x0000' bus='0x00' slot='0x05' function='0x0'/> >>    </controller> >>    <controller type='ide' index='0'> >>      <alias name='ide'/> >>      <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x1'/> >>    </controller> >>    <controller type='virtio-serial' index='0' ports='16'> >>      <alias name='ua-8fe74299-b60f-4778-8e80-db05393a9489'/> >>      <address type='pci' domain='0x0000' bus='0x00' slot='0x06' function='0x0'/> >>    </controller> >>    <controller type='usb' index='0' model='piix3-uhci'> >>      <alias name='usb'/> >>      <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x2'/> >>    </controller> >>    <controller type='pci' index='0' model='pci-root'> >>      <alias name='pci.0'/> >>    </controller> >>    <lease> >>      <lockspace>c17c1934-332f-464c-8f89-ad72463c00b3</lockspace> >>      <key>ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</key> >>      <target path='/dev/c17c1934-332f-464c-8f89-ad72463c00b3/leases' offset='108003328'/> >>    </lease> >>    <interface type='bridge'> >>      <mac address='00:16:3e:50:c1:97'/> >>      <source bridge='ovirtmgmt'/> >>      <target dev='vnet0'/> >>      <model type='virtio'/> >>      <driver name='vhost' queues='4'/> >>      <filterref filter='vdsm-no-mac-spoofing'/> >>      <link state='up'/> >>      <mtu size='1500'/> >>      <alias name='ua-fada74ee-2338-4cde-a7ba-43a9a636ad6e'/> >>      <address type='pci' domain='0x0000' bus='0x00' slot='0x03' function='0x0'/> >>    </interface> >>    <interface type='bridge'> >>      <mac address='56:6f:9b:b0:00:01'/> >>      <source bridge='external'/> >>      <target dev='vnet1'/> >>      <model type='virtio'/> >>      <driver name='vhost' queues='4'/> >>      <filterref filter='vdsm-no-mac-spoofing'/> >>      <link state='up'/> >>      <mtu size='1500'/> >>      <alias name='ua-f7b4c949-1f9f-4355-811d-88428c88ce4e'/> >>      <address type='pci' domain='0x0000' bus='0x00' slot='0x0a' function='0x0'/> >>    </interface> >>    <serial type='unix'> >>      <source mode='bind' path='/var/run/ovirt-vmconsole-console/b4e805ff-556d-42bd-a6df-02f5902fd01c.sock'/> >>      <target type='isa-serial' port='0'> >>        <model name='isa-serial'/> >>      </target> >>      <alias name='serial0'/> >>    </serial> >>    <console type='unix'> >>      <source mode='bind' path='/var/run/ovirt-vmconsole-console/b4e805ff-556d-42bd-a6df-02f5902fd01c.sock'/> >>      <target type='serial' port='0'/> >>      <alias name='serial0'/> >>    </console> >>    <channel type='unix'> >>      <source mode='bind' path='/var/lib/libvirt/qemu/channels/b4e805ff-556d-42bd-a6df-02f5902fd01c.ovirt-guest-agent.0'/> >>      <target type='virtio' name='ovirt-guest-agent.0' state='connected'/> >>      <alias name='channel0'/> >>      <address type='virtio-serial' controller='0' bus='0' port='1'/> >>    </channel> >>    <channel type='unix'> >>      <source mode='bind' path='/var/lib/libvirt/qemu/channels/b4e805ff-556d-42bd-a6df-02f5902fd01c.org.qemu.guest_agent.0'/> >>      <target type='virtio' name='org.qemu.guest_agent.0' state='connected'/> >>      <alias name='channel1'/> >>      <address type='virtio-serial' controller='0' bus='0' port='2'/> >>    </channel> >>    <channel type='spicevmc'> >>      <target type='virtio' name='com.redhat.spice.0' state='disconnected'/> >>      <alias name='channel2'/> >>      <address type='virtio-serial' controller='0' bus='0' port='3'/> >>    </channel> >>    <channel type='unix'> >>      <source mode='bind' path='/var/lib/libvirt/qemu/channels/b4e805ff-556d-42bd-a6df-02f5902fd01c.org.ovirt.hosted-engine-setup.0'/> >>      <target type='virtio' name='org.ovirt.hosted-engine-setup.0' state='disconnected'/> >>      <alias name='channel3'/> >>      <address type='virtio-serial' controller='0' bus='0' port='4'/> >>    </channel> >>    <input type='tablet' bus='usb'> >>      <alias name='input0'/> >>      <address type='usb' bus='0' port='1'/> >>    </input> >>    <input type='mouse' bus='ps2'> >>      <alias name='input1'/> >>    </input> >>    <input type='keyboard' bus='ps2'> >>      <alias name='input2'/> >>    </input> >>    <graphics type='vnc' port='5900' autoport='yes' listen='192.168.1.22' keymap='en-us' passwdValidTo='1970-01-01T00:00:01'> >>      <listen type='network' address='192.168.1.22' network='vdsm-external'/> >>    </graphics> >>    <graphics type='spice' port='5901' tlsPort='5902' autoport='yes' listen='192.168.1.22' passwdValidTo='1970-01-01T00:00:01'> >>      <listen type='network' address='192.168.1.22' network='vdsm-external'/> >>      <channel name='main' mode='secure'/> >>      <channel name='display' mode='secure'/> >>      <channel name='inputs' mode='secure'/> >>      <channel name='cursor' mode='secure'/> >>      <channel name='playback' mode='secure'/> >>      <channel name='record' mode='secure'/> >>      <channel name='smartcard' mode='secure'/> >>      <channel name='usbredir' mode='secure'/> >>    </graphics> >>    <sound model='ich6'> >>      <alias name='ua-bd287767-9b83-4e44-ac6f-8b527f9632b8'/> >>      <address type='pci' domain='0x0000' bus='0x00' slot='0x04' function='0x0'/> >>    </sound> >>    <video> >>      <model type='qxl' ram='65536' vram='32768' vgamem='16384' heads='1' primary='yes'/> >>      <alias name='ua-bcfb6b4b-0b3c-4d5b-ba2d-8ce40a65facd'/> >>      <address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x0'/> >>    </video> >>    <memballoon model='virtio'> >>      <stats period='5'/> >>      <alias name='ua-39d36063-8808-47db-9fef-a0baad9f9661'/> >>      <address type='pci' domain='0x0000' bus='0x00' slot='0x08' function='0x0'/> >>    </memballoon> >>    <rng model='virtio'> >>      <backend model='random'>/dev/urandom</backend> >>      <alias name='ua-75516d34-dd8f-4f0f-8496-e1f222a359a8'/> >>      <address type='pci' domain='0x0000' bus='0x00' slot='0x09' function='0x0'/> >>    </rng> >>  </devices> >>  <seclabel type='dynamic' model='selinux' relabel='yes'> >>    <label>system_u:system_r:svirt_t:s0:c162,c716</label> >>    <imagelabel>system_u:object_r:svirt_image_t:s0:c162,c716</imagelabel> >>  </seclabel> >>  <seclabel type='dynamic' model='dac' relabel='yes'> >>    <label>+107:+107</label> >>    <imagelabel>+107:+107</imagelabel> >>  </seclabel> >></domain> >> >>hosts: >> >><capabilities> >>  <host> >>    <uuid>b25adcec-eef2-49a5-8663-7cdcfc50891b</uuid> >>    <cpu> >>      <arch>x86_64</arch> >>      <model>Westmere</model> >>      <vendor>Intel</vendor> >>      <microcode version='34'/> >>      <counter name='tsc' frequency='3699996000' scaling='no'/> >>      <topology sockets='1' cores='2' threads='2'/> >>      <feature name='vme'/> >>      <feature name='ds'/> >>      <feature name='acpi'/> >>      <feature name='ss'/> >>      <feature name='ht'/> >>      <feature name='tm'/> >>      <feature name='pbe'/> >>      <feature name='pclmuldq'/> >>      <feature name='dtes64'/> >>      <feature name='monitor'/> >>      <feature name='ds_cpl'/> >>      <feature name='vmx'/> >>      <feature name='est'/> >>      <feature name='tm2'/> >>      <feature name='fma'/> >>      <feature name='xtpr'/> >>      <feature name='pdcm'/> >>      <feature name='pcid'/> >>      <feature name='movbe'/> >>      <feature name='tsc-deadline'/> >>      <feature name='xsave'/> >>      <feature name='osxsave'/> >>      <feature name='avx'/> >>      <feature name='f16c'/> >>      <feature name='rdrand'/> >>      <feature name='arat'/> >>      <feature name='fsgsbase'/> >>      <feature name='tsc_adjust'/> >>      <feature name='bmi1'/> >>      <feature name='avx2'/> >>      <feature name='smep'/> >>      <feature name='bmi2'/> >>      <feature name='erms'/> >>      <feature name='invpcid'/> >>      <feature name='xsaveopt'/> >>      <feature name='pdpe1gb'/> >>      <feature name='rdtscp'/> >>      <feature name='abm'/> >>      <feature name='invtsc'/> >>      <pages unit='KiB' size='4'/> >>      <pages unit='KiB' size='2048'/> >>      <pages unit='KiB' size='1048576'/> >>    </cpu> >>    <power_management> >>      <suspend_mem/> >>      <suspend_disk/> >>      <suspend_hybrid/> >>    </power_management> >>    <iommu support='no'/> >>    <migration_features> >>      <live/> >>      <uri_transports> >>        <uri_transport>tcp</uri_transport> >>        <uri_transport>rdma</uri_transport> >>      </uri_transports> >>    </migration_features> >>    <topology> >>      <cells num='1'> >>        <cell id='0'> >>          <memory unit='KiB'>32903488</memory> >>          <pages unit='KiB' size='4'>8225872</pages> >>          <pages unit='KiB' size='2048'>0</pages> >>          <pages unit='KiB' size='1048576'>0</pages> >>          <distances> >>            <sibling id='0' value='10'/> >>          </distances> >>          <cpus num='4'> >>            <cpu id='0' socket_id='0' core_id='0' siblings='0,2'/> >>            <cpu id='1' socket_id='0' core_id='1' siblings='1,3'/> >>            <cpu id='2' socket_id='0' core_id='0' siblings='0,2'/> >>            <cpu id='3' socket_id='0' core_id='1' siblings='1,3'/> >>          </cpus> >>        </cell> >>      </cells> >>    </topology> >>    <cache> >>      <bank id='0' level='3' type='both' size='3' unit='MiB' cpus='0-3'/> >>    </cache> >>    <secmodel> >>      <model>selinux</model> >>      <doi>0</doi> >>      <baselabel type='kvm'>system_u:system_r:svirt_t:s0</baselabel> >>      <baselabel type='qemu'>system_u:system_r:svirt_tcg_t:s0</baselabel> >>    </secmodel> >>    <secmodel> >>      <model>dac</model> >>      <doi>0</doi> >>      <baselabel type='kvm'>+107:+107</baselabel> >>      <baselabel type='qemu'>+107:+107</baselabel> >>    </secmodel> >>  </host> >>  <guest> >>    <os_type>hvm</os_type> >>    <arch name='i686'> >>      <wordsize>32</wordsize> >>      <emulator>/usr/libexec/qemu-kvm</emulator> >>      <machine maxCpus='240'>pc-i440fx-rhel7.6.0</machine> >>      <machine canonical='pc-i440fx-rhel7.6.0' maxCpus='240'>pc</machine> >>      <machine maxCpus='240'>pc-i440fx-rhel7.0.0</machine> >>      <machine maxCpus='384'>pc-q35-rhel7.6.0</machine> >>      <machine canonical='pc-q35-rhel7.6.0' maxCpus='384'>q35</machine> >>      <machine maxCpus='240'>rhel6.3.0</machine> >>      <machine maxCpus='240'>rhel6.4.0</machine> >>      <machine maxCpus='240'>rhel6.0.0</machine> >>      <machine maxCpus='240'>pc-i440fx-rhel7.5.0</machine> >>      <machine maxCpus='240'>pc-i440fx-rhel7.1.0</machine> >>      <machine maxCpus='240'>pc-i440fx-rhel7.2.0</machine> >>      <machine maxCpus='255'>pc-q35-rhel7.3.0</machine> >>      <machine maxCpus='240'>rhel6.5.0</machine> >>      <machine maxCpus='384'>pc-q35-rhel7.4.0</machine> >>      <machine maxCpus='240'>rhel6.6.0</machine> >>      <machine maxCpus='240'>rhel6.1.0</machine> >>      <machine maxCpus='240'>rhel6.2.0</machine> >>      <machine maxCpus='240'>pc-i440fx-rhel7.3.0</machine> >>      <machine maxCpus='240'>pc-i440fx-rhel7.4.0</machine> >>      <machine maxCpus='384'>pc-q35-rhel7.5.0</machine> >>      <domain type='qemu'/> >>      <domain type='kvm'> >>        <emulator>/usr/libexec/qemu-kvm</emulator> >>      </domain> >>    </arch> >>    <features> >>      <cpuselection/> >>      <deviceboot/> >>      <disksnapshot default='on' toggle='no'/> >>      <acpi default='on' toggle='yes'/> >>      <apic default='on' toggle='no'/> >>      <pae/> >>      <nonpae/> >>    </features> >>  </guest> >>  <guest> >>    <os_type>hvm</os_type> >>    <arch name='x86_64'> >>      <wordsize>64</wordsize> >>      <emulator>/usr/libexec/qemu-kvm</emulator> >>      <machine maxCpus='240'>pc-i440fx-rhel7.6.0</machine> >>      <machine canonical='pc-i440fx-rhel7.6.0' maxCpus='240'>pc</machine> >>      <machine maxCpus='240'>pc-i440fx-rhel7.0.0</machine> >>      <machine maxCpus='384'>pc-q35-rhel7.6.0</machine> >>      <machine canonical='pc-q35-rhel7.6.0' maxCpus='384'>q35</machine> >>      <machine maxCpus='240'>rhel6.3.0</machine> >>      <machine maxCpus='240'>rhel6.4.0</machine> >>      <machine maxCpus='240'>rhel6.0.0</machine> >>      <machine maxCpus='240'>pc-i440fx-rhel7.5.0</machine> >>      <machine maxCpus='240'>pc-i440fx-rhel7.1.0</machine> >>      <machine maxCpus='240'>pc-i440fx-rhel7.2.0</machine> >>      <machine maxCpus='255'>pc-q35-rhel7.3.0</machine> >>      <machine maxCpus='240'>rhel6.5.0</machine> >>      <machine maxCpus='384'>pc-q35-rhel7.4.0</machine> >>      <machine maxCpus='240'>rhel6.6.0</machine> >>      <machine maxCpus='240'>rhel6.1.0</machine> >>      <machine maxCpus='240'>rhel6.2.0</machine> >>      <machine maxCpus='240'>pc-i440fx-rhel7.3.0</machine> >>      <machine maxCpus='240'>pc-i440fx-rhel7.4.0</machine> >>      <machine maxCpus='384'>pc-q35-rhel7.5.0</machine> >>      <domain type='qemu'/> >>      <domain type='kvm'> >>        <emulator>/usr/libexec/qemu-kvm</emulator> >>      </domain> >>    </arch> >>    <features> >>      <cpuselection/> >>      <deviceboot/> >>      <disksnapshot default='on' toggle='no'/> >>      <acpi default='on' toggle='yes'/> >>      <apic default='on' toggle='no'/> >>    </features> >>  </guest> >></capabilities> >> >> >>在 2020-09-17 12:00:19,"Strahil Nikolov" <hunter86_bg@yahoo.com> 写道: >>>It would be easier if you posted the whole xml. >>> >>>What about the sections (in HE xml) starting with: >>>feature policy= >>> >>>Also the hosts have a section which contains: >>> >>><feature name= >>> >>>If you can share a VM's xml sections for a good VM. >>> >>>Best Regards, >>>Strahil Nikolov >>> >>> >>> >>> >>> >>> >>>В четвъртък, 17 септември 2020 г., 05:54:12 Гринуич+3, ddqlo <ddqlo@126.com> написа: >>> >>> >>> >>> >>> >>>HostedEngine: >>>...... >>><model fallback='forbid'>Haswell-noTSX</model> >>>...... >>> >>>both of the hosts: >>>...... >>><model>Westmere</model> >>>...... >>> >>>others vms which can be migrated: >>>...... >>><model fallback='forbid'>Haswell-noTSX</model> >>>...... >>> >>> >>> >>>在 2020-09-17 03:03:24,"Strahil Nikolov" <hunter86_bg@yahoo.com> 写道: >>>>Can you verify the HostedEngine's CPU ? >>>> >>>>1. ssh to the host hosting the HE >>>>2. alias virsh='virsh -c qemu:///system?authfile=/etc/ovirt-hosted-engine/virsh_auth.conf' >>>>3. virsh dumpxml HostedEngine >>>> >>>> >>>>Then set the alias for virsh on all Hosts and 'virsh capabilites' should show the Hosts' <cpu><model> . >>>> >>>>Best Regards, >>>>Strahil Nikolov >>>> >>>> >>>> >>>> >>>> >>>> >>>>В сряда, 16 септември 2020 г., 10:16:08 Гринуич+3, ddqlo <ddqlo@126.com> написа: >>>> >>>> >>>> >>>> >>>> >>>>My gateway was not pingable. I have fixed this problem and now both nodes have a score(3400). >>>>Yet, hosted engine could not be migrated. Same log in engine.log: >>>>host filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'CPU' >>>> >>>> >>>>在 2020-09-16 02:11:09,"Strahil Nikolov" <hunter86_bg@yahoo.com> 写道: >>>>>Both nodes have a lower than the usual score (should be 3400 ). >>>>>Based on the score you are probably suffering from gateway-score-penalty [1][2]. >>>>>Check if your gateway is pingable. >>>>> >>>>>Best Regards, >>>>>Strahil Nikolov >>>>> >>>>>1 - https://www.ovirt.org/images/Hosted-Engine-4.3-deep-dive.pdf(page 8) >>>>>2 - /etc/ovirt-hosted-engine-ha/agent.conf  >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>>В вторник, 15 септември 2020 г., 04:49:48 Гринуич+3, ddqlo <ddqlo@126.com> написа: >>>>> >>>>> >>>>> >>>>> >>>>> >>>>>--== Host node28 (id: 1) status ==-- >>>>> >>>>>conf_on_shared_storage             : True >>>>>Status up-to-date                  : True >>>>>Hostname                           : node28 >>>>>Host ID                            : 1 >>>>>Engine status                      : {"reason": "vm not running on this host", "health": "bad", "vm": "down_unexpected", "detail": "unknown"} >>>>>Score                              : 1800 >>>>>stopped                            : False >>>>>Local maintenance                  : False >>>>>crc32                              : 4ac6105b >>>>>local_conf_timestamp               : 1794597 >>>>>Host timestamp                     : 1794597 >>>>>Extra metadata (valid at timestamp): >>>>>        metadata_parse_version=1 >>>>>        metadata_feature_version=1 >>>>>        timestamp=1794597 (Tue Sep 15 09:47:17 2020) >>>>>        host-id=1 >>>>>        score=1800 >>>>>        vm_conf_refresh_time=1794597 (Tue Sep 15 09:47:17 2020) >>>>>        conf_on_shared_storage=True >>>>>        maintenance=False >>>>>        state=EngineDown >>>>>        stopped=False >>>>> >>>>> >>>>>--== Host node22 (id: 2) status ==-- >>>>> >>>>>conf_on_shared_storage             : True >>>>>Status up-to-date                  : True >>>>>Hostname                           : node22 >>>>>Host ID                            : 2 >>>>>Engine status                      : {"health": "good", "vm": "up", "detail": "Up"} >>>>>Score                              : 1800 >>>>>stopped                            : False >>>>>Local maintenance                  : False >>>>>crc32                              : ffc41893 >>>>>local_conf_timestamp               : 1877876 >>>>>Host timestamp                     : 1877876 >>>>>Extra metadata (valid at timestamp): >>>>>        metadata_parse_version=1 >>>>>        metadata_feature_version=1 >>>>>        timestamp=1877876 (Tue Sep 15 09:47:13 2020) >>>>>        host-id=2 >>>>>        score=1800 >>>>>        vm_conf_refresh_time=1877876 (Tue Sep 15 09:47:13 2020) >>>>>        conf_on_shared_storage=True >>>>>        maintenance=False >>>>>        state=EngineUp >>>>>        stopped=False >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>>在 2020-09-09 01:32:55,"Strahil Nikolov" <hunter86_bg@yahoo.com> 写道: >>>>>>What is the output of 'hosted-engine --vm-status' on the node where the HostedEngine is running ? >>>>>> >>>>>> >>>>>>Best Regards, >>>>>>Strahil Nikolov >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>>В понеделник, 7 септември 2020 г., 03:53:13 Гринуич+3, ddqlo <ddqlo@126.com> написа: >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>>I could not find any logs because the migration button is disabled in the web UI. It seems that the engine migration operation is prevented at first. Any other ideas? Thanks! >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>>在 2020-09-01 00:06:19,"Strahil Nikolov" <hunter86_bg@yahoo.com> 写道: >>>>>>>I'm running oVirt 4.3.10 and I can migrate my Engine from node to node. >>>>>>>I had one similar issue , but powering off and on the HE has fixed it. >>>>>>> >>>>>>>You have to check the vdsm log on the source and on destination in order to figure out what is going on. >>>>>>>Also you might consider checking the libvirt logs on the destination. >>>>>>> >>>>>>>Best Regards, >>>>>>>Strahil Nikolov >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>>В понеделник, 31 август 2020 г., 10:47:22 Гринуич+3, ddqlo <ddqlo@126.com> написа: >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>>Thanks! The scores of all nodes are not '0'. I find that someone has already asked a question like this. It seems that  this feature has been disabled in 4.3. I am not sure if it is enabled in 4.4. >>>>>>> >>>>>>> >>>>>>>在 2020-08-29 02:27:03,"Strahil Nikolov" <hunter86_bg@yahoo.com> : >>>>>>>>Have you checked under a shell the output of 'hosted-engine --vm-status' . Check the Score of the hosts. Maybe there is a node with score of '0' ? >>>>>>>> >>>>>>>>Best Regards, >>>>>>>>Strahil Nikolov >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>>В вторник, 25 август 2020 г., 13:46:18 Гринуич+3, 董青龙 <ddqlo@126.com> написа: >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>>Hi all, >>>>>>>>        I have an ovirt4.3.10.4 environment of 2 hosts. Normal vms in this environment could be migrated, but the hosted engine vm could not be migrated. Anyone can help? Thanks a lot! >>>>>>>> >>>>>>>>hosts status: >>>>>>>> >>>>>>>>normal vm migration: >>>>>>>> >>>>>>>>hosted engine vm migration: >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>>  >>>>>>>>_______________________________________________ >>>>>>>>Users mailing list -- users@ovirt.org >>>>>>>>To unsubscribe send an email to users-leave@ovirt.org >>>>>>>>Privacy Statement: https://www.ovirt.org/privacy-policy.html >>>>>>>>oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ >>>>>>>>List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZXHE2AJX42HNHOMYHTDCUUIU3VQTQTLF/ >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>>  >>>>>>>_______________________________________________ >>>>>>>Users mailing list -- users@ovirt.org >>>>>>>To unsubscribe send an email to users-leave@ovirt.org >>>>>>>Privacy Statement: https://www.ovirt.org/privacy-policy.html >>>>>>>oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ >>>>>>>List Archives: >>>>>>>https://lists.ovirt.org/archives/list/users@ovirt.org/message/IAYLFLC6K42OUPZSZU3P3ZYAU66LGSCD/ >>>>>> >>>>>> >>>>>> >>>>>> >>>>>>  >>>>>>_______________________________________________ >>>>>>Users mailing list -- users@ovirt.org >>>>>>To unsubscribe send an email to users-leave@ovirt.org >>>>>>Privacy Statement: https://www.ovirt.org/privacy-policy.html >>>>>>oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ >>>>>>List Archives: >>>>>>https://lists.ovirt.org/archives/list/users@ovirt.org/message/23ZMAP5K7N5KKX3HOKUEJNVCBH7CY4ZL/ >>>>> >>>>> >>>>> >>>>> >>>>>  >>>> >>>> >>>> >>>> >>>>  >>> >>> >>> >>> >>>  >> >> >> >> >>  > >_______________________________________________ >Users mailing list -- users@ovirt.org >To unsubscribe send an email to users-leave@ovirt.org >Privacy Statement: https://www.ovirt.org/privacy-policy.html >oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ >List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/BM3QAMWVBKUATSMHAOEKNJTXLNKLNDBM/