在 2020-09-22 03:08:18,"Strahil Nikolov" <hunter86_bg@yahoo.com> 写道: >Can you put 1 host in maintenance and use the "Installation" -> "Reinstall" and enable the HE deployment from one of the tabs ? > >Best Regards, >Strahil Nikolov > > > > > > >В понеделник, 21 септември 2020 г., 06:38:06 Гринуич+3, ddqlo <ddqlo@126.com> написа: > > > > > >so strange! After I set global maintenance, powered off and started H The cpu of HE became 'Westmere'(did not change anything). But HE still could not be migrated. > >HE xml: > <cpu mode='custom' match='exact' check='full'> > <model fallback='forbid'>Westmere</model> > <topology sockets='16' cores='4' threads='1'/> > <feature policy='require' name='vme'/> > <feature policy='require' name='pclmuldq'/> > <feature policy='require' name='x2apic'/> > <feature policy='require' name='hypervisor'/> > <feature policy='require' name='arat'/> > <numa> > <cell id='0' cpus='0-3' memory='16777216' unit='KiB'/> > </numa> > </cpu> > >host capabilities: ><model>Westmere</model> > >cluster cpu type (UI): > > >host cpu type (UI): > > >HE cpu type (UI): > > > > > > > >在 2020-09-19 13:27:35,"Strahil Nikolov" <hunter86_bg@yahoo.com> 写道: >>Hm... interesting. >> >>The VM is using 'Haswell-noTSX' while the host is 'Westmere'. >> >>In my case I got no difference: >> >>[root@ovirt1 ~]# virsh dumpxml HostedEngine | grep Opteron >> <model fallback='forbid'>Opteron_G5</model> >>[root@ovirt1 ~]# virsh capabilities | grep Opteron >> <model>Opteron_G5</model> >> >>Did you update the cluster holding the Hosted Engine ? >> >> >>I guess you can try to: >> >>- Set global maintenance >>- Power off the HostedEngine VM >>- virsh dumpxml HostedEngine > /root/HE.xml >>- use virsh edit to change the cpu of the HE (non-permanent) change >>- try to power on the modified HE >> >>If it powers on , you can try to migrate it and if it succeeds - then you should make it permanent. >> >> >> >> >> >>Best Regards, >>Strahil Nikolov >> >>В петък, 18 септември 2020 г., 04:40:39 Гринуич+3, ddqlo <ddqlo@126.com> написа: >> >> >> >> >> >>HE: >> >><domain type='kvm' id='1'> >> <name>HostedEngine</name> >> <uuid>b4e805ff-556d-42bd-a6df-02f5902fd01c</uuid> >> <metadata xmlns:ns0="http://ovirt.org/vm/tune/1.0" xmlns:ovirt-vm="http://ovirt.org/vm/1.0"> >> <ns0:qos/> >> <ovirt-vm:vm xmlns:ovirt-vm="http://ovirt.org/vm/1.0"> >> <ovirt-vm:clusterVersion>4.3</ovirt-vm:clusterVersion> >> <ovirt-vm:destroy_on_reboot type="bool">False</ovirt-vm:destroy_on_reboot> >> <ovirt-vm:launchPaused>false</ovirt-vm:launchPaused> >> <ovirt-vm:memGuaranteedSize type="int">1024</ovirt-vm:memGuaranteedSize> >> <ovirt-vm:minGuaranteedMemoryMb type="int">1024</ovirt-vm:minGuaranteedMemoryMb> >> <ovirt-vm:resumeBehavior>auto_resume</ovirt-vm:resumeBehavior> >> <ovirt-vm:startTime type="float">1600307555.19</ovirt-vm:startTime> >> <ovirt-vm:device mac_address="56:6f:9b:b0:00:01"> >> <ovirt-vm:network>external</ovirt-vm:network> >> <ovirt-vm:custom> >> <ovirt-vm:queues>4</ovirt-vm:queues> >> </ovirt-vm:custom> >> </ovirt-vm:device> >> <ovirt-vm:device mac_address="00:16:3e:50:c1:97"> >> <ovirt-vm:network>ovirtmgmt</ovirt-vm:network> >> <ovirt-vm:custom> >> <ovirt-vm:queues>4</ovirt-vm:queues> >> </ovirt-vm:custom> >> </ovirt-vm:device> >> <ovirt-vm:device devtype="disk" name="vda"> >> <ovirt-vm:domainID>c17c1934-332f-464c-8f89-ad72463c00b3</ovirt-vm:domainID> >> <ovirt-vm:guestName>/dev/vda2</ovirt-vm:guestName> >> <ovirt-vm:imageID>8eca143a-4535-4421-bd35-9f5764d67d70</ovirt-vm:imageID> >> <ovirt-vm:poolID>00000000-0000-0000-0000-000000000000</ovirt-vm:poolID> >> <ovirt-vm:shared>exclusive</ovirt-vm:shared> >> <ovirt-vm:volumeID>ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</ovirt-vm:volumeID> >> <ovirt-vm:specParams> >> <ovirt-vm:pinToIoThread>1</ovirt-vm:pinToIoThread> >> </ovirt-vm:specParams> >> <ovirt-vm:volumeChain> >> <ovirt-vm:volumeChainNode> >> <ovirt-vm:domainID>c17c1934-332f-464c-8f89-ad72463c00b3</ovirt-vm:domainID> >> <ovirt-vm:imageID>8eca143a-4535-4421-bd35-9f5764d67d70</ovirt-vm:imageID> >> <ovirt-vm:leaseOffset type="int">108003328</ovirt-vm:leaseOffset> >> <ovirt-vm:leasePath>/dev/c17c1934-332f-464c-8f89-ad72463c00b3/leases</ovirt-vm:leasePath> >> <ovirt-vm:path>/rhev/data-center/mnt/blockSD/c17c1934-332f-464c-8f89-ad72463c00b3/images/8eca143a-4535-4421-bd35-9f5764d67d70/ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</ovirt-vm:path> >> <ovirt-vm:volumeID>ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</ovirt-vm:volumeID> >> </ovirt-vm:volumeChainNode> >> </ovirt-vm:volumeChain> >> </ovirt-vm:device> >> <ovirt-vm:device devtype="disk" name="hdc"/> >></ovirt-vm:vm> >> </metadata> >> <maxMemory slots='16' unit='KiB'>67108864</maxMemory> >> <memory unit='KiB'>16777216</memory> >> <currentMemory unit='KiB'>16777216</currentMemory> >> <vcpu placement='static' current='4'>64</vcpu> >> <iothreads>1</iothreads> >> <resource> >> <partition>/machine</partition> >> </resource> >> <sysinfo type='smbios'> >> <system> >> <entry name='manufacturer'>oVirt</entry> >> <entry name='product'>oVirt Node</entry> >> <entry name='version'>7-5.1804.el7.centos</entry> >> <entry name='serial'>00000000-0000-0000-0000-0CC47A6B3160</entry> >> <entry name='uuid'>b4e805ff-556d-42bd-a6df-02f5902fd01c</entry> >> </system> >> </sysinfo> >> <os> >> <type arch='x86_64' machine='pc-i440fx-rhel7.6.0'>hvm</type> >> <boot dev='hd'/> >> <bios useserial='yes'/> >> <smbios mode='sysinfo'/> >> </os> >> <features> >> <acpi/> >> </features> >> <cpu mode='custom' match='exact' check='full'> >> <model fallback='forbid'>Haswell-noTSX</model> >> <topology sockets='16' cores='4' threads='1'/> >> <feature policy='require' name='vme'/> >> <feature policy='require' name='f16c'/> >> <feature policy='require' name='rdrand'/> >> <feature policy='require' name='hypervisor'/> >> <feature policy='require' name='arat'/> >> <feature policy='require' name='xsaveopt'/> >> <feature policy='require' name='abm'/> >> <numa> >> <cell id='0' cpus='0-3' memory='16777216' unit='KiB'/> >> </numa> >> </cpu> >> <clock offset='variable' adjustment='0' basis='utc'> >> <timer name='rtc' tickpolicy='catchup'/> >> <timer name='pit' tickpolicy='delay'/> >> <timer name='hpet' present='no'/> >> </clock> >> <on_poweroff>destroy</on_poweroff> >> <on_reboot>destroy</on_reboot> >> <on_crash>destroy</on_crash> >> <pm> >> <suspend-to-mem enabled='no'/> >> <suspend-to-disk enabled='no'/> >> </pm> >> <devices> >> <emulator>/usr/libexec/qemu-kvm</emulator> >> <disk type='file' device='cdrom'> >> <driver name='qemu' error_policy='report'/> >> <source startupPolicy='optional'/> >> <target dev='hdc' bus='ide'/> >> <readonly/> >> <alias name='ua-80fde7d5-ee7f-4201-9118-11bc6c3b8530'/> >> <address type='drive' controller='0' bus='1' target='0' unit='0'/> >> </disk> >> <disk type='block' device='disk' snapshot='no'> >> <driver name='qemu' type='raw' cache='none' error_policy='stop' io='native' iothread='1'/> >> <source dev='/var/run/vdsm/storage/c17c1934-332f-464c-8f89-ad72463c00b3/8eca143a-4535-4421-bd35-9f5764d67d70/ae961104-c3b3-4a43-9f46-7fa6bdc2ac33'> >> <seclabel model='dac' relabel='no'/> >> </source> >> <backingStore/> >> <target dev='vda' bus='virtio'/> >> <serial>8eca143a-4535-4421-bd35-9f5764d67d70</serial> >> <alias name='ua-8eca143a-4535-4421-bd35-9f5764d67d70'/> >> <address type='pci' domain='0x0000' bus='0x00' slot='0x07' function='0x0'/> >> </disk> >> <controller type='scsi' index='0' model='virtio-scsi'> >> <driver iothread='1'/> >> <alias name='ua-27331e83-03f4-42a3-9554-c41649c02ba4'/> >> <address type='pci' domain='0x0000' bus='0x00' slot='0x05' function='0x0'/> >> </controller> >> <controller type='ide' index='0'> >> <alias name='ide'/> >> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x1'/> >> </controller> >> <controller type='virtio-serial' index='0' ports='16'> >> <alias name='ua-8fe74299-b60f-4778-8e80-db05393a9489'/> >> <address type='pci' domain='0x0000' bus='0x00' slot='0x06' function='0x0'/> >> </controller> >> <controller type='usb' index='0' model='piix3-uhci'> >> <alias name='usb'/> >> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x2'/> >> </controller> >> <controller type='pci' index='0' model='pci-root'> >> <alias name='pci.0'/> >> </controller> >> <lease> >> <lockspace>c17c1934-332f-464c-8f89-ad72463c00b3</lockspace> >> <key>ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</key> >> <target path='/dev/c17c1934-332f-464c-8f89-ad72463c00b3/leases' offset='108003328'/> >> </lease> >> <interface type='bridge'> >> <mac address='00:16:3e:50:c1:97'/> >> <source bridge='ovirtmgmt'/> >> <target dev='vnet0'/> >> <model type='virtio'/> >> <driver name='vhost' queues='4'/> >> <filterref filter='vdsm-no-mac-spoofing'/> >> <link state='up'/> >> <mtu size='1500'/> >> <alias name='ua-fada74ee-2338-4cde-a7ba-43a9a636ad6e'/> >> <address type='pci' domain='0x0000' bus='0x00' slot='0x03' function='0x0'/> >> </interface> >> <interface type='bridge'> >> <mac address='56:6f:9b:b0:00:01'/> >> <source bridge='external'/> >> <target dev='vnet1'/> >> <model type='virtio'/> >> <driver name='vhost' queues='4'/> >> <filterref filter='vdsm-no-mac-spoofing'/> >> <link state='up'/> >> <mtu size='1500'/> >> <alias name='ua-f7b4c949-1f9f-4355-811d-88428c88ce4e'/> >> <address type='pci' domain='0x0000' bus='0x00' slot='0x0a' function='0x0'/> >> </interface> >> <serial type='unix'> >> <source mode='bind' path='/var/run/ovirt-vmconsole-console/b4e805ff-556d-42bd-a6df-02f5902fd01c.sock'/> >> <target type='isa-serial' port='0'> >> <model name='isa-serial'/> >> </target> >> <alias name='serial0'/> >> </serial> >> <console type='unix'> >> <source mode='bind' path='/var/run/ovirt-vmconsole-console/b4e805ff-556d-42bd-a6df-02f5902fd01c.sock'/> >> <target type='serial' port='0'/> >> <alias name='serial0'/> >> </console> >> <channel type='unix'> >> <source mode='bind' path='/var/lib/libvirt/qemu/channels/b4e805ff-556d-42bd-a6df-02f5902fd01c.ovirt-guest-agent.0'/> >> <target type='virtio' name='ovirt-guest-agent.0' state='connected'/> >> <alias name='channel0'/> >> <address type='virtio-serial' controller='0' bus='0' port='1'/> >> </channel> >> <channel type='unix'> >> <source mode='bind' path='/var/lib/libvirt/qemu/channels/b4e805ff-556d-42bd-a6df-02f5902fd01c.org.qemu.guest_agent.0'/> >> <target type='virtio' name='org.qemu.guest_agent.0' state='connected'/> >> <alias name='channel1'/> >> <address type='virtio-serial' controller='0' bus='0' port='2'/> >> </channel> >> <channel type='spicevmc'> >> <target type='virtio' name='com.redhat.spice.0' state='disconnected'/> >> <alias name='channel2'/> >> <address type='virtio-serial' controller='0' bus='0' port='3'/> >> </channel> >> <channel type='unix'> >> <source mode='bind' path='/var/lib/libvirt/qemu/channels/b4e805ff-556d-42bd-a6df-02f5902fd01c.org.ovirt.hosted-engine-setup.0'/> >> <target type='virtio' name='org.ovirt.hosted-engine-setup.0' state='disconnected'/> >> <alias name='channel3'/> >> <address type='virtio-serial' controller='0' bus='0' port='4'/> >> </channel> >> <input type='tablet' bus='usb'> >> <alias name='input0'/> >> <address type='usb' bus='0' port='1'/> >> </input> >> <input type='mouse' bus='ps2'> >> <alias name='input1'/> >> </input> >> <input type='keyboard' bus='ps2'> >> <alias name='input2'/> >> </input> >> <graphics type='vnc' port='5900' autoport='yes' listen='192.168.1.22' keymap='en-us' passwdValidTo='1970-01-01T00:00:01'> >> <listen type='network' address='192.168.1.22' network='vdsm-external'/> >> </graphics> >> <graphics type='spice' port='5901' tlsPort='5902' autoport='yes' listen='192.168.1.22' passwdValidTo='1970-01-01T00:00:01'> >> <listen type='network' address='192.168.1.22' network='vdsm-external'/> >> <channel name='main' mode='secure'/> >> <channel name='display' mode='secure'/> >> <channel name='inputs' mode='secure'/> >> <channel name='cursor' mode='secure'/> >> <channel name='playback' mode='secure'/> >> <channel name='record' mode='secure'/> >> <channel name='smartcard' mode='secure'/> >> <channel name='usbredir' mode='secure'/> >> </graphics> >> <sound model='ich6'> >> <alias name='ua-bd287767-9b83-4e44-ac6f-8b527f9632b8'/> >> <address type='pci' domain='0x0000' bus='0x00' slot='0x04' function='0x0'/> >> </sound> >> <video> >> <model type='qxl' ram='65536' vram='32768' vgamem='16384' heads='1' primary='yes'/> >> <alias name='ua-bcfb6b4b-0b3c-4d5b-ba2d-8ce40a65facd'/> >> <address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x0'/> >> </video> >> <memballoon model='virtio'> >> <stats period='5'/> >> <alias name='ua-39d36063-8808-47db-9fef-a0baad9f9661'/> >> <address type='pci' domain='0x0000' bus='0x00' slot='0x08' function='0x0'/> >> </memballoon> >> <rng model='virtio'> >> <backend model='random'>/dev/urandom</backend> >> <alias name='ua-75516d34-dd8f-4f0f-8496-e1f222a359a8'/> >> <address type='pci' domain='0x0000' bus='0x00' slot='0x09' function='0x0'/> >> </rng> >> </devices> >> <seclabel type='dynamic' model='selinux' relabel='yes'> >> <label>system_u:system_r:svirt_t:s0:c162,c716</label> >> <imagelabel>system_u:object_r:svirt_image_t:s0:c162,c716</imagelabel> >> </seclabel> >> <seclabel type='dynamic' model='dac' relabel='yes'> >> <label>+107:+107</label> >> <imagelabel>+107:+107</imagelabel> >> </seclabel> >></domain> >> >>hosts: >> >><capabilities> >> <host> >> <uuid>b25adcec-eef2-49a5-8663-7cdcfc50891b</uuid> >> <cpu> >> <arch>x86_64</arch> >> <model>Westmere</model> >> <vendor>Intel</vendor> >> <microcode version='34'/> >> <counter name='tsc' frequency='3699996000' scaling='no'/> >> <topology sockets='1' cores='2' threads='2'/> >> <feature name='vme'/> >> <feature name='ds'/> >> <feature name='acpi'/> >> <feature name='ss'/> >> <feature name='ht'/> >> <feature name='tm'/> >> <feature name='pbe'/> >> <feature name='pclmuldq'/> >> <feature name='dtes64'/> >> <feature name='monitor'/> >> <feature name='ds_cpl'/> >> <feature name='vmx'/> >> <feature name='est'/> >> <feature name='tm2'/> >> <feature name='fma'/> >> <feature name='xtpr'/> >> <feature name='pdcm'/> >> <feature name='pcid'/> >> <feature name='movbe'/> >> <feature name='tsc-deadline'/> >> <feature name='xsave'/> >> <feature name='osxsave'/> >> <feature name='avx'/> >> <feature name='f16c'/> >> <feature name='rdrand'/> >> <feature name='arat'/> >> <feature name='fsgsbase'/> >> <feature name='tsc_adjust'/> >> <feature name='bmi1'/> >> <feature name='avx2'/> >> <feature name='smep'/> >> <feature name='bmi2'/> >> <feature name='erms'/> >> <feature name='invpcid'/> >> <feature name='xsaveopt'/> >> <feature name='pdpe1gb'/> >> <feature name='rdtscp'/> >> <feature name='abm'/> >> <feature name='invtsc'/> >> <pages unit='KiB' size='4'/> >> <pages unit='KiB' size='2048'/> >> <pages unit='KiB' size='1048576'/> >> </cpu> >> <power_management> >> <suspend_mem/> >> <suspend_disk/> >> <suspend_hybrid/> >> </power_management> >> <iommu support='no'/> >> <migration_features> >> <live/> >> <uri_transports> >> <uri_transport>tcp</uri_transport> >> <uri_transport>rdma</uri_transport> >> </uri_transports> >> </migration_features> >> <topology> >> <cells num='1'> >> <cell id='0'> >> <memory unit='KiB'>32903488</memory> >> <pages unit='KiB' size='4'>8225872</pages> >> <pages unit='KiB' size='2048'>0</pages> >> <pages unit='KiB' size='1048576'>0</pages> >> <distances> >> <sibling id='0' value='10'/> >> </distances> >> <cpus num='4'> >> <cpu id='0' socket_id='0' core_id='0' siblings='0,2'/> >> <cpu id='1' socket_id='0' core_id='1' siblings='1,3'/> >> <cpu id='2' socket_id='0' core_id='0' siblings='0,2'/> >> <cpu id='3' socket_id='0' core_id='1' siblings='1,3'/> >> </cpus> >> </cell> >> </cells> >> </topology> >> <cache> >> <bank id='0' level='3' type='both' size='3' unit='MiB' cpus='0-3'/> >> </cache> >> <secmodel> >> <model>selinux</model> >> <doi>0</doi> >> <baselabel type='kvm'>system_u:system_r:svirt_t:s0</baselabel> >> <baselabel type='qemu'>system_u:system_r:svirt_tcg_t:s0</baselabel> >> </secmodel> >> <secmodel> >> <model>dac</model> >> <doi>0</doi> >> <baselabel type='kvm'>+107:+107</baselabel> >> <baselabel type='qemu'>+107:+107</baselabel> >> </secmodel> >> </host> >> <guest> >> <os_type>hvm</os_type> >> <arch name='i686'> >> <wordsize>32</wordsize> >> <emulator>/usr/libexec/qemu-kvm</emulator> >> <machine maxCpus='240'>pc-i440fx-rhel7.6.0</machine> >> <machine canonical='pc-i440fx-rhel7.6.0' maxCpus='240'>pc</machine> >> <machine maxCpus='240'>pc-i440fx-rhel7.0.0</machine> >> <machine maxCpus='384'>pc-q35-rhel7.6.0</machine> >> <machine canonical='pc-q35-rhel7.6.0' maxCpus='384'>q35</machine> >> <machine maxCpus='240'>rhel6.3.0</machine> >> <machine maxCpus='240'>rhel6.4.0</machine> >> <machine maxCpus='240'>rhel6.0.0</machine> >> <machine maxCpus='240'>pc-i440fx-rhel7.5.0</machine> >> <machine maxCpus='240'>pc-i440fx-rhel7.1.0</machine> >> <machine maxCpus='240'>pc-i440fx-rhel7.2.0</machine> >> <machine maxCpus='255'>pc-q35-rhel7.3.0</machine> >> <machine maxCpus='240'>rhel6.5.0</machine> >> <machine maxCpus='384'>pc-q35-rhel7.4.0</machine> >> <machine maxCpus='240'>rhel6.6.0</machine> >> <machine maxCpus='240'>rhel6.1.0</machine> >> <machine maxCpus='240'>rhel6.2.0</machine> >> <machine maxCpus='240'>pc-i440fx-rhel7.3.0</machine> >> <machine maxCpus='240'>pc-i440fx-rhel7.4.0</machine> >> <machine maxCpus='384'>pc-q35-rhel7.5.0</machine> >> <domain type='qemu'/> >> <domain type='kvm'> >> <emulator>/usr/libexec/qemu-kvm</emulator> >> </domain> >> </arch> >> <features> >> <cpuselection/> >> <deviceboot/> >> <disksnapshot default='on' toggle='no'/> >> <acpi default='on' toggle='yes'/> >> <apic default='on' toggle='no'/> >> <pae/> >> <nonpae/> >> </features> >> </guest> >> <guest> >> <os_type>hvm</os_type> >> <arch name='x86_64'> >> <wordsize>64</wordsize> >> <emulator>/usr/libexec/qemu-kvm</emulator> >> <machine maxCpus='240'>pc-i440fx-rhel7.6.0</machine> >> <machine canonical='pc-i440fx-rhel7.6.0' maxCpus='240'>pc</machine> >> <machine maxCpus='240'>pc-i440fx-rhel7.0.0</machine> >> <machine maxCpus='384'>pc-q35-rhel7.6.0</machine> >> <machine canonical='pc-q35-rhel7.6.0' maxCpus='384'>q35</machine> >> <machine maxCpus='240'>rhel6.3.0</machine> >> <machine maxCpus='240'>rhel6.4.0</machine> >> <machine maxCpus='240'>rhel6.0.0</machine> >> <machine maxCpus='240'>pc-i440fx-rhel7.5.0</machine> >> <machine maxCpus='240'>pc-i440fx-rhel7.1.0</machine> >> <machine maxCpus='240'>pc-i440fx-rhel7.2.0</machine> >> <machine maxCpus='255'>pc-q35-rhel7.3.0</machine> >> <machine maxCpus='240'>rhel6.5.0</machine> >> <machine maxCpus='384'>pc-q35-rhel7.4.0</machine> >> <machine maxCpus='240'>rhel6.6.0</machine> >> <machine maxCpus='240'>rhel6.1.0</machine> >> <machine maxCpus='240'>rhel6.2.0</machine> >> <machine maxCpus='240'>pc-i440fx-rhel7.3.0</machine> >> <machine maxCpus='240'>pc-i440fx-rhel7.4.0</machine> >> <machine maxCpus='384'>pc-q35-rhel7.5.0</machine> >> <domain type='qemu'/> >> <domain type='kvm'> >> <emulator>/usr/libexec/qemu-kvm</emulator> >> </domain> >> </arch> >> <features> >> <cpuselection/> >> <deviceboot/> >> <disksnapshot default='on' toggle='no'/> >> <acpi default='on' toggle='yes'/> >> <apic default='on' toggle='no'/> >> </features> >> </guest> >></capabilities> >> >> >>在 2020-09-17 12:00:19,"Strahil Nikolov" <hunter86_bg@yahoo.com> 写道: >>>It would be easier if you posted the whole xml. >>> >>>What about the sections (in HE xml) starting with: >>>feature policy= >>> >>>Also the hosts have a section which contains: >>> >>><feature name= >>> >>>If you can share a VM's xml sections for a good VM. >>> >>>Best Regards, >>>Strahil Nikolov >>> >>> >>> >>> >>> >>> >>>В четвъртък, 17 септември 2020 г., 05:54:12 Гринуич+3, ddqlo <ddqlo@126.com> написа: >>> >>> >>> >>> >>> >>>HostedEngine: >>>...... >>><model fallback='forbid'>Haswell-noTSX</model> >>>...... >>> >>>both of the hosts: >>>...... >>><model>Westmere</model> >>>...... >>> >>>others vms which can be migrated: >>>...... >>><model fallback='forbid'>Haswell-noTSX</model> >>>...... >>> >>> >>> >>>在 2020-09-17 03:03:24,"Strahil Nikolov" <hunter86_bg@yahoo.com> 写道: >>>>Can you verify the HostedEngine's CPU ? >>>> >>>>1. ssh to the host hosting the HE >>>>2. alias virsh='virsh -c qemu:///system?authfile=/etc/ovirt-hosted-engine/virsh_auth.conf' >>>>3. virsh dumpxml HostedEngine >>>> >>>> >>>>Then set the alias for virsh on all Hosts and 'virsh capabilites' should show the Hosts' <cpu><model> . >>>> >>>>Best Regards, >>>>Strahil Nikolov >>>> >>>> >>>> >>>> >>>> >>>> >>>>В сряда, 16 септември 2020 г., 10:16:08 Гринуич+3, ddqlo <ddqlo@126.com> написа: >>>> >>>> >>>> >>>> >>>> >>>>My gateway was not pingable. I have fixed this problem and now both nodes have a score(3400). >>>>Yet, hosted engine could not be migrated. Same log in engine.log: >>>>host filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'CPU' >>>> >>>> >>>>在 2020-09-16 02:11:09,"Strahil Nikolov" <hunter86_bg@yahoo.com> 写道: >>>>>Both nodes have a lower than the usual score (should be 3400 ). >>>>>Based on the score you are probably suffering from gateway-score-penalty [1][2]. >>>>>Check if your gateway is pingable. >>>>> >>>>>Best Regards, >>>>>Strahil Nikolov >>>>> >>>>>1 - https://www.ovirt.org/images/Hosted-Engine-4.3-deep-dive.pdf(page 8) >>>>>2 - /etc/ovirt-hosted-engine-ha/agent.conf >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>>В вторник, 15 септември 2020 г., 04:49:48 Гринуич+3, ddqlo <ddqlo@126.com> написа: >>>>> >>>>> >>>>> >>>>> >>>>> >>>>>--== Host node28 (id: 1) status ==-- >>>>> >>>>>conf_on_shared_storage : True >>>>>Status up-to-date : True >>>>>Hostname : node28 >>>>>Host ID : 1 >>>>>Engine status : {"reason": "vm not running on this host", "health": "bad", "vm": "down_unexpected", "detail": "unknown"} >>>>>Score : 1800 >>>>>stopped : False >>>>>Local maintenance : False >>>>>crc32 : 4ac6105b >>>>>local_conf_timestamp : 1794597 >>>>>Host timestamp : 1794597 >>>>>Extra metadata (valid at timestamp): >>>>> metadata_parse_version=1 >>>>> metadata_feature_version=1 >>>>> timestamp=1794597 (Tue Sep 15 09:47:17 2020) >>>>> host-id=1 >>>>> score=1800 >>>>> vm_conf_refresh_time=1794597 (Tue Sep 15 09:47:17 2020) >>>>> conf_on_shared_storage=True >>>>> maintenance=False >>>>> state=EngineDown >>>>> stopped=False >>>>> >>>>> >>>>>--== Host node22 (id: 2) status ==-- >>>>> >>>>>conf_on_shared_storage : True >>>>>Status up-to-date : True >>>>>Hostname : node22 >>>>>Host ID : 2 >>>>>Engine status : {"health": "good", "vm": "up", "detail": "Up"} >>>>>Score : 1800 >>>>>stopped : False >>>>>Local maintenance : False >>>>>crc32 : ffc41893 >>>>>local_conf_timestamp : 1877876 >>>>>Host timestamp : 1877876 >>>>>Extra metadata (valid at timestamp): >>>>> metadata_parse_version=1 >>>>> metadata_feature_version=1 >>>>> timestamp=1877876 (Tue Sep 15 09:47:13 2020) >>>>> host-id=2 >>>>> score=1800 >>>>> vm_conf_refresh_time=1877876 (Tue Sep 15 09:47:13 2020) >>>>> conf_on_shared_storage=True >>>>> maintenance=False >>>>> state=EngineUp >>>>> stopped=False >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>>在 2020-09-09 01:32:55,"Strahil Nikolov" <hunter86_bg@yahoo.com> 写道: >>>>>>What is the output of 'hosted-engine --vm-status' on the node where the HostedEngine is running ? >>>>>> >>>>>> >>>>>>Best Regards, >>>>>>Strahil Nikolov >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>>В понеделник, 7 септември 2020 г., 03:53:13 Гринуич+3, ddqlo <ddqlo@126.com> написа: >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>>I could not find any logs because the migration button is disabled in the web UI. It seems that the engine migration operation is prevented at first. Any other ideas? Thanks! >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>>在 2020-09-01 00:06:19,"Strahil Nikolov" <hunter86_bg@yahoo.com> 写道: >>>>>>>I'm running oVirt 4.3.10 and I can migrate my Engine from node to node. >>>>>>>I had one similar issue , but powering off and on the HE has fixed it. >>>>>>> >>>>>>>You have to check the vdsm log on the source and on destination in order to figure out what is going on. >>>>>>>Also you might consider checking the libvirt logs on the destination. >>>>>>> >>>>>>>Best Regards, >>>>>>>Strahil Nikolov >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>>В понеделник, 31 август 2020 г., 10:47:22 Гринуич+3, ddqlo <ddqlo@126.com> написа: >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>>Thanks! The scores of all nodes are not '0'. I find that someone has already asked a question like this. It seems that this feature has been disabled in 4.3. I am not sure if it is enabled in 4.4. >>>>>>> >>>>>>> >>>>>>>在 2020-08-29 02:27:03,"Strahil Nikolov" <hunter86_bg@yahoo.com> : >>>>>>>>Have you checked under a shell the output of 'hosted-engine --vm-status' . Check the Score of the hosts. Maybe there is a node with score of '0' ? >>>>>>>> >>>>>>>>Best Regards, >>>>>>>>Strahil Nikolov >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>>В вторник, 25 август 2020 г., 13:46:18 Гринуич+3, 董青龙 <ddqlo@126.com> написа: >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>>Hi all, >>>>>>>> I have an ovirt4.3.10.4 environment of 2 hosts. Normal vms in this environment could be migrated, but the hosted engine vm could not be migrated. Anyone can help? Thanks a lot! >>>>>>>> >>>>>>>>hosts status: >>>>>>>> >>>>>>>>normal vm migration: >>>>>>>> >>>>>>>>hosted engine vm migration: >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>>_______________________________________________ >>>>>>>>Users mailing list -- users@ovirt.org >>>>>>>>To unsubscribe send an email to users-leave@ovirt.org >>>>>>>>Privacy Statement: https://www.ovirt.org/privacy-policy.html >>>>>>>>oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ >>>>>>>>List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZXHE2AJX42HNHOMYHTDCUUIU3VQTQTLF/ >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>>_______________________________________________ >>>>>>>Users mailing list -- users@ovirt.org >>>>>>>To unsubscribe send an email to users-leave@ovirt.org >>>>>>>Privacy Statement: https://www.ovirt.org/privacy-policy.html >>>>>>>oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ >>>>>>>List Archives: >>>>>>>https://lists.ovirt.org/archives/list/users@ovirt.org/message/IAYLFLC6K42OUPZSZU3P3ZYAU66LGSCD/ >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>>_______________________________________________ >>>>>>Users mailing list -- users@ovirt.org >>>>>>To unsubscribe send an email to users-leave@ovirt.org >>>>>>Privacy Statement: https://www.ovirt.org/privacy-policy.html >>>>>>oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ >>>>>>List Archives: >>>>>>https://lists.ovirt.org/archives/list/users@ovirt.org/message/23ZMAP5K7N5KKX3HOKUEJNVCBH7CY4ZL/ >>>>> >>>>> >>>>> >>>>> >>>>> >>>> >>>> >>>> >>>> >>>> >>> >>> >>> >>> >>> >> >> >> >> >> > >_______________________________________________ >Users mailing list -- users@ovirt.org >To unsubscribe send an email to users-leave@ovirt.org >Privacy Statement: https://www.ovirt.org/privacy-policy.html >oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ >List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/BM3QAMWVBKUATSMHAOEKNJTXLNKLNDBM/