so strange! After I set global maintenance, powered off and started H The cpu of HE became 'Westmere'(did not change anything). But HE still could not be migrated.
在 2020-09-19 13:27:35,"Strahil Nikolov" <hunter86_bg@yahoo.com> 写道: >Hm... interesting. > >The VM is using 'Haswell-noTSX' while the host is 'Westmere'. > >In my case I got no difference: > >[root@ovirt1 ~]# virsh dumpxml HostedEngine | grep Opteron > <model fallback='forbid'>Opteron_G5</model> >[root@ovirt1 ~]# virsh capabilities | grep Opteron > <model>Opteron_G5</model> > >Did you update the cluster holding the Hosted Engine ? > > >I guess you can try to: > >- Set global maintenance >- Power off the HostedEngine VM >- virsh dumpxml HostedEngine > /root/HE.xml >- use virsh edit to change the cpu of the HE (non-permanent) change >- try to power on the modified HE > >If it powers on , you can try to migrate it and if it succeeds - then you should make it permanent. > > > > > >Best Regards, >Strahil Nikolov > >В петък, 18 септември 2020 г., 04:40:39 Гринуич+3, ddqlo <ddqlo@126.com> написа: > > > > > >HE: > ><domain type='kvm' id='1'> > <name>HostedEngine</name> > <uuid>b4e805ff-556d-42bd-a6df-02f5902fd01c</uuid> > <metadata xmlns:ns0="http://ovirt.org/vm/tune/1.0" xmlns:ovirt-vm="http://ovirt.org/vm/1.0"> > <ns0:qos/> > <ovirt-vm:vm xmlns:ovirt-vm="http://ovirt.org/vm/1.0"> > <ovirt-vm:clusterVersion>4.3</ovirt-vm:clusterVersion> > <ovirt-vm:destroy_on_reboot type="bool">False</ovirt-vm:destroy_on_reboot> > <ovirt-vm:launchPaused>false</ovirt-vm:launchPaused> > <ovirt-vm:memGuaranteedSize type="int">1024</ovirt-vm:memGuaranteedSize> > <ovirt-vm:minGuaranteedMemoryMb type="int">1024</ovirt-vm:minGuaranteedMemoryMb> > <ovirt-vm:resumeBehavior>auto_resume</ovirt-vm:resumeBehavior> > <ovirt-vm:startTime type="float">1600307555.19</ovirt-vm:startTime> > <ovirt-vm:device mac_address="56:6f:9b:b0:00:01"> > <ovirt-vm:network>external</ovirt-vm:network> > <ovirt-vm:custom> > <ovirt-vm:queues>4</ovirt-vm:queues> > </ovirt-vm:custom> > </ovirt-vm:device> > <ovirt-vm:device mac_address="00:16:3e:50:c1:97"> > <ovirt-vm:network>ovirtmgmt</ovirt-vm:network> > <ovirt-vm:custom> > <ovirt-vm:queues>4</ovirt-vm:queues> > </ovirt-vm:custom> > </ovirt-vm:device> > <ovirt-vm:device devtype="disk" name="vda"> > <ovirt-vm:domainID>c17c1934-332f-464c-8f89-ad72463c00b3</ovirt-vm:domainID> > <ovirt-vm:guestName>/dev/vda2</ovirt-vm:guestName> > <ovirt-vm:imageID>8eca143a-4535-4421-bd35-9f5764d67d70</ovirt-vm:imageID> > <ovirt-vm:poolID>00000000-0000-0000-0000-000000000000</ovirt-vm:poolID> > <ovirt-vm:shared>exclusive</ovirt-vm:shared> > <ovirt-vm:volumeID>ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</ovirt-vm:volumeID> > <ovirt-vm:specParams> > <ovirt-vm:pinToIoThread>1</ovirt-vm:pinToIoThread> > </ovirt-vm:specParams> > <ovirt-vm:volumeChain> > <ovirt-vm:volumeChainNode> > <ovirt-vm:domainID>c17c1934-332f-464c-8f89-ad72463c00b3</ovirt-vm:domainID> > <ovirt-vm:imageID>8eca143a-4535-4421-bd35-9f5764d67d70</ovirt-vm:imageID> > <ovirt-vm:leaseOffset type="int">108003328</ovirt-vm:leaseOffset> > <ovirt-vm:leasePath>/dev/c17c1934-332f-464c-8f89-ad72463c00b3/leases</ovirt-vm:leasePath> > <ovirt-vm:path>/rhev/data-center/mnt/blockSD/c17c1934-332f-464c-8f89-ad72463c00b3/images/8eca143a-4535-4421-bd35-9f5764d67d70/ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</ovirt-vm:path> > <ovirt-vm:volumeID>ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</ovirt-vm:volumeID> > </ovirt-vm:volumeChainNode> > </ovirt-vm:volumeChain> > </ovirt-vm:device> > <ovirt-vm:device devtype="disk" name="hdc"/> ></ovirt-vm:vm> > </metadata> > <maxMemory slots='16' unit='KiB'>67108864</maxMemory> > <memory unit='KiB'>16777216</memory> > <currentMemory unit='KiB'>16777216</currentMemory> > <vcpu placement='static' current='4'>64</vcpu> > <iothreads>1</iothreads> > <resource> > <partition>/machine</partition> > </resource> > <sysinfo type='smbios'> > <system> > <entry name='manufacturer'>oVirt</entry> > <entry name='product'>oVirt Node</entry> > <entry name='version'>7-5.1804.el7.centos</entry> > <entry name='serial'>00000000-0000-0000-0000-0CC47A6B3160</entry> > <entry name='uuid'>b4e805ff-556d-42bd-a6df-02f5902fd01c</entry> > </system> > </sysinfo> > <os> > <type arch='x86_64' machine='pc-i440fx-rhel7.6.0'>hvm</type> > <boot dev='hd'/> > <bios useserial='yes'/> > <smbios mode='sysinfo'/> > </os> > <features> > <acpi/> > </features> > <cpu mode='custom' match='exact' check='full'> > <model fallback='forbid'>Haswell-noTSX</model> > <topology sockets='16' cores='4' threads='1'/> > <feature policy='require' name='vme'/> > <feature policy='require' name='f16c'/> > <feature policy='require' name='rdrand'/> > <feature policy='require' name='hypervisor'/> > <feature policy='require' name='arat'/> > <feature policy='require' name='xsaveopt'/> > <feature policy='require' name='abm'/> > <numa> > <cell id='0' cpus='0-3' memory='16777216' unit='KiB'/> > </numa> > </cpu> > <clock offset='variable' adjustment='0' basis='utc'> > <timer name='rtc' tickpolicy='catchup'/> > <timer name='pit' tickpolicy='delay'/> > <timer name='hpet' present='no'/> > </clock> > <on_poweroff>destroy</on_poweroff> > <on_reboot>destroy</on_reboot> > <on_crash>destroy</on_crash> > <pm> > <suspend-to-mem enabled='no'/> > <suspend-to-disk enabled='no'/> > </pm> > <devices> > <emulator>/usr/libexec/qemu-kvm</emulator> > <disk type='file' device='cdrom'> > <driver name='qemu' error_policy='report'/> > <source startupPolicy='optional'/> > <target dev='hdc' bus='ide'/> > <readonly/> > <alias name='ua-80fde7d5-ee7f-4201-9118-11bc6c3b8530'/> > <address type='drive' controller='0' bus='1' target='0' unit='0'/> > </disk> > <disk type='block' device='disk' snapshot='no'> > <driver name='qemu' type='raw' cache='none' error_policy='stop' io='native' iothread='1'/> > <source dev='/var/run/vdsm/storage/c17c1934-332f-464c-8f89-ad72463c00b3/8eca143a-4535-4421-bd35-9f5764d67d70/ae961104-c3b3-4a43-9f46-7fa6bdc2ac33'> > <seclabel model='dac' relabel='no'/> > </source> > <backingStore/> > <target dev='vda' bus='virtio'/> > <serial>8eca143a-4535-4421-bd35-9f5764d67d70</serial> > <alias name='ua-8eca143a-4535-4421-bd35-9f5764d67d70'/> > <address type='pci' domain='0x0000' bus='0x00' slot='0x07' function='0x0'/> > </disk> > <controller type='scsi' index='0' model='virtio-scsi'> > <driver iothread='1'/> > <alias name='ua-27331e83-03f4-42a3-9554-c41649c02ba4'/> > <address type='pci' domain='0x0000' bus='0x00' slot='0x05' function='0x0'/> > </controller> > <controller type='ide' index='0'> > <alias name='ide'/> > <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x1'/> > </controller> > <controller type='virtio-serial' index='0' ports='16'> > <alias name='ua-8fe74299-b60f-4778-8e80-db05393a9489'/> > <address type='pci' domain='0x0000' bus='0x00' slot='0x06' function='0x0'/> > </controller> > <controller type='usb' index='0' model='piix3-uhci'> > <alias name='usb'/> > <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x2'/> > </controller> > <controller type='pci' index='0' model='pci-root'> > <alias name='pci.0'/> > </controller> > <lease> > <lockspace>c17c1934-332f-464c-8f89-ad72463c00b3</lockspace> > <key>ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</key> > <target path='/dev/c17c1934-332f-464c-8f89-ad72463c00b3/leases' offset='108003328'/> > </lease> > <interface type='bridge'> > <mac address='00:16:3e:50:c1:97'/> > <source bridge='ovirtmgmt'/> > <target dev='vnet0'/> > <model type='virtio'/> > <driver name='vhost' queues='4'/> > <filterref filter='vdsm-no-mac-spoofing'/> > <link state='up'/> > <mtu size='1500'/> > <alias name='ua-fada74ee-2338-4cde-a7ba-43a9a636ad6e'/> > <address type='pci' domain='0x0000' bus='0x00' slot='0x03' function='0x0'/> > </interface> > <interface type='bridge'> > <mac address='56:6f:9b:b0:00:01'/> > <source bridge='external'/> > <target dev='vnet1'/> > <model type='virtio'/> > <driver name='vhost' queues='4'/> > <filterref filter='vdsm-no-mac-spoofing'/> > <link state='up'/> > <mtu size='1500'/> > <alias name='ua-f7b4c949-1f9f-4355-811d-88428c88ce4e'/> > <address type='pci' domain='0x0000' bus='0x00' slot='0x0a' function='0x0'/> > </interface> > <serial type='unix'> > <source mode='bind' path='/var/run/ovirt-vmconsole-console/b4e805ff-556d-42bd-a6df-02f5902fd01c.sock'/> > <target type='isa-serial' port='0'> > <model name='isa-serial'/> > </target> > <alias name='serial0'/> > </serial> > <console type='unix'> > <source mode='bind' path='/var/run/ovirt-vmconsole-console/b4e805ff-556d-42bd-a6df-02f5902fd01c.sock'/> > <target type='serial' port='0'/> > <alias name='serial0'/> > </console> > <channel type='unix'> > <source mode='bind' path='/var/lib/libvirt/qemu/channels/b4e805ff-556d-42bd-a6df-02f5902fd01c.ovirt-guest-agent.0'/> > <target type='virtio' name='ovirt-guest-agent.0' state='connected'/> > <alias name='channel0'/> > <address type='virtio-serial' controller='0' bus='0' port='1'/> > </channel> > <channel type='unix'> > <source mode='bind' path='/var/lib/libvirt/qemu/channels/b4e805ff-556d-42bd-a6df-02f5902fd01c.org.qemu.guest_agent.0'/> > <target type='virtio' name='org.qemu.guest_agent.0' state='connected'/> > <alias name='channel1'/> > <address type='virtio-serial' controller='0' bus='0' port='2'/> > </channel> > <channel type='spicevmc'> > <target type='virtio' name='com.redhat.spice.0' state='disconnected'/> > <alias name='channel2'/> > <address type='virtio-serial' controller='0' bus='0' port='3'/> > </channel> > <channel type='unix'> > <source mode='bind' path='/var/lib/libvirt/qemu/channels/b4e805ff-556d-42bd-a6df-02f5902fd01c.org.ovirt.hosted-engine-setup.0'/> > <target type='virtio' name='org.ovirt.hosted-engine-setup.0' state='disconnected'/> > <alias name='channel3'/> > <address type='virtio-serial' controller='0' bus='0' port='4'/> > </channel> > <input type='tablet' bus='usb'> > <alias name='input0'/> > <address type='usb' bus='0' port='1'/> > </input> > <input type='mouse' bus='ps2'> > <alias name='input1'/> > </input> > <input type='keyboard' bus='ps2'> > <alias name='input2'/> > </input> > <graphics type='vnc' port='5900' autoport='yes' listen='192.168.1.22' keymap='en-us' passwdValidTo='1970-01-01T00:00:01'> > <listen type='network' address='192.168.1.22' network='vdsm-external'/> > </graphics> > <graphics type='spice' port='5901' tlsPort='5902' autoport='yes' listen='192.168.1.22' passwdValidTo='1970-01-01T00:00:01'> > <listen type='network' address='192.168.1.22' network='vdsm-external'/> > <channel name='main' mode='secure'/> > <channel name='display' mode='secure'/> > <channel name='inputs' mode='secure'/> > <channel name='cursor' mode='secure'/> > <channel name='playback' mode='secure'/> > <channel name='record' mode='secure'/> > <channel name='smartcard' mode='secure'/> > <channel name='usbredir' mode='secure'/> > </graphics> > <sound model='ich6'> > <alias name='ua-bd287767-9b83-4e44-ac6f-8b527f9632b8'/> > <address type='pci' domain='0x0000' bus='0x00' slot='0x04' function='0x0'/> > </sound> > <video> > <model type='qxl' ram='65536' vram='32768' vgamem='16384' heads='1' primary='yes'/> > <alias name='ua-bcfb6b4b-0b3c-4d5b-ba2d-8ce40a65facd'/> > <address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x0'/> > </video> > <memballoon model='virtio'> > <stats period='5'/> > <alias name='ua-39d36063-8808-47db-9fef-a0baad9f9661'/> > <address type='pci' domain='0x0000' bus='0x00' slot='0x08' function='0x0'/> > </memballoon> > <rng model='virtio'> > <backend model='random'>/dev/urandom</backend> > <alias name='ua-75516d34-dd8f-4f0f-8496-e1f222a359a8'/> > <address type='pci' domain='0x0000' bus='0x00' slot='0x09' function='0x0'/> > </rng> > </devices> > <seclabel type='dynamic' model='selinux' relabel='yes'> > <label>system_u:system_r:svirt_t:s0:c162,c716</label> > <imagelabel>system_u:object_r:svirt_image_t:s0:c162,c716</imagelabel> > </seclabel> > <seclabel type='dynamic' model='dac' relabel='yes'> > <label>+107:+107</label> > <imagelabel>+107:+107</imagelabel> > </seclabel> ></domain> > >hosts: > ><capabilities> > <host> > <uuid>b25adcec-eef2-49a5-8663-7cdcfc50891b</uuid> > <cpu> > <arch>x86_64</arch> > <model>Westmere</model> > <vendor>Intel</vendor> > <microcode version='34'/> > <counter name='tsc' frequency='3699996000' scaling='no'/> > <topology sockets='1' cores='2' threads='2'/> > <feature name='vme'/> > <feature name='ds'/> > <feature name='acpi'/> > <feature name='ss'/> > <feature name='ht'/> > <feature name='tm'/> > <feature name='pbe'/> > <feature name='pclmuldq'/> > <feature name='dtes64'/> > <feature name='monitor'/> > <feature name='ds_cpl'/> > <feature name='vmx'/> > <feature name='est'/> > <feature name='tm2'/> > <feature name='fma'/> > <feature name='xtpr'/> > <feature name='pdcm'/> > <feature name='pcid'/> > <feature name='movbe'/> > <feature name='tsc-deadline'/> > <feature name='xsave'/> > <feature name='osxsave'/> > <feature name='avx'/> > <feature name='f16c'/> > <feature name='rdrand'/> > <feature name='arat'/> > <feature name='fsgsbase'/> > <feature name='tsc_adjust'/> > <feature name='bmi1'/> > <feature name='avx2'/> > <feature name='smep'/> > <feature name='bmi2'/> > <feature name='erms'/> > <feature name='invpcid'/> > <feature name='xsaveopt'/> > <feature name='pdpe1gb'/> > <feature name='rdtscp'/> > <feature name='abm'/> > <feature name='invtsc'/> > <pages unit='KiB' size='4'/> > <pages unit='KiB' size='2048'/> > <pages unit='KiB' size='1048576'/> > </cpu> > <power_management> > <suspend_mem/> > <suspend_disk/> > <suspend_hybrid/> > </power_management> > <iommu support='no'/> > <migration_features> > <live/> > <uri_transports> > <uri_transport>tcp</uri_transport> > <uri_transport>rdma</uri_transport> > </uri_transports> > </migration_features> > <topology> > <cells num='1'> > <cell id='0'> > <memory unit='KiB'>32903488</memory> > <pages unit='KiB' size='4'>8225872</pages> > <pages unit='KiB' size='2048'>0</pages> > <pages unit='KiB' size='1048576'>0</pages> > <distances> > <sibling id='0' value='10'/> > </distances> > <cpus num='4'> > <cpu id='0' socket_id='0' core_id='0' siblings='0,2'/> > <cpu id='1' socket_id='0' core_id='1' siblings='1,3'/> > <cpu id='2' socket_id='0' core_id='0' siblings='0,2'/> > <cpu id='3' socket_id='0' core_id='1' siblings='1,3'/> > </cpus> > </cell> > </cells> > </topology> > <cache> > <bank id='0' level='3' type='both' size='3' unit='MiB' cpus='0-3'/> > </cache> > <secmodel> > <model>selinux</model> > <doi>0</doi> > <baselabel type='kvm'>system_u:system_r:svirt_t:s0</baselabel> > <baselabel type='qemu'>system_u:system_r:svirt_tcg_t:s0</baselabel> > </secmodel> > <secmodel> > <model>dac</model> > <doi>0</doi> > <baselabel type='kvm'>+107:+107</baselabel> > <baselabel type='qemu'>+107:+107</baselabel> > </secmodel> > </host> > <guest> > <os_type>hvm</os_type> > <arch name='i686'> > <wordsize>32</wordsize> > <emulator>/usr/libexec/qemu-kvm</emulator> > <machine maxCpus='240'>pc-i440fx-rhel7.6.0</machine> > <machine canonical='pc-i440fx-rhel7.6.0' maxCpus='240'>pc</machine> > <machine maxCpus='240'>pc-i440fx-rhel7.0.0</machine> > <machine maxCpus='384'>pc-q35-rhel7.6.0</machine> > <machine canonical='pc-q35-rhel7.6.0' maxCpus='384'>q35</machine> > <machine maxCpus='240'>rhel6.3.0</machine> > <machine maxCpus='240'>rhel6.4.0</machine> > <machine maxCpus='240'>rhel6.0.0</machine> > <machine maxCpus='240'>pc-i440fx-rhel7.5.0</machine> > <machine maxCpus='240'>pc-i440fx-rhel7.1.0</machine> > <machine maxCpus='240'>pc-i440fx-rhel7.2.0</machine> > <machine maxCpus='255'>pc-q35-rhel7.3.0</machine> > <machine maxCpus='240'>rhel6.5.0</machine> > <machine maxCpus='384'>pc-q35-rhel7.4.0</machine> > <machine maxCpus='240'>rhel6.6.0</machine> > <machine maxCpus='240'>rhel6.1.0</machine> > <machine maxCpus='240'>rhel6.2.0</machine> > <machine maxCpus='240'>pc-i440fx-rhel7.3.0</machine> > <machine maxCpus='240'>pc-i440fx-rhel7.4.0</machine> > <machine maxCpus='384'>pc-q35-rhel7.5.0</machine> > <domain type='qemu'/> > <domain type='kvm'> > <emulator>/usr/libexec/qemu-kvm</emulator> > </domain> > </arch> > <features> > <cpuselection/> > <deviceboot/> > <disksnapshot default='on' toggle='no'/> > <acpi default='on' toggle='yes'/> > <apic default='on' toggle='no'/> > <pae/> > <nonpae/> > </features> > </guest> > <guest> > <os_type>hvm</os_type> > <arch name='x86_64'> > <wordsize>64</wordsize> > <emulator>/usr/libexec/qemu-kvm</emulator> > <machine maxCpus='240'>pc-i440fx-rhel7.6.0</machine> > <machine canonical='pc-i440fx-rhel7.6.0' maxCpus='240'>pc</machine> > <machine maxCpus='240'>pc-i440fx-rhel7.0.0</machine> > <machine maxCpus='384'>pc-q35-rhel7.6.0</machine> > <machine canonical='pc-q35-rhel7.6.0' maxCpus='384'>q35</machine> > <machine maxCpus='240'>rhel6.3.0</machine> > <machine maxCpus='240'>rhel6.4.0</machine> > <machine maxCpus='240'>rhel6.0.0</machine> > <machine maxCpus='240'>pc-i440fx-rhel7.5.0</machine> > <machine maxCpus='240'>pc-i440fx-rhel7.1.0</machine> > <machine maxCpus='240'>pc-i440fx-rhel7.2.0</machine> > <machine maxCpus='255'>pc-q35-rhel7.3.0</machine> > <machine maxCpus='240'>rhel6.5.0</machine> > <machine maxCpus='384'>pc-q35-rhel7.4.0</machine> > <machine maxCpus='240'>rhel6.6.0</machine> > <machine maxCpus='240'>rhel6.1.0</machine> > <machine maxCpus='240'>rhel6.2.0</machine> > <machine maxCpus='240'>pc-i440fx-rhel7.3.0</machine> > <machine maxCpus='240'>pc-i440fx-rhel7.4.0</machine> > <machine maxCpus='384'>pc-q35-rhel7.5.0</machine> > <domain type='qemu'/> > <domain type='kvm'> > <emulator>/usr/libexec/qemu-kvm</emulator> > </domain> > </arch> > <features> > <cpuselection/> > <deviceboot/> > <disksnapshot default='on' toggle='no'/> > <acpi default='on' toggle='yes'/> > <apic default='on' toggle='no'/> > </features> > </guest> ></capabilities> > > >在 2020-09-17 12:00:19,"Strahil Nikolov" <hunter86_bg@yahoo.com> 写道: >>It would be easier if you posted the whole xml. >> >>What about the sections (in HE xml) starting with: >>feature policy= >> >>Also the hosts have a section which contains: >> >><feature name= >> >>If you can share a VM's xml sections for a good VM. >> >>Best Regards, >>Strahil Nikolov >> >> >> >> >> >> >>В четвъртък, 17 септември 2020 г., 05:54:12 Гринуич+3, ddqlo <ddqlo@126.com> написа: >> >> >> >> >> >>HostedEngine: >>...... >><model fallback='forbid'>Haswell-noTSX</model> >>...... >> >>both of the hosts: >>...... >><model>Westmere</model> >>...... >> >>others vms which can be migrated: >>...... >><model fallback='forbid'>Haswell-noTSX</model> >>...... >> >> >> >>在 2020-09-17 03:03:24,"Strahil Nikolov" <hunter86_bg@yahoo.com> 写道: >>>Can you verify the HostedEngine's CPU ? >>> >>>1. ssh to the host hosting the HE >>>2. alias virsh='virsh -c qemu:///system?authfile=/etc/ovirt-hosted-engine/virsh_auth.conf' >>>3. virsh dumpxml HostedEngine >>> >>> >>>Then set the alias for virsh on all Hosts and 'virsh capabilites' should show the Hosts' <cpu><model> . >>> >>>Best Regards, >>>Strahil Nikolov >>> >>> >>> >>> >>> >>> >>>В сряда, 16 септември 2020 г., 10:16:08 Гринуич+3, ddqlo <ddqlo@126.com> написа: >>> >>> >>> >>> >>> >>>My gateway was not pingable. I have fixed this problem and now both nodes have a score(3400). >>>Yet, hosted engine could not be migrated. Same log in engine.log: >>>host filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'CPU' >>> >>> >>>在 2020-09-16 02:11:09,"Strahil Nikolov" <hunter86_bg@yahoo.com> 写道: >>>>Both nodes have a lower than the usual score (should be 3400 ). >>>>Based on the score you are probably suffering from gateway-score-penalty [1][2]. >>>>Check if your gateway is pingable. >>>> >>>>Best Regards, >>>>Strahil Nikolov >>>> >>>>1 - https://www.ovirt.org/images/Hosted-Engine-4.3-deep-dive.pdf(page 8) >>>>2 - /etc/ovirt-hosted-engine-ha/agent.conf >>>> >>>> >>>> >>>> >>>> >>>> >>>>В вторник, 15 септември 2020 г., 04:49:48 Гринуич+3, ddqlo <ddqlo@126.com> написа: >>>> >>>> >>>> >>>> >>>> >>>>--== Host node28 (id: 1) status ==-- >>>> >>>>conf_on_shared_storage : True >>>>Status up-to-date : True >>>>Hostname : node28 >>>>Host ID : 1 >>>>Engine status : {"reason": "vm not running on this host", "health": "bad", "vm": "down_unexpected", "detail": "unknown"} >>>>Score : 1800 >>>>stopped : False >>>>Local maintenance : False >>>>crc32 : 4ac6105b >>>>local_conf_timestamp : 1794597 >>>>Host timestamp : 1794597 >>>>Extra metadata (valid at timestamp): >>>> metadata_parse_version=1 >>>> metadata_feature_version=1 >>>> timestamp=1794597 (Tue Sep 15 09:47:17 2020) >>>> host-id=1 >>>> score=1800 >>>> vm_conf_refresh_time=1794597 (Tue Sep 15 09:47:17 2020) >>>> conf_on_shared_storage=True >>>> maintenance=False >>>> state=EngineDown >>>> stopped=False >>>> >>>> >>>>--== Host node22 (id: 2) status ==-- >>>> >>>>conf_on_shared_storage : True >>>>Status up-to-date : True >>>>Hostname : node22 >>>>Host ID : 2 >>>>Engine status : {"health": "good", "vm": "up", "detail": "Up"} >>>>Score : 1800 >>>>stopped : False >>>>Local maintenance : False >>>>crc32 : ffc41893 >>>>local_conf_timestamp : 1877876 >>>>Host timestamp : 1877876 >>>>Extra metadata (valid at timestamp): >>>> metadata_parse_version=1 >>>> metadata_feature_version=1 >>>> timestamp=1877876 (Tue Sep 15 09:47:13 2020) >>>> host-id=2 >>>> score=1800 >>>> vm_conf_refresh_time=1877876 (Tue Sep 15 09:47:13 2020) >>>> conf_on_shared_storage=True >>>> maintenance=False >>>> state=EngineUp >>>> stopped=False >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>>在 2020-09-09 01:32:55,"Strahil Nikolov" <hunter86_bg@yahoo.com> 写道: >>>>>What is the output of 'hosted-engine --vm-status' on the node where the HostedEngine is running ? >>>>> >>>>> >>>>>Best Regards, >>>>>Strahil Nikolov >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>>В понеделник, 7 септември 2020 г., 03:53:13 Гринуич+3, ddqlo <ddqlo@126.com> написа: >>>>> >>>>> >>>>> >>>>> >>>>> >>>>>I could not find any logs because the migration button is disabled in the web UI. It seems that the engine migration operation is prevented at first. Any other ideas? Thanks! >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>>在 2020-09-01 00:06:19,"Strahil Nikolov" <hunter86_bg@yahoo.com> 写道: >>>>>>I'm running oVirt 4.3.10 and I can migrate my Engine from node to node. >>>>>>I had one similar issue , but powering off and on the HE has fixed it. >>>>>> >>>>>>You have to check the vdsm log on the source and on destination in order to figure out what is going on. >>>>>>Also you might consider checking the libvirt logs on the destination. >>>>>> >>>>>>Best Regards, >>>>>>Strahil Nikolov >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>>В понеделник, 31 август 2020 г., 10:47:22 Гринуич+3, ddqlo <ddqlo@126.com> написа: >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>>Thanks! The scores of all nodes are not '0'. I find that someone has already asked a question like this. It seems that this feature has been disabled in 4.3. I am not sure if it is enabled in 4.4. >>>>>> >>>>>> >>>>>>在 2020-08-29 02:27:03,"Strahil Nikolov" <hunter86_bg@yahoo.com> : >>>>>>>Have you checked under a shell the output of 'hosted-engine --vm-status' . Check the Score of the hosts. Maybe there is a node with score of '0' ? >>>>>>> >>>>>>>Best Regards, >>>>>>>Strahil Nikolov >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>>В вторник, 25 август 2020 г., 13:46:18 Гринуич+3, 董青龙 <ddqlo@126.com> написа: >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>>Hi all, >>>>>>> I have an ovirt4.3.10.4 environment of 2 hosts. Normal vms in this environment could be migrated, but the hosted engine vm could not be migrated. Anyone can help? Thanks a lot! >>>>>>> >>>>>>>hosts status: >>>>>>> >>>>>>>normal vm migration: >>>>>>> >>>>>>>hosted engine vm migration: >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>>_______________________________________________ >>>>>>>Users mailing list -- users@ovirt.org >>>>>>>To unsubscribe send an email to users-leave@ovirt.org >>>>>>>Privacy Statement: https://www.ovirt.org/privacy-policy.html >>>>>>>oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ >>>>>>>List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZXHE2AJX42HNHOMYHTDCUUIU3VQTQTLF/ >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>>_______________________________________________ >>>>>>Users mailing list -- users@ovirt.org >>>>>>To unsubscribe send an email to users-leave@ovirt.org >>>>>>Privacy Statement: https://www.ovirt.org/privacy-policy.html >>>>>>oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ >>>>>>List Archives: >>>>>>https://lists.ovirt.org/archives/list/users@ovirt.org/message/IAYLFLC6K42OUPZSZU3P3ZYAU66LGSCD/ >>>>> >>>>> >>>>> >>>>> >>>>> >>>>>_______________________________________________ >>>>>Users mailing list -- users@ovirt.org >>>>>To unsubscribe send an email to users-leave@ovirt.org >>>>>Privacy Statement: https://www.ovirt.org/privacy-policy.html >>>>>oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ >>>>>List Archives: >>>>>https://lists.ovirt.org/archives/list/users@ovirt.org/message/23ZMAP5K7N5KKX3HOKUEJNVCBH7CY4ZL/ >>>> >>>> >>>> >>>> >>>> >>> >>> >>> >>> >>> >> >> >> >> >> > > > > >