[ovirt-users] VMs with multiple vdisks don't migrate
Maor Lipchuk
mlipchuk at redhat.com
Wed Feb 14 10:04:10 UTC 2018
Hi Frank,
I already replied on your last email.
Can you provide the VDSM logs from the time of the migration failure for
both hosts:
ginger.local.systea.f <http://ginger.local.systea.fr/>r and v
ictor.local.systea.fr
Thanks,
Maor
On Wed, Feb 14, 2018 at 11:23 AM, fsoyer <fsoyer at systea.fr> wrote:
> Hi all,
> I discovered yesterday a problem when migrating VM with more than one
> vdisk.
> On our test servers (oVirt4.1, shared storage with Gluster), I created 2
> VMs needed for a test, from a template with a 20G vdisk. On this VMs I
> added a 100G vdisk (for this tests I didn't want to waste time to extend
> the existing vdisks... But I lost time finally...). The VMs with the 2
> vdisks works well.
> Now I saw some updates waiting on the host. I tried to put it in
> maintenance... But it stopped on the two VM. They were marked "migrating",
> but no more accessible. Other (small) VMs with only 1 vdisk was migrated
> without problem at the same time.
> I saw that a kvm process for the (big) VMs was launched on the source AND
> destination host, but after tens of minutes, the migration and the VMs was
> always freezed. I tried to cancel the migration for the VMs : failed. The
> only way to stop it was to poweroff the VMs : the kvm process died on the 2
> hosts and the GUI alerted on a failed migration.
> In doubt, I tried to delete the second vdisk on one of this VMs : it
> migrates then without error ! And no access problem.
> I tried to extend the first vdisk of the second VM, the delete the second
> vdisk : it migrates now without problem !
>
> So after another test with a VM with 2 vdisks, I can say that this blocked
> the migration process :(
>
> In engine.log, for a VMs with 1 vdisk migrating well, we see :
>
> 2018-02-12 16:46:29,705+01 INFO [org.ovirt.engine.core.bll.MigrateVmToServerCommand]
> (default task-28) [2f712024-5982-46a8-82c8-fd8293da5725] Lock Acquired to
> object 'EngineLock:{exclusiveLocks='[3f57e669-5e4c-4d10-85cc-d573004a099d=VM]',
> sharedLocks=''}'
> 2018-02-12 16:46:29,955+01 INFO [org.ovirt.engine.core.bll.MigrateVmToServerCommand]
> (org.ovirt.thread.pool-6-thread-32) [2f712024-5982-46a8-82c8-fd8293da5725]
> Running command: MigrateVmToServerCommand internal: false. Entities
> affected : ID: 3f57e669-5e4c-4d10-85cc-d573004a099d Type: VMAction group
> MIGRATE_VM with role type USER
> 2018-02-12 16:46:30,261+01 INFO [org.ovirt.engine.core.vdsbroker.MigrateVDSCommand]
> (org.ovirt.thread.pool-6-thread-32) [2f712024-5982-46a8-82c8-fd8293da5725]
> START, MigrateVDSCommand( MigrateVDSCommandParameters:{runAsync='true',
> hostId='ce3938b1-b23f-4d22-840a-f17d7cd87bb1',
> vmId='3f57e669-5e4c-4d10-85cc-d573004a099d', srcHost='192.168.0.6',
> dstVdsId='d569c2dd-8f30-4878-8aea-858db285cf69', dstHost='
> 192.168.0.5:54321', migrationMethod='ONLINE', tunnelMigration='false',
> migrationDowntime='0', autoConverge='true', migrateCompressed='false',
> consoleAddress='null', maxBandwidth='500', enableGuestEvents='true',
> maxIncomingMigrations='2', maxOutgoingMigrations='2',
> convergenceSchedule='[init=[{name=setDowntime, params=[100]}],
> stalling=[{limit=1, action={name=setDowntime, params=[150]}}, {limit=2,
> action={name=setDowntime, params=[200]}}, {limit=3,
> action={name=setDowntime, params=[300]}}, {limit=4,
> action={name=setDowntime, params=[400]}}, {limit=6,
> action={name=setDowntime, params=[500]}}, {limit=-1, action={name=abort,
> params=[]}}]]'}), log id: 14f61ee0
> 2018-02-12 16:46:30,262+01 INFO [org.ovirt.engine.core.
> vdsbroker.vdsbroker.MigrateBrokerVDSCommand] (org.ovirt.thread.pool-6-thread-32)
> [2f712024-5982-46a8-82c8-fd8293da5725] START, MigrateBrokerVDSCommand(HostName
> = victor.local.systea.fr, MigrateVDSCommandParameters:{runAsync='true',
> hostId='ce3938b1-b23f-4d22-840a-f17d7cd87bb1',
> vmId='3f57e669-5e4c-4d10-85cc-d573004a099d', srcHost='192.168.0.6',
> dstVdsId='d569c2dd-8f30-4878-8aea-858db285cf69', dstHost='
> 192.168.0.5:54321', migrationMethod='ONLINE', tunnelMigration='false',
> migrationDowntime='0', autoConverge='true', migrateCompressed='false',
> consoleAddress='null', maxBandwidth='500', enableGuestEvents='true',
> maxIncomingMigrations='2', maxOutgoingMigrations='2',
> convergenceSchedule='[init=[{name=setDowntime, params=[100]}],
> stalling=[{limit=1, action={name=setDowntime, params=[150]}}, {limit=2,
> action={name=setDowntime, params=[200]}}, {limit=3,
> action={name=setDowntime, params=[300]}}, {limit=4,
> action={name=setDowntime, params=[400]}}, {limit=6,
> action={name=setDowntime, params=[500]}}, {limit=-1, action={name=abort,
> params=[]}}]]'}), log id: 775cd381
> 2018-02-12 16:46:30,277+01 INFO [org.ovirt.engine.core.
> vdsbroker.vdsbroker.MigrateBrokerVDSCommand] (org.ovirt.thread.pool-6-thread-32)
> [2f712024-5982-46a8-82c8-fd8293da5725] FINISH, MigrateBrokerVDSCommand,
> log id: 775cd381
> 2018-02-12 16:46:30,285+01 INFO [org.ovirt.engine.core.vdsbroker.MigrateVDSCommand]
> (org.ovirt.thread.pool-6-thread-32) [2f712024-5982-46a8-82c8-fd8293da5725]
> FINISH, MigrateVDSCommand, return: MigratingFrom, log id: 14f61ee0
> 2018-02-12 16:46:30,301+01 INFO [org.ovirt.engine.core.dal.
> dbbroker.auditloghandling.AuditLogDirector] (org.ovirt.thread.pool-6-thread-32)
> [2f712024-5982-46a8-82c8-fd8293da5725] EVENT_ID: VM_MIGRATION_START(62),
> Correlation ID: 2f712024-5982-46a8-82c8-fd8293da5725, Job ID:
> 4bd19aa9-cc99-4d02-884e-5a1e857a7738, Call Stack: null, Custom ID: null,
> Custom Event ID: -1, Message: Migration started (VM: Oracle_SECONDARY,
> Source: victor.local.systea.fr, Destination: ginger.local.systea.fr,
> User: admin at internal-authz).
> 2018-02-12 16:46:31,106+01 INFO [org.ovirt.engine.core.
> vdsbroker.vdsbroker.FullListVDSCommand] (DefaultQuartzScheduler9)
> [54a65b66] START, FullListVDSCommand(HostName = victor.local.systea.fr,
> FullListVDSCommandParameters:{runAsync='true', hostId='ce3938b1-b23f-4d22-840a-f17d7cd87bb1',
> vmIds='[3f57e669-5e4c-4d10-85cc-d573004a099d]'}), log id: 54b4b435
> 2018-02-12 16:46:31,147+01 INFO [org.ovirt.engine.core.
> vdsbroker.vdsbroker.FullListVDSCommand] (DefaultQuartzScheduler9)
> [54a65b66] FINISH, FullListVDSCommand, return: [{acpiEnable=true,
> emulatedMachine=pc-i440fx-rhel7.3.0, tabletEnable=true, pid=1493,
> guestDiskMapping={0QEMU_QEMU_HARDDISK_d890fa68-fba4-4f49-9={name=/dev/sda},
> QEMU_DVD-ROM_QM00003={name=/dev/sr0}}, transparentHugePages=true,
> timeOffset=0, cpuType=Nehalem, smp=2, pauseCode=NOERR,
> guestNumaNodes=[Ljava.lang.Object;@1d9042cd, smartcardEnable=false,
> custom={device_fbddd528-7d93-49c6-a286-180e021cb274device_
> 879c93ab-4df1-435c-af02-565039fcc254=VmDevice:{id='VmDeviceId:{deviceId='
> 879c93ab-4df1-435c-af02-565039fcc254', vmId='3f57e669-5e4c-4d10-85cc-d573004a099d'}',
> device='unix', type='CHANNEL', bootOrder='0', specParams='[]',
> address='{bus=0, controller=0, type=virtio-serial, port=1}',
> managed='false', plugged='true', readOnly='false', deviceAlias='channel0',
> customProperties='[]', snapshotId='null', logicalName='null',
> hostDevice='null'}, device_fbddd528-7d93-49c6-a286-180e021cb274device_
> 879c93ab-4df1-435c-af02-565039fcc254device_8945f61a-abbe-4156-8485-
> a4aa6f1908dbdevice_017b5e59-01c4-4aac-bf0c-b5d9557284d6=
> VmDevice:{id='VmDeviceId:{deviceId='017b5e59-01c4-4aac-bf0c-b5d9557284d6',
> vmId='3f57e669-5e4c-4d10-85cc-d573004a099d'}', device='tablet',
> type='UNKNOWN', bootOrder='0', specParams='[]', address='{bus=0, type=usb,
> port=1}', managed='false', plugged='true', readOnly='false',
> deviceAlias='input0', customProperties='[]', snapshotId='null',
> logicalName='null', hostDevice='null'}, device_fbddd528-7d93-49c6-
> a286-180e021cb274=VmDevice:{id='VmDeviceId:{deviceId='
> fbddd528-7d93-49c6-a286-180e021cb274', vmId='3f57e669-5e4c-4d10-85cc-d573004a099d'}',
> device='ide', type='CONTROLLER', bootOrder='0', specParams='[]',
> address='{slot=0x01, bus=0x00, domain=0x0000, type=pci, function=0x1}',
> managed='false', plugged='true', readOnly='false', deviceAlias='ide',
> customProperties='[]', snapshotId='null', logicalName='null',
> hostDevice='null'}, device_fbddd528-7d93-49c6-a286-180e021cb274device_
> 879c93ab-4df1-435c-af02-565039fcc254device_8945f61a-
> abbe-4156-8485-a4aa6f1908db=VmDevice:{id='VmDeviceId:{
> deviceId='8945f61a-abbe-4156-8485-a4aa6f1908db',
> vmId='3f57e669-5e4c-4d10-85cc-d573004a099d'}', device='unix',
> type='CHANNEL', bootOrder='0', specParams='[]', address='{bus=0,
> controller=0, type=virtio-serial, port=2}', managed='false',
> plugged='true', readOnly='false', deviceAlias='channel1',
> customProperties='[]', snapshotId='null', logicalName='null',
> hostDevice='null'}}, vmType=kvm, memSize=8192, smpCoresPerSocket=1,
> vmName=Oracle_SECONDARY, nice=0, status=Migration Source, maxMemSize=32768,
> bootMenuEnable=false, vmId=3f57e669-5e4c-4d10-85cc-d573004a099d,
> numOfIoThreads=2, smpThreadsPerCore=1, memGuaranteedSize=8192,
> kvmEnable=true, pitReinjection=false, displayNetwork=ovirtmgmt,
> devices=[Ljava.lang.Object;@28ae66d7, display=vnc, maxVCpus=16,
> clientIp=, statusTime=4299484520, maxMemSlots=16}], log id: 54b4b435
> 2018-02-12 16:46:31,150+01 INFO [org.ovirt.engine.core.
> vdsbroker.monitoring.VmsStatisticsFetcher] (DefaultQuartzScheduler1)
> [27fac647] Fetched 3 VMs from VDS 'd569c2dd-8f30-4878-8aea-858db285cf69'
> 2018-02-12 16:46:31,151+01 INFO [org.ovirt.engine.core.
> vdsbroker.monitoring.VmDevicesMonitoring] (DefaultQuartzScheduler9)
> [54a65b66] Received a vnc Device without an address when processing VM
> 3f57e669-5e4c-4d10-85cc-d573004a099d devices, skipping device:
> {device=vnc, specParams={displayNetwork=ovirtmgmt, keyMap=fr,
> displayIp=192.168.0.6}, type=graphics, deviceId=813957b1-446a-4e88-9e40-9fe76d2c442d,
> port=5901}
> 2018-02-12 16:46:31,151+01 INFO [org.ovirt.engine.core.
> vdsbroker.monitoring.VmDevicesMonitoring] (DefaultQuartzScheduler9)
> [54a65b66] Received a lease Device without an address when processing VM
> 3f57e669-5e4c-4d10-85cc-d573004a099d devices, skipping device:
> {lease_id=3f57e669-5e4c-4d10-85cc-d573004a099d,
> sd_id=1e51cecc-eb2e-47d0-b185-920fdc7afa16, deviceId={uuid=a09949aa-5642-4b6d-94a4-8b0d04257be5},
> offset=6291456, device=lease, path=/rhev/data-center/mnt/
> glusterSD/192.168.0.6:_DATA01/1e51cecc-eb2e-47d0-b185-920fdc7afa16/dom_md/xleases,
> type=lease}
> 2018-02-12 16:46:31,152+01 INFO [org.ovirt.engine.core.
> vdsbroker.monitoring.VmAnalyzer] (DefaultQuartzScheduler1) [27fac647] VM
> '3f57e669-5e4c-4d10-85cc-d573004a099d'(Oracle_SECONDARY) was unexpectedly
> detected as 'MigratingTo' on VDS 'd569c2dd-8f30-4878-8aea-858db285cf69'(
> ginger.local.systea.fr) (expected on 'ce3938b1-b23f-4d22-840a-
> f17d7cd87bb1')
> 2018-02-12 16:46:31,152+01 INFO [org.ovirt.engine.core.
> vdsbroker.monitoring.VmAnalyzer] (DefaultQuartzScheduler1) [27fac647] VM
> '3f57e669-5e4c-4d10-85cc-d573004a099d' is migrating to VDS
> 'd569c2dd-8f30-4878-8aea-858db285cf69'(ginger.local.systea.fr) ignoring
> it in the refresh until migration is done
> ....
> 2018-02-12 16:46:41,631+01 INFO [org.ovirt.engine.core.
> vdsbroker.monitoring.VmAnalyzer] (ForkJoinPool-1-worker-11) [] VM
> '3f57e669-5e4c-4d10-85cc-d573004a099d' was reported as Down on VDS
> 'ce3938b1-b23f-4d22-840a-f17d7cd87bb1'(victor.local.systea.fr)
> 2018-02-12 16:46:41,632+01 INFO [org.ovirt.engine.core.
> vdsbroker.vdsbroker.DestroyVDSCommand] (ForkJoinPool-1-worker-11) []
> START, DestroyVDSCommand(HostName = victor.local.systea.fr,
> DestroyVmVDSCommandParameters:{runAsync='true',
> hostId='ce3938b1-b23f-4d22-840a-f17d7cd87bb1',
> vmId='3f57e669-5e4c-4d10-85cc-d573004a099d', force='false',
> secondsToWait='0', gracefully='false', reason='', ignoreNoVm='true'}), log
> id: 560eca57
> 2018-02-12 16:46:41,650+01 INFO [org.ovirt.engine.core.
> vdsbroker.vdsbroker.DestroyVDSCommand] (ForkJoinPool-1-worker-11) []
> FINISH, DestroyVDSCommand, log id: 560eca57
> 2018-02-12 16:46:41,650+01 INFO [org.ovirt.engine.core.
> vdsbroker.monitoring.VmAnalyzer] (ForkJoinPool-1-worker-11) [] VM
> '3f57e669-5e4c-4d10-85cc-d573004a099d'(Oracle_SECONDARY) moved from
> 'MigratingFrom' --> 'Down'
> 2018-02-12 16:46:41,651+01 INFO [org.ovirt.engine.core.
> vdsbroker.monitoring.VmAnalyzer] (ForkJoinPool-1-worker-11) [] Handing
> over VM '3f57e669-5e4c-4d10-85cc-d573004a099d'(Oracle_SECONDARY) to Host
> 'd569c2dd-8f30-4878-8aea-858db285cf69'. Setting VM to status 'MigratingTo'
> 2018-02-12 16:46:42,163+01 INFO [org.ovirt.engine.core.
> vdsbroker.monitoring.VmAnalyzer] (ForkJoinPool-1-worker-4) [] VM
> '3f57e669-5e4c-4d10-85cc-d573004a099d'(Oracle_SECONDARY) moved from
> 'MigratingTo' --> 'Up'
> 2018-02-12 16:46:42,169+01 INFO [org.ovirt.engine.core.
> vdsbroker.vdsbroker.MigrateStatusVDSCommand] (ForkJoinPool-1-worker-4) []
> START, MigrateStatusVDSCommand(HostName = ginger.local.systea.fr,
> MigrateStatusVDSCommandParameters:{runAsync='true',
> hostId='d569c2dd-8f30-4878-8aea-858db285cf69',
> vmId='3f57e669-5e4c-4d10-85cc-d573004a099d'}), log id: 7a25c281
> 2018-02-12 16:46:42,174+01 INFO [org.ovirt.engine.core.
> vdsbroker.vdsbroker.MigrateStatusVDSCommand] (ForkJoinPool-1-worker-4) []
> FINISH, MigrateStatusVDSCommand, log id: 7a25c281
> 2018-02-12 16:46:42,194+01 INFO [org.ovirt.engine.core.dal.
> dbbroker.auditloghandling.AuditLogDirector] (ForkJoinPool-1-worker-4) []
> EVENT_ID: VM_MIGRATION_DONE(63), Correlation ID: 2f712024-5982-46a8-82c8-fd8293da5725,
> Job ID: 4bd19aa9-cc99-4d02-884e-5a1e857a7738, Call Stack: null, Custom
> ID: null, Custom Event ID: -1, Message: Migration completed (VM:
> Oracle_SECONDARY, Source: victor.local.systea.fr, Destination:
> ginger.local.systea.fr, Duration: 11 seconds, Total: 11 seconds, Actual
> downtime: (N/A))
> 2018-02-12 16:46:42,201+01 INFO [org.ovirt.engine.core.bll.MigrateVmToServerCommand]
> (ForkJoinPool-1-worker-4) [] Lock freed to object
> 'EngineLock:{exclusiveLocks='[3f57e669-5e4c-4d10-85cc-d573004a099d=VM]',
> sharedLocks=''}'
> 2018-02-12 16:46:42,203+01 INFO [org.ovirt.engine.core.
> vdsbroker.vdsbroker.FullListVDSCommand] (ForkJoinPool-1-worker-4) []
> START, FullListVDSCommand(HostName = ginger.local.systea.fr,
> FullListVDSCommandParameters:{runAsync='true', hostId='d569c2dd-8f30-4878-8aea-858db285cf69',
> vmIds='[3f57e669-5e4c-4d10-85cc-d573004a099d]'}), log id: 7cc65298
> 2018-02-12 16:46:42,254+01 INFO [org.ovirt.engine.core.
> vdsbroker.vdsbroker.FullListVDSCommand] (ForkJoinPool-1-worker-4) []
> FINISH, FullListVDSCommand, return: [{acpiEnable=true,
> emulatedMachine=pc-i440fx-rhel7.3.0, afterMigrationStatus=,
> tabletEnable=true, pid=18748, guestDiskMapping={},
> transparentHugePages=true, timeOffset=0, cpuType=Nehalem, smp=2,
> guestNumaNodes=[Ljava.lang.Object;@760085fd, custom={device_fbddd528-7d93-
> 49c6-a286-180e021cb274device_879c93ab-4df1-435c-af02-
> 565039fcc254=VmDevice:{id='VmDeviceId:{deviceId='879c93ab-4df1-435c-af02-565039fcc254',
> vmId='3f57e669-5e4c-4d10-85cc-d573004a099d'}', device='unix',
> type='CHANNEL', bootOrder='0', specParams='[]', address='{bus=0,
> controller=0, type=virtio-serial, port=1}', managed='false',
> plugged='true', readOnly='false', deviceAlias='channel0',
> customProperties='[]', snapshotId='null', logicalName='null',
> hostDevice='null'}, device_fbddd528-7d93-49c6-a286-180e021cb274device_
> 879c93ab-4df1-435c-af02-565039fcc254device_8945f61a-abbe-4156-8485-
> a4aa6f1908dbdevice_017b5e59-01c4-4aac-bf0c-b5d9557284d6=
> VmDevice:{id='VmDeviceId:{deviceId='017b5e59-01c4-4aac-bf0c-b5d9557284d6',
> vmId='3f57e669-5e4c-4d10-85cc-d573004a099d'}', device='tablet',
> type='UNKNOWN', bootOrder='0', specParams='[]', address='{bus=0, type=usb,
> port=1}', managed='false', plugged='true', readOnly='false',
> deviceAlias='input0', customProperties='[]', snapshotId='null',
> logicalName='null', hostDevice='null'}, device_fbddd528-7d93-49c6-
> a286-180e021cb274=VmDevice:{id='VmDeviceId:{deviceId='
> fbddd528-7d93-49c6-a286-180e021cb274', vmId='3f57e669-5e4c-4d10-85cc-d573004a099d'}',
> device='ide', type='CONTROLLER', bootOrder='0', specParams='[]',
> address='{slot=0x01, bus=0x00, domain=0x0000, type=pci, function=0x1}',
> managed='false', plugged='true', readOnly='false', deviceAlias='ide',
> customProperties='[]', snapshotId='null', logicalName='null',
> hostDevice='null'}, device_fbddd528-7d93-49c6-a286-180e021cb274device_
> 879c93ab-4df1-435c-af02-565039fcc254device_8945f61a-
> abbe-4156-8485-a4aa6f1908db=VmDevice:{id='VmDeviceId:{
> deviceId='8945f61a-abbe-4156-8485-a4aa6f1908db',
> vmId='3f57e669-5e4c-4d10-85cc-d573004a099d'}', device='unix',
> type='CHANNEL', bootOrder='0', specParams='[]', address='{bus=0,
> controller=0, type=virtio-serial, port=2}', managed='false',
> plugged='true', readOnly='false', deviceAlias='channel1',
> customProperties='[]', snapshotId='null', logicalName='null',
> hostDevice='null'}}, vmType=kvm, memSize=8192, smpCoresPerSocket=1,
> vmName=Oracle_SECONDARY, nice=0, status=Up, maxMemSize=32768,
> bootMenuEnable=false, vmId=3f57e669-5e4c-4d10-85cc-d573004a099d,
> numOfIoThreads=2, smpThreadsPerCore=1, smartcardEnable=false,
> maxMemSlots=16, kvmEnable=true, pitReinjection=false,
> displayNetwork=ovirtmgmt, devices=[Ljava.lang.Object;@2e4d3dd3,
> memGuaranteedSize=8192, maxVCpus=16, clientIp=, statusTime=4304259600
> <(430)%20425-9600>, display=vnc}], log id: 7cc65298
> 2018-02-12 16:46:42,257+01 INFO [org.ovirt.engine.core.
> vdsbroker.monitoring.VmDevicesMonitoring] (ForkJoinPool-1-worker-4) []
> Received a vnc Device without an address when processing VM
> 3f57e669-5e4c-4d10-85cc-d573004a099d devices, skipping device:
> {device=vnc, specParams={displayNetwork=ovirtmgmt, keyMap=fr,
> displayIp=192.168.0.5}, type=graphics, deviceId=813957b1-446a-4e88-9e40-9fe76d2c442d,
> port=5901}
> 2018-02-12 16:46:42,257+01 INFO [org.ovirt.engine.core.
> vdsbroker.monitoring.VmDevicesMonitoring] (ForkJoinPool-1-worker-4) []
> Received a lease Device without an address when processing VM
> 3f57e669-5e4c-4d10-85cc-d573004a099d devices, skipping device:
> {lease_id=3f57e669-5e4c-4d10-85cc-d573004a099d,
> sd_id=1e51cecc-eb2e-47d0-b185-920fdc7afa16, deviceId={uuid=a09949aa-5642-4b6d-94a4-8b0d04257be5},
> offset=6291456, device=lease, path=/rhev/data-center/mnt/
> glusterSD/192.168.0.6:_DATA01/1e51cecc-eb2e-47d0-b185-920fdc7afa16/dom_md/xleases,
> type=lease}
> 2018-02-12 16:46:46,260+01 INFO [org.ovirt.engine.core.
> vdsbroker.vdsbroker.FullListVDSCommand] (DefaultQuartzScheduler5)
> [7fcb200a] FINISH, FullListVDSCommand, return: [{acpiEnable=true,
> emulatedMachine=pc-i440fx-rhel7.3.0, afterMigrationStatus=,
> tabletEnable=true, pid=18748, guestDiskMapping={0QEMU_QEMU_
> HARDDISK_d890fa68-fba4-4f49-9={name=/dev/sda},
> QEMU_DVD-ROM_QM00003={name=/dev/sr0}}, transparentHugePages=true,
> timeOffset=0, cpuType=Nehalem, smp=2, guestNumaNodes=[Ljava.lang.
> Object;@77951faf, custom={device_fbddd528-7d93-
> 49c6-a286-180e021cb274device_879c93ab-4df1-435c-af02-
> 565039fcc254=VmDevice:{id='VmDeviceId:{deviceId='879c93ab-4df1-435c-af02-565039fcc254',
> vmId='3f57e669-5e4c-4d10-85cc-d573004a099d'}', device='unix',
> type='CHANNEL', bootOrder='0', specParams='[]', address='{bus=0,
> controller=0, type=virtio-serial, port=1}', managed='false',
> plugged='true', readOnly='false', deviceAlias='channel0',
> customProperties='[]', snapshotId='null', logicalName='null',
> hostDevice='null'}, device_fbddd528-7d93-49c6-a286-180e021cb274device_
> 879c93ab-4df1-435c-af02-565039fcc254device_8945f61a-abbe-4156-8485-
> a4aa6f1908dbdevice_017b5e59-01c4-4aac-bf0c-b5d9557284d6=
> VmDevice:{id='VmDeviceId:{deviceId='017b5e59-01c4-4aac-bf0c-b5d9557284d6',
> vmId='3f57e669-5e4c-4d10-85cc-d573004a099d'}', device='tablet',
> type='UNKNOWN', bootOrder='0', specParams='[]', address='{bus=0, type=usb,
> port=1}', managed='false', plugged='true', readOnly='false',
> deviceAlias='input0', customProperties='[]', snapshotId='null',
> logicalName='null', hostDevice='null'}, device_fbddd528-7d93-49c6-
> a286-180e021cb274=VmDevice:{id='VmDeviceId:{deviceId='
> fbddd528-7d93-49c6-a286-180e021cb274', vmId='3f57e669-5e4c-4d10-85cc-d573004a099d'}',
> device='ide', type='CONTROLLER', bootOrder='0', specParams='[]',
> address='{slot=0x01, bus=0x00, domain=0x0000, type=pci, function=0x1}',
> managed='false', plugged='true', readOnly='false', deviceAlias='ide',
> customProperties='[]', snapshotId='null', logicalName='null',
> hostDevice='null'}, device_fbddd528-7d93-49c6-a286-180e021cb274device_
> 879c93ab-4df1-435c-af02-565039fcc254device_8945f61a-
> abbe-4156-8485-a4aa6f1908db=VmDevice:{id='VmDeviceId:{
> deviceId='8945f61a-abbe-4156-8485-a4aa6f1908db',
> vmId='3f57e669-5e4c-4d10-85cc-d573004a099d'}', device='unix',
> type='CHANNEL', bootOrder='0', specParams='[]', address='{bus=0,
> controller=0, type=virtio-serial, port=2}', managed='false',
> plugged='true', readOnly='false', deviceAlias='channel1',
> customProperties='[]', snapshotId='null', logicalName='null',
> hostDevice='null'}}, vmType=kvm, memSize=8192, smpCoresPerSocket=1,
> vmName=Oracle_SECONDARY, nice=0, status=Up, maxMemSize=32768,
> bootMenuEnable=false, vmId=3f57e669-5e4c-4d10-85cc-d573004a099d,
> numOfIoThreads=2, smpThreadsPerCore=1, smartcardEnable=false,
> maxMemSlots=16, kvmEnable=true, pitReinjection=false,
> displayNetwork=ovirtmgmt, devices=[Ljava.lang.Object;@286410fd,
> memGuaranteedSize=8192, maxVCpus=16, clientIp=, statusTime=4304263620
> <(430)%20426-3620>, display=vnc}], log id: 58cdef4c
> 2018-02-12 16:46:46,267+01 INFO [org.ovirt.engine.core.
> vdsbroker.monitoring.VmDevicesMonitoring] (DefaultQuartzScheduler5)
> [7fcb200a] Received a vnc Device without an address when processing VM
> 3f57e669-5e4c-4d10-85cc-d573004a099d devices, skipping device:
> {device=vnc, specParams={displayNetwork=ovirtmgmt, keyMap=fr,
> displayIp=192.168.0.5}, type=graphics, deviceId=813957b1-446a-4e88-9e40-9fe76d2c442d,
> port=5901}
> 2018-02-12 16:46:46,268+01 INFO [org.ovirt.engine.core.
> vdsbroker.monitoring.VmDevicesMonitoring] (DefaultQuartzScheduler5)
> [7fcb200a] Received a lease Device without an address when processing VM
> 3f57e669-5e4c-4d10-85cc-d573004a099d devices, skipping device:
> {lease_id=3f57e669-5e4c-4d10-85cc-d573004a099d,
> sd_id=1e51cecc-eb2e-47d0-b185-920fdc7afa16, deviceId={uuid=a09949aa-5642-4b6d-94a4-8b0d04257be5},
> offset=6291456, device=lease, path=/rhev/data-center/mnt/
> glusterSD/192.168.0.6:_DATA01/1e51cecc-eb2e-47d0-b185-920fdc7afa16/dom_md/xleases,
> type=lease}
>
>
>
>
> For the VM with 2 vdisks we see :
>
> 2018-02-12 16:49:06,112+01 INFO [org.ovirt.engine.core.bll.MigrateVmToServerCommand]
> (default task-50) [92b5af33-cb87-4142-b8fe-8b838dd7458e] Lock Acquired to
> object 'EngineLock:{exclusiveLocks='[f7d4ec12-627a-4b83-b59e-886400d55474=VM]',
> sharedLocks=''}'
> 2018-02-12 16:49:06,407+01 INFO [org.ovirt.engine.core.bll.MigrateVmToServerCommand]
> (org.ovirt.thread.pool-6-thread-49) [92b5af33-cb87-4142-b8fe-8b838dd7458e]
> Running command: MigrateVmToServerCommand internal: false. Entities
> affected : ID: f7d4ec12-627a-4b83-b59e-886400d55474 Type: VMAction group
> MIGRATE_VM with role type USER
> 2018-02-12 16:49:06,712+01 INFO [org.ovirt.engine.core.vdsbroker.MigrateVDSCommand]
> (org.ovirt.thread.pool-6-thread-49) [92b5af33-cb87-4142-b8fe-8b838dd7458e]
> START, MigrateVDSCommand( MigrateVDSCommandParameters:{runAsync='true',
> hostId='d569c2dd-8f30-4878-8aea-858db285cf69',
> vmId='f7d4ec12-627a-4b83-b59e-886400d55474', srcHost='192.168.0.5',
> dstVdsId='ce3938b1-b23f-4d22-840a-f17d7cd87bb1', dstHost='
> 192.168.0.6:54321', migrationMethod='ONLINE', tunnelMigration='false',
> migrationDowntime='0', autoConverge='true', migrateCompressed='false',
> consoleAddress='null', maxBandwidth='500', enableGuestEvents='true',
> maxIncomingMigrations='2', maxOutgoingMigrations='2',
> convergenceSchedule='[init=[{name=setDowntime, params=[100]}],
> stalling=[{limit=1, action={name=setDowntime, params=[150]}}, {limit=2,
> action={name=setDowntime, params=[200]}}, {limit=3,
> action={name=setDowntime, params=[300]}}, {limit=4,
> action={name=setDowntime, params=[400]}}, {limit=6,
> action={name=setDowntime, params=[500]}}, {limit=-1, action={name=abort,
> params=[]}}]]'}), log id: 3702a9e0
> 2018-02-12 16:49:06,713+01 INFO [org.ovirt.engine.core.
> vdsbroker.vdsbroker.MigrateBrokerVDSCommand] (org.ovirt.thread.pool-6-thread-49)
> [92b5af33-cb87-4142-b8fe-8b838dd7458e] START, MigrateBrokerVDSCommand(HostName
> = ginger.local.systea.fr, MigrateVDSCommandParameters:{runAsync='true',
> hostId='d569c2dd-8f30-4878-8aea-858db285cf69',
> vmId='f7d4ec12-627a-4b83-b59e-886400d55474', srcHost='192.168.0.5',
> dstVdsId='ce3938b1-b23f-4d22-840a-f17d7cd87bb1', dstHost='
> 192.168.0.6:54321', migrationMethod='ONLINE', tunnelMigration='false',
> migrationDowntime='0', autoConverge='true', migrateCompressed='false',
> consoleAddress='null', maxBandwidth='500', enableGuestEvents='true',
> maxIncomingMigrations='2', maxOutgoingMigrations='2',
> convergenceSchedule='[init=[{name=setDowntime, params=[100]}],
> stalling=[{limit=1, action={name=setDowntime, params=[150]}}, {limit=2,
> action={name=setDowntime, params=[200]}}, {limit=3,
> action={name=setDowntime, params=[300]}}, {limit=4,
> action={name=setDowntime, params=[400]}}, {limit=6,
> action={name=setDowntime, params=[500]}}, {limit=-1, action={name=abort,
> params=[]}}]]'}), log id: 1840069c
> 2018-02-12 16:49:06,724+01 INFO [org.ovirt.engine.core.
> vdsbroker.vdsbroker.MigrateBrokerVDSCommand] (org.ovirt.thread.pool-6-thread-49)
> [92b5af33-cb87-4142-b8fe-8b838dd7458e] FINISH, MigrateBrokerVDSCommand,
> log id: 1840069c
> 2018-02-12 16:49:06,732+01 INFO [org.ovirt.engine.core.vdsbroker.MigrateVDSCommand]
> (org.ovirt.thread.pool-6-thread-49) [92b5af33-cb87-4142-b8fe-8b838dd7458e]
> FINISH, MigrateVDSCommand, return: MigratingFrom, log id: 3702a9e0
> 2018-02-12 16:49:06,753+01 INFO [org.ovirt.engine.core.dal.
> dbbroker.auditloghandling.AuditLogDirector] (org.ovirt.thread.pool-6-thread-49)
> [92b5af33-cb87-4142-b8fe-8b838dd7458e] EVENT_ID: VM_MIGRATION_START(62),
> Correlation ID: 92b5af33-cb87-4142-b8fe-8b838dd7458e, Job ID:
> f4f54054-f7c8-4481-8eda-d5a15c383061, Call Stack: null, Custom ID: null,
> Custom Event ID: -1, Message: Migration started (VM: Oracle_PRIMARY,
> Source: ginger.local.systea.fr, Destination: victor.local.systea.fr,
> User: admin at internal-authz).
> ...
> 2018-02-12 16:49:16,453+01 INFO [org.ovirt.engine.core.
> vdsbroker.monitoring.VmsStatisticsFetcher] (DefaultQuartzScheduler4)
> [162a5bc3] Fetched 2 VMs from VDS 'ce3938b1-b23f-4d22-840a-f17d7cd87bb1'
> 2018-02-12 16:49:16,455+01 INFO [org.ovirt.engine.core.
> vdsbroker.monitoring.VmAnalyzer] (DefaultQuartzScheduler4) [162a5bc3] VM
> 'f7d4ec12-627a-4b83-b59e-886400d55474'(Oracle_PRIMARY) was unexpectedly
> detected as 'MigratingTo' on VDS 'ce3938b1-b23f-4d22-840a-f17d7cd87bb1'(
> victor.local.systea.fr) (expected on 'd569c2dd-8f30-4878-8aea-
> 858db285cf69')
> 2018-02-12 16:49:16,455+01 INFO [org.ovirt.engine.core.
> vdsbroker.monitoring.VmAnalyzer] (DefaultQuartzScheduler4) [162a5bc3] VM
> 'f7d4ec12-627a-4b83-b59e-886400d55474' is migrating to VDS
> 'ce3938b1-b23f-4d22-840a-f17d7cd87bb1'(victor.local.systea.fr) ignoring
> it in the refresh until migration is done
> ...
> 2018-02-12 16:49:31,484+01 INFO [org.ovirt.engine.core.
> vdsbroker.monitoring.VmAnalyzer] (DefaultQuartzScheduler5) [11a7619a] VM
> 'f7d4ec12-627a-4b83-b59e-886400d55474'(Oracle_PRIMARY) was unexpectedly
> detected as 'MigratingTo' on VDS 'ce3938b1-b23f-4d22-840a-f17d7cd87bb1'(
> victor.local.systea.fr) (expected on 'd569c2dd-8f30-4878-8aea-
> 858db285cf69')
> 2018-02-12 16:49:31,484+01 INFO [org.ovirt.engine.core.
> vdsbroker.monitoring.VmAnalyzer] (DefaultQuartzScheduler5) [11a7619a] VM
> 'f7d4ec12-627a-4b83-b59e-886400d55474' is migrating to VDS
> 'ce3938b1-b23f-4d22-840a-f17d7cd87bb1'(victor.local.systea.fr) ignoring
> it in the refresh until migration is done
>
>
>
> and so on, last lines repeated indefinitly for hours since we poweroff the
> VM...
> Is this something known ? Any idea about that ?
>
> Thanks
>
> Ovirt 4.1.6, updated last at feb-13. Gluster 3.12.1.
>
> --
>
> Cordialement,
>
> *Frank Soyer *
>
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20180214/0cb56555/attachment.html>
More information about the Users
mailing list