I have a hardware node that for whatever reason most of the VM
status was "?", even though VM was up and running fine.
node was running ovirt 3.6 so I decided to upgrade it to 4.1.0.4-1
like most of the rest of the cluster is, including engine.
After upgrade some VMs I'm able to start up fine on the upgraded
hosts, other fail to start on that host and migrate themselves to
another host.
If I try to manually migrate them to the recently upgraded host
migration fails. I have yet to find in the logs where it says why
they fail.
I am able to "runonce" the VM and tell it to start on this host and
it starts fine.
Why is migration failing?
ovirt-engine-4.1.0.4-1.el7.centos.noarch
vdsm-4.19.4-1.el7.centos.x86_64
2017-03-04 10:16:51,716-08 ERROR
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(org.ovirt.thread.pool-7-thread-48)
[519095c3-83ed-4478-8dd4-f432db6db140] EVENT_ID:
VM_MIGRATION_TO_SERVER_FAILED(120), Correlation ID: 2
6d1ecba-8711-408d-b9d5-6461a4aab4e5, Job ID:
546e0e11-9b4b-48d4-bd50-2f525d049ae2, Call Stack: null, Custom Event
ID: -1, Message: Migration failedĀ (VM: j2es2.test.j2noc.com,
Source: ovirt6.test.j2noc.com, Destination: ovirt4.test.j2noc.com)
(engine.log attached)
Thanks.