Hello,
I Tried this solution , but to some vms not was resolved.
Logs:
src logs:
Thread-12::DEBUG::2016-12-06
08:50:58,112::check::327::storage.check::(_check_completed) FINISH check
u'/rhev/data-center/mnt/192.168.144.6:_home_iso/b5fa054f-0d3d-458b-a891-13fd9383ee7d/dom_md/metadata'
rc=0 err=bytearray(b'0+1 records in\n0+1 records out\n328 bytes (328 B)
copied, 0.000474535 s, 691 kB/s\n') elapsed=0.08
Thread-2374888::WARNING::2016-12-06
08:50:58,815::migration::671::virt.vm::(monitor_migration)
vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::The migration took 520 seconds
which is exceeding the configured maximum time for migrations of 512
seconds. The migration will be aborted.
Thread-2374888::DEBUG::2016-12-06
08:50:58,816::migration::715::virt.vm::(stop)
vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopping migration monitor
thread
Thread-2374888::DEBUG::2016-12-06
08:50:58,816::migration::570::virt.vm::(stop)
vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopping migration downtime
thread
Thread-2374888::DEBUG::2016-12-06
08:50:58,817::migration::629::virt.vm::(run)
vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopped migration monitor
thread
Thread-2374886::DEBUG::2016-12-06
08:50:59,098::migration::715::virt.vm::(stop)
vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopping migration monitor
thread
Thread-2374886::ERROR::2016-12-06
08:50:59,098::migration::252::virt.vm::(_recover)
vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::operation aborted: migration
job: canceled by client
Thread-2374886::DEBUG::2016-12-06
08:50:59,098::stompreactor::408::jsonrpc.AsyncoreClient::(send) Sending
response
Thread-2374886::DEBUG::2016-12-06
08:50:59,321::__init__::208::jsonrpc.Notification::(emit) Sending event
{"params": {"notify_time": 6040272640,
"9b5ab7b4-1045-4858-b24c-1f5a9f6172c3": {"status": "Migration
Source"}},
"jsonrpc": "2.0", "method":
"|virt|VM_status|9b5ab7b4-1045-4858-b24c-1f5a9f6172c3"}
Thread-2374886::ERROR::2016-12-06
08:50:59,322::migration::381::virt.vm::(run)
vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::Failed to migrate
Traceback (most recent call last):
File "/usr/share/vdsm/virt/migration.py", line 363, in run
self._startUnderlyingMigration(time.time())
File "/usr/share/vdsm/virt/migration.py", line 438, in
_startUnderlyingMigration
self._perform_with_downtime_thread(duri, muri)
File "/usr/share/vdsm/virt/migration.py", line 489, in
_perform_with_downtime_thread
self._perform_migration(duri, muri)
File "/usr/share/vdsm/virt/migration.py", line 476, in _perform_migration
self._vm._dom.migrateToURI3(duri, params, flags)
File "/usr/lib/python2.7/site-packages/vdsm/virt/virdomain.py", line 69,
in f
ret = attr(*args, **kwargs)
File "/usr/lib/python2.7/site-packages/vdsm/libvirtconnection.py", line
123, in wrapper
ret = f(*args, **kwargs)
File "/usr/lib/python2.7/site-packages/vdsm/utils.py", line 916, in
wrapper
return func(inst, *args, **kwargs)
File "/usr/lib64/python2.7/site-packages/libvirt.py", line 1836, in
migrateToURI3
if ret == -1: raise libvirtError ('virDomainMigrateToURI3() failed',
dom=self)
libvirtError: operation aborted: migration job: canceled by client
Thread-12::DEBUG::2016-12-06
08:50:59,875::check::296::storage.check::(_start_process) START check
'/dev/c2dc0101-748e-4a7b-9913-47993eaa52bd/metadata'
cmd=['/usr/bin/taskset', '--cpu-list', '0-31',
'/usr/bin/dd',
'if=/dev/c2dc0101-748e-4a7b-9913-47993eaa52bd/metadata', 'of=/dev/null',
'bs=4096', 'count=1', 'iflag=direct'] delay=0.00
mailbox.SPMMonitor::DEBUG::2016-12-06
08:50:59,914::storage_mailbox::733::Storage.Misc.excCmd::(_checkForMail)
/usr/bin/taskset --cpu-list 0-31 dd
if=/rhev/data-center/77e24b20-9d21-4952-a089-3c5c592b4e6d/mastersd/dom_md/inbox
iflag=direct,fullblock count=1 bs=1024000 (cwd None)
dst log:
libvirtEventLoop::DEBUG::2016-12-06
08:50:59,080::task::995::Storage.TaskManager.Task::(_decref)
Task=`7446f040-c5f4-497c-b4a5-8934921a7b89`::ref 1 aborting False
libvirtEventLoop::DEBUG::2016-12-06
08:50:59,080::fileUtils::190::Storage.fileUtils::(cleanupdir) Removing
directory: /var/run/vdsm/storage/6e5cce71-3438-4045-9d54-6071
23e0557e/413a560d-4919-4870-88d6-f7fedbb77523
libvirtEventLoop::DEBUG::2016-12-06
08:50:59,080::lvm::288::Storage.Misc.excCmd::(cmd) /usr/bin/taskset
--cpu-list 0-31 /usr/bin/sudo -n /usr/sbin/lvm lvs --config ' de
vices { preferred_names = ["^/dev/mapper/"] ignore_suspended_devices=1
write_cache_state=0 disable_after_error_count=3 filter = [
'\''a|/dev/mapper/36005076300810a4db80
0000000000002|'\'', '\''r|.*|'\'' ] } global {
locking_type=1
prioritise_write_locks=1 wait_for_locks=1 use_lvmetad=0 } backup {
retain_min = 50 retain_days = 0
} ' --noheadings --units b --nosuffix --separator '|'
--ignoreskippedcluster -o
uuid,name,vg_name,attr,size,seg_start_pe,devices,tags
6e5cce71-3438-4045-9d54-607123e05
57e (cwd None)
jsonrpc.Executor/6::DEBUG::2016-12-06
08:50:59,090::__init__::530::jsonrpc.JsonRpcServer::(_handle_request)
Calling 'VM.destroy' in bridge with {u'vmID': u'9b5ab7b4-104
5-4858-b24c-1f5a9f6172c3'}
jsonrpc.Executor/6::DEBUG::2016-12-06
08:50:59,090::API::314::vds::(destroy) About to destroy VM
9b5ab7b4-1045-4858-b24c-1f5a9f6172c3
jsonrpc.Executor/6::DEBUG::2016-12-06
08:50:59,091::vm::4171::virt.vm::(destroy)
vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::destroy Called
Thread-51550::ERROR::2016-12-06
08:50:59,091::vm::759::virt.vm::(_startUnderlyingVm)
vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::Failed to start a migration
destinatio
n vm
Traceback (most recent call last):
File "/usr/share/vdsm/virt/vm.py", line 725, in _startUnderlyingVm
self._completeIncomingMigration()
File "/usr/share/vdsm/virt/vm.py", line 3071, in
_completeIncomingMigration
self._incomingMigrationFinished.isSet(), usedTimeout)
File "/usr/share/vdsm/virt/vm.py", line 3154, in
_attachLibvirtDomainAfterMigration
raise MigrationError(e.get_error_message())
MigrationError: Domain not found: no domain with matching uuid
'9b5ab7b4-1045-4858-b24c-1f5a9f6172c3'
Thread-51550::INFO::2016-12-06
08:50:59,093::vm::1308::virt.vm::(setDownStatus)
vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::Changed state to Down: VM
failed to migrate (code=8)
Logs Attached.
any ideas?
Very thanks.
2016-12-06 4:16 GMT-03:00 Yedidyah Bar David <didi(a)redhat.com>:
On Mon, Dec 5, 2016 at 10:25 PM, Marcelo Leandro
<marceloltmm(a)gmail.com>
wrote:
> Hello
> I am with problem, I can not migrate vm.
> Can someone help me?
>
> Message in log:
>
> Src vdsm log:
>
> Thread-89514 :: WARNING :: 2016-12-05 17: 01: 49,542 :: migration ::
683 ::
> virt.vm ::: monitor_migration vmId = `f38b9f7d-5bd0-4bdf-885c-
e03e8d6bc70e`
> :: Migration Stalling: remaining
> (56MiB)> lowmark (2MiB). Refer to RHBZ # 919201.
> Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,543 :: migration :: 689 ::
> virt.vm ::: monitor_migration vmId = `f38b9f7d-5bd0-4bdf-885c-
e03e8d6bc70e`
> :: new Iteration detected: 15
> Thread-89514 :: WARNING :: 2016-12-05 17: 01: 49,543 :: migration :: 704
::
> virt.vm ::: monitor_migration vmId = `f38b9f7d-5bd0-4bdf-885c-
e03e8d6bc70e`
> :: Migration Is stuck: Has not pro
> Gressed in 240.071660042 seconds. Aborting.
This is usually a result of a too-busy VM, changing its memory faster than
the migration process can copy the changes to the destination.
You can try changing the cluster migration policy to "suspend workload
if needed".
For more details/background, see also:
https://www.ovirt.org/develop/release-management/features/
migration-enhancements/
Best,
> Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,544 :: migration :: 715
::
> virt.vm ::: stop} vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` ::
stopping
> Migration monitor thread
> Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,545 :: migration :: 570 ::
> virt.vm ::: stop) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` ::
stopping
> Migration downtime thread
> Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,545 :: migration :: 629 ::
> virt.vm ::: (run) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` ::
stopped
> Migration monitor thread
> Thread-89513 :: DEBUG :: 2016-12-05 17: 01: 49,766 :: migration :: 715 ::
> virt.vm ::: stop} vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` ::
stopping
> Migration monitor thread
> Thread-89513 :: ERROR :: 2016-12-05 17: 01: 49,767 :: migration :: 252 ::
> virt.vm :: (_ recover) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` ::
> operation Aborted: migration job: cancel
> D by cliente
>
> Dst vdsm.log:
>
> Periodic / 13 :: WARNING :: 2016-12-05 17: 01: 49,791 :: sampling :: 483
::
> virt.sampling.StatsCache: :( put) dropped stale old sample: sampled
> 4303678.080000 stored 4303693.070000
> Periodic / 13 :: DEBUG :: 2016-12-05 17: 01: 49,791 :: executor :: 221 ::
> Executor :: (_ run) Worker was discarded
> Jsonrpc.Executor / 0 :: DEBUG :: 2016-12-05 17: 01: 49,792 :: __ init __
::
> 530 :: jsonrpc.JsonRpcServer :: (_ handle_request) Calling 'VM.destroy'
in
> bridge with {u'vmID ' : U'f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e '}
> Jsonrpc.Executor / 0 :: DEBUG :: 2016-12-05 17: 01: 49,793 :: API :: 314
::
> vds :::( destroy) About to destroy VM f38b9f7d-5bd0-4bdf-885c-
e03e8d6bc70e
> Jsonrpc.Executor / 0 :: DEBUG :: 2016-12-05 17:01:49,793 :: vm :: 4171 ::
> virt.vm :::( destroy) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e`: :
> Destroy Called
> Thread-483 :: ERROR :: 2016-12-05 17: 01: 49,793 :: vm :: 759 :: virt.vm
::
> (_ startUnderlyingVm) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` ::
> Failed To start a migration destination vm
> Traceback (most recent call last):
> File "/usr/share/vdsm/virt/vm.py", line 725, in _startUnderlyingVm
> Self._completeIncomingMigration ()
> File "/usr/share/vdsm/virt/vm.py", line 3071, in
> _completeIncomingMigration
> Self._incomingMigrationFinished.isSet (), usedTimeout)
> File "/usr/share/vdsm/virt/vm.py", line 3154, in
> _attachLibvirtDomainAfterMigration
> Raise MigrationError (e.get_error_message ())
> MigrationError: Domain not found: no domain with matching uuid
> 'f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e'
>
> The logs attached.
> Thanks.
>
> _______________________________________________
> Users mailing list
> Users(a)ovirt.org
>
http://lists.ovirt.org/mailman/listinfo/users
>
--
Didi