
Hello, I Tried this solution , but to some vms not was resolved. Logs: src logs: Thread-12::DEBUG::2016-12-06 08:50:58,112::check::327::storage.check::(_check_completed) FINISH check u'/rhev/data-center/mnt/192.168.144.6:_home_iso/b5fa054f-0d3d-458b-a891-13fd9383ee7d/dom_md/metadata' rc=0 err=bytearray(b'0+1 records in\n0+1 records out\n328 bytes (328 B) copied, 0.000474535 s, 691 kB/s\n') elapsed=0.08 Thread-2374888::WARNING::2016-12-06 08:50:58,815::migration::671::virt.vm::(monitor_migration) vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::The migration took 520 seconds which is exceeding the configured maximum time for migrations of 512 seconds. The migration will be aborted. Thread-2374888::DEBUG::2016-12-06 08:50:58,816::migration::715::virt.vm::(stop) vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopping migration monitor thread Thread-2374888::DEBUG::2016-12-06 08:50:58,816::migration::570::virt.vm::(stop) vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopping migration downtime thread Thread-2374888::DEBUG::2016-12-06 08:50:58,817::migration::629::virt.vm::(run) vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopped migration monitor thread Thread-2374886::DEBUG::2016-12-06 08:50:59,098::migration::715::virt.vm::(stop) vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopping migration monitor thread Thread-2374886::ERROR::2016-12-06 08:50:59,098::migration::252::virt.vm::(_recover) vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::operation aborted: migration job: canceled by client Thread-2374886::DEBUG::2016-12-06 08:50:59,098::stompreactor::408::jsonrpc.AsyncoreClient::(send) Sending response Thread-2374886::DEBUG::2016-12-06 08:50:59,321::__init__::208::jsonrpc.Notification::(emit) Sending event {"params": {"notify_time": 6040272640, "9b5ab7b4-1045-4858-b24c-1f5a9f6172c3": {"status": "Migration Source"}}, "jsonrpc": "2.0", "method": "|virt|VM_status|9b5ab7b4-1045-4858-b24c-1f5a9f6172c3"} Thread-2374886::ERROR::2016-12-06 08:50:59,322::migration::381::virt.vm::(run) vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::Failed to migrate Traceback (most recent call last): File "/usr/share/vdsm/virt/migration.py", line 363, in run self._startUnderlyingMigration(time.time()) File "/usr/share/vdsm/virt/migration.py", line 438, in _startUnderlyingMigration self._perform_with_downtime_thread(duri, muri) File "/usr/share/vdsm/virt/migration.py", line 489, in _perform_with_downtime_thread self._perform_migration(duri, muri) File "/usr/share/vdsm/virt/migration.py", line 476, in _perform_migration self._vm._dom.migrateToURI3(duri, params, flags) File "/usr/lib/python2.7/site-packages/vdsm/virt/virdomain.py", line 69, in f ret = attr(*args, **kwargs) File "/usr/lib/python2.7/site-packages/vdsm/libvirtconnection.py", line 123, in wrapper ret = f(*args, **kwargs) File "/usr/lib/python2.7/site-packages/vdsm/utils.py", line 916, in wrapper return func(inst, *args, **kwargs) File "/usr/lib64/python2.7/site-packages/libvirt.py", line 1836, in migrateToURI3 if ret == -1: raise libvirtError ('virDomainMigrateToURI3() failed', dom=self) libvirtError: operation aborted: migration job: canceled by client Thread-12::DEBUG::2016-12-06 08:50:59,875::check::296::storage.check::(_start_process) START check '/dev/c2dc0101-748e-4a7b-9913-47993eaa52bd/metadata' cmd=['/usr/bin/taskset', '--cpu-list', '0-31', '/usr/bin/dd', 'if=/dev/c2dc0101-748e-4a7b-9913-47993eaa52bd/metadata', 'of=/dev/null', 'bs=4096', 'count=1', 'iflag=direct'] delay=0.00 mailbox.SPMMonitor::DEBUG::2016-12-06 08:50:59,914::storage_mailbox::733::Storage.Misc.excCmd::(_checkForMail) /usr/bin/taskset --cpu-list 0-31 dd if=/rhev/data-center/77e24b20-9d21-4952-a089-3c5c592b4e6d/mastersd/dom_md/inbox iflag=direct,fullblock count=1 bs=1024000 (cwd None) dst log: libvirtEventLoop::DEBUG::2016-12-06 08:50:59,080::task::995::Storage.TaskManager.Task::(_decref) Task=`7446f040-c5f4-497c-b4a5-8934921a7b89`::ref 1 aborting False libvirtEventLoop::DEBUG::2016-12-06 08:50:59,080::fileUtils::190::Storage.fileUtils::(cleanupdir) Removing directory: /var/run/vdsm/storage/6e5cce71-3438-4045-9d54-6071 23e0557e/413a560d-4919-4870-88d6-f7fedbb77523 libvirtEventLoop::DEBUG::2016-12-06 08:50:59,080::lvm::288::Storage.Misc.excCmd::(cmd) /usr/bin/taskset --cpu-list 0-31 /usr/bin/sudo -n /usr/sbin/lvm lvs --config ' de vices { preferred_names = ["^/dev/mapper/"] ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3 filter = [ '\''a|/dev/mapper/36005076300810a4db80 0000000000002|'\'', '\''r|.*|'\'' ] } global { locking_type=1 prioritise_write_locks=1 wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days = 0 } ' --noheadings --units b --nosuffix --separator '|' --ignoreskippedcluster -o uuid,name,vg_name,attr,size,seg_start_pe,devices,tags 6e5cce71-3438-4045-9d54-607123e05 57e (cwd None) jsonrpc.Executor/6::DEBUG::2016-12-06 08:50:59,090::__init__::530::jsonrpc.JsonRpcServer::(_handle_request) Calling 'VM.destroy' in bridge with {u'vmID': u'9b5ab7b4-104 5-4858-b24c-1f5a9f6172c3'} jsonrpc.Executor/6::DEBUG::2016-12-06 08:50:59,090::API::314::vds::(destroy) About to destroy VM 9b5ab7b4-1045-4858-b24c-1f5a9f6172c3 jsonrpc.Executor/6::DEBUG::2016-12-06 08:50:59,091::vm::4171::virt.vm::(destroy) vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::destroy Called Thread-51550::ERROR::2016-12-06 08:50:59,091::vm::759::virt.vm::(_startUnderlyingVm) vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::Failed to start a migration destinatio n vm Traceback (most recent call last): File "/usr/share/vdsm/virt/vm.py", line 725, in _startUnderlyingVm self._completeIncomingMigration() File "/usr/share/vdsm/virt/vm.py", line 3071, in _completeIncomingMigration self._incomingMigrationFinished.isSet(), usedTimeout) File "/usr/share/vdsm/virt/vm.py", line 3154, in _attachLibvirtDomainAfterMigration raise MigrationError(e.get_error_message()) MigrationError: Domain not found: no domain with matching uuid '9b5ab7b4-1045-4858-b24c-1f5a9f6172c3' Thread-51550::INFO::2016-12-06 08:50:59,093::vm::1308::virt.vm::(setDownStatus) vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::Changed state to Down: VM failed to migrate (code=8) Logs Attached. any ideas? Very thanks. 2016-12-06 4:16 GMT-03:00 Yedidyah Bar David <didi@redhat.com>:
On Mon, Dec 5, 2016 at 10:25 PM, Marcelo Leandro <marceloltmm@gmail.com> wrote:
Hello I am with problem, I can not migrate vm. Can someone help me?
Message in log:
Src vdsm log:
Thread-89514 :: WARNING :: 2016-12-05 17: 01: 49,542 :: migration :: 683 :: virt.vm ::: monitor_migration vmId = `f38b9f7d-5bd0-4bdf-885c- e03e8d6bc70e` :: Migration Stalling: remaining (56MiB)> lowmark (2MiB). Refer to RHBZ # 919201. Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,543 :: migration :: 689 :: virt.vm ::: monitor_migration vmId = `f38b9f7d-5bd0-4bdf-885c- e03e8d6bc70e` :: new Iteration detected: 15 Thread-89514 :: WARNING :: 2016-12-05 17: 01: 49,543 :: migration :: 704 :: virt.vm ::: monitor_migration vmId = `f38b9f7d-5bd0-4bdf-885c- e03e8d6bc70e` :: Migration Is stuck: Has not pro Gressed in 240.071660042 seconds. Aborting.
This is usually a result of a too-busy VM, changing its memory faster than the migration process can copy the changes to the destination.
You can try changing the cluster migration policy to "suspend workload if needed".
For more details/background, see also:
https://www.ovirt.org/develop/release-management/features/ migration-enhancements/
Best,
Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,544 :: migration :: 715 :: virt.vm ::: stop} vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` :: stopping Migration monitor thread Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,545 :: migration :: 570 :: virt.vm ::: stop) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` :: stopping Migration downtime thread Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,545 :: migration :: 629 :: virt.vm ::: (run) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` :: stopped Migration monitor thread Thread-89513 :: DEBUG :: 2016-12-05 17: 01: 49,766 :: migration :: 715 :: virt.vm ::: stop} vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` :: stopping Migration monitor thread Thread-89513 :: ERROR :: 2016-12-05 17: 01: 49,767 :: migration :: 252 :: virt.vm :: (_ recover) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` :: operation Aborted: migration job: cancel D by cliente
Dst vdsm.log:
Periodic / 13 :: WARNING :: 2016-12-05 17: 01: 49,791 :: sampling :: 483 :: virt.sampling.StatsCache: :( put) dropped stale old sample: sampled 4303678.080000 stored 4303693.070000 Periodic / 13 :: DEBUG :: 2016-12-05 17: 01: 49,791 :: executor :: 221 :: Executor :: (_ run) Worker was discarded Jsonrpc.Executor / 0 :: DEBUG :: 2016-12-05 17: 01: 49,792 :: __ init __ :: 530 :: jsonrpc.JsonRpcServer :: (_ handle_request) Calling 'VM.destroy' in bridge with {u'vmID ' : U'f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e '} Jsonrpc.Executor / 0 :: DEBUG :: 2016-12-05 17: 01: 49,793 :: API :: 314 :: vds :::( destroy) About to destroy VM f38b9f7d-5bd0-4bdf-885c- e03e8d6bc70e Jsonrpc.Executor / 0 :: DEBUG :: 2016-12-05 17:01:49,793 :: vm :: 4171 :: virt.vm :::( destroy) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e`: : Destroy Called Thread-483 :: ERROR :: 2016-12-05 17: 01: 49,793 :: vm :: 759 :: virt.vm :: (_ startUnderlyingVm) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` :: Failed To start a migration destination vm Traceback (most recent call last): File "/usr/share/vdsm/virt/vm.py", line 725, in _startUnderlyingVm Self._completeIncomingMigration () File "/usr/share/vdsm/virt/vm.py", line 3071, in _completeIncomingMigration Self._incomingMigrationFinished.isSet (), usedTimeout) File "/usr/share/vdsm/virt/vm.py", line 3154, in _attachLibvirtDomainAfterMigration Raise MigrationError (e.get_error_message ()) MigrationError: Domain not found: no domain with matching uuid 'f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e'
The logs attached. Thanks.
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
-- Didi