This snippet is not enough, and the attached logs are too new. Can youOn Tue, Dec 6, 2016 at 2:14 PM, Marcelo Leandro <marceloltmm@gmail.com> wrote:
> Hello,
>
> I Tried this solution , but to some vms not was resolved.
> Logs:
>
> src logs:
>
> Thread-12::DEBUG::2016-12-06
> 08:50:58,112::check::327::storage.check::(_check_ completed) FINISH check
> u'/rhev/data-center/mnt/192.168.144.6:_home_iso/b5fa054f- 0d3d-458b-a891-13fd9383ee7d/ dom_md/metadata'
> rc=0 err=bytearray(b'0+1 records in\n0+1 records out\n328 bytes (328 B)
> copied, 0.000474535 s, 691 kB/s\n') elapsed=0.08
> Thread-2374888::WARNING::2016-12-06
> 08:50:58,815::migration::671::virt.vm::(monitor_migration)
> vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::The migration took 520 seconds
> which is exceeding the configured maximum time for migrations of 512
> seconds. The migration will be aborted.
> Thread-2374888::DEBUG::2016-12-06
> 08:50:58,816::migration::715::virt.vm::(stop)
> vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopping migration monitor
> thread
> Thread-2374888::DEBUG::2016-12-06
> 08:50:58,816::migration::570::virt.vm::(stop)
> vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopping migration downtime
> thread
> Thread-2374888::DEBUG::2016-12-06
> 08:50:58,817::migration::629::virt.vm::(run)
> vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopped migration monitor
> thread
> Thread-2374886::DEBUG::2016-12-06
> 08:50:59,098::migration::715::virt.vm::(stop)
> vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopping migration monitor
> thread
> Thread-2374886::ERROR::2016-12-06
> 08:50:59,098::migration::252::virt.vm::(_recover)
> vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::operation aborted: migration
> job: canceled by client
> Thread-2374886::DEBUG::2016-12-06
> 08:50:59,098::stompreactor::408::jsonrpc.AsyncoreClient::( send) Sending
> response
> Thread-2374886::DEBUG::2016-12-06
> 08:50:59,321::__init__::208::jsonrpc.Notification::(emit) Sending event
> {"params": {"notify_time": 6040272640,
> "9b5ab7b4-1045-4858-b24c-1f5a9f6172c3": {"status": "Migration Source"}},
> "jsonrpc": "2.0", "method":
> "|virt|VM_status|9b5ab7b4-1045-4858-b24c-1f5a9f6172c3"}
> Thread-2374886::ERROR::2016-12-06
> 08:50:59,322::migration::381::virt.vm::(run)
> vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::Failed to migrate
> Traceback (most recent call last):
> File "/usr/share/vdsm/virt/migration.py", line 363, in run
> self._startUnderlyingMigration(time. time())
> File "/usr/share/vdsm/virt/migration.py", line 438, in
> _startUnderlyingMigration
> self._perform_with_downtime_thread(duri, muri)
> File "/usr/share/vdsm/virt/migration.py", line 489, in
> _perform_with_downtime_thread
> self._perform_migration(duri, muri)
> File "/usr/share/vdsm/virt/migration.py", line 476, in _perform_migration
> self._vm._dom.migrateToURI3(duri, params, flags)
> File "/usr/lib/python2.7/site-packages/vdsm/virt/virdomain. py", line 69,
> in f
> ret = attr(*args, **kwargs)
> File "/usr/lib/python2.7/site-packages/vdsm/ libvirtconnection.py", line
> 123, in wrapper
> ret = f(*args, **kwargs)
> File "/usr/lib/python2.7/site-packages/vdsm/utils.py", line 916, in
> wrapper
> return func(inst, *args, **kwargs)
> File "/usr/lib64/python2.7/site-packages/libvirt.py", line 1836, in
> migrateToURI3
> if ret == -1: raise libvirtError ('virDomainMigrateToURI3() failed',
> dom=self)
> libvirtError: operation aborted: migration job: canceled by client
> Thread-12::DEBUG::2016-12-06
> 08:50:59,875::check::296::storage.check::(_start_ process) START check
> '/dev/c2dc0101-748e-4a7b-9913-47993eaa52bd/metadata'
> cmd=['/usr/bin/taskset', '--cpu-list', '0-31', '/usr/bin/dd',
> 'if=/dev/c2dc0101-748e-4a7b-9913-47993eaa52bd/metadata', 'of=/dev/null',
> 'bs=4096', 'count=1', 'iflag=direct'] delay=0.00
> mailbox.SPMMonitor::DEBUG::2016-12-06
> 08:50:59,914::storage_mailbox::733::Storage.Misc.excCmd::(_ checkForMail)
> /usr/bin/taskset --cpu-list 0-31 dd
> if=/rhev/data-center/77e24b20-9d21-4952-a089-3c5c592b4e6d/ mastersd/dom_md/inbox
> iflag=direct,fullblock count=1 bs=1024000 (cwd None)
check/share
more of the relevant log? Thanks.
--
>
> dst log:
>
> libvirtEventLoop::DEBUG::2016-12-06
> 08:50:59,080::task::995::Storage.TaskManager.Task::(_ decref)
> Task=`7446f040-c5f4-497c-b4a5-8934921a7b89`::ref 1 aborting False
> libvirtEventLoop::DEBUG::2016-12-06
> 08:50:59,080::fileUtils::190::Storage.fileUtils::( cleanupdir) Removing
> directory: /var/run/vdsm/storage/6e5cce71-3438-4045-9d54-6071
> 23e0557e/413a560d-4919-4870-88d6-f7fedbb77523
> libvirtEventLoop::DEBUG::2016-12-06
> 08:50:59,080::lvm::288::Storage.Misc.excCmd::(cmd) /usr/bin/taskset
> --cpu-list 0-31 /usr/bin/sudo -n /usr/sbin/lvm lvs --config ' de
> vices { preferred_names = ["^/dev/mapper/"] ignore_suspended_devices=1
> write_cache_state=0 disable_after_error_count=3 filter = [
> '\''a|/dev/mapper/36005076300810a4db80
> 0000000000002|'\'', '\''r|.*|'\'' ] } global { locking_type=1
> prioritise_write_locks=1 wait_for_locks=1 use_lvmetad=0 } backup {
> retain_min = 50 retain_days = 0
> } ' --noheadings --units b --nosuffix --separator '|'
> --ignoreskippedcluster -o
> uuid,name,vg_name,attr,size,seg_start_pe,devices,tags
> 6e5cce71-3438-4045-9d54-607123e05
> 57e (cwd None)
> jsonrpc.Executor/6::DEBUG::2016-12-06
> 08:50:59,090::__init__::530::jsonrpc.JsonRpcServer::(_ handle_request)
> Calling 'VM.destroy' in bridge with {u'vmID': u'9b5ab7b4-104
> 5-4858-b24c-1f5a9f6172c3'}
> jsonrpc.Executor/6::DEBUG::2016-12-06 08:50:59,090::API::314::vds::( destroy)
> About to destroy VM 9b5ab7b4-1045-4858-b24c-1f5a9f6172c3
> jsonrpc.Executor/6::DEBUG::2016-12-06
> 08:50:59,091::vm::4171::virt.vm::(destroy)
> vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::destroy Called
> Thread-51550::ERROR::2016-12-06
> 08:50:59,091::vm::759::virt.vm::(_startUnderlyingVm)
> vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::Failed to start a migration
> destinatio
> n vm
> Traceback (most recent call last):
> File "/usr/share/vdsm/virt/vm.py", line 725, in _startUnderlyingVm
> self._completeIncomingMigration()
> File "/usr/share/vdsm/virt/vm.py", line 3071, in
> _completeIncomingMigration
> self._incomingMigrationFinished. isSet(), usedTimeout)
> File "/usr/share/vdsm/virt/vm.py", line 3154, in
> _attachLibvirtDomainAfterMigrat ion
> raise MigrationError(e.get_error_message())
> MigrationError: Domain not found: no domain with matching uuid
> '9b5ab7b4-1045-4858-b24c-1f5a9f6172c3'
> Thread-51550::INFO::2016-12-06
> 08:50:59,093::vm::1308::virt.vm::(setDownStatus)
> vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::Changed state to Down: VM
> failed to migrate (code=8)
>
>
>
> Logs Attached.
>
> any ideas?
>
> Very thanks.
>
>
> 2016-12-06 4:16 GMT-03:00 Yedidyah Bar David <didi@redhat.com>:
>>
>> On Mon, Dec 5, 2016 at 10:25 PM, Marcelo Leandro <marceloltmm@gmail.com>
>> wrote:
>> > Hello
>> > I am with problem, I can not migrate vm.
>> > Can someone help me?
>> >
>> > Message in log:
>> >
>> > Src vdsm log:
>> >
>> > Thread-89514 :: WARNING :: 2016-12-05 17: 01: 49,542 :: migration :: 683
>> > ::
>> > virt.vm ::: monitor_migration vmId =
>> > `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e`
>> > :: Migration Stalling: remaining
>> > (56MiB)> lowmark (2MiB). Refer to RHBZ # 919201.
>> > Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,543 :: migration :: 689
>> > ::
>> > virt.vm ::: monitor_migration vmId =
>> > `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e`
>> > :: new Iteration detected: 15
>> > Thread-89514 :: WARNING :: 2016-12-05 17: 01: 49,543 :: migration :: 704
>> > ::
>> > virt.vm ::: monitor_migration vmId =
>> > `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e`
>> > :: Migration Is stuck: Has not pro
>> > Gressed in 240.071660042 seconds. Aborting.
>>
>> This is usually a result of a too-busy VM, changing its memory faster than
>> the migration process can copy the changes to the destination.
>>
>> You can try changing the cluster migration policy to "suspend workload
>> if needed".
>>
>> For more details/background, see also:
>>
>>
>> https://www.ovirt.org/develop/release-management/features/ migration-enhancements/
>>
>> Best,
>>
>> > Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,544 :: migration :: 715
>> > ::
>> > virt.vm ::: stop} vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` ::
>> > stopping
>> > Migration monitor thread
>> > Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,545 :: migration :: 570
>> > ::
>> > virt.vm ::: stop) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` ::
>> > stopping
>> > Migration downtime thread
>> > Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,545 :: migration :: 629
>> > ::
>> > virt.vm ::: (run) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` ::
>> > stopped
>> > Migration monitor thread
>> > Thread-89513 :: DEBUG :: 2016-12-05 17: 01: 49,766 :: migration :: 715
>> > ::
>> > virt.vm ::: stop} vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` ::
>> > stopping
>> > Migration monitor thread
>> > Thread-89513 :: ERROR :: 2016-12-05 17: 01: 49,767 :: migration :: 252
>> > ::
>> > virt.vm :: (_ recover) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` ::
>> > operation Aborted: migration job: cancel
>> > D by cliente
>> >
>> > Dst vdsm.log:
>> >
>> > Periodic / 13 :: WARNING :: 2016-12-05 17: 01: 49,791 :: sampling :: 483
>> > ::
>> > virt.sampling.StatsCache: :( put) dropped stale old sample: sampled
>> > 4303678.080000 stored 4303693.070000
>> > Periodic / 13 :: DEBUG :: 2016-12-05 17: 01: 49,791 :: executor :: 221
>> > ::
>> > Executor :: (_ run) Worker was discarded
>> > Jsonrpc.Executor / 0 :: DEBUG :: 2016-12-05 17: 01: 49,792 :: __ init __
>> > ::
>> > 530 :: jsonrpc.JsonRpcServer :: (_ handle_request) Calling 'VM.destroy'
>> > in
>> > bridge with {u'vmID ' : U'f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e '}
>> > Jsonrpc.Executor / 0 :: DEBUG :: 2016-12-05 17: 01: 49,793 :: API :: 314
>> > ::
>> > vds :::( destroy) About to destroy VM
>> > f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e
>> > Jsonrpc.Executor / 0 :: DEBUG :: 2016-12-05 17:01:49,793 :: vm :: 4171
>> > ::
>> > virt.vm :::( destroy) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e`: :
>> > Destroy Called
>> > Thread-483 :: ERROR :: 2016-12-05 17: 01: 49,793 :: vm :: 759 :: virt.vm
>> > ::
>> > (_ startUnderlyingVm) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` ::
>> > Failed To start a migration destination vm
>> > Traceback (most recent call last):
>> > File "/usr/share/vdsm/virt/vm.py", line 725, in _startUnderlyingVm
>> > Self._completeIncomingMigration ()
>> > File "/usr/share/vdsm/virt/vm.py", line 3071, in
>> > _completeIncomingMigration
>> > Self._incomingMigrationFinished. isSet (), usedTimeout)
>> > File "/usr/share/vdsm/virt/vm.py", line 3154, in
>> > _attachLibvirtDomainAfterMigrat ion
>> > Raise MigrationError (e.get_error_message ())
>> > MigrationError: Domain not found: no domain with matching uuid
>> > 'f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e'
>> >
>> > The logs attached.
>> > Thanks.
>> >
>> > _______________________________________________
>> > Users mailing list
>> > Users@ovirt.org
>> > http://lists.ovirt.org/mailman/listinfo/users
>> >
>>
>>
>>
>> --
>> Didi
>
>
Didi