On Tue, Dec 6, 2016 at 2:14 PM, Marcelo Leandro <marceloltmm(a)gmail.com> wrote:
Hello,
I Tried this solution , but to some vms not was resolved.
Logs:
src logs:
Thread-12::DEBUG::2016-12-06
08:50:58,112::check::327::storage.check::(_check_completed) FINISH check
u'/rhev/data-center/mnt/192.168.144.6:_home_iso/b5fa054f-0d3d-458b-a891-13fd9383ee7d/dom_md/metadata'
rc=0 err=bytearray(b'0+1 records in\n0+1 records out\n328 bytes (328 B)
copied, 0.000474535 s, 691 kB/s\n') elapsed=0.08
Thread-2374888::WARNING::2016-12-06
08:50:58,815::migration::671::virt.vm::(monitor_migration)
vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::The migration took 520 seconds
which is exceeding the configured maximum time for migrations of 512
seconds. The migration will be aborted.
Thread-2374888::DEBUG::2016-12-06
08:50:58,816::migration::715::virt.vm::(stop)
vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopping migration monitor
thread
Thread-2374888::DEBUG::2016-12-06
08:50:58,816::migration::570::virt.vm::(stop)
vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopping migration downtime
thread
Thread-2374888::DEBUG::2016-12-06
08:50:58,817::migration::629::virt.vm::(run)
vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopped migration monitor
thread
Thread-2374886::DEBUG::2016-12-06
08:50:59,098::migration::715::virt.vm::(stop)
vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopping migration monitor
thread
Thread-2374886::ERROR::2016-12-06
08:50:59,098::migration::252::virt.vm::(_recover)
vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::operation aborted: migration
job: canceled by client
Thread-2374886::DEBUG::2016-12-06
08:50:59,098::stompreactor::408::jsonrpc.AsyncoreClient::(send) Sending
response
Thread-2374886::DEBUG::2016-12-06
08:50:59,321::__init__::208::jsonrpc.Notification::(emit) Sending event
{"params": {"notify_time": 6040272640,
"9b5ab7b4-1045-4858-b24c-1f5a9f6172c3": {"status": "Migration
Source"}},
"jsonrpc": "2.0", "method":
"|virt|VM_status|9b5ab7b4-1045-4858-b24c-1f5a9f6172c3"}
Thread-2374886::ERROR::2016-12-06
08:50:59,322::migration::381::virt.vm::(run)
vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::Failed to migrate
Traceback (most recent call last):
File "/usr/share/vdsm/virt/migration.py", line 363, in run
self._startUnderlyingMigration(time.time())
File "/usr/share/vdsm/virt/migration.py", line 438, in
_startUnderlyingMigration
self._perform_with_downtime_thread(duri, muri)
File "/usr/share/vdsm/virt/migration.py", line 489, in
_perform_with_downtime_thread
self._perform_migration(duri, muri)
File "/usr/share/vdsm/virt/migration.py", line 476, in _perform_migration
self._vm._dom.migrateToURI3(duri, params, flags)
File "/usr/lib/python2.7/site-packages/vdsm/virt/virdomain.py", line 69,
in f
ret = attr(*args, **kwargs)
File "/usr/lib/python2.7/site-packages/vdsm/libvirtconnection.py", line
123, in wrapper
ret = f(*args, **kwargs)
File "/usr/lib/python2.7/site-packages/vdsm/utils.py", line 916, in
wrapper
return func(inst, *args, **kwargs)
File "/usr/lib64/python2.7/site-packages/libvirt.py", line 1836, in
migrateToURI3
if ret == -1: raise libvirtError ('virDomainMigrateToURI3() failed',
dom=self)
libvirtError: operation aborted: migration job: canceled by client
Thread-12::DEBUG::2016-12-06
08:50:59,875::check::296::storage.check::(_start_process) START check
'/dev/c2dc0101-748e-4a7b-9913-47993eaa52bd/metadata'
cmd=['/usr/bin/taskset', '--cpu-list', '0-31',
'/usr/bin/dd',
'if=/dev/c2dc0101-748e-4a7b-9913-47993eaa52bd/metadata', 'of=/dev/null',
'bs=4096', 'count=1', 'iflag=direct'] delay=0.00
mailbox.SPMMonitor::DEBUG::2016-12-06
08:50:59,914::storage_mailbox::733::Storage.Misc.excCmd::(_checkForMail)
/usr/bin/taskset --cpu-list 0-31 dd
if=/rhev/data-center/77e24b20-9d21-4952-a089-3c5c592b4e6d/mastersd/dom_md/inbox
iflag=direct,fullblock count=1 bs=1024000 (cwd None)
This snippet is not enough, and the attached logs are too new. Can you
check/share
more of the relevant log? Thanks.
dst log:
libvirtEventLoop::DEBUG::2016-12-06
08:50:59,080::task::995::Storage.TaskManager.Task::(_decref)
Task=`7446f040-c5f4-497c-b4a5-8934921a7b89`::ref 1 aborting False
libvirtEventLoop::DEBUG::2016-12-06
08:50:59,080::fileUtils::190::Storage.fileUtils::(cleanupdir) Removing
directory: /var/run/vdsm/storage/6e5cce71-3438-4045-9d54-6071
23e0557e/413a560d-4919-4870-88d6-f7fedbb77523
libvirtEventLoop::DEBUG::2016-12-06
08:50:59,080::lvm::288::Storage.Misc.excCmd::(cmd) /usr/bin/taskset
--cpu-list 0-31 /usr/bin/sudo -n /usr/sbin/lvm lvs --config ' de
vices { preferred_names = ["^/dev/mapper/"] ignore_suspended_devices=1
write_cache_state=0 disable_after_error_count=3 filter = [
'\''a|/dev/mapper/36005076300810a4db80
0000000000002|'\'', '\''r|.*|'\'' ] } global {
locking_type=1
prioritise_write_locks=1 wait_for_locks=1 use_lvmetad=0 } backup {
retain_min = 50 retain_days = 0
} ' --noheadings --units b --nosuffix --separator '|'
--ignoreskippedcluster -o
uuid,name,vg_name,attr,size,seg_start_pe,devices,tags
6e5cce71-3438-4045-9d54-607123e05
57e (cwd None)
jsonrpc.Executor/6::DEBUG::2016-12-06
08:50:59,090::__init__::530::jsonrpc.JsonRpcServer::(_handle_request)
Calling 'VM.destroy' in bridge with {u'vmID': u'9b5ab7b4-104
5-4858-b24c-1f5a9f6172c3'}
jsonrpc.Executor/6::DEBUG::2016-12-06 08:50:59,090::API::314::vds::(destroy)
About to destroy VM 9b5ab7b4-1045-4858-b24c-1f5a9f6172c3
jsonrpc.Executor/6::DEBUG::2016-12-06
08:50:59,091::vm::4171::virt.vm::(destroy)
vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::destroy Called
Thread-51550::ERROR::2016-12-06
08:50:59,091::vm::759::virt.vm::(_startUnderlyingVm)
vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::Failed to start a migration
destinatio
n vm
Traceback (most recent call last):
File "/usr/share/vdsm/virt/vm.py", line 725, in _startUnderlyingVm
self._completeIncomingMigration()
File "/usr/share/vdsm/virt/vm.py", line 3071, in
_completeIncomingMigration
self._incomingMigrationFinished.isSet(), usedTimeout)
File "/usr/share/vdsm/virt/vm.py", line 3154, in
_attachLibvirtDomainAfterMigration
raise MigrationError(e.get_error_message())
MigrationError: Domain not found: no domain with matching uuid
'9b5ab7b4-1045-4858-b24c-1f5a9f6172c3'
Thread-51550::INFO::2016-12-06
08:50:59,093::vm::1308::virt.vm::(setDownStatus)
vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::Changed state to Down: VM
failed to migrate (code=8)
Logs Attached.
any ideas?
Very thanks.
2016-12-06 4:16 GMT-03:00 Yedidyah Bar David <didi(a)redhat.com>:
>
> On Mon, Dec 5, 2016 at 10:25 PM, Marcelo Leandro <marceloltmm(a)gmail.com>
> wrote:
> > Hello
> > I am with problem, I can not migrate vm.
> > Can someone help me?
> >
> > Message in log:
> >
> > Src vdsm log:
> >
> > Thread-89514 :: WARNING :: 2016-12-05 17: 01: 49,542 :: migration :: 683
> > ::
> > virt.vm ::: monitor_migration vmId =
> > `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e`
> > :: Migration Stalling: remaining
> > (56MiB)> lowmark (2MiB). Refer to RHBZ # 919201.
> > Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,543 :: migration :: 689
> > ::
> > virt.vm ::: monitor_migration vmId =
> > `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e`
> > :: new Iteration detected: 15
> > Thread-89514 :: WARNING :: 2016-12-05 17: 01: 49,543 :: migration :: 704
> > ::
> > virt.vm ::: monitor_migration vmId =
> > `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e`
> > :: Migration Is stuck: Has not pro
> > Gressed in 240.071660042 seconds. Aborting.
>
> This is usually a result of a too-busy VM, changing its memory faster than
> the migration process can copy the changes to the destination.
>
> You can try changing the cluster migration policy to "suspend workload
> if needed".
>
> For more details/background, see also:
>
>
>
https://www.ovirt.org/develop/release-management/features/migration-enhan...
>
> Best,
>
> > Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,544 :: migration :: 715
> > ::
> > virt.vm ::: stop} vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` ::
> > stopping
> > Migration monitor thread
> > Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,545 :: migration :: 570
> > ::
> > virt.vm ::: stop) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` ::
> > stopping
> > Migration downtime thread
> > Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,545 :: migration :: 629
> > ::
> > virt.vm ::: (run) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` ::
> > stopped
> > Migration monitor thread
> > Thread-89513 :: DEBUG :: 2016-12-05 17: 01: 49,766 :: migration :: 715
> > ::
> > virt.vm ::: stop} vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` ::
> > stopping
> > Migration monitor thread
> > Thread-89513 :: ERROR :: 2016-12-05 17: 01: 49,767 :: migration :: 252
> > ::
> > virt.vm :: (_ recover) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` ::
> > operation Aborted: migration job: cancel
> > D by cliente
> >
> > Dst vdsm.log:
> >
> > Periodic / 13 :: WARNING :: 2016-12-05 17: 01: 49,791 :: sampling :: 483
> > ::
> > virt.sampling.StatsCache: :( put) dropped stale old sample: sampled
> > 4303678.080000 stored 4303693.070000
> > Periodic / 13 :: DEBUG :: 2016-12-05 17: 01: 49,791 :: executor :: 221
> > ::
> > Executor :: (_ run) Worker was discarded
> > Jsonrpc.Executor / 0 :: DEBUG :: 2016-12-05 17: 01: 49,792 :: __ init __
> > ::
> > 530 :: jsonrpc.JsonRpcServer :: (_ handle_request) Calling 'VM.destroy'
> > in
> > bridge with {u'vmID ' : U'f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e
'}
> > Jsonrpc.Executor / 0 :: DEBUG :: 2016-12-05 17: 01: 49,793 :: API :: 314
> > ::
> > vds :::( destroy) About to destroy VM
> > f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e
> > Jsonrpc.Executor / 0 :: DEBUG :: 2016-12-05 17:01:49,793 :: vm :: 4171
> > ::
> > virt.vm :::( destroy) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e`: :
> > Destroy Called
> > Thread-483 :: ERROR :: 2016-12-05 17: 01: 49,793 :: vm :: 759 :: virt.vm
> > ::
> > (_ startUnderlyingVm) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` ::
> > Failed To start a migration destination vm
> > Traceback (most recent call last):
> > File "/usr/share/vdsm/virt/vm.py", line 725, in _startUnderlyingVm
> > Self._completeIncomingMigration ()
> > File "/usr/share/vdsm/virt/vm.py", line 3071, in
> > _completeIncomingMigration
> > Self._incomingMigrationFinished.isSet (), usedTimeout)
> > File "/usr/share/vdsm/virt/vm.py", line 3154, in
> > _attachLibvirtDomainAfterMigration
> > Raise MigrationError (e.get_error_message ())
> > MigrationError: Domain not found: no domain with matching uuid
> > 'f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e'
> >
> > The logs attached.
> > Thanks.
> >
> > _______________________________________________
> > Users mailing list
> > Users(a)ovirt.org
> >
http://lists.ovirt.org/mailman/listinfo/users
> >
>
>
>
> --
> Didi
--
Didi