[ovirt-users] can not migrate vm

Yedidyah Bar David didi at redhat.com
Tue Dec 6 14:33:23 UTC 2016


On Tue, Dec 6, 2016 at 4:02 PM, Marcelo Leandro <marceloltmm at gmail.com> wrote:
> Hello.
>
> Logs after rotate.

Seem like the same files, only with more newer stuff. We need older ones,
from the time frame when the migration failed. Thanks.

>
> Thanks
>
>
> 2016-12-06 10:50 GMT-03:00 Yedidyah Bar David <didi at redhat.com>:
>>
>> On Tue, Dec 6, 2016 at 2:14 PM, Marcelo Leandro <marceloltmm at gmail.com>
>> wrote:
>> > Hello,
>> >
>> > I Tried this solution , but to some vms not was resolved.
>> > Logs:
>> >
>> > src logs:
>> >
>> > Thread-12::DEBUG::2016-12-06
>> > 08:50:58,112::check::327::storage.check::(_check_completed) FINISH check
>> >
>> > u'/rhev/data-center/mnt/192.168.144.6:_home_iso/b5fa054f-0d3d-458b-a891-13fd9383ee7d/dom_md/metadata'
>> > rc=0 err=bytearray(b'0+1 records in\n0+1 records out\n328 bytes (328 B)
>> > copied, 0.000474535 s, 691 kB/s\n') elapsed=0.08
>> > Thread-2374888::WARNING::2016-12-06
>> > 08:50:58,815::migration::671::virt.vm::(monitor_migration)
>> > vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::The migration took 520
>> > seconds
>> > which is exceeding the configured maximum time for migrations of 512
>> > seconds. The migration will be aborted.
>> > Thread-2374888::DEBUG::2016-12-06
>> > 08:50:58,816::migration::715::virt.vm::(stop)
>> > vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopping migration monitor
>> > thread
>> > Thread-2374888::DEBUG::2016-12-06
>> > 08:50:58,816::migration::570::virt.vm::(stop)
>> > vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopping migration downtime
>> > thread
>> > Thread-2374888::DEBUG::2016-12-06
>> > 08:50:58,817::migration::629::virt.vm::(run)
>> > vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopped migration monitor
>> > thread
>> > Thread-2374886::DEBUG::2016-12-06
>> > 08:50:59,098::migration::715::virt.vm::(stop)
>> > vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::stopping migration monitor
>> > thread
>> > Thread-2374886::ERROR::2016-12-06
>> > 08:50:59,098::migration::252::virt.vm::(_recover)
>> > vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::operation aborted:
>> > migration
>> > job: canceled by client
>> > Thread-2374886::DEBUG::2016-12-06
>> > 08:50:59,098::stompreactor::408::jsonrpc.AsyncoreClient::(send) Sending
>> > response
>> > Thread-2374886::DEBUG::2016-12-06
>> > 08:50:59,321::__init__::208::jsonrpc.Notification::(emit) Sending event
>> > {"params": {"notify_time": 6040272640,
>> > "9b5ab7b4-1045-4858-b24c-1f5a9f6172c3": {"status": "Migration Source"}},
>> > "jsonrpc": "2.0", "method":
>> > "|virt|VM_status|9b5ab7b4-1045-4858-b24c-1f5a9f6172c3"}
>> > Thread-2374886::ERROR::2016-12-06
>> > 08:50:59,322::migration::381::virt.vm::(run)
>> > vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::Failed to migrate
>> > Traceback (most recent call last):
>> >   File "/usr/share/vdsm/virt/migration.py", line 363, in run
>> >     self._startUnderlyingMigration(time.time())
>> >   File "/usr/share/vdsm/virt/migration.py", line 438, in
>> > _startUnderlyingMigration
>> >     self._perform_with_downtime_thread(duri, muri)
>> >   File "/usr/share/vdsm/virt/migration.py", line 489, in
>> > _perform_with_downtime_thread
>> >     self._perform_migration(duri, muri)
>> >   File "/usr/share/vdsm/virt/migration.py", line 476, in
>> > _perform_migration
>> >     self._vm._dom.migrateToURI3(duri, params, flags)
>> >   File "/usr/lib/python2.7/site-packages/vdsm/virt/virdomain.py", line
>> > 69,
>> > in f
>> >     ret = attr(*args, **kwargs)
>> >   File "/usr/lib/python2.7/site-packages/vdsm/libvirtconnection.py",
>> > line
>> > 123, in wrapper
>> >     ret = f(*args, **kwargs)
>> >   File "/usr/lib/python2.7/site-packages/vdsm/utils.py", line 916, in
>> > wrapper
>> >     return func(inst, *args, **kwargs)
>> >   File "/usr/lib64/python2.7/site-packages/libvirt.py", line 1836, in
>> > migrateToURI3
>> >     if ret == -1: raise libvirtError ('virDomainMigrateToURI3() failed',
>> > dom=self)
>> > libvirtError: operation aborted: migration job: canceled by client
>> > Thread-12::DEBUG::2016-12-06
>> > 08:50:59,875::check::296::storage.check::(_start_process) START check
>> > '/dev/c2dc0101-748e-4a7b-9913-47993eaa52bd/metadata'
>> > cmd=['/usr/bin/taskset', '--cpu-list', '0-31', '/usr/bin/dd',
>> > 'if=/dev/c2dc0101-748e-4a7b-9913-47993eaa52bd/metadata', 'of=/dev/null',
>> > 'bs=4096', 'count=1', 'iflag=direct'] delay=0.00
>> > mailbox.SPMMonitor::DEBUG::2016-12-06
>> > 08:50:59,914::storage_mailbox::733::Storage.Misc.excCmd::(_checkForMail)
>> > /usr/bin/taskset --cpu-list 0-31 dd
>> >
>> > if=/rhev/data-center/77e24b20-9d21-4952-a089-3c5c592b4e6d/mastersd/dom_md/inbox
>> > iflag=direct,fullblock count=1 bs=1024000 (cwd None)
>>
>> This snippet is not enough, and the attached logs are too new. Can you
>> check/share
>> more of the relevant log? Thanks.
>>
>> >
>> > dst log:
>> >
>> > libvirtEventLoop::DEBUG::2016-12-06
>> > 08:50:59,080::task::995::Storage.TaskManager.Task::(_decref)
>> > Task=`7446f040-c5f4-497c-b4a5-8934921a7b89`::ref 1 aborting False
>> > libvirtEventLoop::DEBUG::2016-12-06
>> > 08:50:59,080::fileUtils::190::Storage.fileUtils::(cleanupdir) Removing
>> > directory: /var/run/vdsm/storage/6e5cce71-3438-4045-9d54-6071
>> > 23e0557e/413a560d-4919-4870-88d6-f7fedbb77523
>> > libvirtEventLoop::DEBUG::2016-12-06
>> > 08:50:59,080::lvm::288::Storage.Misc.excCmd::(cmd) /usr/bin/taskset
>> > --cpu-list 0-31 /usr/bin/sudo -n /usr/sbin/lvm lvs --config ' de
>> > vices { preferred_names = ["^/dev/mapper/"] ignore_suspended_devices=1
>> > write_cache_state=0 disable_after_error_count=3 filter = [
>> > '\''a|/dev/mapper/36005076300810a4db80
>> > 0000000000002|'\'', '\''r|.*|'\'' ] }  global {  locking_type=1
>> > prioritise_write_locks=1  wait_for_locks=1  use_lvmetad=0 }  backup {
>> > retain_min = 50  retain_days = 0
>> >  } ' --noheadings --units b --nosuffix --separator '|'
>> > --ignoreskippedcluster -o
>> > uuid,name,vg_name,attr,size,seg_start_pe,devices,tags
>> > 6e5cce71-3438-4045-9d54-607123e05
>> > 57e (cwd None)
>> > jsonrpc.Executor/6::DEBUG::2016-12-06
>> > 08:50:59,090::__init__::530::jsonrpc.JsonRpcServer::(_handle_request)
>> > Calling 'VM.destroy' in bridge with {u'vmID': u'9b5ab7b4-104
>> > 5-4858-b24c-1f5a9f6172c3'}
>> > jsonrpc.Executor/6::DEBUG::2016-12-06
>> > 08:50:59,090::API::314::vds::(destroy)
>> > About to destroy VM 9b5ab7b4-1045-4858-b24c-1f5a9f6172c3
>> > jsonrpc.Executor/6::DEBUG::2016-12-06
>> > 08:50:59,091::vm::4171::virt.vm::(destroy)
>> > vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::destroy Called
>> > Thread-51550::ERROR::2016-12-06
>> > 08:50:59,091::vm::759::virt.vm::(_startUnderlyingVm)
>> > vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::Failed to start a migration
>> > destinatio
>> > n vm
>> > Traceback (most recent call last):
>> >   File "/usr/share/vdsm/virt/vm.py", line 725, in _startUnderlyingVm
>> >     self._completeIncomingMigration()
>> >   File "/usr/share/vdsm/virt/vm.py", line 3071, in
>> > _completeIncomingMigration
>> >     self._incomingMigrationFinished.isSet(), usedTimeout)
>> >   File "/usr/share/vdsm/virt/vm.py", line 3154, in
>> > _attachLibvirtDomainAfterMigration
>> >     raise MigrationError(e.get_error_message())
>> > MigrationError: Domain not found: no domain with matching uuid
>> > '9b5ab7b4-1045-4858-b24c-1f5a9f6172c3'
>> > Thread-51550::INFO::2016-12-06
>> > 08:50:59,093::vm::1308::virt.vm::(setDownStatus)
>> > vmId=`9b5ab7b4-1045-4858-b24c-1f5a9f6172c3`::Changed state to Down: VM
>> > failed to migrate (code=8)
>> >
>> >
>> >
>> > Logs Attached.
>> >
>> > any ideas?
>> >
>> > Very thanks.
>> >
>> >
>> > 2016-12-06 4:16 GMT-03:00 Yedidyah Bar David <didi at redhat.com>:
>> >>
>> >> On Mon, Dec 5, 2016 at 10:25 PM, Marcelo Leandro
>> >> <marceloltmm at gmail.com>
>> >> wrote:
>> >> > Hello
>> >> > I am with problem, I can not migrate vm.
>> >> > Can someone help me?
>> >> >
>> >> > Message in log:
>> >> >
>> >> > Src vdsm log:
>> >> >
>> >> > Thread-89514 :: WARNING :: 2016-12-05 17: 01: 49,542 :: migration ::
>> >> > 683
>> >> > ::
>> >> > virt.vm ::: monitor_migration vmId =
>> >> > `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e`
>> >> > :: Migration Stalling: remaining
>> >> > (56MiB)> lowmark (2MiB). Refer to RHBZ # 919201.
>> >> > Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,543 :: migration ::
>> >> > 689
>> >> > ::
>> >> > virt.vm ::: monitor_migration vmId =
>> >> > `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e`
>> >> > :: new Iteration detected: 15
>> >> > Thread-89514 :: WARNING :: 2016-12-05 17: 01: 49,543 :: migration ::
>> >> > 704
>> >> > ::
>> >> > virt.vm ::: monitor_migration vmId =
>> >> > `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e`
>> >> > :: Migration Is stuck: Has not pro
>> >> > Gressed in 240.071660042 seconds. Aborting.
>> >>
>> >> This is usually a result of a too-busy VM, changing its memory faster
>> >> than
>> >> the migration process can copy the changes to the destination.
>> >>
>> >> You can try changing the cluster migration policy to "suspend workload
>> >> if needed".
>> >>
>> >> For more details/background, see also:
>> >>
>> >>
>> >>
>> >> https://www.ovirt.org/develop/release-management/features/migration-enhancements/
>> >>
>> >> Best,
>> >>
>> >> > Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,544 :: migration ::
>> >> > 715
>> >> > ::
>> >> > virt.vm ::: stop} vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` ::
>> >> > stopping
>> >> > Migration monitor thread
>> >> > Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,545 :: migration ::
>> >> > 570
>> >> > ::
>> >> > virt.vm ::: stop) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` ::
>> >> > stopping
>> >> > Migration downtime thread
>> >> > Thread-89514 :: DEBUG :: 2016-12-05 17: 01: 49,545 :: migration ::
>> >> > 629
>> >> > ::
>> >> > virt.vm ::: (run) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` ::
>> >> > stopped
>> >> > Migration monitor thread
>> >> > Thread-89513 :: DEBUG :: 2016-12-05 17: 01: 49,766 :: migration ::
>> >> > 715
>> >> > ::
>> >> > virt.vm ::: stop} vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e` ::
>> >> > stopping
>> >> > Migration monitor thread
>> >> > Thread-89513 :: ERROR :: 2016-12-05 17: 01: 49,767 :: migration ::
>> >> > 252
>> >> > ::
>> >> > virt.vm :: (_ recover) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e`
>> >> > ::
>> >> > operation Aborted: migration job: cancel
>> >> > D by cliente
>> >> >
>> >> > Dst vdsm.log:
>> >> >
>> >> > Periodic / 13 :: WARNING :: 2016-12-05 17: 01: 49,791 :: sampling ::
>> >> > 483
>> >> > ::
>> >> > virt.sampling.StatsCache: :( put) dropped stale old sample: sampled
>> >> > 4303678.080000 stored 4303693.070000
>> >> > Periodic / 13 :: DEBUG :: 2016-12-05 17: 01: 49,791 :: executor ::
>> >> > 221
>> >> > ::
>> >> > Executor :: (_ run) Worker was discarded
>> >> > Jsonrpc.Executor / 0 :: DEBUG :: 2016-12-05 17: 01: 49,792 :: __ init
>> >> > __
>> >> > ::
>> >> > 530 :: jsonrpc.JsonRpcServer :: (_ handle_request) Calling
>> >> > 'VM.destroy'
>> >> > in
>> >> > bridge with {u'vmID ' : U'f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e '}
>> >> > Jsonrpc.Executor / 0 :: DEBUG :: 2016-12-05 17: 01: 49,793 :: API ::
>> >> > 314
>> >> > ::
>> >> > vds :::( destroy) About to destroy VM
>> >> > f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e
>> >> > Jsonrpc.Executor / 0 :: DEBUG :: 2016-12-05 17:01:49,793 :: vm ::
>> >> > 4171
>> >> > ::
>> >> > virt.vm :::( destroy) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e`:
>> >> > :
>> >> > Destroy Called
>> >> > Thread-483 :: ERROR :: 2016-12-05 17: 01: 49,793 :: vm :: 759 ::
>> >> > virt.vm
>> >> > ::
>> >> > (_ startUnderlyingVm) vmId = `f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e`
>> >> > ::
>> >> > Failed To start a migration destination vm
>> >> > Traceback (most recent call last):
>> >> >   File "/usr/share/vdsm/virt/vm.py", line 725, in _startUnderlyingVm
>> >> >     Self._completeIncomingMigration ()
>> >> >   File "/usr/share/vdsm/virt/vm.py", line 3071, in
>> >> > _completeIncomingMigration
>> >> >     Self._incomingMigrationFinished.isSet (), usedTimeout)
>> >> >   File "/usr/share/vdsm/virt/vm.py", line 3154, in
>> >> > _attachLibvirtDomainAfterMigration
>> >> >     Raise MigrationError (e.get_error_message ())
>> >> > MigrationError: Domain not found: no domain with matching uuid
>> >> > 'f38b9f7d-5bd0-4bdf-885c-e03e8d6bc70e'
>> >> >
>> >> > The logs attached.
>> >> > Thanks.
>> >> >
>> >> > _______________________________________________
>> >> > Users mailing list
>> >> > Users at ovirt.org
>> >> > http://lists.ovirt.org/mailman/listinfo/users
>> >> >
>> >>
>> >>
>> >>
>> >> --
>> >> Didi
>> >
>> >
>>
>>
>>
>> --
>> Didi
>
>



-- 
Didi



More information about the Users mailing list