[Users] Migration Failed

Dafna Ron dron at redhat.com
Tue Jan 7 18:13:33 UTC 2014


Ok... several things :)

1. for migration we need to see vdsm logs from both src and dst.

2. Is it possible that the vm has an iso attached? because I see that 
you are having problems with the iso domain:

2014-01-07 14:26:27,714 ERROR 
[org.ovirt.engine.core.vdsbroker.irsbroker.IrsBrokerCommand] 
(pool-6-thread-48) Domain e9ab725d-69c1-4a59-b225-b995d095c289:bla-iso 
was reported with error code 358


Thread-1165153::DEBUG::2014-01-07 
13:39:42,460::libvirtconnection::108::libvirtconnection::(wrapper) 
Unknown libvirterror: ecode: 42 edom: 10 level: 2 message: Domain not 
found: no domain with matching uuid '63da7faa-f92a-4652-90f2-b6660a4fb7b3'

hread-19::ERROR::2014-01-07 
13:01:02,621::sdc::143::Storage.StorageDomainCache::(_findDomain) domain 
e9ab725d-69c1-4a59-b225-b995d095c289 not found
Traceback (most recent call last):
   File "/usr/share/vdsm/storage/sdc.py", line 141, in _findDomain
     dom = findMethod(sdUUID)
   File "/usr/share/vdsm/storage/sdc.py", line 171, in _findUnfetchedDomain
     raise se.StorageDomainDoesNotExist(sdUUID)
StorageDomainDoesNotExist: Storage domain does not exist: 
(u'e9ab725d-69c1-4a59-b225-b995d095c289',)
Thread-19::ERROR::2014-01-07 
13:01:02,622::domainMonitor::225::Storage.DomainMonitorThread::(_monitorDomain) 
Error while collecting domain e9ab725d-69c1-4a59-b225-b995d095c289 
monitoring information
Traceback (most recent call last):
   File "/usr/share/vdsm/storage/domainMonitor.py", line 190, in 
_monitorDomain
     self.domain = sdCache.produce(self.sdUUID)
   File "/usr/share/vdsm/storage/sdc.py", line 98, in produce
     domain.getRealDomain()
   File "/usr/share/vdsm/storage/sdc.py", line 52, in getRealDomain
     return self._cache._realProduce(self._sdUUID)
   File "/usr/share/vdsm/storage/sdc.py", line 122, in _realProduce
     domain = self._findDomain(sdUUID)
   File "/usr/share/vdsm/storage/sdc.py", line 141, in _findDomain
     dom = findMethod(sdUUID)
   File "/usr/share/vdsm/storage/sdc.py", line 171, in _findUnfetchedDomain
     raise se.StorageDomainDoesNotExist(sdUUID)
StorageDomainDoesNotExist: Storage domain does not exist: 
(u'e9ab725d-69c1-4a59-b225-b995d095c289',)
Dummy-29013::DEBUG::2014-01-07 
13:01:03,507::storage_mailbox::733::Storage.Misc.excCmd::(_checkForMail) 
'dd 
if=/rhev/data-center/28adaf38-a4f6-11e1-a859-cb68949043e4/mastersd/dom_md/inbox 
iflag=direct,fullblock count=1 bs=1024000' (cwd N
one)

3. The migration fails with libvirt error but we need the trace from the 
second log:

Thread-1165153::DEBUG::2014-01-07 
13:39:42,451::sampling::292::vm.Vm::(stop) 
vmId=`63da7faa-f92a-4652-90f2-b6660a4fb7b3`::Stop statistics collection
Thread-1163583::DEBUG::2014-01-07 
13:39:42,452::sampling::323::vm.Vm::(run) 
vmId=`63da7faa-f92a-4652-90f2-b6660a4fb7b3`::Stats thread finished
Thread-1165153::DEBUG::2014-01-07 
13:39:42,460::libvirtconnection::108::libvirtconnection::(wrapper) 
Unknown libvirterror: ecode: 42 edom: 10 level: 2 message: Domain not 
found: no domain with matching uuid '63da7faa-f92a-4652-90f2-b6660
a4fb7b3'


4. But I am worried about this and would more info about this vm...

Thread-247::ERROR::2014-01-07 
15:35:14,868::sampling::355::vm.Vm::(collect) 
vmId=`63da7faa-f92a-4652-90f2-b6660a4fb7b3`::Stats function failed: 
<AdvancedStatsFunction _highWrite at 0x2ce0998>
Traceback (most recent call last):
   File "/usr/share/vdsm/sampling.py", line 351, in collect
     statsFunction()
   File "/usr/share/vdsm/sampling.py", line 226, in __call__
     retValue = self._function(*args, **kwargs)
   File "/usr/share/vdsm/vm.py", line 509, in _highWrite
     if not vmDrive.blockDev or vmDrive.format != 'cow':
AttributeError: 'Drive' object has no attribute 'format'

How did you create this vm? was it from the UI? was it from a script? 
what are the parameters you used?

Thanks,

Dafna


On 01/07/2014 04:34 PM, Neil wrote:
> Hi Elad,
>
> Thanks for assisting me, yes the same condition exists, if I try to
> migrate Tux it says "The VM Tux is being migrated".
>
>
> Below are the details requested.
>
>
> [root at node01 ~]# virsh -r list
>   Id    Name                           State
> ----------------------------------------------------
>   1     adam                           running
>
> [root at node01 ~]# pgrep qemu
> 11232
> [root at node01 ~]# vdsClient -s 0 list table
> 63da7faa-f92a-4652-90f2-b6660a4fb7b3  11232  adam                 Up
>
>
> [root at node03 ~]# virsh -r list
>   Id    Name                           State
> ----------------------------------------------------
>   7     tux                            running
>
> [root at node03 ~]# pgrep qemu
> 32333
> [root at node03 ~]# vdsClient -s 0 list table
> 2736197b-6dc3-4155-9a29-9306ca64881d  32333  tux                  Up
>
> Thanks.
>
> Regards.
>
> Neil Wilson.
>
>
> On Tue, Jan 7, 2014 at 4:43 PM, Elad Ben Aharon <ebenahar at redhat.com> wrote:
>> Is it still in the same condition?
>> If yes, please add the outputs from both hosts for:
>>
>> #virsh -r list
>> #pgrep qemu
>> #vdsClient -s 0 list table     (or 'vdsClient 0 list table' if you are working in insecure mode)
>>
>>
>> Thnaks,
>>
>> Elad Ben Aharon
>> RHEV-QE storage team
>>
>>
>>
>>
>> ----- Original Message -----
>> From: "Neil" <nwilson123 at gmail.com>
>> To: users at ovirt.org
>> Sent: Tuesday, January 7, 2014 4:21:43 PM
>> Subject: [Users] Migration Failed
>>
>> Hi guys,
>>
>> I've tried to migrate a VM from one host(node03) to another(node01),
>> and it failed to migrate, and the VM(tux) remained on the original
>> host. I've now tried to migrate the same VM again, and it picks up
>> that the previous migration is still in progress and refuses to
>> migrate.
>>
>> I've checked for the KVM process on each of the hosts and the VM is
>> definitely still running on node03 so there doesn't appear to be any
>> chance of the VM trying to run on both hosts (which I've had before
>> which is very scary).
>>
>> These are my versions... and attached are my engine.log and my vdsm.log
>>
>> Centos 6.5
>> ovirt-iso-uploader-3.3.1-1.el6.noarch
>> ovirt-host-deploy-1.1.2-1.el6.noarch
>> ovirt-release-el6-9-1.noarch
>> ovirt-engine-setup-3.3.1-2.el6.noarch
>> ovirt-engine-3.3.1-2.el6.noarch
>> ovirt-host-deploy-java-1.1.2-1.el6.noarch
>> ovirt-image-uploader-3.3.1-1.el6.noarch
>> ovirt-engine-dbscripts-3.3.1-2.el6.noarch
>> ovirt-engine-cli-3.3.0.6-1.el6.noarch
>> ovirt-engine-websocket-proxy-3.3.1-2.el6.noarch
>> ovirt-engine-userportal-3.3.1-2.el6.noarch
>> ovirt-log-collector-3.3.1-1.el6.noarch
>> ovirt-engine-tools-3.3.1-2.el6.noarch
>> ovirt-engine-lib-3.3.1-2.el6.noarch
>> ovirt-engine-webadmin-portal-3.3.1-2.el6.noarch
>> ovirt-engine-backend-3.3.1-2.el6.noarch
>> ovirt-engine-sdk-python-3.3.0.8-1.el6.noarch
>> ovirt-engine-restapi-3.3.1-2.el6.noarch
>>
>>
>> vdsm-python-4.13.0-11.el6.x86_64
>> vdsm-cli-4.13.0-11.el6.noarch
>> vdsm-xmlrpc-4.13.0-11.el6.noarch
>> vdsm-4.13.0-11.el6.x86_64
>> vdsm-python-cpopen-4.13.0-11.el6.x86_64
>>
>> I've had a few issues with this particular installation in the past,
>> as it's from a very old pre release of ovirt, then upgrading to the
>> dreyou repo, then finally moving to the official Centos ovirt repo.
>>
>> Thanks, any help is greatly appreciated.
>>
>> Regards.
>>
>> Neil Wilson.
>>
>> _______________________________________________
>> Users mailing list
>> Users at ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users


-- 
Dafna Ron



More information about the Users mailing list