[Users] Migration Failed

Dafna Ron dron at redhat.com
Wed Jan 8 10:23:17 UTC 2014


Thread-847747::INFO::2014-01-07 
14:30:32,353::logUtils::44::dispatcher::(wrapper) Run and protect: 
inappropriateDevices(thiefId='63da7faa-f92a-4652-90f2-b6660a4fb7b3')
Thread-847747::INFO::2014-01-07 
14:30:32,354::logUtils::47::dispatcher::(wrapper) Run and protect: 
inappropriateDevices, Return response: None

Please check if the vm's were booted with a cd...


bject at 0x7fb1f00cbbd0>> log:<logUtils.SimpleLogAdapter instance at 
0x7fb1f00be7e8> name:hdc networkDev:False path: readonly:True reqsize:0 
serial: truesize:0 *type:cdrom* volExtensionChunk:1024 
watermarkLimit:536870912
Traceback (most recent call last):
   File "/usr/share/vdsm/clientIF.py", line 356, in teardownVolumePath
     res = self.irs.teardownImage(drive['domainID'],
   File "/usr/share/vdsm/vm.py", line 1386, in __getitem__
     raise KeyError(key)
KeyError: 'domainID'
Thread-847747::WARNING::2014-01-07 
14:30:32,351::clientIF::362::vds::(teardownVolumePath) Drive is not a 
vdsm image: VOLWM_CHUNK_MB:1024 VOLWM_CHUNK_REPLICATE_MULT:2 
VOLWM_FREE_PCT:50 _blockDev:True _checkIoTuneCategories:<bound method D
rive._checkIoTuneCategories of <vm.Drive object at 0x7fb1f00cbc10>> 
_customize:<bound method Drive._customize of <vm.Drive object at 
0x7fb1f00cbc10>> _deviceXML:<disk device="disk" snapshot="no" type="block">
       <driver cache="none" error_policy="stop" io="native" name="qemu" 
type="raw"/>
       <source 
dev="/rhev/data-center/28adaf38-a4f6-11e1-a859-cb68949043e4/0e6991ae-6238-4c61-96d2-ca8fed35161e/images/9f16f896-1da3-4f9a-a305-ac9c4f51a482/e04c6600-abb9-4ebc-a9b3-77b6c536e258"/>
       <target bus="ide" dev="hda"/>
<serial>9f16f896-1da3-4f9a-a305-ac9c4f51a482</serial>
       <alias name="ide0-0-0"/>
       <address bus="0" controller="0" target="0" type="drive" unit="0"/>


On 01/08/2014 06:28 AM, Neil wrote:
> Hi Dafna,
>
> Thanks for the reply.
>
> Attached is the log from the source server (node03).
>
> I'll reply to your other questions as soon as I'm back in the office
> this afternoon, have to run off to a meeting.
>
> Regards.
>
> Neil Wilson.
>
>
> On Tue, Jan 7, 2014 at 8:13 PM, Dafna Ron <dron at redhat.com> wrote:
>> Ok... several things :)
>>
>> 1. for migration we need to see vdsm logs from both src and dst.
>>
>> 2. Is it possible that the vm has an iso attached? because I see that you
>> are having problems with the iso domain:
>>
>> 2014-01-07 14:26:27,714 ERROR
>> [org.ovirt.engine.core.vdsbroker.irsbroker.IrsBrokerCommand]
>> (pool-6-thread-48) Domain e9ab725d-69c1-4a59-b225-b995d095c289:bla-iso was
>> reported with error code 358
>>
>>
>> Thread-1165153::DEBUG::2014-01-07
>> 13:39:42,460::libvirtconnection::108::libvirtconnection::(wrapper) Unknown
>> libvirterror: ecode: 42 edom: 10 level: 2 message: Domain not found: no
>> domain with matching uuid '63da7faa-f92a-4652-90f2-b6660a4fb7b3'
>>
>> hread-19::ERROR::2014-01-07
>> 13:01:02,621::sdc::143::Storage.StorageDomainCache::(_findDomain) domain
>> e9ab725d-69c1-4a59-b225-b995d095c289 not found
>> Traceback (most recent call last):
>>    File "/usr/share/vdsm/storage/sdc.py", line 141, in _findDomain
>>      dom = findMethod(sdUUID)
>>    File "/usr/share/vdsm/storage/sdc.py", line 171, in _findUnfetchedDomain
>>      raise se.StorageDomainDoesNotExist(sdUUID)
>> StorageDomainDoesNotExist: Storage domain does not exist:
>> (u'e9ab725d-69c1-4a59-b225-b995d095c289',)
>> Thread-19::ERROR::2014-01-07
>> 13:01:02,622::domainMonitor::225::Storage.DomainMonitorThread::(_monitorDomain)
>> Error while collecting domain e9ab725d-69c1-4a59-b225-b995d095c289
>> monitoring information
>> Traceback (most recent call last):
>>    File "/usr/share/vdsm/storage/domainMonitor.py", line 190, in
>> _monitorDomain
>>      self.domain = sdCache.produce(self.sdUUID)
>>    File "/usr/share/vdsm/storage/sdc.py", line 98, in produce
>>      domain.getRealDomain()
>>    File "/usr/share/vdsm/storage/sdc.py", line 52, in getRealDomain
>>      return self._cache._realProduce(self._sdUUID)
>>    File "/usr/share/vdsm/storage/sdc.py", line 122, in _realProduce
>>      domain = self._findDomain(sdUUID)
>>    File "/usr/share/vdsm/storage/sdc.py", line 141, in _findDomain
>>      dom = findMethod(sdUUID)
>>    File "/usr/share/vdsm/storage/sdc.py", line 171, in _findUnfetchedDomain
>>      raise se.StorageDomainDoesNotExist(sdUUID)
>> StorageDomainDoesNotExist: Storage domain does not exist:
>> (u'e9ab725d-69c1-4a59-b225-b995d095c289',)
>> Dummy-29013::DEBUG::2014-01-07
>> 13:01:03,507::storage_mailbox::733::Storage.Misc.excCmd::(_checkForMail) 'dd
>> if=/rhev/data-center/28adaf38-a4f6-11e1-a859-cb68949043e4/mastersd/dom_md/inbox
>> iflag=direct,fullblock count=1 bs=1024000' (cwd N
>> one)
>>
>> 3. The migration fails with libvirt error but we need the trace from the
>> second log:
>>
>> Thread-1165153::DEBUG::2014-01-07 13:39:42,451::sampling::292::vm.Vm::(stop)
>> vmId=`63da7faa-f92a-4652-90f2-b6660a4fb7b3`::Stop statistics collection
>> Thread-1163583::DEBUG::2014-01-07 13:39:42,452::sampling::323::vm.Vm::(run)
>> vmId=`63da7faa-f92a-4652-90f2-b6660a4fb7b3`::Stats thread finished
>> Thread-1165153::DEBUG::2014-01-07
>> 13:39:42,460::libvirtconnection::108::libvirtconnection::(wrapper) Unknown
>> libvirterror: ecode: 42 edom: 10 level: 2 message: Domain not found: no
>> domain with matching uuid '63da7faa-f92a-4652-90f2-b6660
>> a4fb7b3'
>>
>>
>> 4. But I am worried about this and would more info about this vm...
>>
>> Thread-247::ERROR::2014-01-07 15:35:14,868::sampling::355::vm.Vm::(collect)
>> vmId=`63da7faa-f92a-4652-90f2-b6660a4fb7b3`::Stats function failed:
>> <AdvancedStatsFunction _highWrite at 0x2ce0998>
>> Traceback (most recent call last):
>>    File "/usr/share/vdsm/sampling.py", line 351, in collect
>>      statsFunction()
>>    File "/usr/share/vdsm/sampling.py", line 226, in __call__
>>      retValue = self._function(*args, **kwargs)
>>    File "/usr/share/vdsm/vm.py", line 509, in _highWrite
>>      if not vmDrive.blockDev or vmDrive.format != 'cow':
>> AttributeError: 'Drive' object has no attribute 'format'
>>
>> How did you create this vm? was it from the UI? was it from a script? what
>> are the parameters you used?
>>
>> Thanks,
>>
>> Dafna
>>
>>
>>
>> On 01/07/2014 04:34 PM, Neil wrote:
>>> Hi Elad,
>>>
>>> Thanks for assisting me, yes the same condition exists, if I try to
>>> migrate Tux it says "The VM Tux is being migrated".
>>>
>>>
>>> Below are the details requested.
>>>
>>>
>>> [root at node01 ~]# virsh -r list
>>>    Id    Name                           State
>>> ----------------------------------------------------
>>>    1     adam                           running
>>>
>>> [root at node01 ~]# pgrep qemu
>>> 11232
>>> [root at node01 ~]# vdsClient -s 0 list table
>>> 63da7faa-f92a-4652-90f2-b6660a4fb7b3  11232  adam                 Up
>>>
>>>
>>> [root at node03 ~]# virsh -r list
>>>    Id    Name                           State
>>> ----------------------------------------------------
>>>    7     tux                            running
>>>
>>> [root at node03 ~]# pgrep qemu
>>> 32333
>>> [root at node03 ~]# vdsClient -s 0 list table
>>> 2736197b-6dc3-4155-9a29-9306ca64881d  32333  tux                  Up
>>>
>>> Thanks.
>>>
>>> Regards.
>>>
>>> Neil Wilson.
>>>
>>>
>>> On Tue, Jan 7, 2014 at 4:43 PM, Elad Ben Aharon <ebenahar at redhat.com>
>>> wrote:
>>>> Is it still in the same condition?
>>>> If yes, please add the outputs from both hosts for:
>>>>
>>>> #virsh -r list
>>>> #pgrep qemu
>>>> #vdsClient -s 0 list table     (or 'vdsClient 0 list table' if you are
>>>> working in insecure mode)
>>>>
>>>>
>>>> Thnaks,
>>>>
>>>> Elad Ben Aharon
>>>> RHEV-QE storage team
>>>>
>>>>
>>>>
>>>>
>>>> ----- Original Message -----
>>>> From: "Neil" <nwilson123 at gmail.com>
>>>> To: users at ovirt.org
>>>> Sent: Tuesday, January 7, 2014 4:21:43 PM
>>>> Subject: [Users] Migration Failed
>>>>
>>>> Hi guys,
>>>>
>>>> I've tried to migrate a VM from one host(node03) to another(node01),
>>>> and it failed to migrate, and the VM(tux) remained on the original
>>>> host. I've now tried to migrate the same VM again, and it picks up
>>>> that the previous migration is still in progress and refuses to
>>>> migrate.
>>>>
>>>> I've checked for the KVM process on each of the hosts and the VM is
>>>> definitely still running on node03 so there doesn't appear to be any
>>>> chance of the VM trying to run on both hosts (which I've had before
>>>> which is very scary).
>>>>
>>>> These are my versions... and attached are my engine.log and my vdsm.log
>>>>
>>>> Centos 6.5
>>>> ovirt-iso-uploader-3.3.1-1.el6.noarch
>>>> ovirt-host-deploy-1.1.2-1.el6.noarch
>>>> ovirt-release-el6-9-1.noarch
>>>> ovirt-engine-setup-3.3.1-2.el6.noarch
>>>> ovirt-engine-3.3.1-2.el6.noarch
>>>> ovirt-host-deploy-java-1.1.2-1.el6.noarch
>>>> ovirt-image-uploader-3.3.1-1.el6.noarch
>>>> ovirt-engine-dbscripts-3.3.1-2.el6.noarch
>>>> ovirt-engine-cli-3.3.0.6-1.el6.noarch
>>>> ovirt-engine-websocket-proxy-3.3.1-2.el6.noarch
>>>> ovirt-engine-userportal-3.3.1-2.el6.noarch
>>>> ovirt-log-collector-3.3.1-1.el6.noarch
>>>> ovirt-engine-tools-3.3.1-2.el6.noarch
>>>> ovirt-engine-lib-3.3.1-2.el6.noarch
>>>> ovirt-engine-webadmin-portal-3.3.1-2.el6.noarch
>>>> ovirt-engine-backend-3.3.1-2.el6.noarch
>>>> ovirt-engine-sdk-python-3.3.0.8-1.el6.noarch
>>>> ovirt-engine-restapi-3.3.1-2.el6.noarch
>>>>
>>>>
>>>> vdsm-python-4.13.0-11.el6.x86_64
>>>> vdsm-cli-4.13.0-11.el6.noarch
>>>> vdsm-xmlrpc-4.13.0-11.el6.noarch
>>>> vdsm-4.13.0-11.el6.x86_64
>>>> vdsm-python-cpopen-4.13.0-11.el6.x86_64
>>>>
>>>> I've had a few issues with this particular installation in the past,
>>>> as it's from a very old pre release of ovirt, then upgrading to the
>>>> dreyou repo, then finally moving to the official Centos ovirt repo.
>>>>
>>>> Thanks, any help is greatly appreciated.
>>>>
>>>> Regards.
>>>>
>>>> Neil Wilson.
>>>>
>>>> _______________________________________________
>>>> Users mailing list
>>>> Users at ovirt.org
>>>> http://lists.ovirt.org/mailman/listinfo/users
>>> _______________________________________________
>>> Users mailing list
>>> Users at ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
>> --
>> Dafna Ron


-- 
Dafna Ron



More information about the Users mailing list