[Users] Live migration of VM's occasionally fails

Dafna Ron dron at redhat.com
Sun Feb 16 12:05:55 UTC 2014


does the vm that fails migration have a live snapshot?
if so how many snapshots does the vm have.
I think that there are newer packages of vdsm, libvirt and qemu - can 
you try to update


On 02/16/2014 12:33 AM, Steve Dainard wrote:
> Versions are the same:
>
> [root at ovirt001 ~]# rpm -qa | egrep 'libvirt|vdsm|qemu' | sort
> gpxe-roms-qemu-0.9.7-6.10.el6.noarch
> libvirt-0.10.2-29.el6_5.3.x86_64
> libvirt-client-0.10.2-29.el6_5.3.x86_64
> libvirt-lock-sanlock-0.10.2-29.el6_5.3.x86_64
> libvirt-python-0.10.2-29.el6_5.3.x86_64
> qemu-img-rhev-0.12.1.2-2.355.el6.5.x86_64
> qemu-kvm-rhev-0.12.1.2-2.355.el6.5.x86_64
> qemu-kvm-rhev-tools-0.12.1.2-2.355.el6.5.x86_64
> vdsm-4.13.3-3.el6.x86_64
> vdsm-cli-4.13.3-3.el6.noarch
> vdsm-gluster-4.13.3-3.el6.noarch
> vdsm-python-4.13.3-3.el6.x86_64
> vdsm-xmlrpc-4.13.3-3.el6.noarch
>
> [root at ovirt002 ~]# rpm -qa | egrep 'libvirt|vdsm|qemu' | sort
> gpxe-roms-qemu-0.9.7-6.10.el6.noarch
> libvirt-0.10.2-29.el6_5.3.x86_64
> libvirt-client-0.10.2-29.el6_5.3.x86_64
> libvirt-lock-sanlock-0.10.2-29.el6_5.3.x86_64
> libvirt-python-0.10.2-29.el6_5.3.x86_64
> qemu-img-rhev-0.12.1.2-2.355.el6.5.x86_64
> qemu-kvm-rhev-0.12.1.2-2.355.el6.5.x86_64
> qemu-kvm-rhev-tools-0.12.1.2-2.355.el6.5.x86_64
> vdsm-4.13.3-3.el6.x86_64
> vdsm-cli-4.13.3-3.el6.noarch
> vdsm-gluster-4.13.3-3.el6.noarch
> vdsm-python-4.13.3-3.el6.x86_64
> vdsm-xmlrpc-4.13.3-3.el6.noarch
>
> Logs attached, thanks.
>
> *Steve Dainard *
> IT Infrastructure Manager
> Miovision <http://miovision.com/> | /Rethink Traffic/
>
> *Blog <http://miovision.com/blog> | **LinkedIn 
> <https://www.linkedin.com/company/miovision-technologies>  | Twitter 
> <https://twitter.com/miovision>  | Facebook 
> <https://www.facebook.com/miovision>*
> ------------------------------------------------------------------------
> Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, 
> ON, Canada | N2C 1L3
> This e-mail may contain information that is privileged or 
> confidential. If you are not the intended recipient, please delete the 
> e-mail and any attachments and notify us immediately.
>
>
> On Sat, Feb 15, 2014 at 6:24 AM, Dafna Ron <dron at redhat.com 
> <mailto:dron at redhat.com>> wrote:
>
>     the migration fails in libvirt:
>
>
>     Thread-153709::ERROR::2014-02-14
>     11:17:40,420::vm::337::vm.Vm::(run)
>     vmId=`08434c90-ffa3-4b63-aa8e-5613f7b0e0cd`::Failed to migrate
>     Traceback (most recent call last):
>       File "/usr/share/vdsm/vm.py", line 323, in run
>         self._startUnderlyingMigration()
>       File "/usr/share/vdsm/vm.py", line 403, in _startUnderlyingMigration
>         None, maxBandwidth)
>       File "/usr/share/vdsm/vm.py", line 841, in f
>         ret = attr(*args, **kwargs)
>       File
>     "/usr/lib64/python2.6/site-packages/vdsm/libvirtconnection.py",
>     line 76, in wrapper
>         ret = f(*args, **kwargs)
>       File "/usr/lib64/python2.6/site-packages/libvirt.py", line 1178,
>     in migrateToURI2
>         if ret == -1: raise libvirtError ('virDomainMigrateToURI2()
>     failed', dom=self)
>     libvirtError: Unable to read from monitor: Connection reset by peer
>     Thread-54041::DEBUG::2014-02-14
>     11:17:41,752::task::579::TaskManager.Task::(_updateState)
>     Task=`094c412a-43dc-4c29-a601-d759486469a8`::moving from state
>     init -> state preparing
>     Thread-54041::INFO::2014-02-14
>     11:17:41,753::logUtils::44::dispatcher::(wrapper) Run and protect:
>     getVolumeSize(sdUUID='a52938f7-2cf4-4771-acb2-0c78d14999e5',
>     spUUID='fcb89071-6cdb-4972-94d1-c9324cebf814',
>     imgUUID='97c9108f-a506-415f-ad2
>     c-370d707cb130', volUUID='61f82f7f-18e4-4ea8-9db3-71ddd9d4e836',
>     options=None)
>
>     Do you have the same libvirt/vdsm/qemu on both your hosts?
>     Please attach the libvirt and vm logs from both hosts.
>
>     Thanks,
>     Dafna
>
>
>
>     On 02/14/2014 04:50 PM, Steve Dainard wrote:
>
>         Quick overview:
>         Ovirt 3.3.2 running on CentOS 6.5
>         Two hosts: ovirt001, ovirt002
>         Migrating two VM's: puppet-agent1, puppet-agent2 from ovirt002
>         to ovirt001.
>
>         The first VM puppet-agent1 migrates successfully. The second
>         VM puppet-agent2 fails with "Migration failed due to Error:
>         Fatal error during migration (VM: puppet-agent2, Source:
>         ovirt002, Destination: ovirt001)."
>
>         I've attached the logs if anyone can help me track down the issue.
>
>         Thanks,
>
>         *Steve Dainard *
>         IT Infrastructure Manager
>         Miovision <http://miovision.com/> | /Rethink Traffic/
>
>         *Blog <http://miovision.com/blog> | **LinkedIn
>         <https://www.linkedin.com/company/miovision-technologies>  |
>         Twitter <https://twitter.com/miovision>  | Facebook
>         <https://www.facebook.com/miovision>*
>         ------------------------------------------------------------------------
>
>
>         Miovision Technologies Inc. | 148 Manitou Drive, Suite 101,
>         Kitchener, ON, Canada | N2C 1L3
>         This e-mail may contain information that is privileged or
>         confidential. If you are not the intended recipient, please
>         delete the e-mail and any attachments and notify us immediately.
>
>
>         _______________________________________________
>         Users mailing list
>         Users at ovirt.org <mailto:Users at ovirt.org>
>         http://lists.ovirt.org/mailman/listinfo/users
>
>
>
>     -- 
>     Dafna Ron
>
>


-- 
Dafna Ron



More information about the Users mailing list