[Users] Live migration of VM's occasionally fails

Dafna Ron dron at redhat.com
Mon Feb 17 16:47:29 UTC 2014


did you install these vm's from a cd? run it as run-once with a special 
monitor?
try to think if there is anything different in the configuration of 
these vm's from the other vm's that succeed to migrate?

On 02/17/2014 04:36 PM, Steve Dainard wrote:
> Hi Dafna,
>
> No snapshots of either of those VM's have been taken, and there are no 
> updates for any of those packages on EL 6.5.
>
> *Steve Dainard *
> IT Infrastructure Manager
> Miovision <http://miovision.com/> | /Rethink Traffic/
>
> *Blog <http://miovision.com/blog> | **LinkedIn 
> <https://www.linkedin.com/company/miovision-technologies>  | Twitter 
> <https://twitter.com/miovision>  | Facebook 
> <https://www.facebook.com/miovision>*
> ------------------------------------------------------------------------
> Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, 
> ON, Canada | N2C 1L3
> This e-mail may contain information that is privileged or 
> confidential. If you are not the intended recipient, please delete the 
> e-mail and any attachments and notify us immediately.
>
>
> On Sun, Feb 16, 2014 at 7:05 AM, Dafna Ron <dron at redhat.com 
> <mailto:dron at redhat.com>> wrote:
>
>     does the vm that fails migration have a live snapshot?
>     if so how many snapshots does the vm have.
>     I think that there are newer packages of vdsm, libvirt and qemu -
>     can you try to update
>
>
>
>     On 02/16/2014 12:33 AM, Steve Dainard wrote:
>
>         Versions are the same:
>
>         [root at ovirt001 ~]# rpm -qa | egrep 'libvirt|vdsm|qemu' | sort
>         gpxe-roms-qemu-0.9.7-6.10.el6.noarch
>         libvirt-0.10.2-29.el6_5.3.x86_64
>         libvirt-client-0.10.2-29.el6_5.3.x86_64
>         libvirt-lock-sanlock-0.10.2-29.el6_5.3.x86_64
>         libvirt-python-0.10.2-29.el6_5.3.x86_64
>         qemu-img-rhev-0.12.1.2-2.355.el6.5.x86_64
>         qemu-kvm-rhev-0.12.1.2-2.355.el6.5.x86_64
>         qemu-kvm-rhev-tools-0.12.1.2-2.355.el6.5.x86_64
>         vdsm-4.13.3-3.el6.x86_64
>         vdsm-cli-4.13.3-3.el6.noarch
>         vdsm-gluster-4.13.3-3.el6.noarch
>         vdsm-python-4.13.3-3.el6.x86_64
>         vdsm-xmlrpc-4.13.3-3.el6.noarch
>
>         [root at ovirt002 ~]# rpm -qa | egrep 'libvirt|vdsm|qemu' | sort
>         gpxe-roms-qemu-0.9.7-6.10.el6.noarch
>         libvirt-0.10.2-29.el6_5.3.x86_64
>         libvirt-client-0.10.2-29.el6_5.3.x86_64
>         libvirt-lock-sanlock-0.10.2-29.el6_5.3.x86_64
>         libvirt-python-0.10.2-29.el6_5.3.x86_64
>         qemu-img-rhev-0.12.1.2-2.355.el6.5.x86_64
>         qemu-kvm-rhev-0.12.1.2-2.355.el6.5.x86_64
>         qemu-kvm-rhev-tools-0.12.1.2-2.355.el6.5.x86_64
>         vdsm-4.13.3-3.el6.x86_64
>         vdsm-cli-4.13.3-3.el6.noarch
>         vdsm-gluster-4.13.3-3.el6.noarch
>         vdsm-python-4.13.3-3.el6.x86_64
>         vdsm-xmlrpc-4.13.3-3.el6.noarch
>
>         Logs attached, thanks.
>
>         *Steve Dainard *
>         IT Infrastructure Manager
>         Miovision <http://miovision.com/> | /Rethink Traffic/
>
>         *Blog <http://miovision.com/blog> | **LinkedIn
>         <https://www.linkedin.com/company/miovision-technologies>  |
>         Twitter <https://twitter.com/miovision>  | Facebook
>         <https://www.facebook.com/miovision>*
>         ------------------------------------------------------------------------
>         Miovision Technologies Inc. | 148 Manitou Drive, Suite 101,
>         Kitchener, ON, Canada | N2C 1L3
>         This e-mail may contain information that is privileged or
>         confidential. If you are not the intended recipient, please
>         delete the e-mail and any attachments and notify us immediately.
>
>
>         On Sat, Feb 15, 2014 at 6:24 AM, Dafna Ron <dron at redhat.com
>         <mailto:dron at redhat.com> <mailto:dron at redhat.com
>         <mailto:dron at redhat.com>>> wrote:
>
>             the migration fails in libvirt:
>
>
>             Thread-153709::ERROR::2014-02-14
>             11:17:40,420::vm::337::vm.Vm::(run)
>             vmId=`08434c90-ffa3-4b63-aa8e-5613f7b0e0cd`::Failed to migrate
>             Traceback (most recent call last):
>               File "/usr/share/vdsm/vm.py", line 323, in run
>                 self._startUnderlyingMigration()
>               File "/usr/share/vdsm/vm.py", line 403, in
>         _startUnderlyingMigration
>                 None, maxBandwidth)
>               File "/usr/share/vdsm/vm.py", line 841, in f
>                 ret = attr(*args, **kwargs)
>               File
>            
>         "/usr/lib64/python2.6/site-packages/vdsm/libvirtconnection.py",
>             line 76, in wrapper
>                 ret = f(*args, **kwargs)
>               File "/usr/lib64/python2.6/site-packages/libvirt.py",
>         line 1178,
>             in migrateToURI2
>                 if ret == -1: raise libvirtError
>         ('virDomainMigrateToURI2()
>             failed', dom=self)
>             libvirtError: Unable to read from monitor: Connection
>         reset by peer
>             Thread-54041::DEBUG::2014-02-14
>             11:17:41,752::task::579::TaskManager.Task::(_updateState)
>             Task=`094c412a-43dc-4c29-a601-d759486469a8`::moving from state
>             init -> state preparing
>             Thread-54041::INFO::2014-02-14
>             11:17:41,753::logUtils::44::dispatcher::(wrapper) Run and
>         protect:
>             getVolumeSize(sdUUID='a52938f7-2cf4-4771-acb2-0c78d14999e5',
>             spUUID='fcb89071-6cdb-4972-94d1-c9324cebf814',
>             imgUUID='97c9108f-a506-415f-ad2
>             c-370d707cb130',
>         volUUID='61f82f7f-18e4-4ea8-9db3-71ddd9d4e836',
>             options=None)
>
>             Do you have the same libvirt/vdsm/qemu on both your hosts?
>             Please attach the libvirt and vm logs from both hosts.
>
>             Thanks,
>             Dafna
>
>
>
>             On 02/14/2014 04:50 PM, Steve Dainard wrote:
>
>                 Quick overview:
>                 Ovirt 3.3.2 running on CentOS 6.5
>                 Two hosts: ovirt001, ovirt002
>                 Migrating two VM's: puppet-agent1, puppet-agent2 from
>         ovirt002
>                 to ovirt001.
>
>                 The first VM puppet-agent1 migrates successfully. The
>         second
>                 VM puppet-agent2 fails with "Migration failed due to
>         Error:
>                 Fatal error during migration (VM: puppet-agent2, Source:
>                 ovirt002, Destination: ovirt001)."
>
>                 I've attached the logs if anyone can help me track
>         down the issue.
>
>                 Thanks,
>
>                 *Steve Dainard *
>                 IT Infrastructure Manager
>                 Miovision <http://miovision.com/> | /Rethink Traffic/
>
>                 *Blog <http://miovision.com/blog> | **LinkedIn
>                
>         <https://www.linkedin.com/company/miovision-technologies>  |
>                 Twitter <https://twitter.com/miovision>  | Facebook
>                 <https://www.facebook.com/miovision>*
>                
>         ------------------------------------------------------------------------
>
>
>                 Miovision Technologies Inc. | 148 Manitou Drive, Suite
>         101,
>                 Kitchener, ON, Canada | N2C 1L3
>                 This e-mail may contain information that is privileged or
>                 confidential. If you are not the intended recipient,
>         please
>                 delete the e-mail and any attachments and notify us
>         immediately.
>
>
>                 _______________________________________________
>                 Users mailing list
>         Users at ovirt.org <mailto:Users at ovirt.org>
>         <mailto:Users at ovirt.org <mailto:Users at ovirt.org>>
>
>         http://lists.ovirt.org/mailman/listinfo/users
>
>
>
>             --     Dafna Ron
>
>
>
>
>     -- 
>     Dafna Ron
>
>


-- 
Dafna Ron



More information about the Users mailing list