[Users] snapshot taken while live VM operates cause unresponsive hypervisor behavior

This is a multi-part message in MIME format. --------------080000040408020604050403 Content-Type: text/plain; charset=ISO-8859-1; format=flowed Content-Transfer-Encoding: 7bit Hi, Using 3.1.0-3.19 from dreyou's repo on Centos 6.3 version, running multiple VM on the hypervisor node and attempting to take a snapshot from one of the running VMs generates errors for VDSM daemon as show below: / //[root@hyper02 ~]# vdsClient -s 0 list// //Traceback (most recent call last):// // File "/usr/share/vdsm/vdsClient.py", line 2275, in <module>// // code, message = commands[command][0](commandArgs)// // File "/usr/share/vdsm/vdsClient.py", line 280, in do_list// // response = self.s.list(True, vms)// // File "/usr/lib64/python2.6/xmlrpclib.py", line 1199, in __call__// // return self.__send(self.__name, args)// // File "/usr/lib64/python2.6/xmlrpclib.py", line 1489, in __request// // verbose=self.__verbose// // File "/usr/lib64/python2.6/xmlrpclib.py", line 1253, in request// // return self._parse_response(h.getfile(), sock)// // File "/usr/lib64/python2.6/xmlrpclib.py", line 1392, in _parse_response// // return u.close()// // File "/usr/lib64/python2.6/xmlrpclib.py", line 838, in close// // raise Fault(**self._stack[0])// //Fault: <Fault 1: "<type 'exceptions.TypeError'>:cannot marshal None unless allow_none is enabled">/ Paste from ovirt admin GUI log: /2013-Feb-07, 15:29:59 VM ipa01 is down. Exit message: User shut down// //2013-Feb-07, 15:29:34 Migration failed due to Error: Fatal error during migration (VM: ipa01, Source Host: Hyper02).// //2013-Feb-07, 15:29:32 Starting migration of VM ipa01 from Host Hyper02 to Host Hyper01 (User: admin@internal.).// //2013-Feb-07, 15:29:10 Migration failed due to Error: Fatal error during migration (VM: ipa01, Source Host: Hyper02).// //2013-Feb-07, 15:29:10 Migration failed due to Error: Fatal error during migration (VM: ipa01, Source Host: Hyper02). Trying to migrate to another Host.// //2013-Feb-07, 15:29:08 Starting migration of VM ipa01 from Host Hyper02 to Host Hyper01 (User: admin@internal.).// //2013-Feb-07, 15:24:58 Detected new Host Hyper02. Host state was set to Up.// //2013-Feb-07, 15:24:53 Host Hyper02 is initializing. Message: Recovering from crash or Initializing// //2013-Feb-07, 15:22:18 VM ipa01 was resumed by admin@internal (Host: Hyper02).// //2013-Feb-07, 15:22:18 VM ipa01 was resumed by admin@internal (Host: Hyper02).// //2013-Feb-07, 15:20:43 Snapshot [initial installation][after ipa server installation] creation for VM ipa01 has been completed.// //2013-Feb-07, 15:20:43 VM ipa01 has paused due to unknown storage error./ Looking through htop command could see that VM were still running and responsive but any vdsclient command failed with the upper code paste I have given. Solution here was to connect to each VM, shut it down and restart vdsmd and power it up again. There's no problem with the storage size as it has aprox 800GB free of space and running through NFS. All nfs mount points were running okay at that time. After the VM has been snapshoot-ed went into pause mode and had to un-pause it from GUI, but the VM was still responsive through network and operations. Has anybody seen something related to this by now or it's just related to 3.1 version ? Mentioning the fact that I did a research on the similar mailing list topics and could not find something related to this error. I would happily provide logs if needed. Regards, Alex --------------080000040408020604050403 Content-Type: text/html; charset=ISO-8859-1 Content-Transfer-Encoding: 7bit <html> <head> <meta http-equiv="content-type" content="text/html; charset=ISO-8859-1"> </head> <body bgcolor="#FFFFFF" text="#000000"> <br> Hi,<br> <br> <br> Using 3.1.0-3.19 from dreyou's repo on Centos 6.3 version, running multiple VM on the hypervisor node and attempting to take a snapshot from one of the running VMs generates errors for VDSM daemon as show below:<br> <small><i><br> </i><i>[root@hyper02 ~]# vdsClient -s 0 list</i><i><br> </i><i>Traceback (most recent call last):</i><i><br> </i><i> File "/usr/share/vdsm/vdsClient.py", line 2275, in <module></i><i><br> </i><i> code, message = commands[command][0](commandArgs)</i><i><br> </i><i> File "/usr/share/vdsm/vdsClient.py", line 280, in do_list</i><i><br> </i><i> response = self.s.list(True, vms)</i><i><br> </i><i> File "/usr/lib64/python2.6/xmlrpclib.py", line 1199, in __call__</i><i><br> </i><i> return self.__send(self.__name, args)</i><i><br> </i><i> File "/usr/lib64/python2.6/xmlrpclib.py", line 1489, in __request</i><i><br> </i><i> verbose=self.__verbose</i><i><br> </i><i> File "/usr/lib64/python2.6/xmlrpclib.py", line 1253, in request</i><i><br> </i><i> return self._parse_response(h.getfile(), sock)</i><i><br> </i><i> File "/usr/lib64/python2.6/xmlrpclib.py", line 1392, in _parse_response</i><i><br> </i><i> return u.close()</i><i><br> </i><i> File "/usr/lib64/python2.6/xmlrpclib.py", line 838, in close</i><i><br> </i><i> raise Fault(**self._stack[0])</i><i><br> </i><i>Fault: <Fault 1: "<type 'exceptions.TypeError'>:cannot marshal None unless allow_none is enabled"></i></small><br> <br> <br> Paste from ovirt admin GUI log:<br> <br> <small><i>2013-Feb-07, 15:29:59 VM ipa01 is down. Exit message: User shut down</i><i><br> </i><i>2013-Feb-07, 15:29:34 Migration failed due to Error: Fatal error during migration (VM: ipa01, Source Host: Hyper02).</i><i><br> </i><i>2013-Feb-07, 15:29:32 Starting migration of VM ipa01 from Host Hyper02 to Host Hyper01 (User: admin@internal.).</i><i><br> </i><i>2013-Feb-07, 15:29:10 Migration failed due to Error: Fatal error during migration (VM: ipa01, Source Host: Hyper02).</i><i><br> </i><i>2013-Feb-07, 15:29:10 Migration failed due to Error: Fatal error during migration (VM: ipa01, Source Host: Hyper02). Trying to migrate to another Host.</i><i><br> </i><i>2013-Feb-07, 15:29:08 Starting migration of VM ipa01 from Host Hyper02 to Host Hyper01 (User: admin@internal.).</i><i><br> </i><i>2013-Feb-07, 15:24:58 Detected new Host Hyper02. Host state was set to Up.</i><i><br> </i><i>2013-Feb-07, 15:24:53 Host Hyper02 is initializing. Message: Recovering from crash or Initializing</i><i><br> </i><i>2013-Feb-07, 15:22:18 VM ipa01 was resumed by admin@internal (Host: Hyper02).</i><i><br> </i><i>2013-Feb-07, 15:22:18 VM ipa01 was resumed by admin@internal (Host: Hyper02).</i><i><br> </i><i>2013-Feb-07, 15:20:43 Snapshot [initial installation][after ipa server installation] creation for VM ipa01 has been completed.</i><i><br> </i><i>2013-Feb-07, 15:20:43 VM ipa01 has paused due to unknown storage error.</i></small><br> <br> Looking through htop command could see that VM were still running and responsive but any vdsclient command failed with the upper code paste I have given. Solution here was to connect to each VM, shut it down and restart vdsmd and power it up again. There's no problem with the storage size as it has aprox 800GB free of space and running through NFS. All nfs mount points were running okay at that time.<br> <br> After the VM has been snapshoot-ed went into pause mode and had to un-pause it from GUI, but the VM was still responsive through network and operations.<br> <br> Has anybody seen something related to this by now or it's just related to 3.1 version ? Mentioning the fact that I did a research on the similar mailing list topics and could not find something related to this error. I would happily provide logs if needed. <br> <br> <br> Regards,<br> Alex<br> </body> </html> --------------080000040408020604050403--

This is a multi-part message in MIME format. --------------040306010800040404050600 Content-Type: text/plain; charset=ISO-8859-1; format=flowed Content-Transfer-Encoding: 7bit The libvirt and qemu version in the VDSM host may help the debugging. Alexandru Vladulescu:
Hi,
Using 3.1.0-3.19 from dreyou's repo on Centos 6.3 version, running multiple VM on the hypervisor node and attempting to take a snapshot from one of the running VMs generates errors for VDSM daemon as show below: / //[root@hyper02 ~]# vdsClient -s 0 list// //Traceback (most recent call last):// // File "/usr/share/vdsm/vdsClient.py", line 2275, in <module>// // code, message = commands[command][0](commandArgs)// // File "/usr/share/vdsm/vdsClient.py", line 280, in do_list// // response = self.s.list(True, vms)// // File "/usr/lib64/python2.6/xmlrpclib.py", line 1199, in __call__// // return self.__send(self.__name, args)// // File "/usr/lib64/python2.6/xmlrpclib.py", line 1489, in __request// // verbose=self.__verbose// // File "/usr/lib64/python2.6/xmlrpclib.py", line 1253, in request// // return self._parse_response(h.getfile(), sock)// // File "/usr/lib64/python2.6/xmlrpclib.py", line 1392, in _parse_response// // return u.close()// // File "/usr/lib64/python2.6/xmlrpclib.py", line 838, in close// // raise Fault(**self._stack[0])// //Fault: <Fault 1: "<type 'exceptions.TypeError'>:cannot marshal None unless allow_none is enabled">/
Paste from ovirt admin GUI log:
/2013-Feb-07, 15:29:59 VM ipa01 is down. Exit message: User shut down// //2013-Feb-07, 15:29:34 Migration failed due to Error: Fatal error during migration (VM: ipa01, Source Host: Hyper02).// //2013-Feb-07, 15:29:32 Starting migration of VM ipa01 from Host Hyper02 to Host Hyper01 (User: admin@internal.).// //2013-Feb-07, 15:29:10 Migration failed due to Error: Fatal error during migration (VM: ipa01, Source Host: Hyper02).// //2013-Feb-07, 15:29:10 Migration failed due to Error: Fatal error during migration (VM: ipa01, Source Host: Hyper02). Trying to migrate to another Host.// //2013-Feb-07, 15:29:08 Starting migration of VM ipa01 from Host Hyper02 to Host Hyper01 (User: admin@internal.).// //2013-Feb-07, 15:24:58 Detected new Host Hyper02. Host state was set to Up.// //2013-Feb-07, 15:24:53 Host Hyper02 is initializing. Message: Recovering from crash or Initializing// //2013-Feb-07, 15:22:18 VM ipa01 was resumed by admin@internal (Host: Hyper02).// //2013-Feb-07, 15:22:18 VM ipa01 was resumed by admin@internal (Host: Hyper02).// //2013-Feb-07, 15:20:43 Snapshot [initial installation][after ipa server installation] creation for VM ipa01 has been completed.// //2013-Feb-07, 15:20:43 VM ipa01 has paused due to unknown storage error./
Looking through htop command could see that VM were still running and responsive but any vdsclient command failed with the upper code paste I have given. Solution here was to connect to each VM, shut it down and restart vdsmd and power it up again. There's no problem with the storage size as it has aprox 800GB free of space and running through NFS. All nfs mount points were running okay at that time.
After the VM has been snapshoot-ed went into pause mode and had to un-pause it from GUI, but the VM was still responsive through network and operations.
Has anybody seen something related to this by now or it's just related to 3.1 version ? Mentioning the fact that I did a research on the similar mailing list topics and could not find something related to this error. I would happily provide logs if needed.
Regards, Alex
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
-- --- ?? Shu Ming Open Virtualization Engineerning; CSTL, IBM Corp. Tel: 86-10-82451626 Tieline: 9051626 E-mail: shuming@cn.ibm.com or shuming@linux.vnet.ibm.com Address: 3/F Ring Building, ZhongGuanCun Software Park, Haidian District, Beijing 100193, PRC --------------040306010800040404050600 Content-Type: text/html; charset=ISO-8859-1 Content-Transfer-Encoding: 7bit <html> <head> <meta content="text/html; charset=ISO-8859-1" http-equiv="Content-Type"> </head> <body text="#000000" bgcolor="#FFFFFF"> <div class="moz-cite-prefix">The libvirt and qemu version in the VDSM host may help the debugging.<br> Alexandru Vladulescu:<br> </div> <blockquote cite="mid:5113AFF8.2040202@bfproject.ro" type="cite"> <meta http-equiv="content-type" content="text/html; charset=ISO-8859-1"> <br> Hi,<br> <br> <br> Using 3.1.0-3.19 from dreyou's repo on Centos 6.3 version, running multiple VM on the hypervisor node and attempting to take a snapshot from one of the running VMs generates errors for VDSM daemon as show below:<br> <small><i><br> </i><i>[root@hyper02 ~]# vdsClient -s 0 list</i><i><br> </i><i>Traceback (most recent call last):</i><i><br> </i><i> File "/usr/share/vdsm/vdsClient.py", line 2275, in <module></i><i><br> </i><i> code, message = commands[command][0](commandArgs)</i><i><br> </i><i> File "/usr/share/vdsm/vdsClient.py", line 280, in do_list</i><i><br> </i><i> response = self.s.list(True, vms)</i><i><br> </i><i> File "/usr/lib64/python2.6/xmlrpclib.py", line 1199, in __call__</i><i><br> </i><i> return self.__send(self.__name, args)</i><i><br> </i><i> File "/usr/lib64/python2.6/xmlrpclib.py", line 1489, in __request</i><i><br> </i><i> verbose=self.__verbose</i><i><br> </i><i> File "/usr/lib64/python2.6/xmlrpclib.py", line 1253, in request</i><i><br> </i><i> return self._parse_response(h.getfile(), sock)</i><i><br> </i><i> File "/usr/lib64/python2.6/xmlrpclib.py", line 1392, in _parse_response</i><i><br> </i><i> return u.close()</i><i><br> </i><i> File "/usr/lib64/python2.6/xmlrpclib.py", line 838, in close</i><i><br> </i><i> raise Fault(**self._stack[0])</i><i><br> </i><i>Fault: <Fault 1: "<type 'exceptions.TypeError'>:cannot marshal None unless allow_none is enabled"></i></small><br> <br> <br> Paste from ovirt admin GUI log:<br> <br> <small><i>2013-Feb-07, 15:29:59 VM ipa01 is down. Exit message: User shut down</i><i><br> </i><i>2013-Feb-07, 15:29:34 Migration failed due to Error: Fatal error during migration (VM: ipa01, Source Host: Hyper02).</i><i><br> </i><i>2013-Feb-07, 15:29:32 Starting migration of VM ipa01 from Host Hyper02 to Host Hyper01 (User: admin@internal.).</i><i><br> </i><i>2013-Feb-07, 15:29:10 Migration failed due to Error: Fatal error during migration (VM: ipa01, Source Host: Hyper02).</i><i><br> </i><i>2013-Feb-07, 15:29:10 Migration failed due to Error: Fatal error during migration (VM: ipa01, Source Host: Hyper02). Trying to migrate to another Host.</i><i><br> </i><i>2013-Feb-07, 15:29:08 Starting migration of VM ipa01 from Host Hyper02 to Host Hyper01 (User: admin@internal.).</i><i><br> </i><i>2013-Feb-07, 15:24:58 Detected new Host Hyper02. Host state was set to Up.</i><i><br> </i><i>2013-Feb-07, 15:24:53 Host Hyper02 is initializing. Message: Recovering from crash or Initializing</i><i><br> </i><i>2013-Feb-07, 15:22:18 VM ipa01 was resumed by admin@internal (Host: Hyper02).</i><i><br> </i><i>2013-Feb-07, 15:22:18 VM ipa01 was resumed by admin@internal (Host: Hyper02).</i><i><br> </i><i>2013-Feb-07, 15:20:43 Snapshot [initial installation][after ipa server installation] creation for VM ipa01 has been completed.</i><i><br> </i><i>2013-Feb-07, 15:20:43 VM ipa01 has paused due to unknown storage error.</i></small><br> <br> Looking through htop command could see that VM were still running and responsive but any vdsclient command failed with the upper code paste I have given. Solution here was to connect to each VM, shut it down and restart vdsmd and power it up again. There's no problem with the storage size as it has aprox 800GB free of space and running through NFS. All nfs mount points were running okay at that time.<br> <br> After the VM has been snapshoot-ed went into pause mode and had to un-pause it from GUI, but the VM was still responsive through network and operations.<br> <br> Has anybody seen something related to this by now or it's just related to 3.1 version ? Mentioning the fact that I did a research on the similar mailing list topics and could not find something related to this error. I would happily provide logs if needed. <br> <br> <br> Regards,<br> Alex<br> <br> <fieldset class="mimeAttachmentHeader"></fieldset> <br> <pre wrap="">_______________________________________________ Users mailing list <a class="moz-txt-link-abbreviated" href="mailto:Users@ovirt.org">Users@ovirt.org</a> <a class="moz-txt-link-freetext" href="http://lists.ovirt.org/mailman/listinfo/users">http://lists.ovirt.org/mailman/listinfo/users</a> </pre> </blockquote> <br> <br> <pre class="moz-signature" cols="72">-- --- 舒明 Shu Ming Open Virtualization Engineerning; CSTL, IBM Corp. Tel: 86-10-82451626 Tieline: 9051626 E-mail: <a class="moz-txt-link-abbreviated" href="mailto:shuming@cn.ibm.com">shuming@cn.ibm.com</a> or <a class="moz-txt-link-abbreviated" href="mailto:shuming@linux.vnet.ibm.com">shuming@linux.vnet.ibm.com</a> Address: 3/F Ring Building, ZhongGuanCun Software Park, Haidian District, Beijing 100193, PRC</pre> </body> </html> --------------040306010800040404050600--

also full engine, vdsm and libvirtd logs On 02/07/2013 05:05 PM, Shu Ming wrote:
The libvirt and qemu version in the VDSM host may help the debugging. Alexandru Vladulescu:
Hi,
Using 3.1.0-3.19 from dreyou's repo on Centos 6.3 version, running multiple VM on the hypervisor node and attempting to take a snapshot from one of the running VMs generates errors for VDSM daemon as show below: / //[root@hyper02 ~]# vdsClient -s 0 list// //Traceback (most recent call last):// //File "/usr/share/vdsm/vdsClient.py", line 2275, in <module>// //code, message = commands[command][0](commandArgs)// //File "/usr/share/vdsm/vdsClient.py", line 280, in do_list// //response = self.s.list(True, vms)// //File "/usr/lib64/python2.6/xmlrpclib.py", line 1199, in __call__// //return self.__send(self.__name, args)// //File "/usr/lib64/python2.6/xmlrpclib.py", line 1489, in __request// //verbose=self.__verbose// //File "/usr/lib64/python2.6/xmlrpclib.py", line 1253, in request// //return self._parse_response(h.getfile(), sock)// //File "/usr/lib64/python2.6/xmlrpclib.py", line 1392, in _parse_response// //return u.close()// //File "/usr/lib64/python2.6/xmlrpclib.py", line 838, in close// //raise Fault(**self._stack[0])// //Fault: <Fault 1: "<type 'exceptions.TypeError'>:cannot marshal None unless allow_none is enabled">/
Paste from ovirt admin GUI log:
/2013-Feb-07, 15:29:59 VM ipa01 is down. Exit message: User shut down// //2013-Feb-07, 15:29:34 Migration failed due to Error: Fatal error during migration (VM: ipa01, Source Host: Hyper02).// //2013-Feb-07, 15:29:32 Starting migration of VM ipa01 from Host Hyper02 to Host Hyper01 (User: admin@internal.).// //2013-Feb-07, 15:29:10 Migration failed due to Error: Fatal error during migration (VM: ipa01, Source Host: Hyper02).// //2013-Feb-07, 15:29:10 Migration failed due to Error: Fatal error during migration (VM: ipa01, Source Host: Hyper02). Trying to migrate to another Host.// //2013-Feb-07, 15:29:08 Starting migration of VM ipa01 from Host Hyper02 to Host Hyper01 (User: admin@internal.).// //2013-Feb-07, 15:24:58 Detected new Host Hyper02. Host state was set to Up.// //2013-Feb-07, 15:24:53 Host Hyper02 is initializing. Message: Recovering from crash or Initializing// //2013-Feb-07, 15:22:18 VM ipa01 was resumed by admin@internal (Host: Hyper02).// //2013-Feb-07, 15:22:18 VM ipa01 was resumed by admin@internal (Host: Hyper02).// //2013-Feb-07, 15:20:43 Snapshot [initial installation][after ipa server installation] creation for VM ipa01 has been completed.// //2013-Feb-07, 15:20:43 VM ipa01 has paused due to unknown storage error./
Looking through htop command could see that VM were still running and responsive but any vdsclient command failed with the upper code paste I have given. Solution here was to connect to each VM, shut it down and restart vdsmd and power it up again. There's no problem with the storage size as it has aprox 800GB free of space and running through NFS. All nfs mount points were running okay at that time.
After the VM has been snapshoot-ed went into pause mode and had to un-pause it from GUI, but the VM was still responsive through network and operations.
Has anybody seen something related to this by now or it's just related to 3.1 version ? Mentioning the fact that I did a research on the similar mailing list topics and could not find something related to this error. I would happily provide logs if needed.
Regards, Alex
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
-- --- 舒明 Shu Ming Open Virtualization Engineerning; CSTL, IBM Corp. Tel: 86-10-82451626 Tieline: 9051626 E-mail: shuming@cn.ibm.com or shuming@linux.vnet.ibm.com Address: 3/F Ring Building, ZhongGuanCun Software Park, Haidian District, Beijing 100193, PRC
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
-- Dafna Ron

Hi I add the same problem on ovrit3.1 on f17 and it is resolved on f18+ovirt3.2 Kevin 2013/2/7 Dafna Ron <dron@redhat.com>
also full engine, vdsm and libvirtd logs
On 02/07/2013 05:05 PM, Shu Ming wrote:
The libvirt and qemu version in the VDSM host may help the debugging. Alexandru Vladulescu:
Hi,
Using 3.1.0-3.19 from dreyou's repo on Centos 6.3 version, running multiple VM on the hypervisor node and attempting to take a snapshot from one of the running VMs generates errors for VDSM daemon as show below: / //[root@hyper02 ~]# vdsClient -s 0 list// //Traceback (most recent call last):// //File "/usr/share/vdsm/vdsClient.py", line 2275, in <module>// //code, message = commands[command][0](commandArgs)// //File "/usr/share/vdsm/vdsClient.py", line 280, in do_list// //response = self.s.list(True, vms)// //File "/usr/lib64/python2.6/xmlrpclib.py", line 1199, in __call__// //return self.__send(self.__name, args)// //File "/usr/lib64/python2.6/xmlrpclib.py", line 1489, in __request// //verbose=self.__verbose// //File "/usr/lib64/python2.6/xmlrpclib.py", line 1253, in request// //return self._parse_response(h.getfile(), sock)// //File "/usr/lib64/python2.6/xmlrpclib.py", line 1392, in _parse_response// //return u.close()// //File "/usr/lib64/python2.6/xmlrpclib.py", line 838, in close// //raise Fault(**self._stack[0])// //Fault: <Fault 1: "<type 'exceptions.TypeError'>:cannot marshal None unless allow_none is enabled">/
Paste from ovirt admin GUI log:
/2013-Feb-07, 15:29:59 VM ipa01 is down. Exit message: User shut down// //2013-Feb-07, 15:29:34 Migration failed due to Error: Fatal error during migration (VM: ipa01, Source Host: Hyper02).// //2013-Feb-07, 15:29:32 Starting migration of VM ipa01 from Host Hyper02 to Host Hyper01 (User: admin@internal.).// //2013-Feb-07, 15:29:10 Migration failed due to Error: Fatal error during migration (VM: ipa01, Source Host: Hyper02).// //2013-Feb-07, 15:29:10 Migration failed due to Error: Fatal error during migration (VM: ipa01, Source Host: Hyper02). Trying to migrate to another Host.// //2013-Feb-07, 15:29:08 Starting migration of VM ipa01 from Host Hyper02 to Host Hyper01 (User: admin@internal.).// //2013-Feb-07, 15:24:58 Detected new Host Hyper02. Host state was set to Up.// //2013-Feb-07, 15:24:53 Host Hyper02 is initializing. Message: Recovering from crash or Initializing// //2013-Feb-07, 15:22:18 VM ipa01 was resumed by admin@internal (Host: Hyper02).// //2013-Feb-07, 15:22:18 VM ipa01 was resumed by admin@internal (Host: Hyper02).// //2013-Feb-07, 15:20:43 Snapshot [initial installation][after ipa server installation] creation for VM ipa01 has been completed.// //2013-Feb-07, 15:20:43 VM ipa01 has paused due to unknown storage error./
Looking through htop command could see that VM were still running and responsive but any vdsclient command failed with the upper code paste I have given. Solution here was to connect to each VM, shut it down and restart vdsmd and power it up again. There's no problem with the storage size as it has aprox 800GB free of space and running through NFS. All nfs mount points were running okay at that time.
After the VM has been snapshoot-ed went into pause mode and had to un-pause it from GUI, but the VM was still responsive through network and operations.
Has anybody seen something related to this by now or it's just related to 3.1 version ? Mentioning the fact that I did a research on the similar mailing list topics and could not find something related to this error. I would happily provide logs if needed.
Regards, Alex
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
-- --- 舒明 Shu Ming Open Virtualization Engineerning; CSTL, IBM Corp. Tel: 86-10-82451626 Tieline: 9051626 E-mail: shuming@cn.ibm.com or shuming@linux.vnet.ibm.com Address: 3/F Ring Building, ZhongGuanCun Software Park, Haidian District, Beijing 100193, PRC
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
-- Dafna Ron _______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
-- Kevin Mazière Responsable Infrastructure Alter Way – Hosting 1 rue Royal - 227 Bureaux de la Colline 92213 Saint-Cloud Cedex Tél : +33 (0)1 41 16 38 41 Mob : +33 (0)7 62 55 57 05 http://www.alterway.fr

Hi, Thanks Kevin for your feedback on this. I will try to attached the conclusive logs for the guys if they consider/think to dig further into it for any new relevant probe or cases. So here are the rpm versions installed: / //[root@hyper02 ~]# rpm -qa | grep vdsm// //vdsm-cli-4.10.0-0.44.14.el6.noarch// //vdsm-4.10.0-0.44.14.el6.x86_64// //vdsm-gluster-4.10.0-0.44.14.el6.noarch// //vdsm-xmlrpc-4.10.0-0.44.14.el6.noarch// //vdsm-bootstrap-4.10.0-0.44.14.el6.noarch// //vdsm-python-4.10.0-0.44.14.el6.x86_64// //[root@hyper02 ~]# rpm -qa | grep qemu// //qemu-img-rhev-0.12.1.2-2.295.el6.8.x86_64// //qemu-kvm-rhev-tools-0.12.1.2-2.295.el6.8.x86_64// //gpxe-roms-qemu-0.9.7-6.9.el6.noarch// //qemu-kvm-rhev-0.12.1.2-2.295.el6.8.x86_64/ Regards, Alex. On 02/08/2013 10:44 AM, Kevin Maziere Aubry wrote:
Hi
I add the same problem on ovrit3.1 on f17 and it is resolved on f18+ovirt3.2
Kevin
2013/2/7 Dafna Ron <dron@redhat.com <mailto:dron@redhat.com>>
also full engine, vdsm and libvirtd logs
On 02/07/2013 05:05 PM, Shu Ming wrote: > The libvirt and qemu version in the VDSM host may help the debugging. > Alexandru Vladulescu: >> >> Hi, >> >> >> Using 3.1.0-3.19 from dreyou's repo on Centos 6.3 version, running >> multiple VM on the hypervisor node and attempting to take a snapshot >> from one of the running VMs generates errors for VDSM daemon as show >> below: >> / >> //[root@hyper02 ~]# vdsClient -s 0 list// >> //Traceback (most recent call last):// >> //File "/usr/share/vdsm/vdsClient.py", line 2275, in <module>// >> //code, message = commands[command][0](commandArgs)// >> //File "/usr/share/vdsm/vdsClient.py", line 280, in do_list// >> //response = self.s.list(True, vms)// >> //File "/usr/lib64/python2.6/xmlrpclib.py", line 1199, in __call__// >> //return self.__send(self.__name, args)// >> //File "/usr/lib64/python2.6/xmlrpclib.py", line 1489, in __request// >> //verbose=self.__verbose// >> //File "/usr/lib64/python2.6/xmlrpclib.py", line 1253, in request// >> //return self._parse_response(h.getfile(), sock)// >> //File "/usr/lib64/python2.6/xmlrpclib.py", line 1392, in >> _parse_response// >> //return u.close()// >> //File "/usr/lib64/python2.6/xmlrpclib.py", line 838, in close// >> //raise Fault(**self._stack[0])// >> //Fault: <Fault 1: "<type 'exceptions.TypeError'>:cannot marshal None >> unless allow_none is enabled">/ >> >> >> Paste from ovirt admin GUI log: >> >> /2013-Feb-07, 15:29:59 VM ipa01 is down. Exit message: User shut down// >> //2013-Feb-07, 15:29:34 Migration failed due to Error: Fatal error >> during migration (VM: ipa01, Source Host: Hyper02).// >> //2013-Feb-07, 15:29:32 Starting migration of VM ipa01 from Host >> Hyper02 to Host Hyper01 (User: admin@internal.).// >> //2013-Feb-07, 15:29:10 Migration failed due to Error: Fatal error >> during migration (VM: ipa01, Source Host: Hyper02).// >> //2013-Feb-07, 15:29:10 Migration failed due to Error: Fatal error >> during migration (VM: ipa01, Source Host: Hyper02). Trying to migrate >> to another Host.// >> //2013-Feb-07, 15:29:08 Starting migration of VM ipa01 from Host >> Hyper02 to Host Hyper01 (User: admin@internal.).// >> //2013-Feb-07, 15:24:58 Detected new Host Hyper02. Host state was set >> to Up.// >> //2013-Feb-07, 15:24:53 Host Hyper02 is initializing. Message: >> Recovering from crash or Initializing// >> //2013-Feb-07, 15:22:18 VM ipa01 was resumed by admin@internal (Host: >> Hyper02).// >> //2013-Feb-07, 15:22:18 VM ipa01 was resumed by admin@internal (Host: >> Hyper02).// >> //2013-Feb-07, 15:20:43 Snapshot [initial installation][after ipa >> server installation] creation for VM ipa01 has been completed.// >> //2013-Feb-07, 15:20:43 VM ipa01 has paused due to unknown storage >> error./ >> >> Looking through htop command could see that VM were still running and >> responsive but any vdsclient command failed with the upper code paste >> I have given. Solution here was to connect to each VM, shut it down >> and restart vdsmd and power it up again. There's no problem with the >> storage size as it has aprox 800GB free of space and running through >> NFS. All nfs mount points were running okay at that time. >> >> After the VM has been snapshoot-ed went into pause mode and had to >> un-pause it from GUI, but the VM was still responsive through network >> and operations. >> >> Has anybody seen something related to this by now or it's just >> related to 3.1 version ? Mentioning the fact that I did a research on >> the similar mailing list topics and could not find something related >> to this error. I would happily provide logs if needed. >> >> >> Regards, >> Alex >> >> >> _______________________________________________ >> Users mailing list >> Users@ovirt.org <mailto:Users@ovirt.org> >> http://lists.ovirt.org/mailman/listinfo/users > > > -- > --- > ?? Shu Ming > Open Virtualization Engineerning; CSTL, IBM Corp. > Tel: 86-10-82451626 Tieline: 9051626 E-mail: shuming@cn.ibm.com <mailto:shuming@cn.ibm.com> or shuming@linux.vnet.ibm.com <mailto:shuming@linux.vnet.ibm.com> > Address: 3/F Ring Building, ZhongGuanCun Software Park, Haidian District, Beijing 100193, PRC > > > _______________________________________________ > Users mailing list > Users@ovirt.org <mailto:Users@ovirt.org> > http://lists.ovirt.org/mailman/listinfo/users
-- Dafna Ron _______________________________________________ Users mailing list Users@ovirt.org <mailto:Users@ovirt.org> http://lists.ovirt.org/mailman/listinfo/users
--
Kevin Mazière Responsable Infrastructure Alter Way -- Hosting 1 rue Royal - 227 Bureaux de la Colline 92213 Saint-Cloud Cedex Tél : +33 (0)1 41 16 38 41 Mob : +33 (0)7 62 55 57 05 http://www.alterway.fr <http://www.alterway.fr/>
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
-- Alexandru Vladulescu Platform System Engineer --------------------------------------------------------------------------------- Bright Future Project Romania Web url : www.bfproject.ro Skype : avladulescu Mobile : +4(0)726.373.098 ---------------------------------------------------------------------------------
participants (4)
-
Alexandru Vladulescu
-
Dafna Ron
-
Kevin Maziere Aubry
-
Shu Ming