[Users] unable to start vm in 3.3 and f19 with gluster

Gianluca Cecchi gianluca.cecchi at gmail.com
Tue Oct 1 17:57:06 EDT 2013


Today I was able to work again on this matter and it seems related to spice
Every time I start the VM (that is defined with spice) it goes in
paused state and after a few minutes the node gets unreachable, down
form the gui, no response also inside its console and forced to power
off it.

So I tried to create a vm taken from oVirt generated xml using gluster
to see if it is the problem in any way...
and start it via virsh and libvirt on the node.

See in attach:
I got the ovirt generated xml file, under /var/run/libvirt/qemu/C6.xml
myvm.xml the derived vm that defined and started and is able to go and
remain running

[root at ovnode02 ~]# virsh define myvm.xml
Please enter your authentication name: vuser
Please enter your password:
Domain myvm defined from myvm.xml

[root at ovnode02 ~]# virsh list --all
Please enter your authentication name: vuser
Please enter your password:
 Id    Name                           State
----------------------------------------------------
 -     myvm                           shut off

[root at ovnode02 ~]# virsh start myvm
Please enter your authentication name: vuser
Please enter your password:
Domain myvm started

[root at ovnode02 ~]# virsh list --all
Please enter your authentication name: vuser
Please enter your password:
 Id    Name                           State
----------------------------------------------------
 2     myvm                           running

In this case

[root at ovnode02 ~]# ps -ef|grep qemu
qemu      1617   574  0 18:24 ?        00:00:00 [python] <defunct>
qemu      6083     1  7 18:40 ?        00:00:14
/usr/bin/qemu-system-x86_64 -machine accel=kvm -name myvm -S -machine
pc-1.0,accel=kvm,usb=off -cpu Nehalem -m 2048 -smp
1,sockets=1,cores=1,threads=1 -uuid
dfadc661-6288-4f21-8faa-012daf29478f -nographic -no-user-config
-nodefaults -chardev
socket,id=charmonitor,path=/var/lib/libvirt/qemu/myvm.monitor,server,nowait
-mon chardev=charmonitor,id=monitor,mode=control -rtc
base=2013-10-01T16:40:55,driftfix=slew -no-shutdown -device
piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -device
virtio-scsi-pci,id=scsi0,bus=pci.0,addr=0x4 -device
virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x5 -drive
file=gluster://ovnode01/gv01/20042e7b-0929-48ca-ad40-2a2aa22f0689/images/d004045e-620b-4d90-8a7f-6c6d26393a08/dff09892-bc60-4de5-85c0-2a1fa215a161,if=none,id=drive-virtio-disk0,format=raw,cache=none
-device virtio-blk-pci,scsi=off,bus=pci.0,addr=0x6,drive=drive-virtio-disk0,id=virtio-disk0,bootindex=1
-device virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x3
root      7137  7083  0 18:44 pts/2    00:00:00 grep --color=auto qemu

If I undefine and redefine myvm adding the spice parts in it, I get
the same bad behaviour as in oVirt

[root at ovnode02 ~]# virsh list --all
Please enter your authentication name: vuser
Please enter your password:
 Id    Name                           State
----------------------------------------------------
 3     myvm                           paused

A confirmation of spice involvement I get if I run once C6 from oVirt
using vnc instead of spice.
Now the VM goes running and I can access its console from vnc and the
qemu commandline is:

qemu     10786     1  9 21:05 ?        00:00:14
/usr/bin/qemu-system-x86_64 -machine accel=kvm -name C6 -S -machine
pc-1.0,accel=kvm,usb=off -cpu Nehalem -m 2048 -smp
1,sockets=1,cores=1,threads=1 -uuid
409c5dbe-5e70-40de-bf73-46ef484ea2d7 -smbios
type=1,manufacturer=oVirt,product=oVirt
Node,version=19-3,serial=421F7170-C703-34E3-9628-4588D841F8B1,uuid=409c5dbe-5e70-40de-bf73-46ef484ea2d7
-no-user-config -nodefaults -chardev
socket,id=charmonitor,path=/var/lib/libvirt/qemu/C6.monitor,server,nowait
-mon chardev=charmonitor,id=monitor,mode=control -rtc
base=2013-10-01T19:05:47,driftfix=slew -no-shutdown -device
piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -device
virtio-scsi-pci,id=scsi0,bus=pci.0,addr=0x4 -device
virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x5 -drive
file=/rhev/data-center/mnt/ovirt.localdomain.local:_var_lib_exports_iso/6e80607d-5437-4fc5-b73c-66794f6381e0/images/11111111-1111-1111-1111-111111111111/CentOS-6.4-x86_64-bin-DVD1.iso,if=none,id=drive-ide0-1-0,readonly=on,format=raw,serial=
-device ide-cd,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0,bootindex=1
-drive file=gluster://ovnode01/gv01/20042e7b-0929-48ca-ad40-2a2aa22f0689/images/d004045e-620b-4d90-8a7f-6c6d26393a08/dff09892-bc60-4de5-85c0-2a1fa215a161,if=none,id=drive-virtio-disk0,format=raw,serial=d004045e-620b-4d90-8a7f-6c6d26393a08,cache=none,werror=stop,rerror=stop,aio=threads
-device virtio-blk-pci,scsi=off,bus=pci.0,addr=0x6,drive=drive-virtio-disk0,id=virtio-disk0,bootindex=2
-netdev tap,fd=27,id=hostnet0,vhost=on,vhostfd=28 -device
virtio-net-pci,netdev=hostnet0,id=net0,mac=00:1a:4a:bb:9f:10,bus=pci.0,addr=0x3,bootindex=3
-chardev socket,id=charchannel0,path=/var/lib/libvirt/qemu/channels/409c5dbe-5e70-40de-bf73-46ef484ea2d7.com.redhat.rhevm.vdsm,server,nowait
-device virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel0,name=com.redhat.rhevm.vdsm
-chardev socket,id=charchannel1,path=/var/lib/libvirt/qemu/channels/409c5dbe-5e70-40de-bf73-46ef484ea2d7.org.qemu.guest_agent.0,server,nowait
-device virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel1,id=channel1,name=org.qemu.guest_agent.0
-device usb-tablet,id=input0 -vnc 0:0,password -k en-us -vga cirrus
-device virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x7


Any info about how to debug?
Anyone got the former logs and see any information?

Take in mind that the environment is nested virtualization inside ESXi
5.1, so it can take a part too in the problem.
Anyone using spice over glusterfs on f19 without any problem?

On nodes:
[root at ovnode01 ~]# rpm -qa|grep vdsm
vdsm-python-4.12.1-2.fc19.x86_64
vdsm-python-cpopen-4.12.1-2.fc19.x86_64
vdsm-gluster-4.12.1-2.fc19.noarch
vdsm-cli-4.12.1-2.fc19.noarch
vdsm-4.12.1-2.fc19.x86_64
vdsm-xmlrpc-4.12.1-2.fc19.noarch

On engine
[root at ovirt ~]# rpm -qa|grep ovirt
ovirt-engine-restapi-3.3.0-4.fc19.noarch
ovirt-engine-sdk-python-3.3.0.6-1.fc19.noarch
ovirt-log-collector-3.3.0-1.fc19.noarch
ovirt-engine-lib-3.3.0-4.fc19.noarch
ovirt-engine-3.3.0-4.fc19.noarch
ovirt-release-fedora-8-1.noarch
ovirt-iso-uploader-3.3.0-1.fc19.noarch
ovirt-engine-cli-3.3.0.4-1.fc19.noarch
ovirt-engine-setup-3.3.0-4.fc19.noarch
ovirt-engine-dbscripts-3.3.0-4.fc19.noarch
ovirt-host-deploy-java-1.1.1-1.fc19.noarch
ovirt-image-uploader-3.3.0-1.fc19.noarch
ovirt-host-deploy-1.1.1-1.fc19.noarch
ovirt-engine-webadmin-portal-3.3.0-4.fc19.noarch
ovirt-engine-backend-3.3.0-4.fc19.noarch
ovirt-engine-userportal-3.3.0-4.fc19.noarch
ovirt-engine-tools-3.3.0-4.fc19.noarch

Thanks in advance,
Gianluca
-------------- next part --------------
A non-text attachment was scrubbed...
Name: debug_ovirt.zip
Type: application/zip
Size: 3889 bytes
Desc: not available
URL: <http://lists.ovirt.org/pipermail/users/attachments/20131001/71f504f9/attachment.zip>


More information about the Users mailing list