Today I was able to work again on this matter and it seems related to spice
Every time I start the VM (that is defined with spice) it goes in
paused state and after a few minutes the node gets unreachable, down
form the gui, no response also inside its console and forced to power
off it.
So I tried to create a vm taken from oVirt generated xml using gluster
to see if it is the problem in any way...
and start it via virsh and libvirt on the node.
See in attach:
I got the ovirt generated xml file, under /var/run/libvirt/qemu/C6.xml
myvm.xml the derived vm that defined and started and is able to go and
remain running
[root@ovnode02 ~]# virsh define myvm.xml
Please enter your authentication name: vuser
Please enter your password:
Domain myvm defined from myvm.xml
[root@ovnode02 ~]# virsh list --all
Please enter your authentication name: vuser
Please enter your password:
Id Name State
----------------------------------------------------
- myvm shut off
[root@ovnode02 ~]# virsh start myvm
Please enter your authentication name: vuser
Please enter your password:
Domain myvm started
[root@ovnode02 ~]# virsh list --all
Please enter your authentication name: vuser
Please enter your password:
Id Name State
----------------------------------------------------
2 myvm running
In this case
[root@ovnode02 ~]# ps -ef|grep qemu
qemu 1617 574 0 18:24 ? 00:00:00 [python] <defunct>
qemu 6083 1 7 18:40 ? 00:00:14
/usr/bin/qemu-system-x86_64 -machine accel=kvm -name myvm -S -machine
pc-1.0,accel=kvm,usb=off -cpu Nehalem -m 2048 -smp
1,sockets=1,cores=1,threads=1 -uuid
dfadc661-6288-4f21-8faa-012daf29478f -nographic -no-user-config
-nodefaults -chardev
socket,id=charmonitor,path=/var/lib/libvirt/qemu/myvm.monitor,server,nowait
-mon chardev=charmonitor,id=monitor,mode=control -rtc
base=2013-10-01T16:40:55,driftfix=slew -no-shutdown -device
piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -device
virtio-scsi-pci,id=scsi0,bus=pci.0,addr=0x4 -device
virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x5 -drive
file=gluster://ovnode01/gv01/20042e7b-0929-48ca-ad40-2a2aa22f0689/images/d004045e-620b-4d90-8a7f-6c6d26393a08/dff09892-bc60-4de5-85c0-2a1fa215a161,if=none,id=drive-virtio-disk0,format=raw,cache=none
-device
virtio-blk-pci,scsi=off,bus=pci.0,addr=0x6,drive=drive-virtio-disk0,id=virtio-disk0,bootindex=1
-device virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x3
root 7137 7083 0 18:44 pts/2 00:00:00 grep --color=auto qemu
If I undefine and redefine myvm adding the spice parts in it, I get
the same bad behaviour as in oVirt
[root@ovnode02 ~]# virsh list --all
Please enter your authentication name: vuser
Please enter your password:
Id Name State
----------------------------------------------------
3 myvm paused
A confirmation of spice involvement I get if I run once C6 from oVirt
using vnc instead of spice.
Now the VM goes running and I can access its console from vnc and the
qemu commandline is:
qemu 10786 1 9 21:05 ? 00:00:14
/usr/bin/qemu-system-x86_64 -machine accel=kvm -name C6 -S -machine
pc-1.0,accel=kvm,usb=off -cpu Nehalem -m 2048 -smp
1,sockets=1,cores=1,threads=1 -uuid
409c5dbe-5e70-40de-bf73-46ef484ea2d7 -smbios
type=1,manufacturer=oVirt,product=oVirt
Node,version=19-3,serial=421F7170-C703-34E3-9628-4588D841F8B1,uuid=409c5dbe-5e70-40de-bf73-46ef484ea2d7
-no-user-config -nodefaults -chardev
socket,id=charmonitor,path=/var/lib/libvirt/qemu/C6.monitor,server,nowait
-mon chardev=charmonitor,id=monitor,mode=control -rtc
base=2013-10-01T19:05:47,driftfix=slew -no-shutdown -device
piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -device
virtio-scsi-pci,id=scsi0,bus=pci.0,addr=0x4 -device
virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x5 -drive
file=/rhev/data-center/mnt/ovirt.localdomain.local:_var_lib_exports_iso/6e80607d-5437-4fc5-b73c-66794f6381e0/images/11111111-1111-1111-1111-111111111111/CentOS-6.4-x86_64-bin-DVD1.iso,if=none,id=drive-ide0-1-0,readonly=on,format=raw,serial=
-device ide-cd,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0,bootindex=1
-drive
file=gluster://ovnode01/gv01/20042e7b-0929-48ca-ad40-2a2aa22f0689/images/d004045e-620b-4d90-8a7f-6c6d26393a08/dff09892-bc60-4de5-85c0-2a1fa215a161,if=none,id=drive-virtio-disk0,format=raw,serial=d004045e-620b-4d90-8a7f-6c6d26393a08,cache=none,werror=stop,rerror=stop,aio=threads
-device
virtio-blk-pci,scsi=off,bus=pci.0,addr=0x6,drive=drive-virtio-disk0,id=virtio-disk0,bootindex=2
-netdev tap,fd=27,id=hostnet0,vhost=on,vhostfd=28 -device
virtio-net-pci,netdev=hostnet0,id=net0,mac=00:1a:4a:bb:9f:10,bus=pci.0,addr=0x3,bootindex=3
-chardev
socket,id=charchannel0,path=/var/lib/libvirt/qemu/channels/409c5dbe-5e70-40de-bf73-46ef484ea2d7.com.redhat.rhevm.vdsm,server,nowait
-device
virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel0,name=com.redhat.rhevm.vdsm
-chardev
socket,id=charchannel1,path=/var/lib/libvirt/qemu/channels/409c5dbe-5e70-40de-bf73-46ef484ea2d7.org.qemu.guest_agent.0,server,nowait
-device
virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel1,id=channel1,name=org.qemu.guest_agent.0
-device usb-tablet,id=input0 -vnc 0:0,password -k en-us -vga cirrus
-device virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x7
Any info about how to debug?
Anyone got the former logs and see any information?
Take in mind that the environment is nested virtualization inside ESXi
5.1, so it can take a part too in the problem.
Anyone using spice over glusterfs on f19 without any problem?
On nodes:
[root@ovnode01 ~]# rpm -qa|grep vdsm
vdsm-python-4.12.1-2.fc19.x86_64
vdsm-python-cpopen-4.12.1-2.fc19.x86_64
vdsm-gluster-4.12.1-2.fc19.noarch
vdsm-cli-4.12.1-2.fc19.noarch
vdsm-4.12.1-2.fc19.x86_64
vdsm-xmlrpc-4.12.1-2.fc19.noarch
On engine
[root@ovirt ~]# rpm -qa|grep ovirt
ovirt-engine-restapi-3.3.0-4.fc19.noarch
ovirt-engine-sdk-python-3.3.0.6-1.fc19.noarch
ovirt-log-collector-3.3.0-1.fc19.noarch
ovirt-engine-lib-3.3.0-4.fc19.noarch
ovirt-engine-3.3.0-4.fc19.noarch
ovirt-release-fedora-8-1.noarch
ovirt-iso-uploader-3.3.0-1.fc19.noarch
ovirt-engine-cli-3.3.0.4-1.fc19.noarch
ovirt-engine-setup-3.3.0-4.fc19.noarch
ovirt-engine-dbscripts-3.3.0-4.fc19.noarch
ovirt-host-deploy-java-1.1.1-1.fc19.noarch
ovirt-image-uploader-3.3.0-1.fc19.noarch
ovirt-host-deploy-1.1.1-1.fc19.noarch
ovirt-engine-webadmin-portal-3.3.0-4.fc19.noarch
ovirt-engine-backend-3.3.0-4.fc19.noarch
ovirt-engine-userportal-3.3.0-4.fc19.noarch
ovirt-engine-tools-3.3.0-4.fc19.noarch
Thanks in advance,
Gianluca