[ovirt-users] Fwd: Re: ***UNCHECKED*** Re: kvm vcpu0 unhandled rdmsr
Yaniv Kaul
ykaul at redhat.com
Sun Apr 10 20:22:27 UTC 2016
On Sun, Apr 10, 2016 at 6:05 PM, gregor <gregor_forum at catrix.at> wrote:
> Hi,
>
> has anybody a last tip. Now the third installed Windows Server 2012 R2
> VM is damaged and I will move tomorrow my host back to VMWare and leave
> oVirt.
>
It's a QEMU/KVM issue - let me see if I can get someone from KVM
development team to get the details from you.
Y.
>
> regards
> gregor
>
> -------- Forwarded Message --------
> Subject: Re: [ovirt-users] ***UNCHECKED*** Re: kvm vcpu0 unhandled rdmsr
> Date: Mon, 4 Apr 2016 15:06:52 +0200
> From: gregor <gregor_forum at catrix.at>
> To: Yaniv Kaul <ykaul at redhat.com>
> CC: users <users at ovirt.org>
>
> Hi,
>
> the host and VM are all up-to-date with latest packages for CentOS 7.*.
>
> In /proc/cpuinfo I see "nx" in the flags list, the full list is on the
> and of the mail.
>
> Is it possible that this problem destroy the Windows Server 2012 R2 VM?
> Now I start the third installation, hopefully this time it will not get
> damaged. If it fails again I have to use another virtualization provider
> and leave oVirt, and I was so happy to leave VMWare :°(
>
> This is the command line for a VM (got with ps aux ...):
> /usr/libexec/qemu-kvm -name srv02 -S -machine
> pc-i440fx-rhel7.2.0,accel=kvm,usb=off -cpu Westmere -m
> size=2097152k,slots=16,maxmem=4294967296k -realtime mlock=off -smp
> 1,maxcpus=16,sockets=16,cores=1,threads=1 -numa
> node,nodeid=0,cpus=0,mem=2048 -uuid 6765fd03-ac0d-49ea-b8ba-cf10c60d3968
> -smbios type=1,manufacturer=oVirt,product=oVirt
>
> Node,version=7-2.1511.el7.centos.2.10,serial=39343937-3439-5A43-3135-353130324542,uuid=6765fd03-ac0d-49ea-b8ba-cf10c60d3968
> -no-user-config -nodefaults -chardev
>
> socket,id=charmonitor,path=/var/lib/libvirt/qemu/domain-srv02/monitor.sock,server,nowait
> -mon chardev=charmonitor,id=monitor,mode=control -rtc
> base=2016-04-03T21:24:06,driftfix=slew -global
> kvm-pit.lost_tick_policy=discard -no-hpet -no-shutdown -boot strict=on
> -device piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -device
> virtio-scsi-pci,id=scsi0,bus=pci.0,addr=0x4 -device
> virtio-serial-pci,id=virtio-serial0,max_ports=16,bus=pci.0,addr=0x5
> -drive if=none,id=drive-ide0-1-0,readonly=on,format=raw,serial= -device
> ide-cd,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0 -drive
>
> file=/rhev/data-center/00000001-0001-0001-0001-00000000033d/4443edf0-54aa-4ef5-84c2-a433813f304a/images/f596f9a8-c6c4-41b8-b547-7f83829807fe/5028abbd-35c8-4dcd-95a0-3d0c61dfc2b7,if=none,id=drive-virtio-disk0,format=raw,serial=f596f9a8-c6c4-41b8-b547-7f83829807fe,cache=none,werror=stop,rerror=stop,aio=threads
> -device
>
> virtio-blk-pci,scsi=off,bus=pci.0,addr=0x6,drive=drive-virtio-disk0,id=virtio-disk0,bootindex=1
> -netdev tap,fd=30,id=hostnet0,vhost=on,vhostfd=31 -device
>
> virtio-net-pci,netdev=hostnet0,id=net0,mac=00:1a:4a:16:01:57,bus=pci.0,addr=0x3
> -chardev
>
> socket,id=charchannel0,path=/var/lib/libvirt/qemu/channels/6765fd03-ac0d-49ea-b8ba-cf10c60d3968.com.redhat.rhevm.vdsm,server,nowait
> -device
>
> virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel0,name=com.redhat.rhevm.vdsm
> -chardev
>
> socket,id=charchannel1,path=/var/lib/libvirt/qemu/channels/6765fd03-ac0d-49ea-b8ba-cf10c60d3968.org.qemu.guest_agent.0,server,nowait
> -device
>
> virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel1,id=channel1,name=org.qemu.guest_agent.0
> -chardev spicevmc,id=charchannel2,name=vdagent -device
>
> virtserialport,bus=virtio-serial0.0,nr=3,chardev=charchannel2,id=channel2,name=com.redhat.spice.0
> -spice
>
> port=5904,tls-port=5905,addr=0,x509-dir=/etc/pki/vdsm/libvirt-spice,seamless-migration=on
> -device
>
> qxl-vga,id=video0,ram_size=67108864,vram_size=33554432,vgamem_mb=16,bus=pci.0,addr=0x2
> -device virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x7 -msg timestamp=on
>
> Here are the full flags list:
> fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36
> clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb
> rdtscp lm constant_tsc arch_perfmon pebs bts rep_good nopl xtopology
> nonstop_tsc aperfmperf eagerfpu pni pclmulqdq dtes64 monitor ds_cpl vmx
> smx est tm2 ssse3 fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe
> popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm arat epb
> pln pts dtherm tpr_shadow vnmi flexpriority ept vpid fsgsbase tsc_adjust
> bmi1 avx2 smep bmi2 erms invpcid cqm xsaveopt cqm_llc cqm_occup_llc
>
> On 04/04/16 08:52, Yaniv Kaul wrote:
> >
> >
> > On Sun, Apr 3, 2016 at 10:07 PM, gregor <gregor_forum at catrix.at
> > <mailto:gregor_forum at catrix.at>> wrote:
> >
> > Update: The problem occur when a VM reboots.
> > When I change the CPU Type from default "Intel Haswell-noTSX" to
> > "Westmere" the error is gone.
> >
> >
> > The error ""kvm ... vcpu0 unhandled rdmsr ..." is quite harmless.
> > I assume you are running the latest qemu/kvm packages.
> > Can you ensure NX is enabled on your host?
> > In any case, this is most likely a qemu/kvm issue - the command line of
> > the VM and information regarding the qemu packages and host versions
> > will be needed.
> > Y.
> >
> >
> >
> > But which CPU type is now the best so I don't lose performance.
> >
> > Host CPU: Intel(R) Xeon(R) CPU E5-2603 v3 @ 1.60GHz
> >
> > regards
> > gregor
> >
> > On 03/04/16 20:36, gregor wrote:
> > > Hi,
> > >
> > > on one Host I get very often the message
> > >
> > > "kvm ... vcpu0 unhandled rdmsr ..."
> > >
> > > When this occurs some VM's are stuck. Really bad is this for an
> > Windows
> > > Server 2012 R2 VM which stucks so heavy that the VM is getting
> corrupt
> > > and the VM is unable to boot anymore and a Windows Recovery in any
> way
> > > didn't help. Therefor I had to reinstall the VM, this works for
> some
> > > day's but now the VM is still damaged. So I can't use Windows
> Server
> > > 2012 R2 on this machine but the customer needs it and I have some
> days
> > > to ship it to my customer. So I have to decide to stay on oVirt or
> use
> > > another product. Besides, oVirt run on my others hosts (without
> > Windows
> > > VM) very well since a long time.
> > > On a CentOS 7 VM I have similar problems where the NIC is getting
> > > offline sometime + the XFS filesystems get some errors and I have
> > to fix
> > > this in recovery mode.
> > >
> > > oVirt: 3.6.4.1-1.el7.centos
> > > machine: HP ProLiant ML110 Gen9
> > > VM's: 3 CentOS 7 and one Windows Server 2012 R2
> > >
> > > I hope somebody can help.
> > >
> > > regards
> > > gregor
> > > _______________________________________________
> > > Users mailing list
> > > Users at ovirt.org <mailto:Users at ovirt.org>
> > > http://lists.ovirt.org/mailman/listinfo/users
> > >
> > _______________________________________________
> > Users mailing list
> > Users at ovirt.org <mailto:Users at ovirt.org>
> > http://lists.ovirt.org/mailman/listinfo/users
> >
> >
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20160410/e7c8e67d/attachment-0001.html>
More information about the Users
mailing list