[ANN] oVirt 4.3.4 Fourth Release Candidate is now available
by Sandro Bonazzola
The oVirt Project is pleased to announce the availability of the oVirt
4.3.4 FourthRelease Candidate, as of June 6th, 2019.
This update is a release candidate of the fourth in a series of
stabilization updates to the 4.3 series.
This is pre-release software. This pre-release should not to be used
inproduction.
This release is available now on x86_64 architecture for:
* Red Hat Enterprise Linux 7.6 or later
* CentOS Linux (or similar) 7.6 or later
This release supports Hypervisor Hosts on x86_64 and ppc64le architectures
for:
* Red Hat Enterprise Linux 7.6 or later
* CentOS Linux (or similar) 7.6 or later
* oVirt Node 4.3 (available for x86_64 only)
Experimental tech preview for x86_64 and s390x architectures for Fedora 28
is also included.
See the release notes [1] for installation / upgrade instructions and a
list of new features and bugs fixed.
Notes:
- oVirt Appliance is already available
- oVirt Node is already available[2]
- oVirt Windows Guest Tools iso is already available [2]
Additional Resources:
* Read more about the oVirt 4.3.4 release highlights:
http://www.ovirt.org/release/4.3.4/
* Get more oVirt Project updates on Twitter: https://twitter.com/ovirt
* Check out the latest project news on the oVirt blog:
http://www.ovirt.org/blog/
[1] http://www.ovirt.org/release/4.3.4/
[2] http://resources.ovirt.org/pub/ovirt-4.3-pre/iso/
--
Sandro Bonazzola
MANAGER, SOFTWARE ENGINEERING, EMEA R&D RHV
Red Hat EMEA <https://www.redhat.com/>
sbonazzo(a)redhat.com
<https://red.ht/sig>
<https://redhat.com/summit>
5 years, 6 months
high number of interface RX errors on ovirtmgmt network
by Jayme
I have a three node HCI setup on Dell R720s running the latest stable
version of 4.3.3
Each hosts has a 1gig link and a 10gig link. The 1gig is used for ovirt
management network and 10gig link is used for backend glusterFS traffic.
I haven't noticed before but after installing ovirt metrics store I'm
seeing that gig interface used for ovirtmgmt on all three hosts are showing
high RX error rates. The 10gig interfaces for glusterFS on all three hosts
appear to be fine.
The 1gig ethernet controllers are: Broadcom Inc. and subsidiaries NetXtreme
II BCM57800 1/10 Gigabit Ethernet (rev 10)
Other physical servers on the same network/switches outside of oVirt have
zero RX errors.
Here is an example of what I'm seeing:
host0:
# ip -s link show em3
4: em3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master
ovirtmgmt state UP mode DEFAULT group default qlen 1000
link/ether b0:83:fe:cc:9a:2d brd ff:ff:ff:ff:ff:ff
RX: bytes packets errors dropped overrun mcast
51777532544474 36233202312 416993 0 0 2062421
TX: bytes packets errors dropped carrier collsns
7284362442704 18685883330 0 0 0 0
host1:
# ip -s link show em3
4: em3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master
ovirtmgmt state UP mode DEFAULT group default qlen 1000
link/ether b0:83:fe:cc:99:31 brd ff:ff:ff:ff:ff:ff
RX: bytes packets errors dropped overrun mcast
9518766859330 14424644226 89638 0 0 2056578
TX: bytes packets errors dropped carrier collsns
27866585257227 22323979969 0 0 0 0
host2:
# ip -s link show em3
4: em3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master
ovirtmgmt state UP mode DEFAULT group default qlen 1000
link/ether b0:83:fe:cc:92:50 brd ff:ff:ff:ff:ff:ff
RX: bytes packets errors dropped overrun mcast
6409138012195 13045254148 14825 0 0 2040655
TX: bytes packets errors dropped carrier collsns
31577745516683 23466818659 0 0 0 0
Anyone have any ideas why the RX error rate on the ovirtmgmt network could
be so high?
5 years, 6 months
Hosted Engine Abruptly Stopped Responding - Unexpected Shutdown
by souvaliotimaria@mail.com
Hello,
I came upon a problem the previous month that I figured it would be good to discuss here. I'm sorry I didn't post here earlier but time slipped me.
I have set up a glustered, hyperconverged oVirt environment for experimental use as a means to see its behaviour and get used to its management and performance before setting it up as a production environment for use in our organization. The environment is up and running since 2018 October. The three nodes are HP ProLiant DL380 G7 and have the following characteristics:
Mem: 22GB
CPU: 2x Hexa Core - Intel Xeon Hexa Core E56xx
HDD: 5x 300GB
Network: BCM5709C with dual-port Gigabit
OS: Linux RedHat 7.5.1804(Core 3.10.0-862.3.2.el7.x86_64 x86_64) - Ovirt Node 4.2.3.1
As I was working on the environment, the engine stopped working.
Not long before the time the HE stopped, I was in the web interface managing my VMs, when the browser froze and the HE was also not responding to ICMP requests.
The first thing I did was to connect via ssh to all nodes and run the command
#hosted-engine --vm-status
which showed that the HE was down in nodes 1 and 2 and up on the 3rd node.
After executing
#virsh -r list
the VM list that was shown contained two of the VMs I had previously created and were up; the HE was nowhere.
I tried to restart the HE with the
#hosted-engine --vm-start
but it didn't work.
I then put all nodes in maintenance mode with the command
#hosted-engine --set-maintenance --mode=global
(I guess I should have done that earlier) and re-run
#hosted-engine --vm-start
that had the same result as it previously did.
After checking the mails the system sent to the root user, I saw there were several mails on the 3rd node (where the HE had been), informing of the HE's state. The messages were changing between EngineDown-EngineStart, EngineStart-EngineStarting, EngineStarting-EngineMaybeAway, EngineMaybeAway-EngineUnexpectedlyDown, EngineUnexpectedlyDown-EngineDown, EngineDown-EngineStart and so forth.
I continued by searching the following logs in all nodes :
/var/log/libvirt/qemu/HostedEngine.log
/var/log/libvirt/qemu/win10.log
/var/log/libvirt/qemu/DNStest.log
/var/log/vdsm/vdsm.log
/var/log/ovirt-hosted-engine-ha/agent.log
After that I spotted and error that had started appearing almost a month ago in node #2:
ERROR Internal server error Traceback (most recent call last): File "/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py", line 606, in _handle_request res = method(**params) File "/usr/lib/python2.7/site-packages/vdsm/rpc/Bridge.py", line 197, in _dynamicMethod result = fn(*methodArgs) File "/usr/lib/python2.7/site-packages/vdsm/gluster/apiwrapper.py", line 85, in logicalVolumeList return self._gluster.logicalVolumeList() File "/usr/lib/python2.7/site-packages/vdsm/gluster/api.py", line 90, in wrapper rv = func(*args, **kwargs) File "/usr/lib/python2.7/site-packages/vdsm/gluster/api.py", line 808, in logicalVolumeList status = self.svdsmProxy.glusterLogicalVolumeList() File "/usr/lib/python2.7/site-packages/vdsm/common/supervdsm.py", line 55, in __call__ return callMethod() File "/usr/lib/python2.7/site-packages/vdsm/common/supervdsm.py", line 52, in <lambda> getattr(self._supervdsmProxy._svdsm, self._funcName)(*args, AttributeError: 'AutoProxy[instance]' object has no attribute 'glusterLogicalVolumeList'
The outputs of the following commands were also checked as a way to see if there was a mandatory process missing/killed, a memory problem or even disk space shortage that led to the sudden death of a process
#ps -A
#top
#free -h
#df -hT
Finally, after some time delving in the logs, the output of the
#journalctl --dmesg
showed the following message
"Out of memory: Kill process 5422 (qemu-kvm) score 514 or sacrifice child.
Killed process 5422 (qemu-kvm) total-vm:17526548kB, anon-rss:9310396kB,
file-rss:2336kB, shmem-rss:12kB"
which after that the ovirtmgmt started not responding.
I tried to restart the vhostd by executing
#/etc/rc.d/init.d/vhostmd start
but it didn't work.
Finally, I decided to run the HE restart command on the other nodes as well (I'd figured that since the HE was last running on the node #3, that's where I should try to restart it). So, I run
#hosted-engine --vm-start
and the output was
"Command VM.getStats with args {'vmID':'...<το ID της HE>....'} failed:
(code=1,message=Virtual machine does not exist: {'vmID':'...<το ID της
HE>....'})"
And then I run the command again and the output was
"VM exists and its status is Powering Up."
After that I executed
#virsh -r list
and the output was the following:
Id Name State
----------------------------------------------------
2 HostedEngine running
After the HE's restart two mails came that stated: ReinitializeFSMEngineStarting and EngineStarting-EngineUp
After that and after checking that we had access to the web interface again, we executed
hosted-engine --set-maintenance --mode=none
to get out of the maintenance mode.
The thing is, I still am not 1000% sure what the problem was that led to the shutdown of the hosted engine and I think that maybe some of the steps I took were not needed. I believe it was because the process qemu-kvm was killed after there was not enough memory for it but is this the real cause? I wasn't doing anything unusual before the shutdown to believe it was because of the new VM that was still in shutdown mode or anything of the sort. Also, I believe it may be because of memory shortage because I hadn't executed the
#sync ; echo 3 > /proc/sys/vm/drop_caches
command for a couple of weeks.
What are your thoughts on this? Could you point me to where to search for more information on the topic or tell me what is the right process to follow when something like this happens?
Also, I have set up a few VMs but only three are Up and they have no users yet, even so the buffers fill almost to the brim when the usage is almost non-existant. If you have an environment that has some users or you use the VMs as virtual servers of some sort, what is the consumption of the memory? What's the optimal size for the memory?
Thank you all very much.
5 years, 6 months
Bond Mode 1 (Active-Backup),vm unreachable for minutes when bond link change
by henaumars@sina.com
Hello,
I've a problem, all my ovirt hosts and vms are linked with a bonding mode 1(Active-Backup)2x10Gbps
ovirt version:4.3
topology:
----------eno2
vm----------ovirtmgmt------bond0-----------eno1
ifcfg-bond0:
# Generated by VDSM version 4.30.9.1
DEVICE=bond0
BONDING_OPTIOS='mode=1 miion=100'
BRIDGE=ovirtmgmt
MACADDR=a4:be:26:16:e9:b2
ONBOOT=yes
MTU=1500
DEFROUTE=no
NM_CONTROLLER=no
IPV6INIT=no
ifcfg-eno1:
# Generated by VDSM version 4.30.9.1
DEVICE=eno1
MASTER=bond0
SLAVE=yes
ONBOOT=yes
MTU=1500
DEFROUTE=no
NM_CONTROLLER=no
IPV6INIT=no
ifcfg-eno2:
# Generated by VDSM version 4.30.9.1
DEVICE=eno2
MASTER=bond0
SLAVE=yes
ONBOOT=yes
MTU=1500
DEFROUTE=no
NM_CONTROLLER=no
IPV6INIT=no
ifcfg-ovirtmgmt:
# Generated by VDSM version 4.30.9.1
DEVICE=ovirtmgmt
TYPE=Brodge
DELAY=0
STP=off
ONBOOT=yes
IPADDR=x.x.x.x
NEYMASK=255.255.255.0
GATEWAY=x.x.x.x
BOOTPROTO=none
MTU=1500
DEFROUTE=yes
NM_CONTROLLER=no
IPV6INIT=yes
IPV6_AUTOCONF=yes
cat /proc/net/bonding/bond0
Ethernet Chanel Bonding Driver:v3.7.1(April 27, 2011)
Bonding Mode:fault-tolerance(active-ackup)
Primary Slave:none
Currently Active Slave:eno1
MII Status:up
MII Polling Intercal (ms):100
Up Delay (ms) : 0
Down Delay (ms) : 0
Slave Interface :eno1
MII Status:up
Speed : 10000 Mbps
Link Failure Count : 0
Permanent HW addr :a4:be:26:16:e9:b2
Slave queue ID: 0
Slave Interface :eno2
MII Status:up
Speed : 10000 Mbps
Link Failure Count : 0
Permanent HW addr :a4:be:26:16:e9:b2
Slave queue ID: 0
ping vm from different subnet.
Eveything is okay if I don't change bond link interface。When I unplug Currently Active Slave eno1,bond link change to eno2 as expected but vm become unreachable until external physical switch MAC Table ageing time expired.It seems that vm doesn't sent gratuitous ARP when bond link change. How can I fix if?
vm os is Centos 7.5
ovirt version 4.2 also tested.
5 years, 6 months
oVirt on RHEL 8.
by vrk5398@gmail.com
Hello
Did anyone managed to install oVirt on RHEL8?
If so, Can you please guide me?
I'm trying to install but it throws an error as "Failed to synchronize repo for ovirt-*".
Please help me.
Thank You.
5 years, 7 months
oVirt Hosted-Engine upgrade filed
by mail@set-pro.net
Hello! Get problems with upgrade oVirt Hosted-Engine from 4.2.8 to 4.3.3. After installing http://resources.ovirt.org/pub/yum-repo/ovirt-release43.rpm on Engine VM i run yum update, then engine-setup and got this error:
--== DATABASE CONFIGURATION ==--
[WARNING] This release requires PostgreSQL server 10.6 but the engine database is currently hosted on PostgreSQL server 9.5.14.
[ INFO ] Verifying PostgreSQL SELinux file context rules
[ ERROR ] Failed to execute stage 'Environment customization': could not stat file "base/13699369/1259_fsm": Permission denied
[ INFO ] Stage: Clean up
Log file is located at /var/log/ovirt-engine/setup/ovirt-engine-setup-20190602110956-iqr6bc.log
[ INFO ] Generating answer file '/var/lib/ovirt-engine/setup/answers/20190602111120-setup.conf'
[ INFO ] Stage: Pre-termination
[ INFO ] Stage: Termination
[ ERROR ] Execution of setup failed
But i don't find any solution how resolve this issue. Maybe i doing something wrong?
5 years, 7 months
Followup: Create New VM from Template Glitch
by Andrei Verovski
Hi !
Followup to my previous post.
Creating template from snapshot in 4.2.8 also suffers from the same bug - allocation policy selected to thin provision even if I select RAW from the popup menu.
Verified twice.
> Looks like there is some strange glitch creating new VM from template on oVirt 4.2.8.2.
>
> Basically I’m need to convert disk image from QCOW2 thin allocated to raw format in an existing VM.
> These are steps performed:
> 1) Create snapshot of an existing VM.
> 2) Make template from snapshot.
> 3) Create new VM using template, in Resource Allocation -> Storage Allocation -> Clone -> Format -> Raw.
>
> Unfortunately, newly created VM still have QCOW2 thin allocated disk.
> Did this operation twice to be sure, and both times things went wrong.
>
> What is done wrong, and how to solve this problem?
> Thanks in advance.
>
5 years, 7 months
Create New VM from Template Glitch
by Andrei Verovski
Hi !
Looks like there is some strange glitch creating new VM from template on oVirt 4.2.8.2.
Basically I’m need to convert disk image from QCOW2 thin allocated to raw format in an existing VM.
These are steps performed:
1) Create snapshot of an existing VM.
2) Make template from snapshot.
3) Create new VM using template, in Resource Allocation -> Storage Allocation -> Clone -> Format -> Raw.
Unfortunately, newly created VM still have QCOW2 thin allocated disk.
Did this operation twice to be sure, and both times things went wrong.
What is done wrong, and how to solve this problem?
Thanks in advance.
Andrei
5 years, 7 months
storware vProtect
by Giulio Casella
Hi everyone,
I'm evaluating some datacenter backup solution, and finally I'm going to
give a chance to storware vprotect for my ovirt DC (2 clusters, 3+32
hypervisors, about 100+ VMs).
How many of you are using it? Any suggestion? Any caveat not immediately
visible? Hints? Drawbacks?
For example I have some problem mounting backup of some VM (pfsense
firewalls), based on UFS filesystem. And I couldn't browse mounted
backup from WebUI (NFS problems between server and node).
I know these are problems I can easily get around, but I'm more
concerned about issue I can't see right now.
TIA,
gc
5 years, 7 months