I had this issue this week as well.
When asked about the glusterfs that you self provisioned you stated "ovirt1.localdomain:/gluster_bricks/engine”
So I am new to gluster but as a client of gluster you can only refer to it via volume name.
Hence maybe try ovirt1.localdomain:/engine.
That did the trick for me. Hope this helps. Cost me a week.
I've using ovirt from 3.4 and we have a strange, rarely problem with live migrations.
- In most of case impossible to do a live migrate of VMs that have a lot of memory (16 gigs and more up to 40 GB) .
- Those are need to try and try until the migration will be successful
- impossible to understand why does working in a case and doesn't in another
- it is happening always if there bigger load on vms, but only on big memory VM-s
- The small VMs (2-4-8 gigs) can migrate without problems
- In very rarely cases possible to migrate on first try.
- We using the memory ballooning function on all of VMs.
We using dedicated migration network with separated vlans, but on same physical lan (1gbe) with ovirtmgmt.
At this moment, we don't use QoS for dedicate bandwidth.
What can I do ?
Thanks in advance,
Hello, after migrating our Fiber Channel to a new Ovirt 4.2
environment, we over the years and through experimenting and mistakes
have ended up with some stale LVM's/disks on our fibre channel storage
that Ovirt no longer is able to manage correctly or is unaware of. I am
looking for a reliable to way to do a few different things.
The first is figuring out precisely what LVM ids belong to a VM and is
being used by that VM.
The second is figuring out if a LVM I have found on the Storage domain
is being used at all by any VM or if ovirt is even aware of it.
I have fumbled around a bit. And using a combination of the following I
have been able to figure out some of them. But now I am finding
information that does not match or may not be correct or I am
interpreting the data wrong. Anyway this is a big deal, because we want
to remove the stale unused LVMs and it would obviously be disastrous if
I deleted the wrong LVM from the FC.
So I know its not recommend but since I am not actually telling
vdsm-client to do anything other than get information I figure its
harmless. So here is what I have found so far.
*vdsm-client Host getVMFullList vmname=<VM_NAME_HERE> | grep volumeID*
*lvs | grep <noted volumeID here>*
With the above I have had some success verifying what LVMs a device is
using. However now I am having trouble figuring out a particular windows
server VMs LVM id's.
what I want to know is there a better way?
Also two more things.
You could call these feature requests, however it would be nice if there
was a way to see all the unused LVMs on a storage domain that are not
Tied to a VM. And it would also be nice to be able to remove un-imported
VMs that reside on a storage domain without importing them.
Anyway trying to get rid of un-imported vms and getting rid of unused
LVMs has been a chore. I wish there was an easier way.
American Alloy Steel
I continue to try to understand my problem between (I suppose) oVirt anf Gluster.
After my recents posts titled 'VMs unexpectidly restarted' that did not provide solution nor search idea, I submit to you another (related ?) problem.
Parallely with the problem of VMs down (that did not reproduce since Oct 16), I have ramdomly some events in the GUI saying "VM xxxxx is not responding." For example, VM "patjoub1" on 2018-11-11 14:34. Never the same hour, not all the days, often this VM patjoub1 but not always : I had it on two others. All VMs disks are on a volume DATA02 (with leases on the same volume).
Searching in engine.log, I found :
2018-11-11 14:34:32,953+01 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer] (EE-ManagedThreadFactory-engineScheduled-Thread-28)  VM '6116fb07-096b-4c7e-97fe-01ecc9a6bd9b'(patjoub1) moved from 'Up' --> 'NotResponding'
2018-11-11 14:34:33,116+01 WARN [org.ovirt.engine.core.vdsbroker.vdsbroker.VdsBrokerObjectsBuilder] (EE-ManagedThreadFactory-engineScheduled-Thread-1)  Invalid or unknown guest architecture type '' received from guest agent
2018-11-11 14:34:33,176+01 WARN [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (EE-ManagedThreadFactory-engineScheduled-Thread-28)  EVENT_ID: VM_NOT_RESPONDING(126), VM patjoub1 is not responding.
2018-11-11 14:34:48,278+01 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer] (EE-ManagedThreadFactory-engineScheduled-Thread-48)  VM '6116fb07-096b-4c7e-97fe-01ecc9a6bd9b'(patjoub1) moved from 'NotResponding' --> 'Up'So it becomes up 15s after, and the VM (and the monitoring) see no downtime.
At this time, I see in vdsm.log of the nodes :
2018-11-11 14:33:49,450+0100 ERROR (check/loop) [storage.Monitor] Error checking path /rhev/data-center/mnt/glusterSD/victor.local.systea.fr:_DATA02/ffc53fd8-c5d1-4070-ae51-2e91835cd937/dom_md/metadata (monitor:498)
Traceback (most recent call last):
File "/usr/lib/python2.7/site-packages/vdsm/storage/monitor.py", line 496, in _pathChecked
delay = result.delay()
File "/usr/lib/python2.7/site-packages/vdsm/storage/check.py", line 391, in delay
raise exception.MiscFileReadException(self.path, self.rc, self.err)
MiscFileReadException: Internal file read failure: (u'/rhev/data-center/mnt/glusterSD/victor.local.systea.fr:_DATA02/ffc53fd8-c5d1-4070-ae51-2e91835cd937/dom_md/metadata', 1, 'Read timeout')
2018-11-11 14:33:49,450+0100 INFO (check/loop) [storage.Monitor] Domain ffc53fd8-c5d1-4070-ae51-2e91835cd937 became INVALID (monitor:469)
2018-11-11 14:33:59,451+0100 WARN (check/loop) [storage.check] Checker u'/rhev/data-center/mnt/glusterSD/victor.local.systea.fr:_DATA02/ffc53fd8-c5d1-4070-ae51-2e91835cd937/dom_md/metadata' is blocked for 20.00 seconds (check:282)
2018-11-11 14:34:09,480+0100 INFO (event/37) [storage.StoragePool] Linking /rhev/data-center/mnt/glusterSD/victor.local.systea.fr:_DATA02/ffc53fd8-c5d1-4070-ae51-2e91835cd937 to /rhev/data-center/6efda7f8-b62f-11e8-9d16-00163e263d21/ffc53fd8-c5d1-4070-ae51-2e91835cd937 (sp:1230)OK : so, DATA02 marked as blocked for 20s ? I definitly have a problem with gluster ? I'll inevitably find the reason in the gluster logs ? Uh : not at all.
Please see gluster logs here :
Unfortunatly I discovered this morning that I have not the sanlock.log for this date. I don't understand why, the log rotate seems OK with "rotate 3", but I have no backups files :(.
But, luck in bad luck, the same event occurs this morning ! Same VM patjoub1, 2018-11-13 08:01:37. So I have added the sanlock.log for today, maybe it can help.
IMPORTANT NOTE : don't forget that Gluster log with on hour shift. For this event at 14:34, search at 13h34 in gluster logs.
I recall my configuration :
3 nodes where the third is arbiter (volumes in replica 2)
The nodes are never overloaded (CPU average 5%, no peak detected at the time of the event, mem 128G used at 15% (only 10 VMs on this cluster)). Network underused, gluster is on a separate network on a bond (2 NICs) 1+1Gb mode 4 = 2Gb, used in peak at 10%.
Here is the configuration for the given volume :
# gluster volume status DATA02
Status of volume: DATA02
Gluster process TCP Port RDMA Port Online Pid
ata02/data02/brick 49158 0 Y 4990
ata02/data02/brick 49153 0 Y 8460
/data01/data02/brick 49158 0 Y 2470
Self-heal Daemon on localhost N/A N/A Y 8771
Self-heal Daemon on eskarinastorage.local.s
ystea.fr N/A N/A Y 11745
Self-heal Daemon on victorstorage.local.sys
tea.fr N/A N/A Y 17055
Task Status of Volume DATA02
There are no active volume tasks
# gluster volume info DATA02
Volume Name: DATA02
Volume ID: 48bf5871-339b-4f39-bea5-9b5848809c83
Snapshot Count: 0
Number of Bricks: 1 x (2 + 1) = 3
Brick3: eskarinastorage.local.systea.fr:/home/data01/data02/brick (arbiter)
So : is there someone around trying to make me understand what append ? Pleeease :/
On Fri, 7 Dec 2018 at 5:35 PM, Matthias Barmeier <
> I don't think so because the known_hosts contain all host keys of all
> gluster nodes before start of setup.
Also, the FQDN that’s used to add the hosts to engine? Is that also in
> Or do I miss something ?
> Am Freitag, den 07.12.2018, 12:17 +0530 schrieb Sahina Bose:
> > I think you may be running into
> > https://bugzilla.redhat.com/show_bug.cgi?id=1651516
> > On Thu, Dec 6, 2018 at 7:30 PM <matthias.barmeier(a)sourcepark.de>
> > wrote:
> > >
> > > Hi,
> > >
> > > tried to setup hyperconverged with glusterfs. I used three i7 with
> > > two 1TB Disks and two NICs. Everythin worked fine till:
> > >
> > > [ INFO ] TASK [Set Engine public key as authorized key without
> > > validating the TLS/SSL certificates]
> > >
> > > appears in the wizards window. Does anyone has a hint on what to
> > > do?
> > > _______________________________________________
> > > Users mailing list -- users(a)ovirt.org
> > > To unsubscribe send an email to users-leave(a)ovirt.org
> > > Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> > > oVirt Code of Conduct: https://www.ovirt.org/community/about/commun
> > > ity-guidelines/
> > > List Archives: https://firstname.lastname@example.org
> > > g/message/BI6Z576SKSTNQZRBCWLYMVFYTRZVZOCE/
tried to setup hyperconverged with glusterfs. I used three i7 with two 1TB Disks and two NICs. Everythin worked fine till:
[ INFO ] TASK [Set Engine public key as authorized key without validating the TLS/SSL certificates]
appears in the wizards window. Does anyone has a hint on what to do?
I am testing vGPU in ovirt4.2. I have a nvidia p100 GPU installed in the host and I have installed the correct dirver. I followed this page(https://www.ovirt.org/documentation/install-guide/appe-Preparing_a_H... but I failed. In the "Host Devices tab" I got an device called "GP100GL [Tesla P100 PCIe 16GB]". I could not find any device of "mdev type". Anyone can help me? Thanks a lot!
I just wondering I would like to make an upgrade from 4.2.5 to latest stable release of ovirt.
Meanwhile centos7.6 was released.
Will that compatible with the current release of ovirt?
How is safe an upgrade now?
Thanks in advance,
The oVirt Project is pleased to announce the availability of the Second
Alpha Release of oVirt 4.3.0, as of December 6th, 2018
This is pre-release software. This pre-release should not to be used in
Please take a look at our community page to learn how to ask questions
and interact with developers and users.
All issues or bugs should be reported via oVirt Bugzilla.
This update is the second alpha release of the 4.3.0 version.
This release brings more than 80 enhancements and more than 330 bug fixes
on top of oVirt 4.2 series.
What's new in oVirt 4.3.0?
* Q35 chipset, support booting using UEFI and Secure Boot
* Skylake-server and AMD EPYC support
* New smbus driver in windows guest tools
* Improved support for v2v
* Tech preview for oVirt on Fedora 28
* Hundreds of bug fixes on top of oVirt 4.2 series
* New VM portal details page (see a preview here:
* New Cluster upgrade UI
* OVN security groups
* IPv6 (static host addresses)
* LLDP Labeler
* Support for 3.6 and 4.0 data centers, clusters and hosts were removed
* Now using PostgreSQL 10
This release is available now on x86_64 architecture for:
* Red Hat Enterprise Linux 7.6 or later
* CentOS Linux (or similar) 7.6 or later
This release supports Hypervisor Hosts on x86_64 and ppc64le architectures
* Red Hat Enterprise Linux 7.6 or later
* CentOS Linux (or similar) 7.6 or later
* oVirt Node 4.3 (available for x86_64 only)
Experimental tech preview for x86_64 and s390x architectures for Fedora 28
is also included.
See the release notes draft  for installation / upgrade instructions and
a list of new features and bugs fixed.
- oVirt Appliance is already available for both CentOS 7 and Fedora 28
- oVirt Node NG is already available for both CentOS 7
- oVirt Node NG for Fedora 28 (tech preview) is facing build issues on
Jenkins and will be released as soon as possible
* Read more about the oVirt 4.3.0 release highlights:
* Get more oVirt project updates on Twitter: https://twitter.com/ovirt
* Check out the latest project news on the oVirt blog:
MANAGER, SOFTWARE ENGINEERING, EMEA R&D RHV
Red Hat EMEA <https://www.redhat.com/>