Re: Weird Memory Leak Issue
by Robert O'Kane
Not for me. After restarting the engine,it doesn't matter if I restart the Hypervisors, I only get the "protocol=gluster" when I restart the VMs... migration is
not enough.
"virsh -r dumpxml <VM-ID>"
Cheers,
Robert O'Kane
On 08/29/2018 06:54 PM, Edward Clay wrote:
> Not sure how you're setup. I only had to migrate the VMs off to another hypervisor then put the hypervisor in maintenance which seems to unmount the one gluster volume. After upgrading, rebooting and activating the HV I could migrate VMs back to it. This seemed to work for me.
>
> ________________________________
> From: Robert O'Kane <okane(a)khm.de>
> Sent: Wednesday, August 29, 2018 10:42:56 AM
> To: users(a)ovirt.org
> Subject: [ovirt-users] Re: Weird Memory Leak Issue
>
> **Security Notice - This external email is NOT from The Hut Group**
>
> Ah, the FUSE mounts... I just saw last week that the upgrade to 4.2 removed the "LibgfApiSupported" flag by default.
>
> That is possibly why the leak simply appeared....
>
> OK, I was wondering where this came from. Tomorrow I will upgrade and test. I still will have to eventually reboot the VMs to get the gluster mounts again.... :-/
>
>
>
> On 08/29/2018 05:18 PM, Cole Johnson wrote:
>> Great! I'll look for the update.
>>
>> On Wed, Aug 29, 2018 at 7:50 AM Darrell Budic <budic(a)onholyground.com> wrote:
>>>
>>> There’s a memory leak in gluster 3.12.9 - 3.12.12 on fuse mounted volumes, sounds like what you’re seeing.
>>>
>>> The fix is in 3.12.13, which should be showing up today or tomorrow in the centos repos (currently available from the testing repo). I’ve been running it overnight on one host to test, looks like they got it.
>>>
>>> ________________________________
>>> From: Cole Johnson <sizzlinsaguaro(a)gmail.com>
>>> Subject: [ovirt-users] Weird Memory Leak Issue
>>> Date: August 29, 2018 at 9:35:39 AM CDT
>>> To: users(a)ovirt.org
>>>
>>> Hello,
>>> I have a hyperconverged, self hosted ovirt cluster with three hosts,
>>> running 4 VM's. The hosts are running the latest ovirt node. The
>>> VM's are Linux, Windows server 2016, and Windows Server 2008r2. The
>>> problem is with any host running the 2008r2 VM will run out of memory
>>> after 8-10 hours, causing any VM on the host to be paused, and making
>>> to host all but unresponsive. This problem seems to only exist with
>>> this specific VM. None of the other running VM's have this problem.
>>> I can resolve the problem by migrating the VM to a different host,
>>> then putting the host into maintenance mode, the activating it back.
>>> The leak appears to be in glusterfsd. Is there anything I can do to
>>> permanently fix this?
>
> ______________________________________________
>> Users mailing list -- users(a)ovirt.org
>> To unsubscribe send an email to users-leave(a)ovirt.org
>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/<https://www.ovirt.org/site/privacy-policy/>
>> oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/<https://www.ovirt.org/community/about/community-guidelines/>
>> List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/CLNILVLZ4D3...<https://lists.ovirt.org/archives/list/users@ovirt.org/message/CLNILVLZ4D3...>
>>
>
> --
> Robert O'Kane
> Systems Administrator
> Kunsthochschule für Medien Köln
> Peter-Welter-Platz 2
> 50676 Köln
>
> fon: +49(221)20189-223
> fax: +49(221)20189-49223
> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/<https://www.ovirt.org/site/privacy-policy/>
> oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/<https://www.ovirt.org/community/about/community-guidelines/>
> List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/F7EY4FX7UOB...<https://lists.ovirt.org/archives/list/users@ovirt.org/message/F7EY4FX7UOB...>
>
> Edward Clay
> Systems Administrator
> The Hut Group<http://www.thehutgroup.com/>
>
> Tel:
> Email: edward.clay(a)uk2group.com<mailto:edward.clay@uk2group.com>
>
> For the purposes of this email, the "company" means The Hut Group Limited, a company registered in England and Wales (company number 6539496) whose registered office is at Fifth Floor, Voyager House, Chicago Avenue, Manchester Airport, M90 3DQ and/or any of its respective subsidiaries.
>
> Confidentiality Notice
> This e-mail is confidential and intended for the use of the named recipient only. If you are not the intended recipient please notify us by telephone immediately on +44(0)1606 811888 or return it to us by e-mail. Please then delete it from your system and note that any use, dissemination, forwarding, printing or copying is strictly prohibited. Any views or opinions are solely those of the author and do not necessarily represent those of the company.
>
> Encryptions and Viruses
> Please note that this e-mail and any attachments have not been encrypted. They may therefore be liable to be compromised. Please also note that it is your responsibility to scan this e-mail and any attachments for viruses. We do not, to the extent permitted by law, accept any liability (whether in contract, negligence or otherwise) for any virus infection and/or external compromise of security and/or confidentiality in relation to transmissions sent by e-mail.
>
> Monitoring
> Activity and use of the company's systems is monitored to secure its effective use and operation and for other lawful business purposes. Communications using these systems will also be monitored and may be recorded to secure effective use and operation and for other lawful business purposes.
>
> hgvyjuv
>
--
Robert O'Kane
Systems Administrator
Kunsthochschule für Medien Köln
Peter-Welter-Platz 2
50676 Köln
fon: +49(221)20189-223
fax: +49(221)20189-49223
6 years, 2 months
Weird Memory Leak Issue
by Cole Johnson
Hello,
I have a hyperconverged, self hosted ovirt cluster with three hosts,
running 4 VM's. The hosts are running the latest ovirt node. The
VM's are Linux, Windows server 2016, and Windows Server 2008r2. The
problem is with any host running the 2008r2 VM will run out of memory
after 8-10 hours, causing any VM on the host to be paused, and making
to host all but unresponsive. This problem seems to only exist with
this specific VM. None of the other running VM's have this problem.
I can resolve the problem by migrating the VM to a different host,
then putting the host into maintenance mode, the activating it back.
The leak appears to be in glusterfsd. Is there anything I can do to
permanently fix this?
6 years, 2 months
Next Gluster Updates?
by Robert O'Kane
I had a bug request in Bugzilla for Gluster being killed due to a memory leak. The Gluster People say it is fixed in gluster-3.12.13
When will Ovirt have this update? I am getting tired of having to restart my hypervisors every week or so...
I currently have ovirt-release42-4.2.5.1-1.el7.noarch and yum check-updates shows me no new gluster versions.....(still 3.12.11)
Cheers,
Robert O'Kane
--
Robert O'Kane
Systems Administrator
Kunsthochschule für Medien Köln
Peter-Welter-Platz 2
50676 Köln
fon: +49(221)20189-223
fax: +49(221)20189-49223
6 years, 2 months
Windows 10 vs others windows
by carl langlois
Hi
Why when en try to do a Windows 10 machine i have a error on the cpu guest
os not supported but not windows 8 or 7?
Thanks,
6 years, 2 months
snapshots upload
by David David
hi all
ovirt engine 4.2.5.2-1.el7
ovirt node:
KVM Version: 2.9.0 - 16.el7_4.14.1
LIBVIRT Version: libvirt-3.2.0-14.el7_4.9
VDSM Version: vdsm-4.20.27.1-1.el7.centos
Can't restore vm by following this instruction
https://ovirt.org/develop/release-management/features/storage/backup-rest...
error message:
# python upload_disk_snapshots.py
Creating disk: 414d6613-5cfe-493c-ae6c-aa29caa32983
Traceback (most recent call last):
File "upload_disk_snapshots.py", line 305, in <module>
disk = create_disk(base_volume, disk_id, sd_name, disks_service)
File "upload_disk_snapshots.py", line 186, in create_disk
name=sd_name
File "/usr/lib64/python2.7/site-packages/ovirtsdk4/services.py", line
6715, in add
return self._internal_add(disk, headers, query, wait)
File "/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py", line 232,
in _internal_add
return future.wait() if wait else future
File "/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py", line 55,
in wait
return self._code(response)
File "/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py", line 229,
in callback
self._check_fault(response)
File "/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py", line 132,
in _check_fault
self._raise_error(response, body)
File "/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py", line 118,
in _raise_error
raise error
ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is
"[Cannot add Virtual Disk. Disk configuration (RAW Sparse) is incompatible
with the storage domain type.]". HTTP response code is 400.
# tree 414d6613-5cfe-493c-ae6c-aa29caa32983/
414d6613-5cfe-493c-ae6c-aa29caa32983/
├── 3610d5fd-6f55-46d9-a226-c06eee8e21e6
└── f77207b2-6e5b-4464-bd6f-5ae6d776435d
414d6613-5cfe-493c-ae6c-aa29caa32983 - disk id
3610d5fd-6f55-46d9-a226-c06eee8e21e6 - base image file
f77207b2-6e5b-4464-bd6f-5ae6d776435d - snapshot1 file
# qemu-img info
414d6613-5cfe-493c-ae6c-aa29caa32983/3610d5fd-6f55-46d9-a226-c06eee8e21e6
image:
414d6613-5cfe-493c-ae6c-aa29caa32983/3610d5fd-6f55-46d9-a226-c06eee8e21e6
file format: qcow2
virtual size: 20G (21474836480 bytes)
disk size: 22G
cluster_size: 65536
Format specific information:
compat: 1.1
lazy refcounts: false
refcount bits: 16
corrupt: false
# qemu-img info
414d6613-5cfe-493c-ae6c-aa29caa32983/f77207b2-6e5b-4464-bd6f-5ae6d776435d
image:
414d6613-5cfe-493c-ae6c-aa29caa32983/f77207b2-6e5b-4464-bd6f-5ae6d776435d
file format: qcow2
virtual size: 20G (21474836480 bytes)
disk size: 1.0G
cluster_size: 65536
backing file: 3610d5fd-6f55-46d9-a226-c06eee8e21e6 (actual path:
414d6613-5cfe-493c-ae6c-aa29caa32983/3610d5fd-6f55-46d9-a226-c06eee8e21e6)
backing file format: qcow2
Format specific information:
compat: 1.1
lazy refcounts: false
refcount bits: 16
corrupt: false
upload_disk_snapshots.py:
==========================
if __name__ == "__main__":
# Set storage domain name
sd_name = 'data_sas3'
# Set OVF file path
ovf_file_path = 'f4fdaf18-b944-4d22-879b-e235145a93f6.ovf'
# Disk to upload
disk_path = '414d6613-5cfe-493c-ae6c-aa29caa32983'
disk_id = os.path.basename(disk_path)
==========================
6 years, 2 months
Guide for setting up GlusterFS geo-replication?
by Jayme
Is there an updated guide for setting up GlusterFS geo-replication? What I
am interested in is having another oVirt setup on a separate server with
glusterFS volume replicated to it. If my primary cluster went down I would
be able to start important VMs on the secondary oVirt build until I'm able
to repair the main cluster.
In oVirt 4.2.5 clicking geo-replication > new in the volume GUI produces no
results. Nothing happens when I click it (using multiple browsers). And
under storage domains clicking remote data-sync -> setup produces a popup
but just has a loading icon in the middle and never completes loading. Why
are neither of these two features working for me?
6 years, 2 months
Turn Off Email Alerts
by Douglas Duckworth
Hi
How do I turn off hosted engine alerts? We are in a testing phase so these
are not needed. I have disabled postfix on all hosts as well as stopped
the ovirt notification daemon on the hosted engine. I kept it running
while putting /dev/null in
/usr/share/ovirt-engine/services/ovirt-engine-notifier/ovirt-engine-notifier.conf
for
mail server. Yet I still get alerts for every thing done such as putting
hosts in maintenance mode. Very confusing.
6 years, 2 months
Re: SSSD on Hosted Engine
by Douglas Duckworth
It says I am not authorized. Can you give me permission to comment? My
bugzilla account would be doug(a)med.cornell.edu.
Thanks,
Douglas Duckworth, MSc, LFCS
HPC System Administrator
Scientific Computing Unit
Weill Cornell Medicine
1300 York - LC-502
E: doug(a)med.cornell.edu
O: 212-746-6305
F: 212-746-8690
On Wed, Aug 22, 2018 at 10:44 AM, Sandro Bonazzola <sbonazzo(a)redhat.com>
wrote:
>
>
> 2018-08-22 14:59 GMT+02:00 Douglas Duckworth <dod2014(a)med.cornell.edu>:
>
>> Yay, I was able to restore nsswitch so things now work.
>>
>> Anyway that oVirt can use SSSD for web auth?
>>
>> Per https://ovirt.org/develop/release-management/features/infra/aaa_faq/
>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__ovirt.org_develop_re...>
>> there's a bug https://bugzilla.redhat.com/show_bug.cgi?id=829292
>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__bugzilla.redhat.com_...>
>> that prevents it from working?
>>
>
> I would suggest to comment on that bug with your use case, it will help
> understanding the needs.
>
>
>
>
>
>
>>
>>
>>
>> Thanks,
>>
>> Douglas Duckworth, MSc, LFCS
>> HPC System Administrator
>> Scientific Computing Unit
>> Weill Cornell Medicine
>> 1300 York - LC-502
>> E: doug(a)med.cornell.edu
>> O: 212-746-6305
>> F: 212-746-8690
>>
>>
>> On Wed, Aug 22, 2018 at 8:51 AM, Douglas Duckworth <
>> dod2014(a)med.cornell.edu> wrote:
>>
>>> Hi
>>>
>>> I am trying to configure sssd on my hosted engine. Essentially we
>>> control host access in LDAP so I want sssd to read that thus allow my
>>> coworkers to login to hosted engine vm.
>>>
>>> For some reason sssd reports backend offline even though it's
>>> resolvable, pingable, with ports open. I see that it's a SELinux issue
>>> which I can resolve. After changing to permissive SSSD works.
>>>
>>> To have system read sssd database I set hosts line in /etc/nsswitch.conf
>>> to:
>>>
>>> hosts files sss
>>>
>>> Though it seems that I did something bad to /etc/nsswitch.conf as now
>>> yum, ping, etc does not work.
>>>
>>> Could someone suggest how to restore this file or could anyone share
>>> theirs?
>>>
>>> Thanks,
>>>
>>> Douglas Duckworth, MSc, LFCS
>>> HPC System Administrator
>>> Scientific Computing Unit
>>> Weill Cornell Medicine
>>> 1300 York - LC-502
>>> E: doug(a)med.cornell.edu
>>> O: 212-746-6305
>>> F: 212-746-8690
>>>
>>>
>>
>> _______________________________________________
>> Users mailing list -- users(a)ovirt.org
>> To unsubscribe send an email to users-leave(a)ovirt.org
>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__www.ovirt.org_site_p...>
>> oVirt Code of Conduct: https://www.ovirt.org/communit
>> y/about/community-guidelines/
>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__www.ovirt.org_commun...>
>> List Archives: https://lists.ovirt.org/archiv
>> es/list/users(a)ovirt.org/message/DSFBEFKBNAWZMLMASRJ7YKZO3PZWKJLV/
>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__lists.ovirt.org_arch...>
>>
>>
>
>
> --
>
> SANDRO BONAZZOLA
>
> MANAGER, SOFTWARE ENGINEERING, EMEA R&D RHV
>
> Red Hat EMEA
> <https://urldefense.proofpoint.com/v2/url?u=https-3A__www.redhat.com_&d=Dw...>
>
> sbonazzo(a)redhat.com
>
> <https://urldefense.proofpoint.com/v2/url?u=https-3A__red.ht_sig&d=DwMFaQ&...>
> <https://urldefense.proofpoint.com/v2/url?u=https-3A__www.redhat.com_en_ev...>
>
6 years, 2 months
mask or disabled lldpad service
by Klaas Demter
Hi,
I have a QLogic Corp. QLogic 2x1GE+2x10GE QL41162HMRJ CNA network
card. This network card comes with it's own lldp implementation inside
their management firmware. Running two lldp agents seems to create
several issues (http://lists.us.dell.com/pipermail/linux-poweredge/2018-July/051860.html).
The QLogic engineers said I need to disable lldpad in the operating
system. This leads me here, what is lldp actually used for within
ovirt? If it is actually needed how could I deal with a card that does
not support lldpad but rather has it's own implementation?
Greetings
Klaas
6 years, 2 months