Just to add my .02 here. I've opened a bug on this issue where HV/host connected to
clusterfs volumes are running out of ram. This seemed to be a bug fixed in gluster 3.13
but that patch doesn't seem to be avaiable any longer and 3.12 is what ovirt is using.
For example I have a host that was showing 72% of memory consumption with 3 VMs running
on it. If I migrate those VMs to another Host memory consumption drops to 52%. If i put
this host into maintenance and then activate it it drops down to 2% or so. Since I ran
into this issue I've been manually watching memory consumption on each host and
migrating VMs from it to others to keep things from dying. I'm hoping with the
announcement of gluster 3.12 end of life and the move to gluster 4.1 that this will get
fixed or that the patch from 3.13 can get backported so this problem will go away.
https://bugzilla.redhat.com/show_bug.cgi?id=1593826
On 07/07/2018 11:49 AM, Jim Kusznir wrote:
**Security Notice - This external email is NOT from The Hut Group**
This host has NO VMs running on it, only 3 running cluster-wide (including the engine,
which is on its own storage):
top - 10:44:41 up 1 day, 17:10, 1 user, load average: 15.86, 14.33, 13.39
Tasks: 381 total, 1 running, 379 sleeping, 1 stopped, 0 zombie
%Cpu(s): 2.7 us, 2.1 sy, 0.0 ni, 89.0 id, 6.1 wa, 0.0 hi, 0.2 si, 0.0 st
KiB Mem : 32764284 total, 338232 free, 842324 used, 31583728 buff/cache
KiB Swap: 12582908 total, 12258660 free, 324248 used. 31076748 avail Mem
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
13279 root 20 0 2380708 37628 4396 S 51.7 0.1 3768:03 glusterfsd
13273 root 20 0 2233212 20460 4380 S 17.2 0.1 105:50.44 glusterfsd
13287 root 20 0 2233212 20608 4340 S 4.3 0.1 34:27.20 glusterfsd
16205 vdsm 0 -20 5048672 88940 13364 S 1.3 0.3 0:32.69 vdsmd
16300 vdsm 20 0 608488 25096 5404 S 1.3 0.1 0:05.78 python
1109 vdsm 20 0 3127696 44228 8552 S 0.7 0.1 18:49.76 ovirt-ha-broker
25555 root 20 0 0 0 0 S 0.7 0.0 0:00.13 kworker/u64:3
10 root 20 0 0 0 0 S 0.3 0.0 4:22.36 rcu_sched
572 root 0 -20 0 0 0 S 0.3 0.0 0:12.02 kworker/1:1H
797 root 20 0 0 0 0 S 0.3 0.0 1:59.59 kdmwork-253:2
877 root 0 -20 0 0 0 S 0.3 0.0 0:11.34 kworker/3:1H
1028 root 20 0 0 0 0 S 0.3 0.0 0:35.35 xfsaild/dm-10
1869 root 20 0 1496472 10540 6564 S 0.3 0.0 2:15.46 python
3747 root 20 0 0 0 0 D 0.3 0.0 0:01.21 kworker/u64:1
10979 root 15 -5 723504 15644 3920 S 0.3 0.0 22:46.27 glusterfs
15085 root 20 0 680884 10792 4328 S 0.3 0.0 0:01.13 glusterd
16102 root 15 -5 1204216 44948 11160 S 0.3 0.1 0:18.61 supervdsmd
At the moment, the engine is barely usable, my other VMs appear to be unresponsive. Two
on one host, one on another, and none on the third.
On Sat, Jul 7, 2018 at 10:38 AM, Jim Kusznir
<jim@palousetech.com<mailto:jim@palousetech.com>> wrote:
I run 4-7 VMs, and most of them are 2GB ram. I have 2 VMs with 4GB.
Ram hasn't been an issue until recent ovirt/gluster upgrades. Storage has always been
slow, especially with these drives. However, even watching network utilization on my
switch, the gig-e links never max out.
The loadavg issues and unresponsive behavior started with yesterday's ovirt updates.
I now have one VM with low I/O that lives on a separate storage volume (data, fully SSD
backed instead of data-hdd, which was having the issues). I moved it to a ovirt host with
no other VMs on it, and that had reshly been rebooted. Before it had this one VM on it,
loadavg was >0.5. Now its up in the 20's, with only one low Disk I/O, 4GB ram VM
on the host.
This to me says there's now a new problem separate from Gluster. I don't have any
non-gluster storage available to test with. I did notice that the last update included a
new kernel, and it appears its the qemu-kvm processes that are consuming way more CPU than
they used to now.
Are there any known issues? I'm going to reboot into my previous kernel to see if its
kernel-caused.
--Jim
On Fri, Jul 6, 2018 at 11:07 PM, Johan Bernhardsson
<johan@kafit.se<mailto:johan@kafit.se>> wrote:
That is a single sata drive that is slow on random I/O and that has to be synced with 2
other servers. Gluster works syncronous so one write has to be written and acknowledged on
all the three nodes.
So you have a bottle neck in io on drives and one on network and depending on how many
virtual servers you have and how much ram they take you might have memory.
Load spikes when you have a wait somewhere and are overusing capacity. But it's now
only CPU that load is counted on. It is waiting for resources so it can be memory or
Network or drives.
How many virtual server do you run and how much ram do they consume?
On July 7, 2018 09:51:42 Jim Kusznir
<jim@palousetech.com<mailto:jim@palousetech.com>> wrote:
In case it matters, the data-hdd gluster volume uses these hard drives:
https://www.amazon.com/gp/product/B01M1NHCZT/ref=oh_aui_detailpage_o05_s0...
This is in a Dell R610 with PERC6/i (one drive per server, configured as a single drive
volume to pass it through as its own /dev/sd* device). Inside the OS, its partitioned
with lvm_thin, then an lvm volume formatted with XFS and mounted as /gluster/brick3, with
the data-hdd volume created inside that.
--Jim
On Fri, Jul 6, 2018 at 10:45 PM, Jim Kusznir
<jim@palousetech.com<mailto:jim@palousetech.com>> wrote:
So, I'm still at a loss...It sounds like its either insufficient ram/swap, or
insufficient network. It seems to be neither now. At this point, it appears that gluster
is just "broke" and killing my systems for no descernable reason. Here's
detals, all from the same system (currently running 3 VMs):
[root@ovirt3 ~]# w
22:26:53 up 36 days, 4:34, 1 user, load average: 42.78, 55.98, 53.31
USER TTY FROM LOGIN@ IDLE JCPU PCPU WHAT
root pts/0 192.168.8.90 22:26 2.00s 0.12s 0.11s w
bwm-ng reports the highest data usage was about 6MB/s during this test (and that was
combined; I have two different gig networks. One gluster network (primary VM storage)
runs on one, the other network handles everything else).
[root@ovirt3 ~]# free -m
total used free shared buff/cache available
Mem: 31996 13236 232 18 18526 18195
Swap: 16383 1475 14908
top - 22:32:56 up 36 days, 4:41, 1 user, load average: 17.99, 39.69, 47.66
Tasks: 407 total, 1 running, 405 sleeping, 1 stopped, 0 zombie
%Cpu(s): 8.6 us, 2.1 sy, 0.0 ni, 87.6 id, 1.6 wa, 0.0 hi, 0.1 si, 0.0 st
KiB Mem : 32764284 total, 228296 free, 13541952 used, 18994036 buff/cache
KiB Swap: 16777212 total, 15246200 free, 1531012 used. 18643960 avail Mem
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
30036 qemu 20 0 6872324 5.2g 13532 S 144.6 16.5 216:14.55 /usr/libexec/qemu-kvm
-name guest=BillingWin,debug-threads=on -S -object
secret,id=masterKey0,format=raw,file=/v+
28501 qemu 20 0 5034968 3.6g 12880 S 16.2 11.7 73:44.99 /usr/libexec/qemu-kvm
-name guest=FusionPBX,debug-threads=on -S -object
secret,id=masterKey0,format=raw,file=/va+
2694 root 20 0 2169224 12164 3108 S 5.0 0.0 3290:42 /usr/sbin/glusterfsd
-s ovirt3.nwfiber.com<http://ovirt3.nwfiber.com> --volfile-id
data.ovirt3.nwfiber.com<http://data.ovirt3.nwfiber.com>.gluster-brick2-data -p
/var/run/+
14293 root 15 -5 944700 13356 4436 S 4.0 0.0 16:32.15 /usr/sbin/glusterfs
--volfile-server=192.168.8.11 --volfile-server=192.168.8.12 --volfile-server=192.168.8.13
--+
25100 vdsm 0 -20 6747440 107868 12836 S 2.3 0.3 21:35.20 /usr/bin/python2
/usr/share/vdsm/vdsmd
28971 qemu 20 0 2842592 1.5g 13548 S 1.7 4.7 241:46.49 /usr/libexec/qemu-kvm
-name guest=unifi.palousetech.com<http://unifi.palousetech.com>,debug-threads=on -S
-object secret,id=masterKey0,format=+
12095 root 20 0 162276 2836 1868 R 1.3 0.0 0:00.25 top
2708 root 20 0 1906040 12404 3080 S 1.0 0.0 1083:33 /usr/sbin/glusterfsd
-s ovirt3.nwfiber.com<http://ovirt3.nwfiber.com> --volfile-id
engine.ovirt3.nwfiber.com<http://engine.ovirt3.nwfiber.com>.gluster-brick1-engine -p
/var/+
28623 qemu 20 0 4749536 1.7g 12896 S 0.7 5.5 4:30.64 /usr/libexec/qemu-kvm
-name guest=billing.nwfiber.com<http://billing.nwfiber.com>,debug-threads=on -S
-object secret,id=masterKey0,format=ra+
10 root 20 0 0 0 0 S 0.3 0.0 215:54.72 [rcu_sched]
1030 sanlock rt 0 773804 27908 2744 S 0.3 0.1 35:55.61 /usr/sbin/sanlock
daemon
1890 zabbix 20 0 83904 1696 1612 S 0.3 0.0 24:30.63
/usr/sbin/zabbix_agentd: collector [idle 1 sec]
2722 root 20 0 1298004 6148 2580 S 0.3 0.0 38:10.82 /usr/sbin/glusterfsd
-s ovirt3.nwfiber.com<http://ovirt3.nwfiber.com> --volfile-id
iso.ovirt3.nwfiber.com<http://iso.ovirt3.nwfiber.com>.gluster-brick4-iso -p
/var/run/gl+
6340 root 20 0 0 0 0 S 0.3 0.0 0:04.30 [kworker/7:0]
10652 root 20 0 0 0 0 S 0.3 0.0 0:00.23 [kworker/u64:2]
14724 root 20 0 1076344 17400 3200 S 0.3 0.1 10:04.13 /usr/sbin/glusterfs -s
localhost --volfile-id gluster/glustershd -p
/var/run/gluster/glustershd/glustershd.pid<http://ustershd.pid> -+
22011 root 20 0 0 0 0 S 0.3 0.0 0:05.04 [kworker/10:1]
Not sure why the system load dropped other than I was trying to take a picture of it :)
In any case, it appears that at this time, I have plenty of swap, ram, and network
capacity, and yet things are still running very sluggish; I'm still getting e-mails
from servers complaining about loss of communication with something or another; I still
get e-mails from the engine about bad engine status, then recovery, etc.
I've shut down 2/3 of my VMs, too....just trying to keep the critical ones operating.
At this point, I don't believe the problem is the memory leak, but it seems to be
triggered by the memory leak, as in all my problems started when I got low ram warnings
from one of my 3 nodes and began recovery efforts from that.
I do really like the idea / concept behind glusterfs, but I really have to figure out why
its been so poor performing from day one, and its caused 95% of my outages (including
several large ones lately). If I can get it stable, reliable, and well performing, then
I'd love to keep it. If I can't, then perhaps NFS is the way to go? I don't
like the single point of failure aspect of it, but my other NAS boxes I run for clients
(central storage for windows boxes) have been very solid; If I could get that kind of
reliability for my ovirt stack, it would be a substantial improvement. Currently, it
seems about every other month I have a gluster-induced outage.
Sometimes I wonder if its just hyperconverged is the issue, but my infrastructure
doesn't justify three servers at the same location...I might be able to do two, but
even that seems like its pushing it.
Looks like I can upgrade to 10G for about $900. I can order a dual-Xeon supermicro
12-disk server, loaded with 2TB WD Enterprise disks and a pair of SSDs for the os, 32GB
ram, 2.67Ghz CPUs for about $720 delivered. I've got to do something to improve my
reliability; I can't keep going the way I have been....
--Jim
On Fri, Jul 6, 2018 at 9:13 PM, Johan Bernhardsson
<johan@kafit.se<mailto:johan@kafit.se>> wrote:
Load like that is mostly io based either the machine is swapping or network is to slow.
Check I/o wait in top.
And the problem where you get oom killer to kill off gluster. That means that you
don't monitor ram usage on the servers? Either it's eating all your ram and swap
gets really io intensive and then is killed off. Or you have the wrong swap settings in
sysctl.conf (there are tons of broken guides that recommends swappines to 0 but that
disables swap on newer kernels. The proper swappines for only swapping when nesseary is 1
or a sufficiently low number like 10 default is 60)
Moving to nfs will not improve things. You will get more memory since gluster isn't
running and that is good. But you will have a single node that can fail with all your
storage and it would still be on 1 gigabit only and your three node cluster would easily
saturate that link.
On July 7, 2018 04:13:13 Jim Kusznir
<jim@palousetech.com<mailto:jim@palousetech.com>> wrote:
So far it does not appear to be helping much. I'm still getting VM's locking up
and all kinds of notices from overt engine about non-responsive hosts. I'm still
seeing load averages in the 20-30 range.
Jim
On Fri, Jul 6, 2018, 3:13 PM Jim Kusznir
<jim@palousetech.com<mailto:jim@palousetech.com>> wrote:
Thank you for the advice and help
I do plan on going 10Gbps networking; haven't quite jumped off that cliff yet,
though.
I did put my data-hdd (main VM storage volume) onto a dedicated 1Gbps network, and
I've watched throughput on that and never seen more than 60GB/s achieved (as reported
by bwm-ng). I have a separate 1Gbps network for communication and ovirt migration, but I
wanted to break that up further (separate out VM traffice from migration/mgmt traffic).
My three SSD-backed gluster volumes run the main network too, as I haven't been able
to get them to move to the new network (which I was trying to use as all gluster). I
tried bonding, but that seamed to reduce performance rather than improve it.
--Jim
On Fri, Jul 6, 2018 at 2:52 PM, Jamie Lawrence
<jlawrence@squaretrade.com<mailto:jlawrence@squaretrade.com>> wrote:
Hi Jim,
I don't have any targeted suggestions, because there isn't much to latch on to. I
can say Gluster replica three (no arbiters) on dedicated servers serving a couple Ovirt
VM clusters here have not had these sorts of issues.
I suspect your long heal times (and the resultant long periods of high load) are at least
partly related to 1G networking. That is just a matter of IO - heals of VMs involve moving
a lot of bits. My cluster uses 10G bonded NICs on the gluster and ovirt boxes for storage
traffic and separate bonded 1G for ovirtmgmt and communication with other machines/people,
and we're occasionally hitting the bandwidth ceiling on the storage network. I'm
starting to think about 40/100G, different ways of splitting up intensive systems, and
considering iSCSI for specific volumes, although I really don't want to go there.
I don't run FreeNAS[1], but I do run FreeBSD as storage servers for their excellent
ZFS implementation, mostly for backups. ZFS will make your `heal` problem go away, but not
your bandwidth problems, which become worse (because of fewer NICS pushing traffic). 10G
hardware is not exactly in the impulse-buy territory, but if you can, I'd recommend
doing some testing using it. I think at least some of your problems are related.
If that's not possible, my next stops would be optimizing everything I could about
sharding, healing and optimizing for serving the shard size to squeeze as much performance
out of 1G as I could, but that will only go so far.
-j
[1] FreeNAS is just a storage-tuned FreeBSD with a GUI.
On Jul 6, 2018, at 1:19 PM, Jim Kusznir
<jim@palousetech.com<mailto:jim@palousetech.com>> wrote:
hi all:
Once again my production ovirt cluster is collapsing in on itself. My servers are
intermittently unavailable or degrading, customers are noticing and calling in. This
seems to be yet another gluster failure that I haven't been able to pin down.
I posted about this a while ago, but didn't get anywhere (no replies that I found).
The problem started out as a glusterfsd process consuming large amounts of ram (up to the
point where ram and swap were exhausted and the kernel OOM killer killed off the
glusterfsd process). For reasons not clear to me at this time, that resulted in any VMs
running on that host and that gluster volume to be paused with I/O error (the glusterfs
process is usually unharmed; why it didn't continue I/O with other servers is
confusing to me).
I have 3 servers and a total of 4 gluster volumes (engine, iso, data, and data-hdd). The
first 3 are replica 2+arb; the 4th (data-hdd) is replica 3. The first 3 are backed by an
LVM partition (some thin provisioned) on an SSD; the 4th is on a seagate hybrid disk (hdd
+ some internal flash for acceleration). data-hdd is the only thing on the disk. Servers
are Dell R610 with the PERC/6i raid card, with the disks individually passed through to
the OS (no raid enabled).
The above RAM usage issue came from the data-hdd volume. Yesterday, I cought one of the
glusterfsd high ram usage before the OOM-Killer had to run. I was able to migrate the VMs
off the machine and for good measure, reboot the entire machine (after taking this
opportunity to run the software updates that ovirt said were pending). Upon booting back
up, the necessary volume healing began. However, this time, the healing caused all three
servers to go to very, very high load averages (I saw just under 200 on one server;
typically they've been 40-70) with top reporting IO Wait at 7-20%. Network for this
volume is a dedicated gig network. According to bwm-ng, initially the network bandwidth
would hit 50MB/s (yes, bytes), but tailed off to mostly in the kB/s for a while. All
machines' load averages were still 40+ and gluster volume heal data-hdd info reported
5 items needing healing. Server's were intermittently experiencing IO issues, even on
the 3 gluster volumes that appeared largely unaffected. Even the OS activities on the
hosts itself (logging in, running commands) would often be very delayed. The ovirt engine
was seemingly randomly throwing engine down / engine up / engine failed notifications.
Responsiveness on ANY VM was horrific most of the time, with random VMs being
inaccessible.
I let the gluster heal run overnight. By morning, there were still 5 items needing
healing, all three servers were still experiencing high load, and servers were still
largely unstable.
I've noticed that all of my ovirt outages (and I've had a lot, way more than is
acceptable for a production cluster) have come from gluster. I still have 3 VMs who's
hard disk images have become corrupted by my last gluster crash that I haven't had
time to repair / rebuild yet (I believe this crash was caused by the OOM issue previously
mentioned, but I didn't know it at the time).
Is gluster really ready for production yet? It seems so unstable to me.... I'm
looking at replacing gluster with a dedicated NFS server likely FreeNAS. Any suggestions?
What is the "right" way to do production storage on this (3 node cluster)? Can
I get this gluster volume stable enough to get my VMs to run reliably again until I can
deploy another storage solution?
--Jim
_______________________________________________
Users mailing list -- users@ovirt.org<mailto:users@ovirt.org>
To unsubscribe send an email to
users-leave@ovirt.org<mailto:users-leave@ovirt.org>
Privacy Statement:
https://www.ovirt.org/site/privacy-policy/<https://www.ovirt.org/site/...
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/<https://ww...
List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/YQX3LQFQQPW...
_______________________________________________
Users mailing list -- users@ovirt.org<mailto:users%40ovirt.org>
To unsubscribe send an email to
users-leave@ovirt.org<mailto:users-leave%40ovirt.org>
Privacy Statement:
https://www.ovirt.org/site/privacy-policy/<https://www.ovirt.org/site/...
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/<https://ww...
List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/O2HIECLFMYG...
_______________________________________________
Users mailing list -- users@ovirt.org<mailto:users@ovirt.org>
To unsubscribe send an email to users-leave@ovirt.org<mailto:users-leave@ovirt.org>
Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/T2M4J3Z7RPS...
Edward Clay
Systems Administrator
The Hut
Group<http://www.thehutgroup.com/>
Tel:
Email: edward.clay@uk2group.com<mailto:edward.clay@uk2group.com>
For the purposes of this email, the "company" means The Hut Group Limited, a
company registered in England and Wales (company number 6539496) whose registered office
is at Fifth Floor, Voyager House, Chicago Avenue, Manchester Airport, M90 3DQ and/or any
of its respective subsidiaries.
Confidentiality Notice
This e-mail is confidential and intended for the use of the named recipient only. If you
are not the intended recipient please notify us by telephone immediately on +44(0)1606
811888 or return it to us by e-mail. Please then delete it from your system and note that
any use, dissemination, forwarding, printing or copying is strictly prohibited. Any views
or opinions are solely those of the author and do not necessarily represent those of the
company.
Encryptions and Viruses
Please note that this e-mail and any attachments have not been encrypted. They may
therefore be liable to be compromised. Please also note that it is your responsibility to
scan this e-mail and any attachments for viruses. We do not, to the extent permitted by
law, accept any liability (whether in contract, negligence or otherwise) for any virus
infection and/or external compromise of security and/or confidentiality in relation to
transmissions sent by e-mail.
Monitoring
Activity and use of the company's systems is monitored to secure its effective use and
operation and for other lawful business purposes. Communications using these systems will
also be monitored and may be recorded to secure effective use and operation and for other
lawful business purposes.
hgvyjuv