
You change the cache mode using a custom property per-VM I believe. I don't know if this would work for the hosted engine. I've already downgraded my system, but once you have the test machine up, perhaps you can try it. The custom property would be:
viodiskcache=writethrough
or
viodiskcache=writeback
https://access.redhat.com/documentation/en-US/Red_Hat_Enterprise_Virt ualization/3.6/html/Virtual_Machine_Management_Guide/appe- Reference_Settings_in_Administration_Portal_and_User_Portal_Windows.h tml#Virtual_Machine_Custom_Properties_settings_explained
Scott
On Thu, Jul 21, 2016 at 12:53 PM David Gossage <dgossage@carouselchec ks.com> wrote:
On Thu, Jul 21, 2016 at 11:47 AM, Scott <romracer@gmail.com> wrote: Hi David, My backend storage is ZFS.
> > > > > > > > > I thought about moving from FUSE to NFS mounts for my Gluster volumes to help test. But since I use hosted engine this would be a real pain. Its difficult to modify the storage domain type/path in the hosted-engine.conf. And I don't want to go
I found this https://bugzilla.redhat.com/show_bug.cgi?id=1347553
Not sure if related.
But I also have zfs backend, another user in gluster mailing list had issues and used zfs backend although she used proxmox and got it working by changing disk to writeback cache I think it was.
> > I also use hosted engine, but I run my gluster volume for HE actually on a LVM separate from zfs on xfs and if i recall it did not have the issues my gluster on zfs did. I'm wondering now if
Hopefully should have a test machone up soon I can play around with
more.
Scott
On Thu, Jul 21, 2016 at 11:36 AM David Gossage <dgossage@carousel
checks.com> wrote:
What back end storage do you run gluster on? xfs/zfs/ext4 etc?
David Gossage
Carousel Checks Inc. | System Administrator
Office 708.613.2284
> > > > > > On Thu, Jul 21, 2016 at 8:18 AM, Scott <romracer@gmail.com> wrote:
> > > > > > > > > I get similar problems with oVirt 4.0.1 and hosted engine. After upgrading all my hosts to Gluster 3.7.13 (client and server), I get the following: $ sudo hosted-engine --set-maintenance --mode=none Traceback (most recent call last): > > > > File "/usr/lib64/python2.7/runpy.py", line 162, in _run_module_as_main "__main__", fname, loader, pkg_name) File "/usr/lib64/python2.7/runpy.py", line 72, in _run_code exec code in run_globals > > > > > > > > > File "/usr/lib/python2.7/site-
if not maintenance.set_mode(sys.argv[1]):
> > > > > > > > > File "/usr/lib/python2.7/site-
value=m_global,
> > > > > > > > > File "/usr/lib/python2.7/site-
str(value))
> > > > > > > > > File "/usr/lib/python2.7/site-
all_stats = broker.get_stats_from_storage(service)
> > > > > > > > > File "/usr/lib/python2.7/site-
result = self._checked_communicate(request)
> > > > > > > > > File "/usr/lib/python2.7/site-
.format(message or response))
> > > > > > > > > ovirt_hosted_engine_ha.lib.exceptions.RequestError: Request failed: failed to read metadata: [Errno 1] Operation not
As per my cluster the problem is on an higher level: you can't activate the domains an FUSE, sanlock can't acquire the lock due to the permission errors visible in brick log. Am Donnerstag, den 21.07.2016, 19:17 +0000 schrieb Scott: through the process of re-deploying hosted engine. the issue was zfs settings. packages/ovirt_hosted_engine_setup/set_maintenance.py", line 73, in <module> packages/ovirt_hosted_engine_setup/set_maintenance.py", line 61, in set_mode packages/ovirt_hosted_engine_ha/client/client.py", line 259, in set_maintenance_mode packages/ovirt_hosted_engine_ha/client/client.py", line 204, in set_global_md_flag packages/ovirt_hosted_engine_ha/lib/brokerlink.py", line 232, in get_stats_from_storage packages/ovirt_hosted_engine_ha/lib/brokerlink.py", line 260, in _checked_communicate permitted
> > > > > > > > > If I only upgrade one host, then things will continue to work
but my nodes are constantly healing shards. My logs are also flooded with:
> > > > > > > > > [2016-07-21 13:15:14.137734] W [fuse-
bridge.c:2227:fuse_readv_cbk] 0-glusterfs-fuse: 274714: READ => -1 gfid=4
> > > > 41f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0041d0 (Operation not permitted) > > > > > > > > > > > > > > > > > > > The message "W [MSGID: 114031] [client-rpc- fops.c:3050:client3_3_readv_cbk] 0-data-client-0: remote operation failed [Operation not permitted]" repeated 6 times between [2016-07-21 13:13:24.134985] and [2016-07-21 13:15:04.132226] > > > > > > > > > > > > > > > > > > > The message "W [MSGID: 114031] [client-rpc- fops.c:3050:client3_3_readv_cbk] 0-data-client-1: remote operation failed [Operation not permitted]" repeated 8 times between [2016-07-21 13:13:34.133116] and [2016-07-21 13:15:14.137178] > > > > > > > > > > > > > > > > > > > The message "W [MSGID: 114031] [client-rpc- fops.c:3050:client3_3_readv_cbk] 0-data-client-2: remote operation failed [Operation not permitted]" repeated 7 times between [2016-07-21 13:13:24.135071] and [2016-07-21 13:15:14.137666] > > > > > > > > > [2016-07-21 13:15:24.134647] W [MSGID: 114031] [client-rpc- fops.c:3050:client3_3_readv_cbk] 0-data-client-0: remote operation failed [Operation not permitted] > > > > > > > > > [2016-07-21 13:15:24.134764] W [MSGID: 114031] [client-rpc- fops.c:3050:client3_3_readv_cbk] 0-data-client-2: remote operation failed [Operation not permitted] > > > > > > > > > > > > > > [2016-07-21 13:15:24.134793] W [fuse- bridge.c:2227:fuse_readv_cbk] 0-glusterfs-fuse: 274741: READ => -1 gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0038f4 (Operation not permitted) > > > > > > > > > > > > > > [2016-07-21 13:15:34.135413] W [fuse- bridge.c:2227:fuse_readv_cbk] 0-glusterfs-fuse: 274756: READ => -1 gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0041d0 (Operation not permitted) > > > > > > > > > > > > > > [2016-07-21 13:15:44.141062] W [fuse- bridge.c:2227:fuse_readv_cbk] 0-glusterfs-fuse: 274818: READ => -1 gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0038f4 (Operation not permitted) > > > > > > > > > [2016-07-21 13:15:54.133582] W [MSGID: 114031] [client-rpc- fops.c:3050:client3_3_readv_cbk] 0-data-client-1: remote operation failed [Operation not permitted] > > > > > > > > > > > > > > [2016-07-21 13:15:54.133629] W [fuse- bridge.c:2227:fuse_readv_cbk] 0-glusterfs-fuse: 274853: READ => -1 gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0036d8 (Operation not permitted) > > > > > > > > > > > > > > [2016-07-21 13:16:04.133666] W [fuse- bridge.c:2227:fuse_readv_cbk] 0-glusterfs-fuse: 274879: READ => -1 gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0041d0 (Operation not permitted) > > > > > > > > > > > > > > [2016-07-21 13:16:14.134954] W [fuse- bridge.c:2227:fuse_readv_cbk] 0-glusterfs-fuse: 274894: READ => -1 gfid=441f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0036d8 (Operation not permitted)
Scott
> > > > On Thu, Jul 21, 2016 at 6:57 AM Frank Rothenstein <f.rothenst ein@bodden-kliniken.de> wrote: Hey Devid, > > > > > > I have the very same problem on my test-cluster, despite on running ovirt 4.0. > > > > > > > > > > > > If you access your volumes via NFS all is fine, problem is FUSE. I stayed on 3.7.13, but have no solution yet, now I use NFS.
Frank
> > > > > > Am Donnerstag, den 21.07.2016, 04:28 -0500 schrieb David Gossage: > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > Anyone running one of recent 3.6.x lines and gluster using 3.7.13? I am looking to upgrade gluster from 3.7.11->3.7.13 for some bug fixes, but have been told by users on gluster mail list due to some gluster changes I'd need to change the disk parameters to use writeback cache. Something to do with aio support being removed. > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > I believe this could be done with custom parameters? But I believe strage tests are done using dd and would they fail with current settings then? Last upgrade to 3.7.13 I had to rollback to 3.7.11 due to stability isues where gluster storage would go into down state and always show N/A as space available/used. Even if hosts saw storage still and VM's were running on it on all 3 hosts. > > > > > > > > > Saw a lot of messages like these that went away once gluster rollback finished > > > > > > > > > > > > > > > > [2016-07-09 15:27:46.935694] I [fuse- bridge.c:4083:fuse_init] 0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.22 kernel 7.22 > > > > > > > > > > > > > > > [2016-07-09 15:27:49.555466] W [MSGID: 114031] [client- rpc-fops.c:3050:client3_3_readv_cbk] 0-GLUSTER1-client-1: remote operation failed [Operation not permitted] > > > > > > > > > > > > > > > [2016-07-09 15:27:49.556574] W [MSGID: 114031] [client- rpc-fops.c:3050:client3_3_readv_cbk] 0-GLUSTER1-client-0: remote operation failed [Operation not permitted] > > > > > > > > > > > > > > > > > > > > > > [2016-07-09 15:27:49.556659] W [fuse- bridge.c:2227:fuse_readv_cbk] 0-glusterfs-fuse: 80: READ => -1 gfid=deb61291-5176-4b81-8315-3f1cf8e3534d fd=0x7f5224002f68 (Operation not permitted) > > > > > > > > > > > > > > > [2016-07-09 15:27:59.612477] W [MSGID: 114031] [client- rpc-fops.c:3050:client3_3_readv_cbk] 0-GLUSTER1-client-1: remote operation failed [Operation not permitted] > > > > > > > > > > > > > > > [2016-07-09 15:27:59.613700] W [MSGID: 114031] [client- rpc-fops.c:3050:client3_3_readv_cbk] 0-GLUSTER1-client-0: remote operation failed [Operation not permitted] > > > > > > > > > > > > > > > > > > > > > > [2016-07-09 15:27:59.613781] W [fuse- bridge.c:2227:fuse_readv_cbk] 0-glusterfs-fuse: 168: READ => -1 gfid=deb61291-5176-4b81-8315-3f1cf8e3534d fd=0x7f5224002f68 (Operation not permitted) > > > > David Gossage > > > Carousel Checks Inc. | System Administrator > > > > > Office 708.613.2284 > > > > > > > > > _______________________________________________ > Users mailing list > Users@ovirt.org > http://lists.ovirt.org/mailman/listinfo/users
> > > > > > ___________________________________________________________
BODDEN-KLINIKEN Ribnitz-Damgarten GmbH
Sandhufe 2
18311 Ribnitz-Damgarten
Telefon: 03821-700-0
Fax: 03821-700-240
> > > > > > E-Mail: info@bodden-kliniken.de Internet: http://www.bodd
en-kliniken.de
> > > > > > > > > > > > Sitz: Ribnitz-
Damgarten, Amtsgericht: Stralsund, HRB 2919, Steuer- Nr.: 079/133/40188
> > > > > > Aufsichtsratsvorsitzende: Carmen Schröter, Geschäftsführer:
Dr. Falko Milski
> > > > > > > > > > > > Der Inhalt dieser E-
Mail ist ausschließlich für den bezeichneten Adressaten bes timmt. Wenn Sie nicht der vorge-
> > > > > > > > > > > > sehene Adressat dieser E-
Mail oder dessen Vertreter sein sollten, beachten Sie bitte , dass jede Form der Veröf-
> > > > > > fentlichung, Vervielfältigung oder Weitergabe des Inhalts d
ieser E-Mail unzulässig ist. Wir bitten Sie, sofort den
Absender zu informieren und die E-Mail zu löschen.
Bodden-Kliniken Ribnitz-Damgarten GmbH 2016
> > > > > > *** Virenfrei durch Kerio Mail Server und Sophos Antivirus
_______________________________________________
Users mailing list
Users@ovirt.org
______________________________________________________________________________ BODDEN-KLINIKEN Ribnitz-Damgarten GmbH Sandhufe 2 18311 Ribnitz-Damgarten Telefon: 03821-700-0 Fax: 03821-700-240 E-Mail: info@bodden-kliniken.de Internet: http://www.bodden-kliniken.de Sitz: Ribnitz-Damgarten, Amtsgericht: Stralsund, HRB 2919, Steuer-Nr.: 079/133/40188 Aufsichtsratsvorsitzende: Carmen Schröter, Geschäftsführer: Dr. Falko Milski Der Inhalt dieser E-Mail ist ausschließlich für den bezeichneten Adressaten bestimmt. Wenn Sie nicht der vorge- sehene Adressat dieser E-Mail oder dessen Vertreter sein sollten, beachten Sie bitte, dass jede Form der Veröf- fentlichung, Vervielfältigung oder Weitergabe des Inhalts dieser E-Mail unzulässig ist. Wir bitten Sie, sofort den Absender zu informieren und die E-Mail zu löschen. Bodden-Kliniken Ribnitz-Damgarten GmbH 2016 *** Virenfrei durch Kerio Mail Server und Sophos Antivirus ***