Had a chance to upgrade my cluster to Gluster 3.7.14 and can confirm
it
works for me too where 3.7.12/13 did not.
I did find that you should NOT turn off network.remote-dio or turn
on performance.strict-o-direct as suggested earlier in the thread. They
will prevent dd (using direct flag) and other things from working
properly. I'd leave those at network.remote-dio=enabled
and performance.strict-o-direct=off.
Those were actually just suggested during a testing phase trying to trace
down the issue. Neither of those 2 I think have ever been suggested as
good practice. At least not for VM storage.
Hopefully we can see Gluster 3.7.14 moved out of testing repo soon.
Scott
On Tue, Aug 2, 2016 at 9:05 AM, David Gossage <dgossage(a)carouselchecks.com
> wrote:
> So far gluster 3.7.14 seems to have resolved issues at least on my test
> box. dd commands that failed previously now work with sharding on zfs
> backend,
>
> Where before I couldn't even mount a new storage domain it now mounted
> and I have a test vm being created.
>
> Still have to let VM run for a few days and make sure no locking freezing
> occurs but looks hopeful so far.
>
> *David Gossage*
> *Carousel Checks Inc. | System Administrator*
> *Office* 708.613.2284
>
> On Tue, Jul 26, 2016 at 8:15 AM, David Gossage <
> dgossage(a)carouselchecks.com> wrote:
>
>> On Tue, Jul 26, 2016 at 4:37 AM, Krutika Dhananjay <kdhananj(a)redhat.com>
>> wrote:
>>
>>> Hi,
>>>
>>> 1. Could you please attach the glustershd logs from all three nodes?
>>>
>>
>> Here are ccgl1 and ccgl2. as previously mentioned ccgl3 third node was
>> down from bad nic so no relevant logs would be on that node.
>>
>>
>>>
>>> 2. Also, so far what we know is that the 'Operation not permitted'
>>> errors are on the main vm image itself and not its individual shards (ex
>>> deb61291-5176-4b81-8315-3f1cf8e3534d). Could you do the following:
>>> Get the inode number of
.glusterfs/de/b6/deb61291-5176-4b81-8315-3f1cf8e3534d
>>> (ls -li) from the first brick. I'll call this number INODE_NUMBER.
>>> Execute `find . -inum INODE_NUMBER` from the brick root on first brick
>>> to print the hard links against the file in the prev step and share the
>>> output.
>>>
>> [dgossage@ccgl1 ~]$ sudo ls -li /gluster1/BRICK1/1/.glusterfs/
>> de/b6/deb61291-5176-4b81-8315-3f1cf8e3534d
>> 16407 -rw-r--r--. 2 36 36 466 Jun 5 16:52 /gluster1/BRICK1/1/.glusterfs/
>> de/b6/deb61291-5176-4b81-8315-3f1cf8e3534d
>> [dgossage@ccgl1 ~]$ cd /gluster1/BRICK1/1/
>> [dgossage@ccgl1 1]$ sudo find . -inum 16407
>> ./7c73a8dd-a72e-4556-ac88-7f6813131e64/dom_md/metadata
>> ./.glusterfs/de/b6/deb61291-5176-4b81-8315-3f1cf8e3534d
>>
>>
>>
>>>
>>> 3. Did you delete any vms at any point before or after the upgrade?
>>>
>>
>> Immediately before or after on same day pretty sure I deleted nothing.
>> During week prior I deleted a few dev vm's that were never setup and some
>> the week after upgrade as I was preparing for moving disks off and on
>> storage to get them sharded and felt it would be easier to just recreate
>> some disks that had no data yet rather than move them off and on later.
>>
>>>
>>> -Krutika
>>>
>>> On Mon, Jul 25, 2016 at 11:30 PM, David Gossage <
>>> dgossage(a)carouselchecks.com> wrote:
>>>
>>>>
>>>> On Mon, Jul 25, 2016 at 9:58 AM, Krutika Dhananjay <
>>>> kdhananj(a)redhat.com> wrote:
>>>>
>>>>> OK, could you try the following:
>>>>>
>>>>> i. Set network.remote-dio to off
>>>>> # gluster volume set <VOL> network.remote-dio off
>>>>>
>>>>> ii. Set performance.strict-o-direct to on
>>>>> # gluster volume set <VOL> performance.strict-o-direct
on
>>>>>
>>>>> iii. Stop the affected vm(s) and start again
>>>>>
>>>>> and tell me if you notice any improvement?
>>>>>
>>>>>
>>>> Previous instll I had issue with is still on gluster 3.7.11
>>>>
>>>> My test install of ovirt 3.6.7 and gluster 3.7.13 with 3 bricks on a
>>>> locak disk right now isn't allowing me to add the gluster storage at
all.
>>>>
>>>> Keep getting some type of UI error
>>>>
>>>> 2016-07-25 12:49:09,277 ERROR [org.ovirt.engine.ui.frontend.
>>>> server.gwt.OvirtRemoteLoggingService] (default task-33) []
>>>> Permutation name: 430985F23DFC1C8BE1C7FDD91EDAA785
>>>> 2016-07-25 12:49:09,277 ERROR [org.ovirt.engine.ui.frontend.
>>>> server.gwt.OvirtRemoteLoggingService] (default task-33) [] Uncaught
>>>> exception: : java.lang.ClassCastException
>>>> at Unknown.ps(https://ccengine2.c
>>>> arouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1
>>>> C7FDD91EDAA785.cache.html@3837) at Unknown.ts(https://ccengine2.c
>>>> arouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1
>>>> C7FDD91EDAA785.cache.html@20) at Unknown.vs(https://ccengine2.c
>>>> arouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1
>>>> C7FDD91EDAA785.cache.html@18) at Unknown.iJf(https://ccengine2.
>>>> carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE
>>>> 1C7FDD91EDAA785.cache.html@19) at Unknown.Xab(https://ccengine2.
>>>> carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE
>>>> 1C7FDD91EDAA785.cache.html@48) at Unknown.P8o(https://ccengine2.
>>>> carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE
>>>> 1C7FDD91EDAA785.cache.html@4447) at Unknown.jQr(https://ccengine2.
>>>> carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE
>>>> 1C7FDD91EDAA785.cache.html@21) at Unknown.A8o(https://ccengine2.
>>>> carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE
>>>> 1C7FDD91EDAA785.cache.html@51) at Unknown.u8o(https://ccengine2.
>>>> carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE
>>>> 1C7FDD91EDAA785.cache.html@101) at Unknown.Eap(https://ccengine2.
>>>> carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE
>>>> 1C7FDD91EDAA785.cache.html@10718) at Unknown.p8n(https://ccengine2.
>>>> carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE
>>>> 1C7FDD91EDAA785.cache.html@161) at Unknown.Cao(https://ccengine2.
>>>> carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE
>>>> 1C7FDD91EDAA785.cache.html@31) at Unknown.Bap(https://ccengine2.
>>>> carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE
>>>> 1C7FDD91EDAA785.cache.html@10469) at Unknown.kRn(https://ccengine2.
>>>> carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE
>>>> 1C7FDD91EDAA785.cache.html@49) at Unknown.nRn(https://ccengine2.
>>>> carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE
>>>> 1C7FDD91EDAA785.cache.html@438) at Unknown.eVn(https://ccengine2.
>>>> carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE
>>>> 1C7FDD91EDAA785.cache.html@40) at Unknown.hVn(https://ccengine2.
>>>> carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE
>>>> 1C7FDD91EDAA785.cache.html@25827) at Unknown.MTn(https://ccengine2.
>>>> carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE
>>>> 1C7FDD91EDAA785.cache.html@25) at Unknown.PTn(https://ccengine2.
>>>> carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE
>>>> 1C7FDD91EDAA785.cache.html@24052) at Unknown.KJe(https://ccengine2.
>>>> carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE
>>>> 1C7FDD91EDAA785.cache.html@21125) at Unknown.Izk(https://ccengine2.
>>>> carouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE
>>>> 1C7FDD91EDAA785.cache.html@10384) at Unknown.P3(https://ccengine2.c
>>>> arouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1
>>>> C7FDD91EDAA785.cache.html@137) at Unknown.g4(https://ccengine2.c
>>>> arouselchecks.local/ovirt-engine/webadmin/430985F23DFC1C8BE1
>>>> C7FDD91EDAA785.cache.html@8271) at
Unknown.<anonymous>(https://cc
>>>> engine2.carouselchecks.local/ovirt-engine/webadmin/430985F2
>>>> 3DFC1C8BE1C7FDD91EDAA785.cache.html@65) at Unknown._t(
>>>>
https://ccengine2.carouselchecks.local/ovirt-engi
>>>> ne/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@29) at
>>>> Unknown.du(https://ccengine2.carouselchecks.local/ovirt-engi
>>>> ne/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@57) at
>>>> Unknown.<anonymous>(https://ccengine2.carouselchecks.local/
>>>> ovirt-engine/webadmin/430985F23DFC1C8BE1C7FDD91EDAA785.cache.html@54)
>>>>
>>>>
>>>>> -Krutika
>>>>>
>>>>> On Mon, Jul 25, 2016 at 4:57 PM, Samuli Heinonen <
>>>>> samppah(a)neutraali.net> wrote:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> > On 25 Jul 2016, at 12:34, David Gossage <
>>>>>> dgossage(a)carouselchecks.com> wrote:
>>>>>> >
>>>>>> > On Mon, Jul 25, 2016 at 1:01 AM, Krutika Dhananjay <
>>>>>> kdhananj(a)redhat.com> wrote:
>>>>>> > Hi,
>>>>>> >
>>>>>> > Thanks for the logs. So I have identified one issue from the
logs
>>>>>> for which the fix is this:
http://review.gluster.org/#/c/14669/.
>>>>>> Because of a bug in the code, ENOENT was getting converted to
EPERM and
>>>>>> being propagated up the stack causing the reads to bail out early
with
>>>>>> 'Operation not permitted' errors.
>>>>>> > I still need to find out two things:
>>>>>> > i) why there was a readv() sent on a non-existent (ENOENT)
file
>>>>>> (this is important since some of the other users have not faced
or reported
>>>>>> this issue on gluster-users with 3.7.13)
>>>>>> > ii) need to see if there's a way to work around this
issue.
>>>>>> >
>>>>>> > Do you mind sharing the steps needed to be executed to run
into
>>>>>> this issue? This is so that we can apply our patches, test and
ensure they
>>>>>> fix the problem.
>>>>>>
>>>>>>
>>>>>> Unfortunately I can’t test this right away nor give exact steps
how
>>>>>> to test this. This is just a theory but please correct me if you
see some
>>>>>> mistakes.
>>>>>>
>>>>>> oVirt uses cache=none settings for VM’s by default which
requires
>>>>>> direct I/O. oVirt also uses dd with iflag=direct to check that
storage has
>>>>>> direct I/O enabled. Problems exist with GlusterFS with sharding
enabled and
>>>>>> bricks running on ZFS on Linux. Everything seems to be fine with
GlusterFS
>>>>>> 3.7.11 and problems exist at least with version .12 and .13.
There has been
>>>>>> some posts saying that GlusterFS 3.8.x is also affected.
>>>>>>
>>>>>> Steps to reproduce:
>>>>>> 1. Sharded file is created with GlusterFS 3.7.11. Everything
works
>>>>>> ok.
>>>>>> 2. GlusterFS is upgraded to 3.7.12+
>>>>>> 3. Sharded file cannot be read or written with direct I/O
enabled.
>>>>>> (Ie. oVirt uses to check storage connection with command
"dd
>>>>>>
if=/rhev/data-center/00000001-0001-0001-0001-0000000002b6/mastersd/dom_md/inbox
>>>>>> iflag=direct,fullblock count=1 bs=1024000”)
>>>>>>
>>>>>> Please let me know if you need more information.
>>>>>>
>>>>>> -samuli
>>>>>>
>>>>>> > Well after upgrade of gluster all I did was start ovirt
hosts up
>>>>>> which launched and started their ha-agent and broker processes.
I don't
>>>>>> believe I started getting any errors till it mounted GLUSTER1. I
had
>>>>>> enabled sharding but had no sharded disk images yet. Not sure if
the check
>>>>>> for shards would have caused that. Unfortunately I can't
just update this
>>>>>> cluster and try and see what caused it as it has sme VM's
users expect to
>>>>>> be available in few hours.
>>>>>> >
>>>>>> > I can see if I can get my test setup to recreate it. I
think I'll
>>>>>> need to de-activate data center so I can detach the storage thats
on xfs
>>>>>> and attach the one thats over zfs with sharding enabled. My test
is 3
>>>>>> bricks on same local machine, with 3 different volumes but I
think im
>>>>>> running into sanlock issue or something as it won't mount
more than one
>>>>>> volume that was created locally.
>>>>>> >
>>>>>> >
>>>>>> > -Krutika
>>>>>> >
>>>>>> > On Fri, Jul 22, 2016 at 7:17 PM, David Gossage <
>>>>>> dgossage(a)carouselchecks.com> wrote:
>>>>>> > Trimmed out the logs to just about when I was shutting down
ovirt
>>>>>> servers for updates which was 14:30 UTC 2016-07-09
>>>>>> >
>>>>>> > Pre-update settings were
>>>>>> >
>>>>>> > Volume Name: GLUSTER1
>>>>>> > Type: Replicate
>>>>>> > Volume ID: 167b8e57-28c3-447a-95cc-8410cbdf3f7f
>>>>>> > Status: Started
>>>>>> > Number of Bricks: 1 x 3 = 3
>>>>>> > Transport-type: tcp
>>>>>> > Bricks:
>>>>>> > Brick1: ccgl1.gl.local:/gluster1/BRICK1/1
>>>>>> > Brick2: ccgl2.gl.local:/gluster1/BRICK1/1
>>>>>> > Brick3: ccgl3.gl.local:/gluster1/BRICK1/1
>>>>>> > Options Reconfigured:
>>>>>> > performance.readdir-ahead: on
>>>>>> > storage.owner-uid: 36
>>>>>> > storage.owner-gid: 36
>>>>>> > performance.quick-read: off
>>>>>> > performance.read-ahead: off
>>>>>> > performance.io-cache: off
>>>>>> > performance.stat-prefetch: off
>>>>>> > cluster.eager-lock: enable
>>>>>> > network.remote-dio: enable
>>>>>> > cluster.quorum-type: auto
>>>>>> > cluster.server-quorum-type: server
>>>>>> > server.allow-insecure: on
>>>>>> > cluster.self-heal-window-size: 1024
>>>>>> > cluster.background-self-heal-count: 16
>>>>>> > performance.strict-write-ordering: off
>>>>>> > nfs.disable: on
>>>>>> > nfs.addr-namelookup: off
>>>>>> > nfs.enable-ino32: off
>>>>>> >
>>>>>> > At the time of updates ccgl3 was offline from bad nic on
server
>>>>>> but had been so for about a week with no issues in volume
>>>>>> >
>>>>>> > Shortly after update I added these settings to enable
sharding but
>>>>>> did not as of yet have any VM images sharded.
>>>>>> > features.shard-block-size: 64MB
>>>>>> > features.shard: on
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>> > David Gossage
>>>>>> > Carousel Checks Inc. | System Administrator
>>>>>> > Office 708.613.2284
>>>>>> >
>>>>>> > On Fri, Jul 22, 2016 at 5:00 AM, Krutika Dhananjay <
>>>>>> kdhananj(a)redhat.com> wrote:
>>>>>> > Hi David,
>>>>>> >
>>>>>> > Could you also share the brick logs from the affected
volume?
>>>>>> They're located at /var/log/glusterfs/bricks/<hyp
>>>>>> henated-path-to-the-brick-directory>.log.
>>>>>> >
>>>>>> > Also, could you share the volume configuration (output of
`gluster
>>>>>> volume info <VOL>`) for the affected volume(s) AND at the
time you actually
>>>>>> saw this issue?
>>>>>> >
>>>>>> > -Krutika
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>> > On Thu, Jul 21, 2016 at 11:23 PM, David Gossage <
>>>>>> dgossage(a)carouselchecks.com> wrote:
>>>>>> > On Thu, Jul 21, 2016 at 11:47 AM, Scott
<romracer(a)gmail.com>
>>>>>> wrote:
>>>>>> > Hi David,
>>>>>> >
>>>>>> > My backend storage is ZFS.
>>>>>> >
>>>>>> > I thought about moving from FUSE to NFS mounts for my
Gluster
>>>>>> volumes to help test. But since I use hosted engine this would
be a real
>>>>>> pain. Its difficult to modify the storage domain type/path in
the
>>>>>> hosted-engine.conf. And I don't want to go through the
process of
>>>>>> re-deploying hosted engine.
>>>>>> >
>>>>>> >
>>>>>> > I found this
>>>>>> >
>>>>>> >
https://bugzilla.redhat.com/show_bug.cgi?id=1347553
>>>>>> >
>>>>>> > Not sure if related.
>>>>>> >
>>>>>> > But I also have zfs backend, another user in gluster mailing
list
>>>>>> had issues and used zfs backend although she used proxmox and got
it
>>>>>> working by changing disk to writeback cache I think it was.
>>>>>> >
>>>>>> > I also use hosted engine, but I run my gluster volume for
HE
>>>>>> actually on a LVM separate from zfs on xfs and if i recall it did
not have
>>>>>> the issues my gluster on zfs did. I'm wondering now if the
issue was zfs
>>>>>> settings.
>>>>>> >
>>>>>> > Hopefully should have a test machone up soon I can play
around
>>>>>> with more.
>>>>>> >
>>>>>> > Scott
>>>>>> >
>>>>>> > On Thu, Jul 21, 2016 at 11:36 AM David Gossage <
>>>>>> dgossage(a)carouselchecks.com> wrote:
>>>>>> > What back end storage do you run gluster on? xfs/zfs/ext4
etc?
>>>>>> >
>>>>>> > David Gossage
>>>>>> > Carousel Checks Inc. | System Administrator
>>>>>> > Office 708.613.2284
>>>>>> >
>>>>>> > On Thu, Jul 21, 2016 at 8:18 AM, Scott
<romracer(a)gmail.com> wrote:
>>>>>> > I get similar problems with oVirt 4.0.1 and hosted engine.
After
>>>>>> upgrading all my hosts to Gluster 3.7.13 (client and server), I
get the
>>>>>> following:
>>>>>> >
>>>>>> > $ sudo hosted-engine --set-maintenance --mode=none
>>>>>> > Traceback (most recent call last):
>>>>>> > File "/usr/lib64/python2.7/runpy.py", line 162,
in
>>>>>> _run_module_as_main
>>>>>> > "__main__", fname, loader, pkg_name)
>>>>>> > File "/usr/lib64/python2.7/runpy.py", line 72,
in _run_code
>>>>>> > exec code in run_globals
>>>>>> > File "/usr/lib/python2.7/site-packa
>>>>>> ges/ovirt_hosted_engine_setup/set_maintenance.py", line 73,
in
>>>>>> <module>
>>>>>> > if not maintenance.set_mode(sys.argv[1]):
>>>>>> > File "/usr/lib/python2.7/site-packa
>>>>>> ges/ovirt_hosted_engine_setup/set_maintenance.py", line 61,
in
>>>>>> set_mode
>>>>>> > value=m_global,
>>>>>> > File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/client/client.py",
>>>>>> line 259, in set_maintenance_mode
>>>>>> > str(value))
>>>>>> > File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/client/client.py",
>>>>>> line 204, in set_global_md_flag
>>>>>> > all_stats = broker.get_stats_from_storage(service)
>>>>>> > File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py",
>>>>>> line 232, in get_stats_from_storage
>>>>>> > result = self._checked_communicate(request)
>>>>>> > File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py",
>>>>>> line 260, in _checked_communicate
>>>>>> > .format(message or response))
>>>>>> > ovirt_hosted_engine_ha.lib.exceptions.RequestError: Request
>>>>>> failed: failed to read metadata: [Errno 1] Operation not
permitted
>>>>>> >
>>>>>> > If I only upgrade one host, then things will continue to
work but
>>>>>> my nodes are constantly healing shards. My logs are also flooded
with:
>>>>>> >
>>>>>> > [2016-07-21 13:15:14.137734] W
[fuse-bridge.c:2227:fuse_readv_cbk]
>>>>>> 0-glusterfs-fuse: 274714: READ => -1 gfid=4
>>>>>> > 41f2789-f6b1-4918-a280-1b9905a11429 fd=0x7f19bc0041d0
(Operation
>>>>>> not permitted)
>>>>>> > The message "W [MSGID: 114031]
[client-rpc-fops.c:3050:client3_3_readv_cbk]
>>>>>> 0-data-client-0: remote operation failed [Operation not
permitted]"
>>>>>> repeated 6 times between [2016-07-21 13:13:24.134985] and
[2016-07-21
>>>>>> 13:15:04.132226]
>>>>>> > The message "W [MSGID: 114031]
[client-rpc-fops.c:3050:client3_3_readv_cbk]
>>>>>> 0-data-client-1: remote operation failed [Operation not
permitted]"
>>>>>> repeated 8 times between [2016-07-21 13:13:34.133116] and
[2016-07-21
>>>>>> 13:15:14.137178]
>>>>>> > The message "W [MSGID: 114031]
[client-rpc-fops.c:3050:client3_3_readv_cbk]
>>>>>> 0-data-client-2: remote operation failed [Operation not
permitted]"
>>>>>> repeated 7 times between [2016-07-21 13:13:24.135071] and
[2016-07-21
>>>>>> 13:15:14.137666]
>>>>>> > [2016-07-21 13:15:24.134647] W [MSGID: 114031]
>>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-data-client-0:
>>>>>> remote operation failed [Operation not permitted]
>>>>>> > [2016-07-21 13:15:24.134764] W [MSGID: 114031]
>>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-data-client-2:
>>>>>> remote operation failed [Operation not permitted]
>>>>>> > [2016-07-21 13:15:24.134793] W
[fuse-bridge.c:2227:fuse_readv_cbk]
>>>>>> 0-glusterfs-fuse: 274741: READ => -1
gfid=441f2789-f6b1-4918-a280-1b9905a11429
>>>>>> fd=0x7f19bc0038f4 (Operation not permitted)
>>>>>> > [2016-07-21 13:15:34.135413] W
[fuse-bridge.c:2227:fuse_readv_cbk]
>>>>>> 0-glusterfs-fuse: 274756: READ => -1
gfid=441f2789-f6b1-4918-a280-1b9905a11429
>>>>>> fd=0x7f19bc0041d0 (Operation not permitted)
>>>>>> > [2016-07-21 13:15:44.141062] W
[fuse-bridge.c:2227:fuse_readv_cbk]
>>>>>> 0-glusterfs-fuse: 274818: READ => -1
gfid=441f2789-f6b1-4918-a280-1b9905a11429
>>>>>> fd=0x7f19bc0038f4 (Operation not permitted)
>>>>>> > [2016-07-21 13:15:54.133582] W [MSGID: 114031]
>>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-data-client-1:
>>>>>> remote operation failed [Operation not permitted]
>>>>>> > [2016-07-21 13:15:54.133629] W
[fuse-bridge.c:2227:fuse_readv_cbk]
>>>>>> 0-glusterfs-fuse: 274853: READ => -1
gfid=441f2789-f6b1-4918-a280-1b9905a11429
>>>>>> fd=0x7f19bc0036d8 (Operation not permitted)
>>>>>> > [2016-07-21 13:16:04.133666] W
[fuse-bridge.c:2227:fuse_readv_cbk]
>>>>>> 0-glusterfs-fuse: 274879: READ => -1
gfid=441f2789-f6b1-4918-a280-1b9905a11429
>>>>>> fd=0x7f19bc0041d0 (Operation not permitted)
>>>>>> > [2016-07-21 13:16:14.134954] W
[fuse-bridge.c:2227:fuse_readv_cbk]
>>>>>> 0-glusterfs-fuse: 274894: READ => -1
gfid=441f2789-f6b1-4918-a280-1b9905a11429
>>>>>> fd=0x7f19bc0036d8 (Operation not permitted)
>>>>>> >
>>>>>> > Scott
>>>>>> >
>>>>>> >
>>>>>> > On Thu, Jul 21, 2016 at 6:57 AM Frank Rothenstein <
>>>>>> f.rothenstein(a)bodden-kliniken.de> wrote:
>>>>>> > Hey Devid,
>>>>>> >
>>>>>> > I have the very same problem on my test-cluster, despite on
>>>>>> running ovirt 4.0.
>>>>>> > If you access your volumes via NFS all is fine, problem is
FUSE. I
>>>>>> stayed on 3.7.13, but have no solution yet, now I use NFS.
>>>>>> >
>>>>>> > Frank
>>>>>> >
>>>>>> > Am Donnerstag, den 21.07.2016, 04:28 -0500 schrieb David
Gossage:
>>>>>> >> Anyone running one of recent 3.6.x lines and gluster
using
>>>>>> 3.7.13? I am looking to upgrade gluster from 3.7.11->3.7.13
for some bug
>>>>>> fixes, but have been told by users on gluster mail list due to
some gluster
>>>>>> changes I'd need to change the disk parameters to use
writeback cache.
>>>>>> Something to do with aio support being removed.
>>>>>> >>
>>>>>> >> I believe this could be done with custom parameters?
But I
>>>>>> believe strage tests are done using dd and would they fail with
current
>>>>>> settings then? Last upgrade to 3.7.13 I had to rollback to 3.7.11
due to
>>>>>> stability isues where gluster storage would go into down state
and always
>>>>>> show N/A as space available/used. Even if hosts saw storage
still and VM's
>>>>>> were running on it on all 3 hosts.
>>>>>> >>
>>>>>> >> Saw a lot of messages like these that went away once
gluster
>>>>>> rollback finished
>>>>>> >>
>>>>>> >> [2016-07-09 15:27:46.935694] I
[fuse-bridge.c:4083:fuse_init]
>>>>>> 0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs
7.22 kernel
>>>>>> 7.22
>>>>>> >> [2016-07-09 15:27:49.555466] W [MSGID: 114031]
>>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk]
0-GLUSTER1-client-1:
>>>>>> remote operation failed [Operation not permitted]
>>>>>> >> [2016-07-09 15:27:49.556574] W [MSGID: 114031]
>>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk]
0-GLUSTER1-client-0:
>>>>>> remote operation failed [Operation not permitted]
>>>>>> >> [2016-07-09 15:27:49.556659] W
[fuse-bridge.c:2227:fuse_readv_cbk]
>>>>>> 0-glusterfs-fuse: 80: READ => -1
gfid=deb61291-5176-4b81-8315-3f1cf8e3534d
>>>>>> fd=0x7f5224002f68 (Operation not permitted)
>>>>>> >> [2016-07-09 15:27:59.612477] W [MSGID: 114031]
>>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk]
0-GLUSTER1-client-1:
>>>>>> remote operation failed [Operation not permitted]
>>>>>> >> [2016-07-09 15:27:59.613700] W [MSGID: 114031]
>>>>>> [client-rpc-fops.c:3050:client3_3_readv_cbk]
0-GLUSTER1-client-0:
>>>>>> remote operation failed [Operation not permitted]
>>>>>> >> [2016-07-09 15:27:59.613781] W
[fuse-bridge.c:2227:fuse_readv_cbk]
>>>>>> 0-glusterfs-fuse: 168: READ => -1
gfid=deb61291-5176-4b81-8315-3f1cf8e3534d
>>>>>> fd=0x7f5224002f68 (Operation not permitted)
>>>>>> >>
>>>>>> >> David Gossage
>>>>>> >> Carousel Checks Inc. | System Administrator
>>>>>> >> Office 708.613.2284
>>>>>> >> _______________________________________________
>>>>>> >> Users mailing list
>>>>>> >>
>>>>>> >> Users(a)ovirt.org
>>>>>> >>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>> >
____________________________________________________________
>>>>>> __________________
>>>>>> > BODDEN-KLINIKEN Ribnitz-Damgarten GmbH
>>>>>> > Sandhufe 2
>>>>>> > 18311 Ribnitz-Damgarten
>>>>>> >
>>>>>> > Telefon: 03821-700-0
>>>>>> > Fax: 03821-700-240
>>>>>> >
>>>>>> > E-Mail: info(a)bodden-kliniken.de Internet:
>>>>>>
http://www.bodden-kliniken.de
>>>>>> >
>>>>>> > Sitz: Ribnitz-Damgarten, Amtsgericht: Stralsund, HRB 2919,
>>>>>> Steuer-Nr.: 079/133/40188
>>>>>> > Aufsichtsratsvorsitzende: Carmen Schröter, Geschäftsführer:
Dr.
>>>>>> Falko Milski
>>>>>> >
>>>>>> > Der Inhalt dieser E-Mail ist ausschließlich für den
bezeichneten
>>>>>> Adressaten bestimmt. Wenn Sie nicht der vorge-
>>>>>> > sehene Adressat dieser E-Mail oder dessen Vertreter sein
sollten,
>>>>>> beachten Sie bitte, dass jede Form der Veröf-
>>>>>> > fentlichung, Vervielfältigung oder Weitergabe des Inhalts
dieser
>>>>>> E-Mail unzulässig ist. Wir bitten Sie, sofort den
>>>>>> > Absender zu informieren und die E-Mail zu löschen.
>>>>>> >
>>>>>> >
>>>>>> > Bodden-Kliniken Ribnitz-Damgarten GmbH 2016
>>>>>> > *** Virenfrei durch Kerio Mail Server und Sophos Antivirus
***
>>>>>> > _______________________________________________
>>>>>> > Users mailing list
>>>>>> > Users(a)ovirt.org
>>>>>> >
http://lists.ovirt.org/mailman/listinfo/users
>>>>>> >
>>>>>> >
>>>>>> > _______________________________________________
>>>>>> > Users mailing list
>>>>>> > Users(a)ovirt.org
>>>>>> >
http://lists.ovirt.org/mailman/listinfo/users
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>> > _______________________________________________
>>>>>> > Users mailing list
>>>>>> > Users(a)ovirt.org
>>>>>> >
http://lists.ovirt.org/mailman/listinfo/users
>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>
> _______________________________________________
> Users mailing list
> Users(a)ovirt.org
>
http://lists.ovirt.org/mailman/listinfo/users
>
>