yes i think this too, but as you see at the top
[root@kvm380 ~]# gluster volume info
...
performance.strict-o-direct: on
...
it was already set
i did a one cluster setup with ovirt and I uses this result
Volume Name: engine
Type: Distribute
Volume ID: a40e848b-a8f1-4990-9d32-133b46db6f1d
Status: Started
Snapshot Count: 0
Number of Bricks: 1
Transport-type: tcp
Bricks:
Brick1: kvm360.durchhalten.intern:/gluster_bricks/engine/engine
Options Reconfigured:
cluster.eager-lock: enable
performance.io-cache: off
performance.read-ahead: off
performance.quick-read: off
user.cifs: off
network.ping-timeout: 30
network.remote-dio: off
performance.strict-o-direct: on
performance.low-prio-threads: 32
features.shard: on
storage.owner-gid: 36
storage.owner-uid: 36
transport.address-family: inet
nfs.disable: on
could there be an other reason?