
yes i think this too, but as you see at the top
[root@kvm380 ~]# gluster volume info ... performance.strict-o-direct: on ... it was already set
i did a one cluster setup with ovirt and I uses this result Volume Name: engine Type: Distribute Volume ID: a40e848b-a8f1-4990-9d32-133b46db6f1d Status: Started Snapshot Count: 0 Number of Bricks: 1 Transport-type: tcp Bricks: Brick1: kvm360.durchhalten.intern:/gluster_bricks/engine/engine Options Reconfigured: cluster.eager-lock: enable performance.io-cache: off performance.read-ahead: off performance.quick-read: off user.cifs: off network.ping-timeout: 30 network.remote-dio: off performance.strict-o-direct: on performance.low-prio-threads: 32 features.shard: on storage.owner-gid: 36 storage.owner-uid: 36 transport.address-family: inet nfs.disable: on could there be an other reason?