
I'm no NFS expert but for development domains I use the following options: rw,sync,no_subtree_check,all_squash,anonuid=36,anongid=36 I wonder if something subtle changed on upgrade that interacts poorly with your configuration? On Wed, May 31, 2017 at 11:34 AM, Moritz Baumann <moritz.baumann@inf.ethz.ch
wrote:
Hi Adam,
Just an idea, but could this be related to stale mounts from when you
rebooted the storage? Please try the following:
1. Place all nodes into maintenance mode 2. Disable the ovirt NFS exports 1. Comment out lines in /etc/exports 2. exportfs -r 3. Reboot your nodes 4. Re-enable the ovirt NFS exports 5. Activate your nodes
all storage domains (data/iso) are down, so is the data center (non-responsive) and no nfs mount is on any of the nodes.
I can however manually mount the data export and touch a file (as root).
So I think stale mounts is not the issue.
However I did the steps and the result is the same.
Best, Mo
-- Adam Litke