I'm no NFS expert but for development domains I use the following options:
rw,sync,no_subtree_check,all_squash,anonuid=36,anongid=36
I wonder if something subtle changed on upgrade that interacts poorly with
your configuration?
On Wed, May 31, 2017 at 11:34 AM, Moritz Baumann <moritz.baumann(a)inf.ethz.ch
wrote:
Hi Adam,
Just an idea, but could this be related to stale mounts from when you
> rebooted the storage? Please try the following:
>
> 1. Place all nodes into maintenance mode
> 2. Disable the ovirt NFS exports
> 1. Comment out lines in /etc/exports
> 2. exportfs -r
> 3. Reboot your nodes
> 4. Re-enable the ovirt NFS exports
> 5. Activate your nodes
>
all storage domains (data/iso) are down, so is the data center
(non-responsive) and no nfs mount is on any of the nodes.
I can however manually mount the data export and touch a file (as root).
So I think stale mounts is not the issue.
However I did the steps and the result is the same.
Best,
Mo
--
Adam Litke