
On Wed, Aug 31, 2016 at 10:43 AM, Rik Theys <Rik.Theys@esat.kuleuven.be> wrote:
On 08/30/2016 04:47 PM, Nir Soffer wrote:
On Tue, Aug 30, 2016 at 3:51 PM, Rik Theys <Rik.Theys@esat.kuleuven.be> wrote:
While rebooting one of the hosts in an oVirt cluster, I noticed that thin_check is run on the thin pool devices of one of the VM's on which the disk is assigned to.
That seems strange to me. I would expect the host to stay clear of any VM disks.
We expect the same thing, but unfortunately systemd and lvm try to auto activate stuff. This may be good idea for desktop system, but probably bad idea for a server and in particular a hypervisor.
We don't have a solution yet, but you can try these:
1. disable lvmetad service
systemctl stop lvm2-lvmetad.service lvm2-lvmetad.socket systemctl mask lvm2-lvmetad.service lvm2-lvmetad.socket
Edit /etc/lvm/lvm.conf:
use_lvmetad = 0
2. disable lvm auto activation
Edit /etc/lvm/lvm.conf:
auto_activation_volume_list = []
3. both 1 and 2
I've now applied both of the above and regenerated the initramfs and rebooted and the host no longer lists the LV's of the VM. Since I rebooted the host before without this issue, I'm not sure a single reboot is enough to conclude it has fully fixed the issue.
You mention that there's no solution yet. Does that mean the above settings are not 100% certain to avoid this behaviour?
Yes, these settings were suggested by lvm developers, but we did not test them yet, and of course did not integrate these settings in vdsm deployment. This will require modifying lvm.conf during configuration and verifying that lvm.conf is configured correctly when starting vdsm.