Re: [Users] AcquireHostId problem

sanlock.log : 2013-12-18 21:23:32+0100 1900 [867]: s1 lockspace b2d69b22-a8b8-466c-bf1f-b6e565228238:250:/rhev/data-center/mnt/lab2.home:_home_vdsm_data/b2d69b22-a8b8-466c-bf1f-b6e565228238/dom_md/ids:0 2013-12-18 21:23:52+0100 1920 [4238]: s1 wdmd_connect failed -111 2013-12-18 21:23:52+0100 1920 [4238]: s1 create_watchdog failed -1 2013-12-18 21:23:53+0100 1921 [867]: s1 add_lockspace fail result -203 2013/12/18 Pascal Jakobi <pascal.jakobi@gmail.com>
Apologies for not replying more quickly - I was cross-checking. The bad news is that it has not been resolved.
I can mount NFS shares outside of ovirt. But I fail creating domains....
Thxs
2013/12/17 Itamar Heim <iheim@redhat.com>
On 12/11/2013 03:57 AM, Pascal Jakobi wrote:
Context : vdsm-4.13.0-11.fc19.x86_64 sanlock-2.8-1.fc19.x86_64
Creating storage domain (NFS) fails with error in engine.log such as Error code AcquireHostIdFailure and error message VDSGenericException: VDSErrorException: Failed to CreateStorage And Cannot acquire host id: ('b4f05e3e-d714-45c4-adf9-8d269418338f', SanlockException(-203, 'Sanlock lockspace add failure', 'Sanlock exception'))
Reading the mail archive, I can see that this is a know issue but I do not see the workaround. Can someone enlight me ?
Also, is there somewhere a coherent instruction set in order to make NFSv3 work (including iptables settings). I had to stop iptables to unblock myself...
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
hi pascal,
was this resolved?
thanks, Itamar
-- *Pascal Jakobi* 116 rue de Stalingrad 93100 Montreuil, France
*+33 6 87 47 58 19 <%2B33%206%2087%2047%2058%2019>*Pascal.Jakobi@gmail.com
-- *Pascal Jakobi* 116 rue de Stalingrad 93100 Montreuil, France *+33 6 87 47 58 19*Pascal.Jakobi@gmail.com

----- Original Message -----
From: "Pascal Jakobi" <pascal.jakobi@gmail.com> To: users@ovirt.org Sent: Wednesday, December 18, 2013 9:32:00 PM Subject: Re: [Users] AcquireHostId problem
sanlock.log :
2013-12-18 21:23:32+0100 1900 [867]: s1 lockspace b2d69b22-a8b8-466c-bf1f-b6e565228238:250:/rhev/data-center/mnt/lab2.home:_home_vdsm_data/b2d69b22-a8b8-466c-bf1f-b6e565228238/dom_md/ids:0 2013-12-18 21:23:52+0100 1920 [4238]: s1 wdmd_connect failed -111 2013-12-18 21:23:52+0100 1920 [4238]: s1 create_watchdog failed -1 2013-12-18 21:23:53+0100 1921 [867]: s1 add_lockspace fail result -203
Hi Pascal, is wdmd up and running? # ps aux | grep wdmd root 1650 0.0 0.2 13552 3320 ? SLs 03:49 0:00 wdmd -G sanlock -- Federico

Federico On may suspect wdmd isn't running as wdmd_connect failed (see sanlock.log). I have ran a "ps" command - no wdmd.... Any idea why wdmd_connect might fail ? 2013/12/19 Federico Simoncelli <fsimonce@redhat.com>
----- Original Message -----
From: "Pascal Jakobi" <pascal.jakobi@gmail.com> To: users@ovirt.org Sent: Wednesday, December 18, 2013 9:32:00 PM Subject: Re: [Users] AcquireHostId problem
sanlock.log :
2013-12-18 21:23:32+0100 1900 [867]: s1 lockspace
b2d69b22-a8b8-466c-bf1f-b6e565228238:250:/rhev/data-center/mnt/lab2.home:_home_vdsm_data/b2d69b22-a8b8-466c-bf1f-b6e565228238/dom_md/ids:0
2013-12-18 21:23:52+0100 1920 [4238]: s1 wdmd_connect failed -111 2013-12-18 21:23:52+0100 1920 [4238]: s1 create_watchdog failed -1 2013-12-18 21:23:53+0100 1921 [867]: s1 add_lockspace fail result -203
Hi Pascal, is wdmd up and running?
# ps aux | grep wdmd root 1650 0.0 0.2 13552 3320 ? SLs 03:49 0:00 wdmd -G sanlock
-- Federico
-- *Pascal Jakobi* 116 rue de Stalingrad 93100 Montreuil, France *+33 6 87 47 58 19*Pascal.Jakobi@gmail.com

----- Original Message -----
From: "Pascal Jakobi" <pascal.jakobi@gmail.com> To: "Federico Simoncelli" <fsimonce@redhat.com> Cc: users@ovirt.org Sent: Thursday, December 19, 2013 4:05:07 PM Subject: Re: [Users] AcquireHostId problem
Federico
On may suspect wdmd isn't running as wdmd_connect failed (see sanlock.log). I have ran a "ps" command - no wdmd....
Any idea why wdmd_connect might fail ?
Are you using fedora? What is the version of sanlock? Do you see any information about wdmd in /var/log/messages? Is the wdmd service started? # service wdmd status # systemctl status wdmd.service Thanks, -- Federico

sanlock-2.8-1.fc19.x86_64 Dec 18 16:46:05 lab2 systemd-wdmd[818]: Starting wdmd: [ OK ] Dec 18 16:46:05 lab2 systemd[1]: Unit wdmd.service entered failed state. Dec 18 20:52:12 lab2 wdmd[887]: wdmd started S0 H1 G179 Dec 18 20:52:12 lab2 wdmd[887]: /dev/watchdog0 failed to set timeout Dec 18 20:52:12 lab2 wdmd[887]: /dev/watchdog0 disarmed Dec 18 20:52:12 lab2 wdmd[887]: /dev/watchdog failed to set timeout Dec 18 20:52:12 lab2 wdmd[887]: /dev/watchdog disarmed Dec 18 20:52:12 lab2 wdmd[887]: no watchdog device, load a watchdog driver Dec 18 20:52:12 lab2 systemd[1]: wdmd.service: main process exited, code=exited, status=255/n/a I do not really understand all this. Furthermore, I could not find reference to it in the documentation. Am I missing sthing ? What does "load a wathdog driver" means in concrete ? Thxs again for your help 2013/12/19 Federico Simoncelli <fsimonce@redhat.com>
----- Original Message -----
From: "Pascal Jakobi" <pascal.jakobi@gmail.com> To: "Federico Simoncelli" <fsimonce@redhat.com> Cc: users@ovirt.org Sent: Thursday, December 19, 2013 4:05:07 PM Subject: Re: [Users] AcquireHostId problem
Federico
On may suspect wdmd isn't running as wdmd_connect failed (see sanlock.log). I have ran a "ps" command - no wdmd....
Any idea why wdmd_connect might fail ?
Are you using fedora? What is the version of sanlock? Do you see any information about wdmd in /var/log/messages? Is the wdmd service started?
# service wdmd status # systemctl status wdmd.service
Thanks, -- Federico
-- *Pascal Jakobi* 116 rue de Stalingrad 93100 Montreuil, France *+33 6 87 47 58 19*Pascal.Jakobi@gmail.com

----- Original Message -----
From: "Pascal Jakobi" <pascal.jakobi@gmail.com> To: "Federico Simoncelli" <fsimonce@redhat.com> Cc: users@ovirt.org Sent: Thursday, December 19, 2013 5:17:59 PM Subject: Re: [Users] AcquireHostId problem
sanlock-2.8-1.fc19.x86_64
Dec 18 16:46:05 lab2 systemd-wdmd[818]: Starting wdmd: [ OK ] Dec 18 16:46:05 lab2 systemd[1]: Unit wdmd.service entered failed state. Dec 18 20:52:12 lab2 wdmd[887]: wdmd started S0 H1 G179 Dec 18 20:52:12 lab2 wdmd[887]: /dev/watchdog0 failed to set timeout Dec 18 20:52:12 lab2 wdmd[887]: /dev/watchdog0 disarmed Dec 18 20:52:12 lab2 wdmd[887]: /dev/watchdog failed to set timeout Dec 18 20:52:12 lab2 wdmd[887]: /dev/watchdog disarmed Dec 18 20:52:12 lab2 wdmd[887]: no watchdog device, load a watchdog driver Dec 18 20:52:12 lab2 systemd[1]: wdmd.service: main process exited, code=exited, status=255/n/a
I do not really understand all this. Furthermore, I could not find reference to it in the documentation. Am I missing sthing ? What does "load a wathdog driver" means in concrete ?
It seems that you have a watchdog device (probably provided by the motherboard of your pc/server) that is not currently usable by wdmd. The temporary workaround to get your system to work is to modprobe the softdog kernel module and restart wdmd: # modprobe softdog # service wdmd restart Depending on the distribution you're using there are different ways to automatically load the module at boot, for example on fedora you can use: http://www.freedesktop.org/software/systemd/man/modules-load.d.html and for centos/rhel 6 you can look at the content of: /etc/sysconfig/modules for an example of how to do the same (e.g. kvm.modules). That said, in order to get your specific watchdog device to work properly I'd need some additional information. Can you provide me the output of these two commands? # lspci -vvv # lsmod Thanks, -- Federico

Here you go ! I am running F19 on a Lenovo S30. Thxs 2013/12/20 Federico Simoncelli <fsimonce@redhat.com>
----- Original Message -----
From: "Pascal Jakobi" <pascal.jakobi@gmail.com> To: "Federico Simoncelli" <fsimonce@redhat.com> Cc: users@ovirt.org Sent: Thursday, December 19, 2013 5:17:59 PM Subject: Re: [Users] AcquireHostId problem
sanlock-2.8-1.fc19.x86_64
Dec 18 16:46:05 lab2 systemd-wdmd[818]: Starting wdmd: [ OK ] Dec 18 16:46:05 lab2 systemd[1]: Unit wdmd.service entered failed state. Dec 18 20:52:12 lab2 wdmd[887]: wdmd started S0 H1 G179 Dec 18 20:52:12 lab2 wdmd[887]: /dev/watchdog0 failed to set timeout Dec 18 20:52:12 lab2 wdmd[887]: /dev/watchdog0 disarmed Dec 18 20:52:12 lab2 wdmd[887]: /dev/watchdog failed to set timeout Dec 18 20:52:12 lab2 wdmd[887]: /dev/watchdog disarmed Dec 18 20:52:12 lab2 wdmd[887]: no watchdog device, load a watchdog driver Dec 18 20:52:12 lab2 systemd[1]: wdmd.service: main process exited, code=exited, status=255/n/a
I do not really understand all this. Furthermore, I could not find reference to it in the documentation. Am I missing sthing ? What does "load a wathdog driver" means in concrete ?
It seems that you have a watchdog device (probably provided by the motherboard of your pc/server) that is not currently usable by wdmd.
The temporary workaround to get your system to work is to modprobe the softdog kernel module and restart wdmd:
# modprobe softdog # service wdmd restart
Depending on the distribution you're using there are different ways to automatically load the module at boot, for example on fedora you can use:
http://www.freedesktop.org/software/systemd/man/modules-load.d.html
and for centos/rhel 6 you can look at the content of:
/etc/sysconfig/modules
for an example of how to do the same (e.g. kvm.modules).
That said, in order to get your specific watchdog device to work properly I'd need some additional information. Can you provide me the output of these two commands?
# lspci -vvv # lsmod
Thanks, -- Federico
-- *Pascal Jakobi* 116 rue de Stalingrad 93100 Montreuil, France *+33 6 87 47 58 19*Pascal.Jakobi@gmail.com

----- Original Message -----
From: "Pascal Jakobi" <pascal.jakobi@gmail.com> To: "Federico Simoncelli" <fsimonce@redhat.com> Cc: users@ovirt.org, "David Teigland" <teigland@redhat.com> Sent: Friday, December 20, 2013 7:19:52 AM Subject: Re: [Users] AcquireHostId problem
Here you go ! I am running F19 on a Lenovo S30. Thxs
Thanks, can you open a bug on this issue? (Attach also the files to the bug). I suppose it will be later split into different ones, one for the failing watchdog device and maybe an RFE to wdmd to automatically load the softdog if there are no usable watchdog devices. -- Federico
participants (2)
-
Federico Simoncelli
-
Pascal Jakobi