[ovirt-users] Can HA Agent control NFS Mount?
Doron Fediuck
dfediuck at redhat.com
Mon May 26 02:38:10 EDT 2014
----- Original Message -----
> From: "Andrew Lau" <andrew at andrewklau.com>
> To: "Bob Doolittle" <bob at doolittle.us.com>
> Cc: "users" <users at ovirt.org>
> Sent: Monday, May 26, 2014 7:30:41 AM
> Subject: Re: [ovirt-users] Can HA Agent control NFS Mount?
>
> On Mon, May 26, 2014 at 5:10 AM, Bob Doolittle <bob at doolittle.us.com> wrote:
> >
> > On 05/25/2014 02:51 PM, Joop wrote:
> >>
> >> On 25-5-2014 19:38, Bob Doolittle wrote:
> >>>
> >>>
> >>> Also curious is that when I say "poweroff" it actually reboots and comes
> >>> up again. Could that be due to the timeouts on the way down?
> >>>
> >> Ah, that's something my F19 host does too. Some more info: if engine
> >> hasn't been started on the host then I can shutdown it and it will
> >> poweroff.
> >> IF engine has been run on it then it will reboot.
> >> Its not vdsm (I think) because my shutdown sequence is (on my f19 host):
> >> service ovirt-agent-ha stop
> >> service ovirt-agent-broker stop
> >> service vdsmd stop
> >> ssh root at engine01 "init 0"
> >> init 0
> >>
> >> I don't use maintenance mode because when I poweron my host (= my desktop)
> >> I want engine to power on automatically which it does most of the time
> >> within 10 min.
> >
> >
> > For comparison, I see this issue and I *do* use maintenance mode (because
> > presumably that's the 'blessed' way to shut things down and I'm scared to
> > mess this complex system up by straying off the beaten path ;). My process
> > is:
> >
> > ssh root at engine "init 0"
> > (wait for "vdsClient -s 0 list | grep Status:" to show the vm as down)
> > hosted-engine --set-maintenance --mode=global
> > poweroff
> >
> > And then on startup:
> > hosted-engine --set-maintenance --mode=none
> > hosted-engine --vm-start
> >
> > There are two issues here. I am not sure if they are related or not.
> > 1. The NFS timeout during shutdown (Joop do you see this also? Or just #2?)
> > 2. The system reboot instead of poweroff (which messes up remote machine
> > management)
> >
>
> For 1. I was wondering if perhaps, we could have an option to specify
> the mount options. If I understand correctly, applying a soft mount
> instead of a hard mount would prevent this from happening. I'm however
> not sure of the implications this would have on the data integrity..
>
> I would really like to see it happen in the ha-agent, as it's the one
> which connects/mounts the storage it should also unmount it on boot.
> However the stability on it, is flaky at best. I've noticed if `df`
> hangs because of another NFS mount having timed-out the agent will
> die. That's not a good sign.. this was what actually caused my
> hosted-engine to run twice in one case.
>
> > Thanks,
> > Bob
> >
> >
> >> I think wdmd or sanlock are causing the reboot instead of poweroff
> >>
> >> Joop
> >>
Great to have your feedback guys!
So just to clarify some of the issues you mentioned;
Hosted engine wasn't designed for a 'single node' use case, as we do
want it to be highly available. This is why it's being restarted
elsewhere or even on the same server if no better alternative.
Having said that, it is possible to set global maintenance mode
as a first step (in the UI: right click engine vm and choose ha-maintenance).
Then you can ssh into the engine vm and init 0.
After a short while, the qemu process should gracefully end and release
its sanlock lease as well as any other resource, which means you can
reboot your hypervisor peacefully.
Doron
More information about the Users
mailing list