[Users] Testing High Availability and Power outages

Doron Fediuck dfediuck at redhat.com
Sun Jan 13 03:54:19 EST 2013


----- Original Message -----

> From: "Alexandru Vladulescu" <avladulescu at bfproject.ro>
> To: "Doron Fediuck" <dfediuck at redhat.com>
> Cc: "users" <users at ovirt.org>
> Sent: Sunday, January 13, 2013 10:46:41 AM
> Subject: Re: [Users] Testing High Availability and Power outages

> Dear Doron,

> I haven't collected the logs from the tests, but I would gladly re-do
> the case and get back to you asap.

> This feature is the main reason of which I have chosen to go with
> Ovirt in the first place, besides other virt environments.

> Could you please inform me what logs should I be focusing on, besides
> the engine log; vdsm maybe or other relevant logs?

> Regards,
> Alex

> --
> Sent from phone.

> On 13.01.2013, at 09:56, Doron Fediuck < dfediuck at redhat.com > wrote:

> > ----- Original Message -----
> 

> > > From: "Alexandru Vladulescu" < avladulescu at bfproject.ro >
> > 
> 
> > > To: "users" < users at ovirt.org >
> > 
> 
> > > Sent: Friday, January 11, 2013 2:47:38 PM
> > 
> 
> > > Subject: [Users] Testing High Availability and Power outages
> > 
> 

> > > Hi,
> > 
> 

> > > Today, I started testing on my Ovirt 3.1 installation (from
> > > dreyou
> > > repos) running on 3 x Centos 6.3 hypervisors the High
> > > Availability
> > > features and the fence mechanism.
> > 
> 

> > > As yesterday, I have reported in a previous email thread, that
> > > the
> > > migration priority queue cannot be increased (bug) in this
> > > current
> > > version, I decided to test what the official documentation says
> > > about the High Availability cases.
> > 
> 

> > > This will be a disaster case scenarios to suffer from if one
> > > hypervisor has a power outage/hardware problem and the VMs
> > > running
> > > on it are not migrating on other spare resources.
> > 
> 

> > > In the official documenation from ovirt.org it is quoted the
> > > following:
> > 
> 
> > > High availability
> > 
> 

> > > Allows critical VMs to be restarted on another host in the event
> > > of
> > > hardware failure with three levels of priority, taking into
> > > account
> > > resiliency policy.
> > 
> 

> > > * Resiliency policy to control high availability VMs at the
> > > cluster
> > > level.
> > 
> 
> > > * Supports application-level high availability with supported
> > > fencing
> > > agents.
> > 
> 

> > > As well as in the Architecture description:
> > 
> 

> > > High Availability - restart guest VMs from failed hosts
> > > automatically
> > > on other hosts
> > 
> 

> > > So the testing went like this -- One VM running a linux box,
> > > having
> > > the check box "High Available" and "Priority for Run/Migration
> > > queue:" set to Low. On Host we have the check box to "Any Host in
> > > Cluster", without "Allow VM migration only upon Admin specific
> > > request" checked.
> > 
> 

> > > My environment:
> > 
> 

> > > Configuration : 2 x Hypervisors (same cluster/hardware
> > > configuration)
> > > ; 1 x Hypervisor + acting as a NAS (NFS) server (different
> > > cluster/hardware configuration)
> > 
> 

> > > Actions: Went and cut-off the power from one of the hypervisors
> > > from
> > > the 2 node clusters, while the VM was running on. This would
> > > translate to a power outage.
> > 
> 

> > > Results: The hypervisor node that suffered from the outage is
> > > showing
> > > in Hosts tab as Non Responsive on Status, and the VM has a
> > > question
> > > mark and cannot be powered off or nothing (therefore it's stuck).
> > 
> 

> > > In the Log console in GUI, I get:
> > 
> 

> > > Host Hyper01 is non-responsive.
> > 
> 
> > > VM Web-Frontend01 was set to the Unknown status.
> > 
> 

> > > There is nothing I could I could do besides clicking on the
> > > Hyper01
> > > "Confirm Host as been rebooted", afterwards the VM starts on the
> > > Hyper02 with a cold reboot of the VM.
> > 
> 

> > > The Log console changes to:
> > 
> 

> > > Vm Web-Frontend01 was shut down due to Hyper01 host reboot or
> > > manual
> > > fence
> > 
> 
> > > All VMs' status on Non-Responsive Host Hyper01 were changed to
> > > 'Down'
> > > by admin at internal
> > 
> 
> > > Manual fencing for host Hyper01 was started.
> > 
> 
> > > VM Web-Frontend01 was restarted on Host Hyper02
> > 
> 

> > > I would like you approach on this problem, reading the
> > > documentation
> > > & features pages on the official website, I suppose that this
> > > would
> > > have been an automatically mechanism working on some sort of a
> > > vdsm
> > > & engine fencing action. Am I missing something regarding it ?
> > 
> 

> > > Thank you for your patience reading this.
> > 
> 

> > > Regards,
> > 
> 
> > > Alex.
> > 
> 

> > > _______________________________________________
> > 
> 
> > > Users mailing list
> > 
> 
> > > Users at ovirt.org
> > 
> 
> > > http://lists.ovirt.org/mailman/listinfo/users
> > 
> 

> > Hi Alex,
> 
> > Can you share with us the engine's log from the relevant time
> > period?
> 

> > Doron
> 

Hi Alex, 
engine log is the important one, as it will indicate on the decision making process. 
VDSM logs should be kept in case something is unclear, but I suggest we begin with 
engine.log. 
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20130113/72c7e3ac/attachment.html>


More information about the Users mailing list