
--_bf9610ce-cb68-4d08-a5d9-84340332e8a6_ Content-Type: text/plain; charset="iso-8859-1" Content-Transfer-Encoding: quoted-printable Hi all=2C
Date: Mon=2C 10 Mar 2014 12:56:19 -0400 From: jbrooks@redhat.com To: msivak@redhat.com CC: users@ovirt.org Subject: Re: [Users] hosted engine help =20 =20 =20 ----- Original Message -----
From: "Martin Sivak" <msivak@redhat.com> To: "Dan Kenigsberg" <danken@redhat.com> Cc: users@ovirt.org Sent: Saturday=2C March 8=2C 2014 11:52:59 PM Subject: Re: [Users] hosted engine help =20 Hi Jason=2C =20 can you please attach the full logs? We had very similar issue before I= we need to see if is the same or not. =20 I may have to recreate it -- I switched back to an all in one engine afte= r my setup started refusing to run the engine at all. It's no fun losing your = engine! =20 This was a migrated-from-standalone setup=2C maybe that caused additional= wrinkles... =20 Jason =20 =20 Thanks
I experienced the exact same symptoms as Jason on a from-scratch installati= on on two physical nodes with CentOS 6.5 (fully up-to-date) using oVirt 3.4= .0_pre (latest test-day release) and GlusterFS 3.5.0beta3 (with Gluster-pro= vided NFS as storage for the self-hosted engine VM only). I roughly followed the guide from Andrew Lau: http://www.andrewklau.com/ovirt-hosted-engine-with-3-4-0-nightly/ with some variations due to newer packages (resolved bugs) and different ha= rdware setup (no VLANs in my setup: physically separated networks=3B custom= second nic added to Engine VM template before deploying etc.) The self-hosted installation on first node + Engine VM (configured for mana= ging both oVirt and the storage=3B Datacenter default set to NFS because no= GlusterFS offered) went apparently smooth=2C but the HA-agent failed to st= art at the very end (same errors in logs as Jason: the storage domain seems= "missing") and I was only able to start it all manually with: hosted-engine --connect-storagehosted-engine --start-poolhosted-engine --vm= -start then the Engine came up and I could use it=2C I even registered the second = node (same final error in HA-agent) and tried to add GlusterFS storage doma= ins for further VMs and ISOs (by the way: the original NFS-GlusterFS domain= for Engine VM only is not present inside the Engine web UI) but it always = failed activating the domains (they remain "Inactive"). Furthermore the engine gets killed some time after starting (from 3 up to 1= 1 hours later) and the only way to get it back is repeating the above comma= nds. I always managed GlusterFS "natively" (not through oVirt) from the commandl= ine and verified that the NFS-exported Engine-VM-only volume gets replicate= d=2C but I obviously failed to try migration because the HA part results in= active and oVirt refuse to migrate the Engine. Since I tried many times=2C with variations and further manual actions betw= een (like trying to manually mount the NFS Engine domain=2C restarting the = HA-agent only etc.)=2C my logs are "cluttered"=2C so I should start from sc= ratch again and pack up all logs in one swipe. Tell me what I should capture and at which points in the whole process and = I will try to follow up as soon as possible. Many thanks=2CGiuseppe
-- Martin Siv=E1k msivak@redhat.com Red Hat Czech RHEV-M SLA / Brno=2C CZ =20 ----- Original Message -----
On Fri=2C Mar 07=2C 2014 at 10:17:43AM +0100=2C Sandro Bonazzola wrot= e:
Il 07/03/2014 01:10=2C Jason Brooks ha scritto:
Hey everyone=2C I've been testing out oVirt 3.4 w/ hosted engine= =2C and while I've managed to bring the engine up=2C I've only been able = to do it manually=2C using "hosted-engine --vm-start". =20 The ovirt-ha-agent service fails reliably for me=2C erroring out = with "RequestError: Request failed: success." =20 I've pasted error passages from the ha agent and vdsm logs below. =20 Any pointers? =20 looks like a VDSM bug=2C Dan? =20 Why? The exception is raised from deep inside the ovirt_hosted_engine= _ha code.
Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users =20
Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users =20
Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users =
http://www.andrewklau.com/ovirt-hosted-engine-with-3-4-0-nightly/</div><di= v><br></div><div>with some variations due to newer packages (resolved bugs)= and different hardware setup (no VLANs in my setup: physically separated n= etworks=3B custom second nic added to Engine VM template before deploying e= tc.)</div><div><br></div><div>The self-hosted installation on first node + = Engine VM (configured for managing both oVirt and the storage=3B Datacenter= default set to NFS because no GlusterFS offered) went apparently smooth=2C= but the HA-agent failed to start at the very end (same errors in logs as J= ason: the storage domain seems "missing") and I was only able to start it a= ll manually with:</div><div><br></div><div><div>hosted-engine --connect-sto= rage</div><div>hosted-engine --start-pool</div></div><div>hosted-engine --v= m-start</div><div><br></div><div>then the Engine came up and I could use it= =2C I even registered the second node (same final error in HA-agent) and tr= ied to add GlusterFS storage domains for further VMs and ISOs (by the way: =
--_bf9610ce-cb68-4d08-a5d9-84340332e8a6_ Content-Type: text/html; charset="iso-8859-1" Content-Transfer-Encoding: quoted-printable <html> <head> <style><!-- .hmmessage P { margin:0px=3B padding:0px } body.hmmessage { font-size: 12pt=3B font-family:Calibri } --></style></head> <body class=3D'hmmessage'><div dir=3D'ltr'><div>Hi all=2C</div><div><br></d= iv><div>>=3B Date: Mon=2C 10 Mar 2014 12:56:19 -0400<br>>=3B From: jbro= oks@redhat.com<br>>=3B To: msivak@redhat.com<br>>=3B CC: users@ovirt.or= g<br>>=3B Subject: Re: [Users] hosted engine help<br>>=3B <br>>=3B <b= r>>=3B <br>>=3B ----- Original Message -----<br>>=3B >=3B From: "Ma= rtin Sivak" <=3Bmsivak@redhat.com>=3B<br>>=3B >=3B To: "Dan Kenigsb= erg" <=3Bdanken@redhat.com>=3B<br>>=3B >=3B Cc: users@ovirt.org<br>= >=3B >=3B Sent: Saturday=2C March 8=2C 2014 11:52:59 PM<br>>=3B >= =3B Subject: Re: [Users] hosted engine help<br>>=3B >=3B <br>>=3B >= =3B Hi Jason=2C<br>>=3B >=3B <br>>=3B >=3B can you please attach th= e full logs? We had very similar issue before I we<br>>=3B >=3B need to= see if is the same or not.<br>>=3B <br>>=3B I may have to recreate it = -- I switched back to an all in one engine after my<br>>=3B setup started= refusing to run the engine at all. It's no fun losing your engine!<br>>= =3B <br>>=3B This was a migrated-from-standalone setup=2C maybe that caus= ed additional wrinkles...<br>>=3B <br>>=3B Jason<br>>=3B <br>>=3B &= gt=3B <br>>=3B >=3B Thanks<br><br></div><div>I experienced the exact sa= me symptoms as Jason on a from-scratch installation on two physical nodes w= ith CentOS 6.5 (fully up-to-date) using oVirt 3.4.0_pre (latest test-day re= lease) and GlusterFS 3.5.0beta3 (with Gluster-provided NFS as =3B<span = style=3D"font-size: 12pt=3B">storage =3B</span><span style=3D"font-size= : 12pt=3B">for the self-hosted engine VM only).</span></div><div><br></div>= <div>I roughly followed the guide from Andrew Lau:</div><div><br></div><div= the original NFS-GlusterFS domain for Engine VM only is not present inside = the Engine web UI) but it always failed activating the domains (they remain= "Inactive").</div><div><br></div><div>Furthermore the engine gets killed s= ome time after starting (from 3 up to 11 hours later) and the only way to g= et it back is repeating the above commands.</div><div><br></div><div>I alwa= ys managed GlusterFS "natively" (not through oVirt) from the commandline an= d verified that the NFS-exported Engine-VM-only volume gets replicated=2C b= ut I obviously failed to try migration because the HA part results inactive= and oVirt refuse to migrate the Engine.</div><div><br></div><div>Since I t= ried many times=2C with variations and further manual actions between (like= trying to manually mount the NFS Engine domain=2C restarting the HA-agent = only etc.)=2C my logs are "cluttered"=2C so I should start from scratch aga= in and pack up all logs in one swipe.</div><div><br></div><div>Tell me what= I should capture and at which points in the whole process and I will try t= o follow up as soon as possible.</div><div><br></div><div>Many thanks=2C</d= iv><div>Giuseppe</div><div><br></div><div>>=3B >=3B --<br>>=3B >=3B= Martin Siv=E1k<br>>=3B >=3B msivak@redhat.com<br>>=3B >=3B Red Hat= Czech<br>>=3B >=3B RHEV-M SLA / Brno=2C CZ<br>>=3B >=3B <br>>=3B= >=3B ----- Original Message -----<br>>=3B >=3B >=3B On Fri=2C Mar = 07=2C 2014 at 10:17:43AM +0100=2C Sandro Bonazzola wrote:<br>>=3B >=3B = >=3B >=3B Il 07/03/2014 01:10=2C Jason Brooks ha scritto:<br>>=3B >= =3B >=3B >=3B >=3B Hey everyone=2C I've been testing out oVirt 3.4 w/= hosted engine=2C and<br>>=3B >=3B >=3B >=3B >=3B while I've mana= ged to bring the engine up=2C I've only been able to do it<br>>=3B >=3B= >=3B >=3B >=3B manually=2C using "hosted-engine --vm-start".<br>>= =3B >=3B >=3B >=3B >=3B <br>>=3B >=3B >=3B >=3B >=3B The = ovirt-ha-agent service fails reliably for me=2C erroring out with<br>>=3B= >=3B >=3B >=3B >=3B "RequestError: Request failed: success."<br>&g= t=3B >=3B >=3B >=3B >=3B <br>>=3B >=3B >=3B >=3B >=3B I'v= e pasted error passages from the ha agent and vdsm logs below.<br>>=3B &g= t=3B >=3B >=3B >=3B <br>>=3B >=3B >=3B >=3B >=3B Any pointe= rs?<br>>=3B >=3B >=3B >=3B <br>>=3B >=3B >=3B >=3B looks li= ke a VDSM bug=2C Dan?<br>>=3B >=3B >=3B <br>>=3B >=3B >=3B Why?= The exception is raised from deep inside the ovirt_hosted_engine_ha<br>>= =3B >=3B >=3B code.<br>>=3B >=3B >=3B ___________________________= ____________________<br>>=3B >=3B >=3B Users mailing list<br>>=3B &= gt=3B >=3B Users@ovirt.org<br>>=3B >=3B >=3B http://lists.ovirt.org= /mailman/listinfo/users<br>>=3B >=3B >=3B <br>>=3B >=3B _________= ______________________________________<br>>=3B >=3B Users mailing list<= br>>=3B >=3B Users@ovirt.org<br>>=3B >=3B http://lists.ovirt.org/ma= ilman/listinfo/users<br>>=3B >=3B <br>>=3B __________________________= _____________________<br>>=3B Users mailing list<br>>=3B Users@ovirt.or= g<br>>=3B http://lists.ovirt.org/mailman/listinfo/users<br></div> = </div></body> </html>= --_bf9610ce-cb68-4d08-a5d9-84340332e8a6_--