[Users] [ovirt-test-day-2] Testing all-in-one feature on f19
Yedidyah Bar David
didi at redhat.com
Wed Feb 12 07:51:02 UTC 2014
----- Original Message -----
> From: "Karli Sjöberg" <Karli.Sjoberg at slu.se>
> To: iheim at redhat.com
> Cc: users at ovirt.org
> Sent: Wednesday, February 12, 2014 9:37:14 AM
> Subject: Re: [Users] [ovirt-test-day-2] Testing all-in-one feature on f19
>
> On Tue, 2014-02-11 at 18:40 +0200, Itamar Heim wrote:
> > On 02/11/2014 06:19 PM, Alexander Wels wrote:
> > > Same issue for me, I did a minimum fedora 19 install added the
> > > appropriate
> > > repositories. Completed almost successfully, host didn't come up with
> > > vdsm
> > > compatibility error message.
> > >
> > > The vdsm on the host is: vdsm-4.13.3-3.fc19.x86_64
> > >
> > > I tried the same 3.3 dc/cluster and the host came up immediately.
> > >
> > > Alexander
> > >
> > > On Tuesday, February 11, 2014 11:02:37 AM Moti Asayag wrote:
> > >> Hi,
> > >>
> > >> In the 3.4 ovirt-test-day-2 I've tested the all-in-one feature.
> > >> I installed the all-in-one setup on a vm.
> > >>
> > >> The installation ended almost successfully, except of the vdsm service,
> > >> where the the host didn't become operational due to lack of support
> > >> in clusterLevel >= 3.4:
> > >>
> > >> Feb 11 15:51:52 localhost vdsm root ERROR VIR_MIGRATE_ABORT_ON_ERROR not
> > >> found in libvirt, support for clusterLevel >= 3.4 is disabled. For
> > >> Fedora
> > >> 19 users, please consider upgrading libvirt from the virt-preview
> > >> repository
> > >>
> > >> Once I created a new 3.3 DC and configured a local storage, the local
> > >> host
> > >> become operational and I was able to create a vm and to run it.
> > >>
> > >> Thanks,
> > >> Moti
> > >> _______________________________________________
> > >> Users mailing list
> > >> Users at ovirt.org
> > >> http://lists.ovirt.org/mailman/listinfo/users
> > >
> > > _______________________________________________
> > > Users mailing list
> > > Users at ovirt.org
> > > http://lists.ovirt.org/mailman/listinfo/users
> > >
> >
> > no one tested on .el6?
> > _______________________________________________
> > Users mailing list
> > Users at ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/users
>
> Was stuck on watch duty first half of the day so was short on time and
> had to continue the rest of the testing today.
>
> It might be worth noting on the wiki page[1] that if you´re a tester
> (like we are) that does change the contents of the '(el6|
> fedora)'-ovirt.repo (like enabling "ovirt-3.4.0-prerelease" e.g.), that
> when you later update the ovirt-release-'(el6|fedora)' rpm, the repo
> file gets saved as an ".rpmnew" instead. Was just luck we saw that in
> yum and move-replaced our "el6-ovirt.repo" with the
> "el6-ovirt.repo.rpmnew" before continuing with the rest of the
> instructions.
>
> Perhaps something simple like:
> "Note: When done with your testing, please remember to change "enabled"
> back to "0" to avoid having duplicate repo files"
>
> As for the upgrade, it was actually done first from 3.3.3-RC to
> 3.3.3-GA, and then up to 3.4.0-beta2. Both of which went through without
> major issues. Minor issue going up to 3.4.0 was that "engine-setup"
> still changed the access rights to the predefined ISO domain in
> "/etc/exports" to "None" (empty python variable?),
Indeed. It's mentioned in "known issues". Will hopefully be fixed in a few days.
> but other than that,
> it went smoothly by.
>
> Tested making a live snapshot on a VM that failed, big surprise. Logs
> are attached.
>
> Will continue testing the new power-saving policy as soon as we can
> manage to find another Host to incorporate into the cluster.
>
> Oh, and we tried using the "ovirt-log-collector" to compile the logs but
> it failed when trying to gather hypervisor data, just after we typed in
> the password for "admin at internal". Even when executing with the option
> "--no-hypervisors" it still tried to do that?!
I think it should work. Can you send logs please?
Thanks!
--
Didi
More information about the Users
mailing list