
If we need that extra server in Production DC (for hosted engine redundancy and to allow maintenance) then lets take the lower end new servers from 17-26 and replace it with the strong one. We need to utilize our servers, I don't think we're at 50% utilization even, looking at the memory consumption last time i checked when all slaves were working
the problem is that we won't have live migration in the Production cluster if we add the new hosts there(because of the multi node NUMA mismatch), we could put 2 new hosts in the Production cluster but then we would have to schedule a downtime in order to move them. I think that coupling the engine upgrade with the new slaves is not necessary. we can start by installing the hook on ovirt-srv11 and spinning up new slaves, that way we can also test the hook in production. because there is no NFS involved, and live migration isn't working with the hook, the host is pretty much self-contained and this can't harm anything. On Thu, May 12, 2016 at 4:59 PM, Anton Marchukov <amarchuk@redhat.com> wrote:
I was talking about using the hooks before installing the ssds, but if
that can be done reliably also before the upgrade it's also a solution that will help scratch our current itches sooner.
That's about it. The hook in 3.6 contains the patch I submitted. It does not work without it at all. Although you can use rpm from 3.5 in 3.6.
^ well, the machines being slaves I think there's no problem having to stop them to migrate them, we have already experience with both fabric and jenkins so I guess it should not be hard to automate with those tools ;)
Yes there is no problem with that. But I am not aware about "stop all slaves on the host" feature in oVirt so that would be either manually clicking or we need to fabricate it. Not a big deal too.
-- Anton Marchukov Senior Software Engineer - RHEV CI - Red Hat
_______________________________________________ Infra mailing list Infra@ovirt.org http://lists.ovirt.org/mailman/listinfo/infra