Dear Simone,
It doesn't seem to want to work:
# Settings
he_fqdn: "he.virt.in.bmrc.ox.ac.uk"
he_ansible_host_name: "virthyp04.virt.in.bmrc.ox.ac.uk"
he_admin_password: <snip>
he_appliance_password: <snip>
# Resources
he_mem_size_MB: "4096"
# Storage
he_domain_type: "nfs"
he_storage_domain_addr: <snip>
he_storage_domain_path: <snip>
# Network
he_vm_ip_addr: "10.141.31.240"
he_vm_ip_prefix: "20"
he_dns_addr:
["10.141.31.251","10.141.31.252","10.141.31.253"]
he_default_gateway_4: "10.141.31.254"
he_gateway: he_default_gateway_4
he_force_ip4: true
he_bridge_if: bond0.910
#he_just_collect_network_interfaces: true
# Email
he_smtp_port: 25
he_smtp_server: smtp.ox.ac.uk<http://smtp.ox.ac.uk>
he_dest_email: rescomp-ops@well.ox.ac.uk<mailto:rescomp-ops@well.ox.ac.uk>
he_source_email: ovirt@bmrc.ox.ac.uk<mailto:ovirt@bmrc.ox.ac.uk>
# Ansible Stuff
ansible_ssh_user: root
ansible_become: false
host_key_checking: false
I've attached the output of the ansible command as a log file, this is what happens
when the IF bond0.910 is assigned the IP and `ovirtmgmt` is not defined on the host.
Regards,
Callum
--
Callum Smith
Research Computing Core
Wellcome Trust Centre for Human Genetics
University of Oxford
e. callum@well.ox.ac.uk<mailto:callum@well.ox.ac.uk>
On 28 Mar 2019, at 16:23, Simone Tiraboschi
<stirabos@redhat.com<mailto:stirabos@redhat.com>> wrote:
On Thu, Mar 28, 2019 at 1:44 PM Callum Smith
<callum@well.ox.ac.uk<mailto:callum@well.ox.ac.uk>> wrote:
Dear Simone,
This is my experience too, but I'm now hitting this error on the hosted-engine install
at the part where it registers the hypervisor as the first host in the engine:
2019-03-28 12:40:50,025Z INFO [org.ovirt.engine.core.bll.host.HostConnectivityChecker]
(EE-ManagedThreadFactory-engine-Thread-1) [49f371c1] Engine managed to communicate with
VDSM
agent on host
'virthyp04.virt.in.bmrc.ox.ac.uk<http://virthyp04.virt.in.bmrc.ox.ac.uk/>'
with address
'virthyp04.virt.in.bmrc.ox.ac.uk<http://virthyp04.virt.in.bmrc.ox.ac.uk/>'
('db571f8a-fc85-40d3-b86f-c0038e3cd7e7')
2019-03-28 12:40:53,111Z WARN [org.ovirt.engine.core.bll.network.NetworkConfigurator]
(EE-ManagedThreadFactory-engine-Thread-1) [49f371c1] Failed to find a valid interface for
the
management network of host
virthyp04.virt.in.bmrc.ox.ac.uk<http://virthyp04.virt.in.bmrc.ox.ac.uk/>. If the
interface ovirtmgmt is a bridge, it should be torn-down manually.
2019-03-28 12:40:53,111Z ERROR
[org.ovirt.engine.core.bll.hostdeploy.InstallVdsInternalCommand]
(EE-ManagedThreadFactory-engine-Thread-1) [49f371c1] Exception: org.ovirt.engine.cor
e.bll.network.NetworkConfigurator$NetworkConfiguratorException: Interface ovirtmgmt is
invalid for management network
The host's ovirtmgmt network connection is a statically assigned IP on a VLAN on a
bond, how should I be configuring this if not manually?
If you need to deploy over vlan 123 over bond0 simply configure a device exactly called
bond0.123 and statically configure your IP address there.
Choose it for hosted-engine deployment, nothing more: ovirtmgmt will be automatically
created over that device and the vlan ID will be set at engine level for the whole
management network.
Regards,
Callum
--
Callum Smith
Research Computing Core
Wellcome Trust Centre for Human Genetics
University of Oxford
e. callum@well.ox.ac.uk<mailto:callum@well.ox.ac.uk>
On 27 Mar 2019, at 17:09, Simone Tiraboschi
<stirabos@redhat.com<mailto:stirabos@redhat.com>> wrote:
On Wed, Mar 27, 2019 at 4:27 PM Callum Smith
<callum@well.ox.ac.uk<mailto:callum@well.ox.ac.uk>> wrote:
It's ok, migrating to 4.3.2 on the oVirt node (from 4.3.0) did the job of fixing it.
It is a bug if you intend on using the ovirtmgmt network to deploy your ansible from
This is a bit tricky: when the engine brings up the host it also creates the management
bridge and this could lead to a temporary network down on the selected interface for the
bridge creation time (a couple of seconds?)
I tried it on a LAN and ansible ssh connection always survived but I 'm not sure
it's always true.
, and you need it to have an IP address already on that range! But - it works as expected
with the ovirtmgmt bridge setup so nothing to worry about.
Regards,
Callum
--
Callum Smith
Research Computing Core
Wellcome Trust Centre for Human Genetics
University of Oxford
e. callum@well.ox.ac.uk<mailto:callum@well.ox.ac.uk>
On 27 Mar 2019, at 14:57, Simone Tiraboschi
<stirabos@redhat.com<mailto:stirabos@redhat.com>> wrote:
On Wed, Mar 27, 2019 at 3:24 PM Callum Smith
<callum@well.ox.ac.uk<mailto:callum@well.ox.ac.uk>> wrote:
Dear All,
We're trying to deploy our hosted engine remotely using the ansible hosted engine
playbook, which has been a rocky road but we're now at the point where it's
installing, and failing. We've got a pre-defined bond/VLAN setup for our interface
which has the correct bond0 bond0.123 and ovirtmgmt bridge on top but we're hitting
the classic error:
Failed to find a valid in
terface for the management network of host
virthyp04.virt.in.bmrc.ox.ac.uk<http://virthyp04.virt.in.bmrc.ox.ac.uk/>. If the
interface ovirtmgmt is a bridge, it should be torn-down manually.
Does this bug still exist in the latest (4.3) version, and is installing using ansible
with this network configuration impossible?
I don't think it's a bug; please avoid manually creating ovirtmgmt and simply
set:
he_bridge_if: "bond0.123", in ansible variable file and the management bridge
will be created for you at host-deploy time.
Regards,
Callum
--
Callum Smith
Research Computing Core
Wellcome Trust Centre for Human Genetics
University of Oxford
e. callum@well.ox.ac.uk<mailto:callum@well.ox.ac.uk>
_______________________________________________
Users mailing list -- users@ovirt.org<mailto:users@ovirt.org>
To unsubscribe send an email to users-leave@ovirt.org<mailto:users-leave@ovirt.org>
Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/SBOZ6FRBRQK...