Den 5 okt. 2018 18:31 skrev Ben Webber <ben.webber@egsgroup.com>:


On 05/10/2018 17:02, Karli Sjöberg wrote:


Den 5 okt. 2018 15:37 skrev Ben Webber <ben.webber@egsgroup.com>:
Hi Miguel,

Thanks for getting back to me so quickly! The pastebin is here:

https://pastebin.com/xNJWiymw

Yes, bond1 and bond2 are 802.3ad bonds and bond0 is an active-backup bond of bond1 and bond2

Ooh, that is really cool, I had no idea that was even possible! I've gotta ask though: why? Is it to have switch redundancy without having to shell out the extra price of having the switches stackable?

/K
Yes, that is exactly why, and that the servers I have contain 2 x 4 port gigabit nics in each of them. At work we've been consolidating some of our systems and I got a bit of old free kit which I'm aiming to use to set up a home lab based on ceph (with the new iscsi gateway) and ovirt. The swtiches I got are fairly basic and don't stack, but I wanted to achieve redundancy throughout so this is the reason for trying to do it with a bond of bonds.

Yeah, for your use-case that makes total sense. And kudos for playing with Ceph, we're thinking of going that route with iSCSI GW, but for VMWare instead.

For large deployments though, I think it's still going to be about as expensive as you'd need to double the amount of ports needed, don't you agree?

/K



Thanks

Ben

On 05/10/2018 14:09, Miguel Duarte de Mora Barroso wrote:
> On Thu, Oct 4, 2018 at 11:49 PM, Ben Webber wrote:
>> Hi,
>>
>> I'm trying to set up ovirt using the hosted-engine --deploy command on CentOS7, but am encountering an error. I am running a slightly unusual network configuration. I have two fairly basic non stacked gigabit switches with port channels connecting the two switches together. I have a lacp bond from the host consisting of 4 ports to each switch (bond1 and bond2). I have then created an active-backup bond (bond0) using the two lacp bonds as slaves in the hope to create ha at the switch layer using my basic switches. There is then a VLAN (101) on bond0.
>>
>> This network configuration runs fine on the host, however, when run, after a short while, the hosted-engine --deploy command outputs the following error:
>>
>> ...
>>
>> [ INFO ] TASK [Force host-deploy in offline mode]
>> [ INFO ] ok: [localhost]
>> [ INFO ] TASK [Add host]
>> [ INFO ] changed: [localhost]
>> [ INFO ] TASK [Wait for the host to be up]
>> [ INFO ] ok: [localhost]
>> [ INFO ] TASK [Check host status]
>> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "The host has been set in non_operational status, please check engine logs, fix accordingly and re-deploy.\n"}
>>
>> ...
>>
>>
>> Looking in /var/log/ovirt-engine/engine.log on the machine created, I can see the following errors logged:
>>
>> ...
>>
>> 2018-10-04 21:51:30,116+01 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.HostSetupNetworksVDSCommand] (EE-ManagedThreadFactory-engine-Thread-1) [59fb360a] START, HostSetupNetworksVDSCommand(HostName = ov1.test.local, HostSetupNetworksVdsCommandParameters:{hostId='7440c9b9-e530-4341-a317-d3a9041dc777', vds='Host[ov1.test.local,7440c9b9-e530-4341-a317-d3a9041dc777]', rollbackOnFailure='true', connectivityTimeout='120', networks='[HostNetwork:{defaultRoute='true', bonding='true', networkName='ovirtmgmt', vdsmName='ovirtmgmt', nicName='bond0', vlan='101', vmNetwork='true', stp='false', properties='null', ipv4BootProtocol='STATIC_IP', ipv4Address='192.168.1.11', ipv4Netmask='255.255.255.0', ipv4Gateway='192.168.1.1', ipv6BootProtocol='AUTOCONF', ipv6Address='null', ipv6Prefix='null', ipv6Gateway='null', nameServers='null'}]', removedNetworks='[]', bonds='[]', removedBonds='[]', clusterSwitchType='LEGACY'}), log id: 4f0c7eaa
>> 2018-10-04 21:51:30,121+01 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.HostSetupNetworksVDSCommand] (EE-ManagedThreadFactory-engine-Thread-1) [59fb360a] FINISH, HostSetupNetworksVDSCommand, log id: 4f0c7eaa
>> 2018-10-04 21:51:30,645+01 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.HostSetupNetworksVDSCommand] (EE-ManagedThreadFactory-engine-Thread-1) [59fb360a] Failed in 'HostSetupNetworksVDS' method
>> 2018-10-04 21:51:30,687+01 ERROR [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (EE-ManagedThreadFactory-engine-Thread-1) [59fb360a] EVENT_ID: VDS_BROKER_COMMAND_FAILURE(10,802), VDSM ov1.test.local command HostSetupNetworksVDS failed: Unknown nics in: ['bond1', 'bond2']
>> 2018-10-04 21:51:30,688+01 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.HostSetupNetworksVDSCommand] (EE-ManagedThreadFactory-engine-Thread-1) [59fb360a] Error: VDSGenericException: VDSErrorException: Failed to HostSetupNetworksVDS, error = Unknown nics in: ['bond1', 'bond2'], code = 23
>>
>> ...
>>
>>
>> It looks like when HostSetupNetworksVDS is run, it is checking that the slave interfaces to the bonds are physical network devices and being as the slaves of bond0 are bond1 and bond2, rather than physical devices, it then throws the error Unknown nics in: ['bond1', 'bond2'].
>>
>> Is there anything I can do or any configuration that I can put anywhere to make it work with this "stacked bond" configuration or does ovirt just not work when bonds are set up like this?
> Forwarding to Simone, who is an ovirt-hosted-engine-setup expert.
>
> Please get us a pastebin with the output of 'ansible-playbook -vvv -i
> localhost, /usr/share/ovirt-hosted-engine-setup/ansible/get_network_interfaces.yml'
> on your engine node.
>
> One thing I want to make sure: your bond1 and bond2 configurations are
> IEEE 802.3ad bonds, please confirm.
>
>> Thanks in advance,
>>
>> Ben
>>
>>
>>
>>
>> _______________________________________________
>> Users mailing list -- users@ovirt.org
>> To unsubscribe send an email to users-leave@ovirt.org
>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>> oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
>> List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/XPHQTPUINKZBSZVUDP2G66UPA5OJL3J7/
> _______________________________________________
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-leave@ovirt.org
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
> List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/RG2SGSFO6XILGKPZH4RLGGEK66NDHPWF/
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/6Q25LC74AYHMCREW7CNWZ4JNCADQD2LN/