[ovirt-devel] node zero Networking
Yedidyah Bar David
didi at redhat.com
Thu Dec 28 11:12:08 UTC 2017
On Mon, Dec 25, 2017 at 4:14 PM, Yedidyah Bar David <didi at redhat.com> wrote:
> Hi all,
>
> I spent quite some time trying to deploy node zero, trying to look at
> https://bugzilla.redhat.com/show_bug.cgi?id=1528253 , and always fail
> around the end, with:
>
> ERROR fatal: [localhost]: FAILED! => {"attempts": 50, "changed": true,
> "cmd": "ip rule list | grep ovirtmgmt | sed s/\\\\[.*\\\\]\\ //g | awk
> '{ print $9 }'", "delta": "0:00:00.008292", "end": "2017-12-25
> 11:51:39.146800", "rc": 0, "start": "2017-12-25 11:51:39.138508",
> "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []}
> ERROR Failed to execute stage 'Closing up': Failed executing ansible-playbook
> ERROR Hosted Engine deployment failed: this system is not reliable,
> please check the issue,fix and redeploy
>
> I use the following setup:
>
> I have a libvirt vm on my laptop, with a single virtual nic eth0.
>
> This nic is connected to a bridge called intbr on my laptop. This
> bridge has no access to outside, and VMs on it have no default route.
> There is a local dhcp+dns serving this bridge, using the address range
> 192.168.3.0/24.
>
> The vm serves as a nested-kvm hosted-engine host.
>
> eth0 gets a static IP address 192.168.3.42 from dhcpd.
>
> There is also (didn't check who/what exactly creates it, I think it's
> libvirt's default) a bridge there called virbr0. virbr0 has IP address
> 192.168.122.1/24 .
>
> When I deploy HE, the engine machine gets also a single virtual nic,
> which is connected to virbr0, and gets an IP address in that range
> (192.168.122.85, currently).
>
> deploy fails when running the task:
>
> - name: Get ovirtmgmt route table id
> shell: ip rule list | grep ovirtmgmt | sed s/\\[.*\\]\ //g | awk
> '{ print $9 }'
> register: ovirtmgmt_table_id
> until: ovirtmgmt_table_id.stdout_lines|length >= 1
> retries: 50
> delay: 10
> changed_when: True
>
> The output of 'ip rule list' is:
>
> 0: from all lookup local
> 32766: from all lookup main
> 32767: from all lookup default
>
> So does not include 'ovirtmgmt'.
>
> I do have:
>
> # brctl show
> bridge name bridge id STP enabled interfaces
> ;vdsmdummy; 8000.000000000000 no
> ovirtmgmt 8000.06d1bd012412 no eth0
> virbr0 8000.525400012499 yes virbr0-nic
> vnet0
> And:
>
> # ip a
> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN qlen 1
> link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
> inet 127.0.0.1/8 scope host lo
> valid_lft forever preferred_lft forever
> inet6 ::1/128 scope host
> valid_lft forever preferred_lft forever
> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
> master ovirtmgmt state UP qlen 1000
> link/ether 06:d1:bd:01:24:12 brd ff:ff:ff:ff:ff:ff
> 18: virbr0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
> state UP qlen 1000
> link/ether 52:54:00:01:24:99 brd ff:ff:ff:ff:ff:ff
> inet 192.168.122.1/24 brd 192.168.122.255 scope global virbr0
> valid_lft forever preferred_lft forever
> 19: virbr0-nic: <BROADCAST,MULTICAST> mtu 1500 qdisc pfifo_fast master
> virbr0 state DOWN qlen 1000
> link/ether 52:54:00:01:24:99 brd ff:ff:ff:ff:ff:ff
> 20: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
> master virbr0 state UNKNOWN qlen 1000
> link/ether fe:d1:bd:01:24:04 brd ff:ff:ff:ff:ff:ff
> inet6 fe80::fcd1:bdff:fe01:2404/64 scope link
> valid_lft forever preferred_lft forever
> 21: ;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
> link/ether 1e:1b:84:c2:51:ff brd ff:ff:ff:ff:ff:ff
> 22: ovirtmgmt: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
> noqueue state UP qlen 1000
> link/ether 06:d1:bd:01:24:12 brd ff:ff:ff:ff:ff:ff
> inet 192.168.3.42/24 brd 192.168.3.255 scope global dynamic ovirtmgmt
> valid_lft 70927sec preferred_lft 70927sec
> inet6 fe80::4d1:bdff:fe01:2412/64 scope link
> valid_lft forever preferred_lft forever
>
> (And of course told deploy that I want to use eth0).
>
> Questions:
>
> 1. Did this already work for anyone at all? If so, can you please
> share details? Specifically, how was networking configured?
>
> 2. It might be that my problems are due to not having a (default)
> route for ovirtmgmt bridge/network. If so, then I consider this a bug,
> but do not mind configuring one for now.
For the record, adding a default router indeed solved this issue.
>
> 3. All of the relevant section of the playbook has a comment preceding it:
>
> # all of the next is a workaroud for the network issue, vdsm
> installation breaks the routing and it needs to be fixed
> # once we'll fix the host installation it could be removed
>
> Do we have specific details/bug/whatever about the problem we are
> working around? Perhaps it's already solved and I can try to remove
> this part?
>
> 4. Both now (with (3.) being worked around) and eventually (when it's
> (what?) fixed), how should this work? Should the engine local vm
> indeed start connected to virbr0, and then move to ovirtmgmt? Or only
> the new engine vm (residing on the shared storage) should be in
> ovirtmgmt?
>
> 5. In particular, what should I supply for the "engine fqdn", and what
> should it resolve to, both in the beginning an eventually?
>
> Thanks,
> --
> Didi
--
Didi
More information about the Devel
mailing list