[ovirt-users] How to setup host network to link ovirtmgmt with eth0 automatically?

Kai Kang kai.7.kang at gmail.com
Tue May 24 07:57:45 UTC 2016


And network configurations on node:

[root at ovirt-node] # brctl show
bridge name     bridge id               STP enabled     interfaces
docker0         8000.0242ae1de711       no
ovirtmgmt               8000.001320ff73aa       no              eth0

[root at ovirt-node] # ifconfig -a
docker0   Link encap:Ethernet  HWaddr 02:42:ae:1d:e7:11
          inet addr:172.17.0.1  Bcast:0.0.0.0  Mask:255.255.0.0
          UP BROADCAST MULTICAST  MTU:1500  Metric:1
          RX packets:0 errors:0 dropped:0 overruns:0 frame:0
          TX packets:0 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:0
          RX bytes:0 (0.0 B)  TX bytes:0 (0.0 B)

eth0      Link encap:Ethernet  HWaddr 00:13:20:ff:73:aa
          UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
          RX packets:6331 errors:0 dropped:0 overruns:0 frame:0
          TX packets:2866 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:1000
          RX bytes:762182 (744.3 KiB)  TX bytes:210611 (205.6 KiB)
          Interrupt:20 Memory:d1100000-d1120000

lo        Link encap:Local Loopback
          inet addr:127.0.0.1  Mask:255.0.0.0
          inet6 addr: ::1/128 Scope:Host
          UP LOOPBACK RUNNING  MTU:65536  Metric:1
          RX packets:36 errors:0 dropped:0 overruns:0 frame:0
          TX packets:36 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:0
          RX bytes:2478 (2.4 KiB)  TX bytes:2478 (2.4 KiB)

ovirtmgmt Link encap:Ethernet  HWaddr 00:13:20:ff:73:aa
          inet addr:128.224.165.170  Bcast:128.224.165.255
 Mask:255.255.255.0
          inet6 addr: fe80::213:20ff:feff:73aa/64 Scope:Link
          inet6 addr: 11:2233:4455:6677:213:20ff:feff:73aa/64 Scope:Global
          UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
          RX packets:6295 errors:0 dropped:6 overruns:0 frame:0
          TX packets:2831 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:0
          RX bytes:644421 (629.3 KiB)  TX bytes:177616 (173.4 KiB)

sit0      Link encap:UNSPEC  HWaddr
00-00-00-00-32-33-33-00-00-00-00-00-00-00-00-00
          NOARP  MTU:1480  Metric:1
          RX packets:0 errors:0 dropped:0 overruns:0 frame:0
          TX packets:0 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:0
          RX bytes:0 (0.0 B)  TX bytes:0 (0.0 B)

wlan0     Link encap:Ethernet  HWaddr 80:86:f2:8b:1d:cf
          BROADCAST MULTICAST  MTU:1500  Metric:1
          RX packets:0 errors:0 dropped:0 overruns:0 frame:0
          TX packets:0 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:1000
          RX bytes:0 (0.0 B)  TX bytes:0 (0.0 B)


Thanks,
Kai



On Tue, May 24, 2016 at 3:36 PM, Kai Kang <kai.7.kang at gmail.com> wrote:

> Hi,
>
> When I install a host, it fails with:
>
> 2016-05-24 07:00:01,749 ERROR
> [org.ovirt.engine.core.bll.SetNonOperationalVdsCommand]
> (DefaultQuartzScheduler_Worker-4) [1bf36cd4] Host 'node3' is set to
> Non-Operational, it is missing the following networks: 'ovirtmgmt'
> 2016-05-24 07:00:01,781 WARN
>  [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
> (DefaultQuartzScheduler_Worker-4) [1bf36cd4] Correlation ID: 1bf36cd4, Job
> ID: db281e8f-67cc-441a-b44c-90b135e509bd, Call Stack: null, Custom Event
> ID: -1, Message: Host node3 does not comply with the cluster Default
> networks, the following networks are missing on host: 'ovirtmgmt'
>
> After I click the "Hosts" -> subtab "Network Interfaces" -> "Setup Host
> Networks" on the web ui and drag ovirtmgmt to "Assigned Logical" to link
> with eth0, then activate host "node3" successfully.
>
> My question is how to make such manual operation automatically?  Then I
> can run some automatic tests.
>
> Thanks a lot.
>
> --Kai
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20160524/ce043d8f/attachment-0001.html>


More information about the Users mailing list