Is the nic to the network staying up or going down for a period?
I'm just thinking, if the network has been configured to block unknown
unicast traffic, I think the VM would need to send a layer 2 frame to the
network before the network would send any frames to that switch port
destined for the VM.
After migration, could you use the VM console to send a packet and then see
if you can SSH in? Is the default Gateway for the VM on the network side? A
ping to the Gateway should be good enough in that case.
On Sat., 24 Aug. 2019, 04:20 Curtis E. Combs Jr., <ej.albany(a)gmail.com>
wrote:
It took a while for my servers to come back on the network this
time.
I think it's due to ovirt continuing to try to migrate the VMs around
like I requested. The 3 servers' names are "swm-01, swm-02 and
swm-03". Eventually (about 2-3 minutes ago) they all came back online.
So I disabled and stopped the lldpad service.
Nope. Started some more migrations and swm-02 and swm-03 disappeared
again. No ping, SSH hung, same as before - almost as soon as the
migration started.
If you wall have any ideas what switch-level setting might be enabled,
let me know, cause I'm stumped. I can add it to the ticket that's
requesting the port configurations. I've already added the port
numbers and switch name that I got from CDP.
Thanks again, I really appreciate the help!
cecjr
On Fri, Aug 23, 2019 at 3:28 PM Dominik Holler <dholler(a)redhat.com> wrote:
>
>
>
> On Fri, Aug 23, 2019 at 9:19 PM Dominik Holler <dholler(a)redhat.com>
wrote:
>>
>>
>>
>> On Fri, Aug 23, 2019 at 8:03 PM Curtis E. Combs Jr. <
ej.albany(a)gmail.com> wrote:
>>>
>>> This little cluster isn't in production or anything like that yet.
>>>
>>> So, I went ahead and used your ethtool commands to disable pause
>>> frames on both interfaces of each server. I then, chose a few VMs to
>>> migrate around at random.
>>>
>>> swm-02 and swm-03 both went out again. Unreachable. Can't ping,
can't
>>> ssh, and the SSH session that I had open was unresponsive.
>>>
>>> Any other ideas?
>>>
>>
>> Sorry, no. Looks like two different NICs with different drivers and
frimware goes down together.
>> This is a strong indication that the root cause is related to the
switch.
>> Maybe you can get some information about the switch config by
>> 'lldptool get-tlv -n -i em1'
>>
>
> Another guess:
> After the optional 'lldptool get-tlv -n -i em1'
> 'systemctl stop lldpad'
> another try to migrate.
>
>
>>
>>
>>>
>>> On Fri, Aug 23, 2019 at 1:50 PM Dominik Holler <dholler(a)redhat.com>
wrote:
>>> >
>>> >
>>> >
>>> > On Fri, Aug 23, 2019 at 6:45 PM Curtis E. Combs Jr. <
ej.albany(a)gmail.com> wrote:
>>> >>
>>> >> Unfortunately, I can't check on the switch. Trust me, I've
tried.
>>> >> These servers are in a Co-Lo and I've put 5 tickets in asking
about
>>> >> the port configuration. They just get ignored - but that's par
for
the
>>> >> coarse for IT here. Only about 2 out of 10 of our tickets get any
>>> >> response and usually the response doesn't help. Then the system
they
>>> >> use auto-closes the ticket. That was why I was suspecting STP
before.
>>> >>
>>> >> I can do ethtool. I do have root on these servers, though. Are you
>>> >> trying to get me to turn off link-speed auto-negotiation? Would
you
>>> >> like me to try that?
>>> >>
>>> >
>>> > It is just a suspicion, that the reason is pause frames.
>>> > Let's start on a NIC which is not used for ovirtmgmt, I guess em1.
>>> > Does 'ethtool -S em1 | grep pause' show something?
>>> > Does 'ethtool em1 | grep pause' indicates support for pause?
>>> > The current config is shown by 'ethtool -a em1'.
>>> > '-A autoneg' "Specifies whether pause autonegotiation
should be
enabled." according to ethtool doc.
>>> > Assuming flow control is enabled by default, I would try to disable
it via
>>> > 'ethtool -A em1 autoneg off rx off tx off'
>>> > and check if it is applied via
>>> > 'ethtool -a em1'
>>> > and check if the behavior under load changes.
>>> >
>>> >
>>> >
>>> >>
>>> >> On Fri, Aug 23, 2019 at 12:24 PM Dominik Holler
<dholler(a)redhat.com>
wrote:
>>> >> >
>>> >> >
>>> >> >
>>> >> > On Fri, Aug 23, 2019 at 5:49 PM Curtis E. Combs Jr. <
ej.albany(a)gmail.com> wrote:
>>> >> >>
>>> >> >> Sure! Right now, I only have a 500gb partition on each
node
shared over NFS, added as storage domains. This is on each node - so,
currently 3.
>>> >> >>
>>> >> >> How can the storage cause a node to drop out?
>>> >> >>
>>> >> >
>>> >> > Thanks, I got it.
>>> >> > All three links go down on load, which causes NFS to fail.
>>> >> >
>>> >> > Can you check in the switch port configuration if there is
some
kind of Ethernet flow control enabled?
>>> >> > Can you try to modify the behavior by changing the settings
of
your host interfaces, e.g.
>>> >> >
>>> >> > ethtool -A em1 autoneg off rx off tx off
>>> >> >
>>> >> > or
>>> >> > ethtool -A em1 autoneg on rx on tx on
>>> >> > ?
>>> >> >
>>> >> >
>>> >> >
>>> >> >
>>> >> >>
>>> >> >> On Fri, Aug 23, 2019, 11:46 AM Dominik Holler <
dholler(a)redhat.com> wrote:
>>> >> >>>
>>> >> >>>
>>> >> >>>
>>> >> >>> On Fri, Aug 23, 2019 at 5:41 PM Curtis E. Combs Jr.
<
ej.albany(a)gmail.com> wrote:
>>> >> >>>>
>>> >> >>>> Also, if it helps, the hosts will sit there,
quietly, for
hours or
>>> >> >>>> days before anything happens. They're up and
working just
fine. But
>>> >> >>>> then, when I manually migrate a VM from one host
to another,
they
>>> >> >>>> become completely inaccessible.
>>> >> >>>>
>>> >> >>>
>>> >> >>> Can you share some details about your storage?
>>> >> >>> Maybe there is a feature used during live migration,
which
triggers the issue.
>>> >> >>>
>>> >> >>>
>>> >> >>>>
>>> >> >>>> These are vanilla-as-possible CentOS7 nodes. Very
basic ovirt
install
>>> >> >>>> and configuration.
>>> >> >>>>
>>> >> >>>> On Fri, Aug 23, 2019 at 11:33 AM Curtis E. Combs
Jr.
>>> >> >>>> <ej.albany(a)gmail.com> wrote:
>>> >> >>>> >
>>> >> >>>> > Hey Dominik,
>>> >> >>>> >
>>> >> >>>> > Thanks for helping. I really want to try to
use ovirt.
>>> >> >>>> >
>>> >> >>>> > When these events happen, I cannot even SSH
to the nodes due
to the
>>> >> >>>> > link being down. After a little while, the
hosts come back...
>>> >> >>>> >
>>> >> >>>> > On Fri, Aug 23, 2019 at 11:30 AM Dominik
Holler <
dholler(a)redhat.com> wrote:
>>> >> >>>> > >
>>> >> >>>> > > Is you storage connected via NFS?
>>> >> >>>> > > Can you manually access the storage on
the host?
>>> >> >>>> > >
>>> >> >>>> > >
>>> >> >>>> > > On Fri, Aug 23, 2019 at 5:19 PM Curtis
E. Combs Jr. <
ej.albany(a)gmail.com> wrote:
>>> >> >>>> > >>
>>> >> >>>> > >> Sorry to dead bump this, but I'm
beginning to suspect
that maybe it's
>>> >> >>>> > >> not STP that's the problem.
>>> >> >>>> > >>
>>> >> >>>> > >> 2 of my hosts just went down when a
few VMs tried to
migrate.
>>> >> >>>> > >>
>>> >> >>>> > >> Do any of you have any idea what
might be going on here?
I don't even
>>> >> >>>> > >> know where to start. I'm going
to include the dmesg in
case it helps.
>>> >> >>>> > >> This happens on both of the hosts
whenever any migration
attempts to start.
>>> >> >>>> > >>
>>> >> >>>> > >>
>>> >> >>>> > >>
>>> >> >>>> > >>
>>> >> >>>> > >>
>>> >> >>>> > >>
>>> >> >>>> > >>
>>> >> >>>> > >>
>>> >> >>>> > >>
>>> >> >>>> > >> [68099.245833] bnx2 0000:01:00.0
em1: NIC Copper Link is
Down
>>> >> >>>> > >> [68099.246055] internal: port 1(em1)
entered disabled
state
>>> >> >>>> > >> [68184.177343] ixgbe 0000:03:00.0
p1p1: NIC Link is Down
>>> >> >>>> > >> [68184.177789] ovirtmgmt: port
1(p1p1) entered disabled
state
>>> >> >>>> > >> [68184.177856] ovirtmgmt: topology
change detected,
propagating
>>> >> >>>> > >> [68277.078671] INFO: task
qemu-kvm:8888 blocked for more
than 120 seconds.
>>> >> >>>> > >> [68277.078700] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
>>> >> >>>> > >> disables this message.
>>> >> >>>> > >> [68277.078723] qemu-kvm D
ffff9db40c359040 0
8888 1 0x000001a0
>>> >> >>>> > >> [68277.078727] Call Trace:
>>> >> >>>> > >> [68277.078738]
[<ffffffff978fd2ac>] ?
avc_has_perm_flags+0xdc/0x1c0
>>> >> >>>> > >> [68277.078743]
[<ffffffff97d69f19>] schedule+0x29/0x70
>>> >> >>>> > >> [68277.078746]
[<ffffffff9785f3d9>]
inode_dio_wait+0xd9/0x100
>>> >> >>>> > >> [68277.078751]
[<ffffffff976c4010>] ?
wake_bit_function+0x40/0x40
>>> >> >>>> > >> [68277.078765]
[<ffffffffc09d6dd6>]
nfs_getattr+0x1b6/0x250 [nfs]
>>> >> >>>> > >> [68277.078768]
[<ffffffff97848109>] vfs_getattr+0x49/0x80
>>> >> >>>> > >> [68277.078769]
[<ffffffff97848185>] vfs_fstat+0x45/0x80
>>> >> >>>> > >> [68277.078771]
[<ffffffff978486f4>]
SYSC_newfstat+0x24/0x60
>>> >> >>>> > >> [68277.078774]
[<ffffffff97d76d21>] ?
system_call_after_swapgs+0xae/0x146
>>> >> >>>> > >> [68277.078778]
[<ffffffff97739f34>] ?
__audit_syscall_entry+0xb4/0x110
>>> >> >>>> > >> [68277.078782]
[<ffffffff9763aaeb>] ?
syscall_trace_enter+0x16b/0x220
>>> >> >>>> > >> [68277.078784]
[<ffffffff97848ace>] SyS_newfstat+0xe/0x10
>>> >> >>>> > >> [68277.078786]
[<ffffffff97d7706b>] tracesys+0xa3/0xc9
>>> >> >>>> > >> [68397.072384] INFO: task
qemu-kvm:8888 blocked for more
than 120 seconds.
>>> >> >>>> > >> [68397.072413] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
>>> >> >>>> > >> disables this message.
>>> >> >>>> > >> [68397.072436] qemu-kvm D
ffff9db40c359040 0
8888 1 0x000001a0
>>> >> >>>> > >> [68397.072439] Call Trace:
>>> >> >>>> > >> [68397.072453]
[<ffffffff978fd2ac>] ?
avc_has_perm_flags+0xdc/0x1c0
>>> >> >>>> > >> [68397.072458]
[<ffffffff97d69f19>] schedule+0x29/0x70
>>> >> >>>> > >> [68397.072462]
[<ffffffff9785f3d9>]
inode_dio_wait+0xd9/0x100
>>> >> >>>> > >> [68397.072467]
[<ffffffff976c4010>] ?
wake_bit_function+0x40/0x40
>>> >> >>>> > >> [68397.072480]
[<ffffffffc09d6dd6>]
nfs_getattr+0x1b6/0x250 [nfs]
>>> >> >>>> > >> [68397.072485]
[<ffffffff97848109>] vfs_getattr+0x49/0x80
>>> >> >>>> > >> [68397.072486]
[<ffffffff97848185>] vfs_fstat+0x45/0x80
>>> >> >>>> > >> [68397.072488]
[<ffffffff978486f4>]
SYSC_newfstat+0x24/0x60
>>> >> >>>> > >> [68397.072491]
[<ffffffff97d76d21>] ?
system_call_after_swapgs+0xae/0x146
>>> >> >>>> > >> [68397.072495]
[<ffffffff97739f34>] ?
__audit_syscall_entry+0xb4/0x110
>>> >> >>>> > >> [68397.072498]
[<ffffffff9763aaeb>] ?
syscall_trace_enter+0x16b/0x220
>>> >> >>>> > >> [68397.072500]
[<ffffffff97848ace>] SyS_newfstat+0xe/0x10
>>> >> >>>> > >> [68397.072502]
[<ffffffff97d7706b>] tracesys+0xa3/0xc9
>>> >> >>>> > >> [68401.573141] bnx2 0000:01:00.0
em1: NIC Copper Link is
Up, 1000 Mbps
>>> >> >>>> > >> full duplex
>>> >> >>>> > >>
>>> >> >>>> > >> [68401.573247] internal: port 1(em1)
entered blocking
state
>>> >> >>>> > >> [68401.573255] internal: port 1(em1)
entered listening
state
>>> >> >>>> > >> [68403.576985] internal: port 1(em1)
entered learning
state
>>> >> >>>> > >> [68405.580907] internal: port 1(em1)
entered forwarding
state
>>> >> >>>> > >> [68405.580916] internal: topology
change detected,
propagating
>>> >> >>>> > >> [68469.565589] nfs: server
swm-01.hpc.moffitt.org not
responding, timed out
>>> >> >>>> > >> [68469.565840] nfs: server
swm-01.hpc.moffitt.org not
responding, timed out
>>> >> >>>> > >> [68487.193932] ixgbe 0000:03:00.0
p1p1: NIC Link is Up 10
Gbps, Flow
>>> >> >>>> > >> Control: RX/TX
>>> >> >>>> > >> [68487.194105] ovirtmgmt: port
1(p1p1) entered blocking
state
>>> >> >>>> > >> [68487.194114] ovirtmgmt: port
1(p1p1) entered listening
state
>>> >> >>>> > >> [68489.196508] ovirtmgmt: port
1(p1p1) entered learning
state
>>> >> >>>> > >> [68491.200400] ovirtmgmt: port
1(p1p1) entered forwarding
state
>>> >> >>>> > >> [68491.200405] ovirtmgmt: topology
change detected,
sending tcn bpdu
>>> >> >>>> > >> [68493.672423] NFS:
nfs4_reclaim_open_state: Lock reclaim
failed!
>>> >> >>>> > >> [68494.777996] NFSD: client
10.15.28.22 testing state ID
with
>>> >> >>>> > >> incorrect client ID
>>> >> >>>> > >> [68494.778580] NFSD: client
10.15.28.22 testing state ID
with
>>> >> >>>> > >> incorrect client ID
>>> >> >>>> > >>
>>> >> >>>> > >>
>>> >> >>>> > >> On Thu, Aug 22, 2019 at 2:53 PM
Curtis E. Combs Jr. <
ej.albany(a)gmail.com> wrote:
>>> >> >>>> > >> >
>>> >> >>>> > >> > Thanks, I'm just going to
revert back to bridges.
>>> >> >>>> > >> >
>>> >> >>>> > >> > On Thu, Aug 22, 2019 at 11:50
AM Dominik Holler <
dholler(a)redhat.com> wrote:
>>> >> >>>> > >> > >
>>> >> >>>> > >> > >
>>> >> >>>> > >> > >
>>> >> >>>> > >> > > On Thu, Aug 22, 2019 at
3:06 PM Curtis E. Combs Jr. <
ej.albany(a)gmail.com> wrote:
>>> >> >>>> > >> > >>
>>> >> >>>> > >> > >> Seems like the STP
options are so common and
necessary that it would
>>> >> >>>> > >> > >> be a priority over
seldom-used bridge_opts. I know
what STP is and I'm
>>> >> >>>> > >> > >> not even a networking
guy - never even heard of half
of the
>>> >> >>>> > >> > >> bridge_opts that have
switches in the UI.
>>> >> >>>> > >> > >>
>>> >> >>>> > >> > >> Anyway. I wanted to
try the openvswitches, so I
reinstalled all of my
>>> >> >>>> > >> > >> nodes and used
"openvswitch (Technology Preview)" as
the engine-setup
>>> >> >>>> > >> > >> option for the first
host. I made a new Cluster for
my nodes, added
>>> >> >>>> > >> > >> them all to the new
cluster, created a new "logical
network" for the
>>> >> >>>> > >> > >> internal network and
attached it to the internal
network ports.
>>> >> >>>> > >> > >>
>>> >> >>>> > >> > >> Now, when I go to
create a new VM, I don't even have
either the
>>> >> >>>> > >> > >> ovirtmgmt switch OR
the internal switch as an
option. The drop-down is
>>> >> >>>> > >> > >> empy as if I don't
have any vnic-profiles.
>>> >> >>>> > >> > >>
>>> >> >>>> > >> > >
>>> >> >>>> > >> > > openvswitch clusters are
limited to ovn networks.
>>> >> >>>> > >> > > You can create one like
described in
>>> >> >>>> > >> > >
https://www.ovirt.org/documentation/admin-guide/chap-External_Providers.h...
>>> >> >>>> > >> > >
>>> >> >>>> > >> > >
>>> >> >>>> > >> > >>
>>> >> >>>> > >> > >> On Thu, Aug 22, 2019
at 7:34 AM Tony Pearce <
tonyppe(a)gmail.com> wrote:
>>> >> >>>> > >> > >> >
>>> >> >>>> > >> > >> > Hi Dominik, would
you mind sharing the use case
for stp via API Only? I am keen to know this.
>>> >> >>>> > >> > >> > Thanks
>>> >> >>>> > >> > >> >
>>> >> >>>> > >> > >> >
>>> >> >>>> > >> > >> > On Thu., 22 Aug.
2019, 19:24 Dominik Holler, <
dholler(a)redhat.com> wrote:
>>> >> >>>> > >> > >> >>
>>> >> >>>> > >> > >> >>
>>> >> >>>> > >> > >> >>
>>> >> >>>> > >> > >> >> On Thu, Aug
22, 2019 at 1:08 PM Miguel Duarte de
Mora Barroso <mdbarroso(a)redhat.com> wrote:
>>> >> >>>> > >> > >> >>>
>>> >> >>>> > >> > >> >>> On Sat,
Aug 17, 2019 at 11:27 AM <
ej.albany(a)gmail.com> wrote:
>>> >> >>>> > >> > >> >>> >
>>> >> >>>> > >> > >> >>> >
Hello. I have been trying to figure out an
issue for a very long time.
>>> >> >>>> > >> > >> >>> > That
issue relates to the ethernet and 10gb fc
links that I have on my
>>> >> >>>> > >> > >> >>> >
cluster being disabled any time a migration
occurs.
>>> >> >>>> > >> > >> >>> >
>>> >> >>>> > >> > >> >>> > I
believe this is because I need to have STP
turned on in order to
>>> >> >>>> > >> > >> >>> >
participate with the switch. However, there
does not seem to be any
>>> >> >>>> > >> > >> >>> > way
to tell oVirt to stop turning it off! Very
frustrating.
>>> >> >>>> > >> > >> >>> >
>>> >> >>>> > >> > >> >>> >
After entering a cronjob that enables stp on
all bridges every 1
>>> >> >>>> > >> > >> >>> >
minute, the migration issue disappears....
>>> >> >>>> > >> > >> >>> >
>>> >> >>>> > >> > >> >>> > Is
there any way at all to do without this
cronjob and set STP to be
>>> >> >>>> > >> > >> >>> > ON
without having to resort to such a silly
solution?
>>> >> >>>> > >> > >> >>>
>>> >> >>>> > >> > >> >>> Vdsm
exposes a per bridge STP knob that you can
use for this. By
>>> >> >>>> > >> > >> >>> default
it is set to false, which is probably
why you had to use this
>>> >> >>>> > >> > >> >>>
shenanigan.
>>> >> >>>> > >> > >> >>>
>>> >> >>>> > >> > >> >>> You can,
for instance:
>>> >> >>>> > >> > >> >>>
>>> >> >>>> > >> > >> >>> # show
present state
>>> >> >>>> > >> > >> >>>
[vagrant@vdsm ~]$ ip a
>>> >> >>>> > >> > >> >>> 1: lo:
<LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc
noqueue state UNKNOWN
>>> >> >>>> > >> > >> >>> group
default qlen 1000
>>> >> >>>> > >> > >> >>>
link/loopback 00:00:00:00:00:00 brd
00:00:00:00:00:00
>>> >> >>>> > >> > >> >>> inet
127.0.0.1/8 scope host lo
>>> >> >>>> > >> > >> >>>
valid_lft forever preferred_lft forever
>>> >> >>>> > >> > >> >>> 2: eth0:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu
1500 qdisc pfifo_fast
>>> >> >>>> > >> > >> >>> state UP
group default qlen 1000
>>> >> >>>> > >> > >> >>>
link/ether 52:54:00:41:fb:37 brd
ff:ff:ff:ff:ff:ff
>>> >> >>>> > >> > >> >>> 3: eth1:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu
1500 qdisc pfifo_fast
>>> >> >>>> > >> > >> >>> state UP
group default qlen 1000
>>> >> >>>> > >> > >> >>>
link/ether 52:54:00:83:5b:6f brd
ff:ff:ff:ff:ff:ff
>>> >> >>>> > >> > >> >>> inet
192.168.50.50/24 brd 192.168.50.255
scope global noprefixroute eth1
>>> >> >>>> > >> > >> >>>
valid_lft forever preferred_lft forever
>>> >> >>>> > >> > >> >>> inet6
fe80::5054:ff:fe83:5b6f/64 scope link
>>> >> >>>> > >> > >> >>>
valid_lft forever preferred_lft forever
>>> >> >>>> > >> > >> >>> 19:
;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500
qdisc noop state DOWN
>>> >> >>>> > >> > >> >>> group
default qlen 1000
>>> >> >>>> > >> > >> >>>
link/ether 8e:5c:2e:87:fa:0b brd
ff:ff:ff:ff:ff:ff
>>> >> >>>> > >> > >> >>>
>>> >> >>>> > >> > >> >>> # show
example bridge configuration - you're
looking for the STP knob here.
>>> >> >>>> > >> > >> >>>
[root@vdsm ~]$ cat bridged_net_with_stp
>>> >> >>>> > >> > >> >>> {
>>> >> >>>> > >> > >> >>>
"bondings": {},
>>> >> >>>> > >> > >> >>>
"networks": {
>>> >> >>>> > >> > >> >>>
"test-network": {
>>> >> >>>> > >> > >> >>>
"nic": "eth0",
>>> >> >>>> > >> > >> >>>
"switch": "legacy",
>>> >> >>>> > >> > >> >>>
"bridged": true,
>>> >> >>>> > >> > >> >>>
"stp": true
>>> >> >>>> > >> > >> >>> }
>>> >> >>>> > >> > >> >>> },
>>> >> >>>> > >> > >> >>>
"options": {
>>> >> >>>> > >> > >> >>>
"connectivityCheck": false
>>> >> >>>> > >> > >> >>> }
>>> >> >>>> > >> > >> >>> }
>>> >> >>>> > >> > >> >>>
>>> >> >>>> > >> > >> >>> # issue
setup networks command:
>>> >> >>>> > >> > >> >>>
[root@vdsm ~]$ vdsm-client -f
bridged_net_with_stp Host setupNetworks
>>> >> >>>> > >> > >> >>> {
>>> >> >>>> > >> > >> >>>
"code": 0,
>>> >> >>>> > >> > >> >>>
"message": "Done"
>>> >> >>>> > >> > >> >>> }
>>> >> >>>> > >> > >> >>>
>>> >> >>>> > >> > >> >>> # show
bridges
>>> >> >>>> > >> > >> >>>
[root@vdsm ~]$ brctl show
>>> >> >>>> > >> > >> >>> bridge
name bridge id STP enabled interfaces
>>> >> >>>> > >> > >> >>>
;vdsmdummy; 8000.000000000000 no
>>> >> >>>> > >> > >> >>>
test-network 8000.52540041fb37 yes eth0
>>> >> >>>> > >> > >> >>>
>>> >> >>>> > >> > >> >>> # show
final state
>>> >> >>>> > >> > >> >>>
[root@vdsm ~]$ ip a
>>> >> >>>> > >> > >> >>> 1: lo:
<LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc
noqueue state UNKNOWN
>>> >> >>>> > >> > >> >>> group
default qlen 1000
>>> >> >>>> > >> > >> >>>
link/loopback 00:00:00:00:00:00 brd
00:00:00:00:00:00
>>> >> >>>> > >> > >> >>> inet
127.0.0.1/8 scope host lo
>>> >> >>>> > >> > >> >>>
valid_lft forever preferred_lft forever
>>> >> >>>> > >> > >> >>> 2: eth0:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu
1500 qdisc pfifo_fast
>>> >> >>>> > >> > >> >>> master
test-network state UP group default qlen
1000
>>> >> >>>> > >> > >> >>>
link/ether 52:54:00:41:fb:37 brd
ff:ff:ff:ff:ff:ff
>>> >> >>>> > >> > >> >>> 3: eth1:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu
1500 qdisc pfifo_fast
>>> >> >>>> > >> > >> >>> state UP
group default qlen 1000
>>> >> >>>> > >> > >> >>>
link/ether 52:54:00:83:5b:6f brd
ff:ff:ff:ff:ff:ff
>>> >> >>>> > >> > >> >>> inet
192.168.50.50/24 brd 192.168.50.255
scope global noprefixroute eth1
>>> >> >>>> > >> > >> >>>
valid_lft forever preferred_lft forever
>>> >> >>>> > >> > >> >>> inet6
fe80::5054:ff:fe83:5b6f/64 scope link
>>> >> >>>> > >> > >> >>>
valid_lft forever preferred_lft forever
>>> >> >>>> > >> > >> >>> 19:
;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500
qdisc noop state DOWN
>>> >> >>>> > >> > >> >>> group
default qlen 1000
>>> >> >>>> > >> > >> >>>
link/ether 8e:5c:2e:87:fa:0b brd
ff:ff:ff:ff:ff:ff
>>> >> >>>> > >> > >> >>> 432:
test-network:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>>> >> >>>> > >> > >> >>> noqueue
state UP group default qlen 1000
>>> >> >>>> > >> > >> >>>
link/ether 52:54:00:41:fb:37 brd
ff:ff:ff:ff:ff:ff
>>> >> >>>> > >> > >> >>>
>>> >> >>>> > >> > >> >>> I
don't think this STP parameter is exposed via
engine UI; @Dominik
>>> >> >>>> > >> > >> >>> Holler ,
could you confirm ? What are our plans
for it ?
>>> >> >>>> > >> > >> >>>
>>> >> >>>> > >> > >> >>
>>> >> >>>> > >> > >> >> STP is only
available via REST-API, see
>>> >> >>>> > >> > >> >>
http://ovirt.github.io/ovirt-engine-api-model/4.3/#types/network
>>> >> >>>> > >> > >> >> please find
an example how to enable STP in
>>> >> >>>> > >> > >> >>
https://gist.github.com/dominikholler/4e70c9ef9929d93b6807f56d43a70b95
>>> >> >>>> > >> > >> >>
>>> >> >>>> > >> > >> >> We have no
plans to add STP to the web ui,
>>> >> >>>> > >> > >> >> but new
feature requests are always welcome on
>>> >> >>>> > >> > >> >>
https://bugzilla.redhat.com/enter_bug.cgi?product=ovirt-engine
>>> >> >>>> > >> > >> >>
>>> >> >>>> > >> > >> >>
>>> >> >>>> > >> > >> >>>
>>> >> >>>> > >> > >> >>> >
>>> >> >>>> > >> > >> >>> > Here
are some details about my systems, if you
need it.
>>> >> >>>> > >> > >> >>> >
>>> >> >>>> > >> > >> >>> >
>>> >> >>>> > >> > >> >>> >
selinux is disabled.
>>> >> >>>> > >> > >> >>> >
>>> >> >>>> > >> > >> >>> >
>>> >> >>>> > >> > >> >>> >
>>> >> >>>> > >> > >> >>> >
>>> >> >>>> > >> > >> >>> >
>>> >> >>>> > >> > >> >>> >
>>> >> >>>> > >> > >> >>> >
>>> >> >>>> > >> > >> >>> >
>>> >> >>>> > >> > >> >>> >
>>> >> >>>> > >> > >> >>> >
[root@swm-02 ~]# rpm -qa | grep ovirt
>>> >> >>>> > >> > >> >>> >
ovirt-imageio-common-1.5.1-0.el7.x86_64
>>> >> >>>> > >> > >> >>> >
ovirt-release43-4.3.5.2-1.el7.noarch
>>> >> >>>> > >> > >> >>> >
ovirt-imageio-daemon-1.5.1-0.el7.noarch
>>> >> >>>> > >> > >> >>> >
ovirt-vmconsole-host-1.0.7-2.el7.noarch
>>> >> >>>> > >> > >> >>> >
ovirt-hosted-engine-setup-2.3.11-1.el7.noarch
>>> >> >>>> > >> > >> >>> >
ovirt-ansible-hosted-engine-setup-1.0.26-1.el7.noarch
>>> >> >>>> > >> > >> >>> >
python2-ovirt-host-deploy-1.8.0-1.el7.noarch
>>> >> >>>> > >> > >> >>> >
ovirt-ansible-engine-setup-1.1.9-1.el7.noarch
>>> >> >>>> > >> > >> >>> >
python2-ovirt-setup-lib-1.2.0-1.el7.noarch
>>> >> >>>> > >> > >> >>> >
cockpit-machines-ovirt-195.1-1.el7.noarch
>>> >> >>>> > >> > >> >>> >
ovirt-hosted-engine-ha-2.3.3-1.el7.noarch
>>> >> >>>> > >> > >> >>> >
ovirt-vmconsole-1.0.7-2.el7.noarch
>>> >> >>>> > >> > >> >>> >
cockpit-ovirt-dashboard-0.13.5-1.el7.noarch
>>> >> >>>> > >> > >> >>> >
ovirt-provider-ovn-driver-1.2.22-1.el7.noarch
>>> >> >>>> > >> > >> >>> >
ovirt-host-deploy-common-1.8.0-1.el7.noarch
>>> >> >>>> > >> > >> >>> >
ovirt-host-4.3.4-1.el7.x86_64
>>> >> >>>> > >> > >> >>> >
python-ovirt-engine-sdk4-4.3.2-2.el7.x86_64
>>> >> >>>> > >> > >> >>> >
ovirt-host-dependencies-4.3.4-1.el7.x86_64
>>> >> >>>> > >> > >> >>> >
ovirt-ansible-repositories-1.1.5-1.el7.noarch
>>> >> >>>> > >> > >> >>> >
[root@swm-02 ~]# cat /etc/redhat-release
>>> >> >>>> > >> > >> >>> >
CentOS Linux release 7.6.1810 (Core)
>>> >> >>>> > >> > >> >>> >
[root@swm-02 ~]# uname -r
>>> >> >>>> > >> > >> >>> >
3.10.0-957.27.2.el7.x86_64
>>> >> >>>> > >> > >> >>> > You
have new mail in /var/spool/mail/root
>>> >> >>>> > >> > >> >>> >
[root@swm-02 ~]# ip a
>>> >> >>>> > >> > >> >>> > 1:
lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc
noqueue state UNKNOWN
>>> >> >>>> > >> > >> >>> >
group default qlen 1000
>>> >> >>>> > >> > >> >>> >
link/loopback 00:00:00:00:00:00 brd
00:00:00:00:00:00
>>> >> >>>> > >> > >> >>> >
inet 127.0.0.1/8 scope host lo
>>> >> >>>> > >> > >> >>> >
valid_lft forever preferred_lft forever
>>> >> >>>> > >> > >> >>> > 2:
em1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu
1500 qdisc mq master
>>> >> >>>> > >> > >> >>> > test
state UP group default qlen 1000
>>> >> >>>> > >> > >> >>> >
link/ether d4:ae:52:8d:50:48 brd
ff:ff:ff:ff:ff:ff
>>> >> >>>> > >> > >> >>> > 3:
em2: <BROADCAST,MULTICAST> mtu 1500 qdisc
noop state DOWN group
>>> >> >>>> > >> > >> >>> >
default qlen 1000
>>> >> >>>> > >> > >> >>> >
link/ether d4:ae:52:8d:50:49 brd
ff:ff:ff:ff:ff:ff
>>> >> >>>> > >> > >> >>> > 4:
p1p1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu
1500 qdisc mq master
>>> >> >>>> > >> > >> >>> >
ovirtmgmt state UP group default qlen 1000
>>> >> >>>> > >> > >> >>> >
link/ether 90:e2:ba:1e:14:80 brd
ff:ff:ff:ff:ff:ff
>>> >> >>>> > >> > >> >>> > 5:
p1p2: <BROADCAST,MULTICAST> mtu 1500 qdisc
noop state DOWN group
>>> >> >>>> > >> > >> >>> >
default qlen 1000
>>> >> >>>> > >> > >> >>> >
link/ether 90:e2:ba:1e:14:81 brd
ff:ff:ff:ff:ff:ff
>>> >> >>>> > >> > >> >>> > 6:
ovs-system: <BROADCAST,MULTICAST> mtu 1500
qdisc noop state DOWN
>>> >> >>>> > >> > >> >>> >
group default qlen 1000
>>> >> >>>> > >> > >> >>> >
link/ether a2:b8:d6:e8:b3:d8 brd
ff:ff:ff:ff:ff:ff
>>> >> >>>> > >> > >> >>> > 7:
br-int: <BROADCAST,MULTICAST> mtu 1500
qdisc noop state DOWN group
>>> >> >>>> > >> > >> >>> >
default qlen 1000
>>> >> >>>> > >> > >> >>> >
link/ether 96:a0:c1:4a:45:4b brd
ff:ff:ff:ff:ff:ff
>>> >> >>>> > >> > >> >>> > 25:
test: <BROADCAST,MULTICAST,UP,LOWER_UP>
mtu 1500 qdisc noqueue
>>> >> >>>> > >> > >> >>> >
state UP group default qlen 1000
>>> >> >>>> > >> > >> >>> >
link/ether d4:ae:52:8d:50:48 brd
ff:ff:ff:ff:ff:ff
>>> >> >>>> > >> > >> >>> >
inet 10.15.11.21/24 brd 10.15.11.255
scope global test
>>> >> >>>> > >> > >> >>> >
valid_lft forever preferred_lft forever
>>> >> >>>> > >> > >> >>> > 26:
ovirtmgmt:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>>> >> >>>> > >> > >> >>> >
noqueue state UP group default qlen 1000
>>> >> >>>> > >> > >> >>> >
link/ether 90:e2:ba:1e:14:80 brd
ff:ff:ff:ff:ff:ff
>>> >> >>>> > >> > >> >>> >
inet 10.15.28.31/24 brd 10.15.28.255
scope global ovirtmgmt
>>> >> >>>> > >> > >> >>> >
valid_lft forever preferred_lft forever
>>> >> >>>> > >> > >> >>> > 27:
;vdsmdummy;: <BROADCAST,MULTICAST> mtu
1500 qdisc noop state DOWN
>>> >> >>>> > >> > >> >>> >
group default qlen 1000
>>> >> >>>> > >> > >> >>> >
link/ether 62:e5:e5:07:99:eb brd
ff:ff:ff:ff:ff:ff
>>> >> >>>> > >> > >> >>> > 29:
vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP>
mtu 1500 qdisc mq master
>>> >> >>>> > >> > >> >>> >
ovirtmgmt state UNKNOWN group default qlen 1000
>>> >> >>>> > >> > >> >>> >
link/ether fe:6f:9c:95:00:02 brd
ff:ff:ff:ff:ff:ff
>>> >> >>>> > >> > >> >>> >
[root@swm-02 ~]# free -m
>>> >> >>>> > >> > >> >>> >
total used free
shared buff/cache available
>>> >> >>>> > >> > >> >>> > Mem:
64413 1873 61804
9 735 62062
>>> >> >>>> > >> > >> >>> >
Swap: 16383 0 16383
>>> >> >>>> > >> > >> >>> >
[root@swm-02 ~]# free -h
>>> >> >>>> > >> > >> >>> >
total used free
shared buff/cache available
>>> >> >>>> > >> > >> >>> > Mem:
62G 1.8G 60G
9.5M 735M 60G
>>> >> >>>> > >> > >> >>> >
Swap: 15G 0B 15G
>>> >> >>>> > >> > >> >>> >
[root@swm-02 ~]# ls
>>> >> >>>> > >> > >> >>> > ls
lsb_release lshw
lslocks
>>> >> >>>> > >> > >> >>> >
lsmod lspci
lssubsys
>>> >> >>>> > >> > >> >>> >
lsusb.py
>>> >> >>>> > >> > >> >>> >
lsattr lscgroup
lsinitrd lslogins
>>> >> >>>> > >> > >> >>> >
lsns lss16toppm
lstopo-no-graphics
>>> >> >>>> > >> > >> >>> >
lsblk lscpu lsipc
lsmem
>>> >> >>>> > >> > >> >>> >
lsof lsscsi
lsusb
>>> >> >>>> > >> > >> >>> >
[root@swm-02 ~]# lscpu
>>> >> >>>> > >> > >> >>> >
Architecture: x86_64
>>> >> >>>> > >> > >> >>> > CPU
op-mode(s): 32-bit, 64-bit
>>> >> >>>> > >> > >> >>> > Byte
Order: Little Endian
>>> >> >>>> > >> > >> >>> >
CPU(s): 16
>>> >> >>>> > >> > >> >>> >
On-line CPU(s) list: 0-15
>>> >> >>>> > >> > >> >>> >
Thread(s) per core: 2
>>> >> >>>> > >> > >> >>> >
Core(s) per socket: 4
>>> >> >>>> > >> > >> >>> >
Socket(s): 2
>>> >> >>>> > >> > >> >>> > NUMA
node(s): 2
>>> >> >>>> > >> > >> >>> >
Vendor ID: GenuineIntel
>>> >> >>>> > >> > >> >>> > CPU
family: 6
>>> >> >>>> > >> > >> >>> >
Model: 44
>>> >> >>>> > >> > >> >>> >
Model name: Intel(R) Xeon(R) CPU
X5672 @ 3.20GHz
>>> >> >>>> > >> > >> >>> >
Stepping: 2
>>> >> >>>> > >> > >> >>> > CPU
MHz: 3192.064
>>> >> >>>> > >> > >> >>> >
BogoMIPS: 6384.12
>>> >> >>>> > >> > >> >>> >
Virtualization: VT-x
>>> >> >>>> > >> > >> >>> > L1d
cache: 32K
>>> >> >>>> > >> > >> >>> > L1i
cache: 32K
>>> >> >>>> > >> > >> >>> > L2
cache: 256K
>>> >> >>>> > >> > >> >>> > L3
cache: 12288K
>>> >> >>>> > >> > >> >>> > NUMA
node0 CPU(s): 0,2,4,6,8,10,12,14
>>> >> >>>> > >> > >> >>> > NUMA
node1 CPU(s): 1,3,5,7,9,11,13,15
>>> >> >>>> > >> > >> >>> >
Flags: fpu vme de pse tsc msr
pae mce cx8 apic sep
>>> >> >>>> > >> > >> >>> > mtrr
pge mca cmov pat pse36 clflush dts acpi
mmx fxsr sse sse2 ss ht
>>> >> >>>> > >> > >> >>> > tm
pbe syscall nx pdpe1gb rdtscp lm
constant_tsc arch_perfmon pebs bts
>>> >> >>>> > >> > >> >>> >
rep_good nopl xtopology nonstop_tsc aperfmperf
eagerfpu pni pclmulqdq
>>> >> >>>> > >> > >> >>> >
dtes64 monitor ds_cpl vmx smx est tm2 ssse3
cx16 xtpr pdcm pcid dca
>>> >> >>>> > >> > >> >>> >
sse4_1 sse4_2 popcnt aes lahf_lm ssbd ibrs
ibpb stibp tpr_shadow vnmi
>>> >> >>>> > >> > >> >>> >
flexpriority ept vpid dtherm ida arat
spec_ctrl intel_stibp flush_l1d
>>> >> >>>> > >> > >> >>> >
[root@swm-02 ~]#
>>> >> >>>> > >> > >> >>> >
_______________________________________________
>>> >> >>>> > >> > >> >>> >
Users mailing list -- users(a)ovirt.org
>>> >> >>>> > >> > >> >>> > To
unsubscribe send an email to
users-leave(a)ovirt.org
>>> >> >>>> > >> > >> >>> >
Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
>>> >> >>>> > >> > >> >>> >
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
>>> >> >>>> > >> > >> >>> > List
Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/MTMZ5MF4CF2...
>>> >> >>>> > >> > >> >>
>>> >> >>>> > >> > >> >>
_______________________________________________
>>> >> >>>> > >> > >> >> Users mailing
list -- users(a)ovirt.org
>>> >> >>>> > >> > >> >> To
unsubscribe send an email to
users-leave(a)ovirt.org
>>> >> >>>> > >> > >> >> Privacy
Statement:
https://www.ovirt.org/site/privacy-policy/
>>> >> >>>> > >> > >> >> oVirt Code of
Conduct:
https://www.ovirt.org/community/about/community-guidelines/
>>> >> >>>> > >> > >> >> List
Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/QBA7NYKAJNR...