On Fri, Aug 23, 2019 at 9:19 PM Dominik Holler <dholler(a)redhat.com> wrote:
On Fri, Aug 23, 2019 at 8:03 PM Curtis E. Combs Jr. <ej.albany(a)gmail.com>
wrote:
> This little cluster isn't in production or anything like that yet.
>
> So, I went ahead and used your ethtool commands to disable pause
> frames on both interfaces of each server. I then, chose a few VMs to
> migrate around at random.
>
> swm-02 and swm-03 both went out again. Unreachable. Can't ping, can't
> ssh, and the SSH session that I had open was unresponsive.
>
> Any other ideas?
>
>
Sorry, no. Looks like two different NICs with different drivers and
frimware goes down together.
This is a strong indication that the root cause is related to the switch.
Maybe you can get some information about the switch config by
'lldptool get-tlv -n -i em1'
Another guess:
After the optional 'lldptool get-tlv -n -i em1'
'systemctl stop lldpad'
another try to migrate.
> On Fri, Aug 23, 2019 at 1:50 PM Dominik Holler <dholler(a)redhat.com>
> wrote:
> >
> >
> >
> > On Fri, Aug 23, 2019 at 6:45 PM Curtis E. Combs Jr. <
> ej.albany(a)gmail.com> wrote:
> >>
> >> Unfortunately, I can't check on the switch. Trust me, I've tried.
> >> These servers are in a Co-Lo and I've put 5 tickets in asking about
> >> the port configuration. They just get ignored - but that's par for the
> >> coarse for IT here. Only about 2 out of 10 of our tickets get any
> >> response and usually the response doesn't help. Then the system they
> >> use auto-closes the ticket. That was why I was suspecting STP before.
> >>
> >> I can do ethtool. I do have root on these servers, though. Are you
> >> trying to get me to turn off link-speed auto-negotiation? Would you
> >> like me to try that?
> >>
> >
> > It is just a suspicion, that the reason is pause frames.
> > Let's start on a NIC which is not used for ovirtmgmt, I guess em1.
> > Does 'ethtool -S em1 | grep pause' show something?
> > Does 'ethtool em1 | grep pause' indicates support for pause?
> > The current config is shown by 'ethtool -a em1'.
> > '-A autoneg' "Specifies whether pause autonegotiation should be
> enabled." according to ethtool doc.
> > Assuming flow control is enabled by default, I would try to disable it
> via
> > 'ethtool -A em1 autoneg off rx off tx off'
> > and check if it is applied via
> > 'ethtool -a em1'
> > and check if the behavior under load changes.
> >
> >
> >
> >>
> >> On Fri, Aug 23, 2019 at 12:24 PM Dominik Holler <dholler(a)redhat.com>
> wrote:
> >> >
> >> >
> >> >
> >> > On Fri, Aug 23, 2019 at 5:49 PM Curtis E. Combs Jr. <
> ej.albany(a)gmail.com> wrote:
> >> >>
> >> >> Sure! Right now, I only have a 500gb partition on each node shared
> over NFS, added as storage domains. This is on each node - so, currently 3.
> >> >>
> >> >> How can the storage cause a node to drop out?
> >> >>
> >> >
> >> > Thanks, I got it.
> >> > All three links go down on load, which causes NFS to fail.
> >> >
> >> > Can you check in the switch port configuration if there is some kind
> of Ethernet flow control enabled?
> >> > Can you try to modify the behavior by changing the settings of your
> host interfaces, e.g.
> >> >
> >> > ethtool -A em1 autoneg off rx off tx off
> >> >
> >> > or
> >> > ethtool -A em1 autoneg on rx on tx on
> >> > ?
> >> >
> >> >
> >> >
> >> >
> >> >>
> >> >> On Fri, Aug 23, 2019, 11:46 AM Dominik Holler
<dholler(a)redhat.com>
> wrote:
> >> >>>
> >> >>>
> >> >>>
> >> >>> On Fri, Aug 23, 2019 at 5:41 PM Curtis E. Combs Jr. <
> ej.albany(a)gmail.com> wrote:
> >> >>>>
> >> >>>> Also, if it helps, the hosts will sit there, quietly, for
hours or
> >> >>>> days before anything happens. They're up and working
just fine.
> But
> >> >>>> then, when I manually migrate a VM from one host to
another, they
> >> >>>> become completely inaccessible.
> >> >>>>
> >> >>>
> >> >>> Can you share some details about your storage?
> >> >>> Maybe there is a feature used during live migration, which
> triggers the issue.
> >> >>>
> >> >>>
> >> >>>>
> >> >>>> These are vanilla-as-possible CentOS7 nodes. Very basic
ovirt
> install
> >> >>>> and configuration.
> >> >>>>
> >> >>>> On Fri, Aug 23, 2019 at 11:33 AM Curtis E. Combs Jr.
> >> >>>> <ej.albany(a)gmail.com> wrote:
> >> >>>> >
> >> >>>> > Hey Dominik,
> >> >>>> >
> >> >>>> > Thanks for helping. I really want to try to use
ovirt.
> >> >>>> >
> >> >>>> > When these events happen, I cannot even SSH to the
nodes due to
> the
> >> >>>> > link being down. After a little while, the hosts come
back...
> >> >>>> >
> >> >>>> > On Fri, Aug 23, 2019 at 11:30 AM Dominik Holler <
> dholler(a)redhat.com> wrote:
> >> >>>> > >
> >> >>>> > > Is you storage connected via NFS?
> >> >>>> > > Can you manually access the storage on the host?
> >> >>>> > >
> >> >>>> > >
> >> >>>> > > On Fri, Aug 23, 2019 at 5:19 PM Curtis E. Combs
Jr. <
> ej.albany(a)gmail.com> wrote:
> >> >>>> > >>
> >> >>>> > >> Sorry to dead bump this, but I'm
beginning to suspect that
> maybe it's
> >> >>>> > >> not STP that's the problem.
> >> >>>> > >>
> >> >>>> > >> 2 of my hosts just went down when a few VMs
tried to migrate.
> >> >>>> > >>
> >> >>>> > >> Do any of you have any idea what might be
going on here? I
> don't even
> >> >>>> > >> know where to start. I'm going to include
the dmesg in case
> it helps.
> >> >>>> > >> This happens on both of the hosts whenever
any migration
> attempts to start.
> >> >>>> > >>
> >> >>>> > >>
> >> >>>> > >>
> >> >>>> > >>
> >> >>>> > >>
> >> >>>> > >>
> >> >>>> > >>
> >> >>>> > >>
> >> >>>> > >>
> >> >>>> > >> [68099.245833] bnx2 0000:01:00.0 em1: NIC
Copper Link is Down
> >> >>>> > >> [68099.246055] internal: port 1(em1) entered
disabled state
> >> >>>> > >> [68184.177343] ixgbe 0000:03:00.0 p1p1: NIC
Link is Down
> >> >>>> > >> [68184.177789] ovirtmgmt: port 1(p1p1)
entered disabled state
> >> >>>> > >> [68184.177856] ovirtmgmt: topology change
detected,
> propagating
> >> >>>> > >> [68277.078671] INFO: task qemu-kvm:8888
blocked for more
> than 120 seconds.
> >> >>>> > >> [68277.078700] "echo 0 >
> /proc/sys/kernel/hung_task_timeout_secs"
> >> >>>> > >> disables this message.
> >> >>>> > >> [68277.078723] qemu-kvm D
ffff9db40c359040 0
> 8888 1 0x000001a0
> >> >>>> > >> [68277.078727] Call Trace:
> >> >>>> > >> [68277.078738] [<ffffffff978fd2ac>] ?
> avc_has_perm_flags+0xdc/0x1c0
> >> >>>> > >> [68277.078743] [<ffffffff97d69f19>]
schedule+0x29/0x70
> >> >>>> > >> [68277.078746] [<ffffffff9785f3d9>]
> inode_dio_wait+0xd9/0x100
> >> >>>> > >> [68277.078751] [<ffffffff976c4010>] ?
> wake_bit_function+0x40/0x40
> >> >>>> > >> [68277.078765] [<ffffffffc09d6dd6>]
nfs_getattr+0x1b6/0x250
> [nfs]
> >> >>>> > >> [68277.078768] [<ffffffff97848109>]
vfs_getattr+0x49/0x80
> >> >>>> > >> [68277.078769] [<ffffffff97848185>]
vfs_fstat+0x45/0x80
> >> >>>> > >> [68277.078771] [<ffffffff978486f4>]
SYSC_newfstat+0x24/0x60
> >> >>>> > >> [68277.078774] [<ffffffff97d76d21>] ?
> system_call_after_swapgs+0xae/0x146
> >> >>>> > >> [68277.078778] [<ffffffff97739f34>] ?
> __audit_syscall_entry+0xb4/0x110
> >> >>>> > >> [68277.078782] [<ffffffff9763aaeb>] ?
> syscall_trace_enter+0x16b/0x220
> >> >>>> > >> [68277.078784] [<ffffffff97848ace>]
SyS_newfstat+0xe/0x10
> >> >>>> > >> [68277.078786] [<ffffffff97d7706b>]
tracesys+0xa3/0xc9
> >> >>>> > >> [68397.072384] INFO: task qemu-kvm:8888
blocked for more
> than 120 seconds.
> >> >>>> > >> [68397.072413] "echo 0 >
> /proc/sys/kernel/hung_task_timeout_secs"
> >> >>>> > >> disables this message.
> >> >>>> > >> [68397.072436] qemu-kvm D
ffff9db40c359040 0
> 8888 1 0x000001a0
> >> >>>> > >> [68397.072439] Call Trace:
> >> >>>> > >> [68397.072453] [<ffffffff978fd2ac>] ?
> avc_has_perm_flags+0xdc/0x1c0
> >> >>>> > >> [68397.072458] [<ffffffff97d69f19>]
schedule+0x29/0x70
> >> >>>> > >> [68397.072462] [<ffffffff9785f3d9>]
> inode_dio_wait+0xd9/0x100
> >> >>>> > >> [68397.072467] [<ffffffff976c4010>] ?
> wake_bit_function+0x40/0x40
> >> >>>> > >> [68397.072480] [<ffffffffc09d6dd6>]
nfs_getattr+0x1b6/0x250
> [nfs]
> >> >>>> > >> [68397.072485] [<ffffffff97848109>]
vfs_getattr+0x49/0x80
> >> >>>> > >> [68397.072486] [<ffffffff97848185>]
vfs_fstat+0x45/0x80
> >> >>>> > >> [68397.072488] [<ffffffff978486f4>]
SYSC_newfstat+0x24/0x60
> >> >>>> > >> [68397.072491] [<ffffffff97d76d21>] ?
> system_call_after_swapgs+0xae/0x146
> >> >>>> > >> [68397.072495] [<ffffffff97739f34>] ?
> __audit_syscall_entry+0xb4/0x110
> >> >>>> > >> [68397.072498] [<ffffffff9763aaeb>] ?
> syscall_trace_enter+0x16b/0x220
> >> >>>> > >> [68397.072500] [<ffffffff97848ace>]
SyS_newfstat+0xe/0x10
> >> >>>> > >> [68397.072502] [<ffffffff97d7706b>]
tracesys+0xa3/0xc9
> >> >>>> > >> [68401.573141] bnx2 0000:01:00.0 em1: NIC
Copper Link is Up,
> 1000 Mbps
> >> >>>> > >> full duplex
> >> >>>> > >>
> >> >>>> > >> [68401.573247] internal: port 1(em1) entered
blocking state
> >> >>>> > >> [68401.573255] internal: port 1(em1) entered
listening state
> >> >>>> > >> [68403.576985] internal: port 1(em1) entered
learning state
> >> >>>> > >> [68405.580907] internal: port 1(em1) entered
forwarding state
> >> >>>> > >> [68405.580916] internal: topology change
detected,
> propagating
> >> >>>> > >> [68469.565589] nfs: server
swm-01.hpc.moffitt.org not
> responding, timed out
> >> >>>> > >> [68469.565840] nfs: server
swm-01.hpc.moffitt.org not
> responding, timed out
> >> >>>> > >> [68487.193932] ixgbe 0000:03:00.0 p1p1: NIC
Link is Up 10
> Gbps, Flow
> >> >>>> > >> Control: RX/TX
> >> >>>> > >> [68487.194105] ovirtmgmt: port 1(p1p1)
entered blocking state
> >> >>>> > >> [68487.194114] ovirtmgmt: port 1(p1p1)
entered listening
> state
> >> >>>> > >> [68489.196508] ovirtmgmt: port 1(p1p1)
entered learning state
> >> >>>> > >> [68491.200400] ovirtmgmt: port 1(p1p1)
entered forwarding
> state
> >> >>>> > >> [68491.200405] ovirtmgmt: topology change
detected, sending
> tcn bpdu
> >> >>>> > >> [68493.672423] NFS: nfs4_reclaim_open_state:
Lock reclaim
> failed!
> >> >>>> > >> [68494.777996] NFSD: client 10.15.28.22
testing state ID with
> >> >>>> > >> incorrect client ID
> >> >>>> > >> [68494.778580] NFSD: client 10.15.28.22
testing state ID with
> >> >>>> > >> incorrect client ID
> >> >>>> > >>
> >> >>>> > >>
> >> >>>> > >> On Thu, Aug 22, 2019 at 2:53 PM Curtis E.
Combs Jr. <
> ej.albany(a)gmail.com> wrote:
> >> >>>> > >> >
> >> >>>> > >> > Thanks, I'm just going to revert
back to bridges.
> >> >>>> > >> >
> >> >>>> > >> > On Thu, Aug 22, 2019 at 11:50 AM Dominik
Holler <
> dholler(a)redhat.com> wrote:
> >> >>>> > >> > >
> >> >>>> > >> > >
> >> >>>> > >> > >
> >> >>>> > >> > > On Thu, Aug 22, 2019 at 3:06 PM
Curtis E. Combs Jr. <
> ej.albany(a)gmail.com> wrote:
> >> >>>> > >> > >>
> >> >>>> > >> > >> Seems like the STP options are
so common and necessary
> that it would
> >> >>>> > >> > >> be a priority over seldom-used
bridge_opts. I know what
> STP is and I'm
> >> >>>> > >> > >> not even a networking guy -
never even heard of half of
> the
> >> >>>> > >> > >> bridge_opts that have switches
in the UI.
> >> >>>> > >> > >>
> >> >>>> > >> > >> Anyway. I wanted to try the
openvswitches, so I
> reinstalled all of my
> >> >>>> > >> > >> nodes and used
"openvswitch (Technology Preview)" as
> the engine-setup
> >> >>>> > >> > >> option for the first host. I
made a new Cluster for my
> nodes, added
> >> >>>> > >> > >> them all to the new cluster,
created a new "logical
> network" for the
> >> >>>> > >> > >> internal network and attached
it to the internal
> network ports.
> >> >>>> > >> > >>
> >> >>>> > >> > >> Now, when I go to create a new
VM, I don't even have
> either the
> >> >>>> > >> > >> ovirtmgmt switch OR the
internal switch as an option.
> The drop-down is
> >> >>>> > >> > >> empy as if I don't have any
vnic-profiles.
> >> >>>> > >> > >>
> >> >>>> > >> > >
> >> >>>> > >> > > openvswitch clusters are limited to
ovn networks.
> >> >>>> > >> > > You can create one like described
in
> >> >>>> > >> > >
>
https://www.ovirt.org/documentation/admin-guide/chap-External_Providers.h...
> >> >>>> > >> > >
> >> >>>> > >> > >
> >> >>>> > >> > >>
> >> >>>> > >> > >> On Thu, Aug 22, 2019 at 7:34 AM
Tony Pearce <
> tonyppe(a)gmail.com> wrote:
> >> >>>> > >> > >> >
> >> >>>> > >> > >> > Hi Dominik, would you mind
sharing the use case for
> stp via API Only? I am keen to know this.
> >> >>>> > >> > >> > Thanks
> >> >>>> > >> > >> >
> >> >>>> > >> > >> >
> >> >>>> > >> > >> > On Thu., 22 Aug. 2019,
19:24 Dominik Holler, <
> dholler(a)redhat.com> wrote:
> >> >>>> > >> > >> >>
> >> >>>> > >> > >> >>
> >> >>>> > >> > >> >>
> >> >>>> > >> > >> >> On Thu, Aug 22, 2019
at 1:08 PM Miguel Duarte de
> Mora Barroso <mdbarroso(a)redhat.com> wrote:
> >> >>>> > >> > >> >>>
> >> >>>> > >> > >> >>> On Sat, Aug 17,
2019 at 11:27 AM <
> ej.albany(a)gmail.com> wrote:
> >> >>>> > >> > >> >>> >
> >> >>>> > >> > >> >>> > Hello. I have
been trying to figure out an issue
> for a very long time.
> >> >>>> > >> > >> >>> > That issue
relates to the ethernet and 10gb fc
> links that I have on my
> >> >>>> > >> > >> >>> > cluster being
disabled any time a migration
> occurs.
> >> >>>> > >> > >> >>> >
> >> >>>> > >> > >> >>> > I believe
this is because I need to have STP
> turned on in order to
> >> >>>> > >> > >> >>> > participate
with the switch. However, there does
> not seem to be any
> >> >>>> > >> > >> >>> > way to tell
oVirt to stop turning it off! Very
> frustrating.
> >> >>>> > >> > >> >>> >
> >> >>>> > >> > >> >>> > After
entering a cronjob that enables stp on all
> bridges every 1
> >> >>>> > >> > >> >>> > minute, the
migration issue disappears....
> >> >>>> > >> > >> >>> >
> >> >>>> > >> > >> >>> > Is there any
way at all to do without this
> cronjob and set STP to be
> >> >>>> > >> > >> >>> > ON without
having to resort to such a silly
> solution?
> >> >>>> > >> > >> >>>
> >> >>>> > >> > >> >>> Vdsm exposes a per
bridge STP knob that you can use
> for this. By
> >> >>>> > >> > >> >>> default it is set
to false, which is probably why
> you had to use this
> >> >>>> > >> > >> >>> shenanigan.
> >> >>>> > >> > >> >>>
> >> >>>> > >> > >> >>> You can, for
instance:
> >> >>>> > >> > >> >>>
> >> >>>> > >> > >> >>> # show present
state
> >> >>>> > >> > >> >>> [vagrant@vdsm ~]$
ip a
> >> >>>> > >> > >> >>> 1: lo:
<LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc
> noqueue state UNKNOWN
> >> >>>> > >> > >> >>> group default qlen
1000
> >> >>>> > >> > >> >>> link/loopback
00:00:00:00:00:00 brd
> 00:00:00:00:00:00
> >> >>>> > >> > >> >>> inet
127.0.0.1/8 scope host lo
> >> >>>> > >> > >> >>> valid_lft
forever preferred_lft forever
> >> >>>> > >> > >> >>> 2: eth0:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500
> qdisc pfifo_fast
> >> >>>> > >> > >> >>> state UP group
default qlen 1000
> >> >>>> > >> > >> >>> link/ether
52:54:00:41:fb:37 brd
> ff:ff:ff:ff:ff:ff
> >> >>>> > >> > >> >>> 3: eth1:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500
> qdisc pfifo_fast
> >> >>>> > >> > >> >>> state UP group
default qlen 1000
> >> >>>> > >> > >> >>> link/ether
52:54:00:83:5b:6f brd
> ff:ff:ff:ff:ff:ff
> >> >>>> > >> > >> >>> inet
192.168.50.50/24 brd 192.168.50.255 scope
> global noprefixroute eth1
> >> >>>> > >> > >> >>> valid_lft
forever preferred_lft forever
> >> >>>> > >> > >> >>> inet6
fe80::5054:ff:fe83:5b6f/64 scope link
> >> >>>> > >> > >> >>> valid_lft
forever preferred_lft forever
> >> >>>> > >> > >> >>> 19: ;vdsmdummy;:
<BROADCAST,MULTICAST> mtu 1500
> qdisc noop state DOWN
> >> >>>> > >> > >> >>> group default qlen
1000
> >> >>>> > >> > >> >>> link/ether
8e:5c:2e:87:fa:0b brd
> ff:ff:ff:ff:ff:ff
> >> >>>> > >> > >> >>>
> >> >>>> > >> > >> >>> # show example
bridge configuration - you're
> looking for the STP knob here.
> >> >>>> > >> > >> >>> [root@vdsm ~]$ cat
bridged_net_with_stp
> >> >>>> > >> > >> >>> {
> >> >>>> > >> > >> >>>
"bondings": {},
> >> >>>> > >> > >> >>>
"networks": {
> >> >>>> > >> > >> >>>
"test-network": {
> >> >>>> > >> > >> >>>
"nic": "eth0",
> >> >>>> > >> > >> >>>
"switch": "legacy",
> >> >>>> > >> > >> >>>
"bridged": true,
> >> >>>> > >> > >> >>>
"stp": true
> >> >>>> > >> > >> >>> }
> >> >>>> > >> > >> >>> },
> >> >>>> > >> > >> >>>
"options": {
> >> >>>> > >> > >> >>>
"connectivityCheck": false
> >> >>>> > >> > >> >>> }
> >> >>>> > >> > >> >>> }
> >> >>>> > >> > >> >>>
> >> >>>> > >> > >> >>> # issue setup
networks command:
> >> >>>> > >> > >> >>> [root@vdsm ~]$
vdsm-client -f bridged_net_with_stp
> Host setupNetworks
> >> >>>> > >> > >> >>> {
> >> >>>> > >> > >> >>>
"code": 0,
> >> >>>> > >> > >> >>>
"message": "Done"
> >> >>>> > >> > >> >>> }
> >> >>>> > >> > >> >>>
> >> >>>> > >> > >> >>> # show bridges
> >> >>>> > >> > >> >>> [root@vdsm ~]$
brctl show
> >> >>>> > >> > >> >>> bridge name bridge
id STP enabled interfaces
> >> >>>> > >> > >> >>> ;vdsmdummy;
8000.000000000000 no
> >> >>>> > >> > >> >>> test-network
8000.52540041fb37 yes eth0
> >> >>>> > >> > >> >>>
> >> >>>> > >> > >> >>> # show final
state
> >> >>>> > >> > >> >>> [root@vdsm ~]$ ip
a
> >> >>>> > >> > >> >>> 1: lo:
<LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc
> noqueue state UNKNOWN
> >> >>>> > >> > >> >>> group default qlen
1000
> >> >>>> > >> > >> >>> link/loopback
00:00:00:00:00:00 brd
> 00:00:00:00:00:00
> >> >>>> > >> > >> >>> inet
127.0.0.1/8 scope host lo
> >> >>>> > >> > >> >>> valid_lft
forever preferred_lft forever
> >> >>>> > >> > >> >>> 2: eth0:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500
> qdisc pfifo_fast
> >> >>>> > >> > >> >>> master
test-network state UP group default qlen 1000
> >> >>>> > >> > >> >>> link/ether
52:54:00:41:fb:37 brd
> ff:ff:ff:ff:ff:ff
> >> >>>> > >> > >> >>> 3: eth1:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500
> qdisc pfifo_fast
> >> >>>> > >> > >> >>> state UP group
default qlen 1000
> >> >>>> > >> > >> >>> link/ether
52:54:00:83:5b:6f brd
> ff:ff:ff:ff:ff:ff
> >> >>>> > >> > >> >>> inet
192.168.50.50/24 brd 192.168.50.255 scope
> global noprefixroute eth1
> >> >>>> > >> > >> >>> valid_lft
forever preferred_lft forever
> >> >>>> > >> > >> >>> inet6
fe80::5054:ff:fe83:5b6f/64 scope link
> >> >>>> > >> > >> >>> valid_lft
forever preferred_lft forever
> >> >>>> > >> > >> >>> 19: ;vdsmdummy;:
<BROADCAST,MULTICAST> mtu 1500
> qdisc noop state DOWN
> >> >>>> > >> > >> >>> group default qlen
1000
> >> >>>> > >> > >> >>> link/ether
8e:5c:2e:87:fa:0b brd
> ff:ff:ff:ff:ff:ff
> >> >>>> > >> > >> >>> 432:
test-network:
> <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
> >> >>>> > >> > >> >>> noqueue state UP
group default qlen 1000
> >> >>>> > >> > >> >>> link/ether
52:54:00:41:fb:37 brd
> ff:ff:ff:ff:ff:ff
> >> >>>> > >> > >> >>>
> >> >>>> > >> > >> >>> I don't think
this STP parameter is exposed via
> engine UI; @Dominik
> >> >>>> > >> > >> >>> Holler , could you
confirm ? What are our plans for
> it ?
> >> >>>> > >> > >> >>>
> >> >>>> > >> > >> >>
> >> >>>> > >> > >> >> STP is only available
via REST-API, see
> >> >>>> > >> > >> >>
>
http://ovirt.github.io/ovirt-engine-api-model/4.3/#types/network
> >> >>>> > >> > >> >> please find an example
how to enable STP in
> >> >>>> > >> > >> >>
>
https://gist.github.com/dominikholler/4e70c9ef9929d93b6807f56d43a70b95
> >> >>>> > >> > >> >>
> >> >>>> > >> > >> >> We have no plans to
add STP to the web ui,
> >> >>>> > >> > >> >> but new feature
requests are always welcome on
> >> >>>> > >> > >> >>
>
https://bugzilla.redhat.com/enter_bug.cgi?product=ovirt-engine
> >> >>>> > >> > >> >>
> >> >>>> > >> > >> >>
> >> >>>> > >> > >> >>>
> >> >>>> > >> > >> >>> >
> >> >>>> > >> > >> >>> > Here are some
details about my systems, if you
> need it.
> >> >>>> > >> > >> >>> >
> >> >>>> > >> > >> >>> >
> >> >>>> > >> > >> >>> > selinux is
disabled.
> >> >>>> > >> > >> >>> >
> >> >>>> > >> > >> >>> >
> >> >>>> > >> > >> >>> >
> >> >>>> > >> > >> >>> >
> >> >>>> > >> > >> >>> >
> >> >>>> > >> > >> >>> >
> >> >>>> > >> > >> >>> >
> >> >>>> > >> > >> >>> >
> >> >>>> > >> > >> >>> >
> >> >>>> > >> > >> >>> > [root@swm-02
~]# rpm -qa | grep ovirt
> >> >>>> > >> > >> >>> >
ovirt-imageio-common-1.5.1-0.el7.x86_64
> >> >>>> > >> > >> >>> >
ovirt-release43-4.3.5.2-1.el7.noarch
> >> >>>> > >> > >> >>> >
ovirt-imageio-daemon-1.5.1-0.el7.noarch
> >> >>>> > >> > >> >>> >
ovirt-vmconsole-host-1.0.7-2.el7.noarch
> >> >>>> > >> > >> >>> >
ovirt-hosted-engine-setup-2.3.11-1.el7.noarch
> >> >>>> > >> > >> >>> >
> ovirt-ansible-hosted-engine-setup-1.0.26-1.el7.noarch
> >> >>>> > >> > >> >>> >
python2-ovirt-host-deploy-1.8.0-1.el7.noarch
> >> >>>> > >> > >> >>> >
ovirt-ansible-engine-setup-1.1.9-1.el7.noarch
> >> >>>> > >> > >> >>> >
python2-ovirt-setup-lib-1.2.0-1.el7.noarch
> >> >>>> > >> > >> >>> >
cockpit-machines-ovirt-195.1-1.el7.noarch
> >> >>>> > >> > >> >>> >
ovirt-hosted-engine-ha-2.3.3-1.el7.noarch
> >> >>>> > >> > >> >>> >
ovirt-vmconsole-1.0.7-2.el7.noarch
> >> >>>> > >> > >> >>> >
cockpit-ovirt-dashboard-0.13.5-1.el7.noarch
> >> >>>> > >> > >> >>> >
ovirt-provider-ovn-driver-1.2.22-1.el7.noarch
> >> >>>> > >> > >> >>> >
ovirt-host-deploy-common-1.8.0-1.el7.noarch
> >> >>>> > >> > >> >>> >
ovirt-host-4.3.4-1.el7.x86_64
> >> >>>> > >> > >> >>> >
python-ovirt-engine-sdk4-4.3.2-2.el7.x86_64
> >> >>>> > >> > >> >>> >
ovirt-host-dependencies-4.3.4-1.el7.x86_64
> >> >>>> > >> > >> >>> >
ovirt-ansible-repositories-1.1.5-1.el7.noarch
> >> >>>> > >> > >> >>> > [root@swm-02
~]# cat /etc/redhat-release
> >> >>>> > >> > >> >>> > CentOS Linux
release 7.6.1810 (Core)
> >> >>>> > >> > >> >>> > [root@swm-02
~]# uname -r
> >> >>>> > >> > >> >>> >
3.10.0-957.27.2.el7.x86_64
> >> >>>> > >> > >> >>> > You have new
mail in /var/spool/mail/root
> >> >>>> > >> > >> >>> > [root@swm-02
~]# ip a
> >> >>>> > >> > >> >>> > 1: lo:
<LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc
> noqueue state UNKNOWN
> >> >>>> > >> > >> >>> > group default
qlen 1000
> >> >>>> > >> > >> >>> >
link/loopback 00:00:00:00:00:00 brd
> 00:00:00:00:00:00
> >> >>>> > >> > >> >>> > inet
127.0.0.1/8 scope host lo
> >> >>>> > >> > >> >>> >
valid_lft forever preferred_lft forever
> >> >>>> > >> > >> >>> > 2: em1:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu
> 1500 qdisc mq master
> >> >>>> > >> > >> >>> > test state UP
group default qlen 1000
> >> >>>> > >> > >> >>> >
link/ether d4:ae:52:8d:50:48 brd
> ff:ff:ff:ff:ff:ff
> >> >>>> > >> > >> >>> > 3: em2:
<BROADCAST,MULTICAST> mtu 1500 qdisc noop
> state DOWN group
> >> >>>> > >> > >> >>> > default qlen
1000
> >> >>>> > >> > >> >>> >
link/ether d4:ae:52:8d:50:49 brd
> ff:ff:ff:ff:ff:ff
> >> >>>> > >> > >> >>> > 4: p1p1:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu
> 1500 qdisc mq master
> >> >>>> > >> > >> >>> > ovirtmgmt
state UP group default qlen 1000
> >> >>>> > >> > >> >>> >
link/ether 90:e2:ba:1e:14:80 brd
> ff:ff:ff:ff:ff:ff
> >> >>>> > >> > >> >>> > 5: p1p2:
<BROADCAST,MULTICAST> mtu 1500 qdisc
> noop state DOWN group
> >> >>>> > >> > >> >>> > default qlen
1000
> >> >>>> > >> > >> >>> >
link/ether 90:e2:ba:1e:14:81 brd
> ff:ff:ff:ff:ff:ff
> >> >>>> > >> > >> >>> > 6:
ovs-system: <BROADCAST,MULTICAST> mtu 1500
> qdisc noop state DOWN
> >> >>>> > >> > >> >>> > group default
qlen 1000
> >> >>>> > >> > >> >>> >
link/ether a2:b8:d6:e8:b3:d8 brd
> ff:ff:ff:ff:ff:ff
> >> >>>> > >> > >> >>> > 7: br-int:
<BROADCAST,MULTICAST> mtu 1500 qdisc
> noop state DOWN group
> >> >>>> > >> > >> >>> > default qlen
1000
> >> >>>> > >> > >> >>> >
link/ether 96:a0:c1:4a:45:4b brd
> ff:ff:ff:ff:ff:ff
> >> >>>> > >> > >> >>> > 25: test:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu
> 1500 qdisc noqueue
> >> >>>> > >> > >> >>> > state UP
group default qlen 1000
> >> >>>> > >> > >> >>> >
link/ether d4:ae:52:8d:50:48 brd
> ff:ff:ff:ff:ff:ff
> >> >>>> > >> > >> >>> > inet
10.15.11.21/24 brd 10.15.11.255 scope
> global test
> >> >>>> > >> > >> >>> >
valid_lft forever preferred_lft forever
> >> >>>> > >> > >> >>> > 26:
ovirtmgmt: <BROADCAST,MULTICAST,UP,LOWER_UP>
> mtu 1500 qdisc
> >> >>>> > >> > >> >>> > noqueue state
UP group default qlen 1000
> >> >>>> > >> > >> >>> >
link/ether 90:e2:ba:1e:14:80 brd
> ff:ff:ff:ff:ff:ff
> >> >>>> > >> > >> >>> > inet
10.15.28.31/24 brd 10.15.28.255 scope
> global ovirtmgmt
> >> >>>> > >> > >> >>> >
valid_lft forever preferred_lft forever
> >> >>>> > >> > >> >>> > 27:
;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500
> qdisc noop state DOWN
> >> >>>> > >> > >> >>> > group default
qlen 1000
> >> >>>> > >> > >> >>> >
link/ether 62:e5:e5:07:99:eb brd
> ff:ff:ff:ff:ff:ff
> >> >>>> > >> > >> >>> > 29: vnet0:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu
> 1500 qdisc mq master
> >> >>>> > >> > >> >>> > ovirtmgmt
state UNKNOWN group default qlen 1000
> >> >>>> > >> > >> >>> >
link/ether fe:6f:9c:95:00:02 brd
> ff:ff:ff:ff:ff:ff
> >> >>>> > >> > >> >>> > [root@swm-02
~]# free -m
> >> >>>> > >> > >> >>> >
total used free
> shared buff/cache available
> >> >>>> > >> > >> >>> > Mem:
64413 1873 61804
> 9 735 62062
> >> >>>> > >> > >> >>> > Swap:
16383 0 16383
> >> >>>> > >> > >> >>> > [root@swm-02
~]# free -h
> >> >>>> > >> > >> >>> >
total used free
> shared buff/cache available
> >> >>>> > >> > >> >>> > Mem:
62G 1.8G 60G
> 9.5M 735M 60G
> >> >>>> > >> > >> >>> > Swap:
15G 0B 15G
> >> >>>> > >> > >> >>> > [root@swm-02
~]# ls
> >> >>>> > >> > >> >>> > ls
lsb_release lshw
> lslocks
> >> >>>> > >> > >> >>> >
lsmod lspci
> lssubsys
> >> >>>> > >> > >> >>> > lsusb.py
> >> >>>> > >> > >> >>> > lsattr
lscgroup lsinitrd
> lslogins
> >> >>>> > >> > >> >>> > lsns
lss16toppm
> lstopo-no-graphics
> >> >>>> > >> > >> >>> > lsblk
lscpu lsipc
> lsmem
> >> >>>> > >> > >> >>> > lsof
lsscsi
> lsusb
> >> >>>> > >> > >> >>> > [root@swm-02
~]# lscpu
> >> >>>> > >> > >> >>> > Architecture:
x86_64
> >> >>>> > >> > >> >>> > CPU
op-mode(s): 32-bit, 64-bit
> >> >>>> > >> > >> >>> > Byte Order:
Little Endian
> >> >>>> > >> > >> >>> > CPU(s):
16
> >> >>>> > >> > >> >>> > On-line
CPU(s) list: 0-15
> >> >>>> > >> > >> >>> > Thread(s) per
core: 2
> >> >>>> > >> > >> >>> > Core(s) per
socket: 4
> >> >>>> > >> > >> >>> > Socket(s):
2
> >> >>>> > >> > >> >>> > NUMA node(s):
2
> >> >>>> > >> > >> >>> > Vendor ID:
GenuineIntel
> >> >>>> > >> > >> >>> > CPU family:
6
> >> >>>> > >> > >> >>> > Model:
44
> >> >>>> > >> > >> >>> > Model name:
Intel(R) Xeon(R) CPU
> X5672 @ 3.20GHz
> >> >>>> > >> > >> >>> > Stepping:
2
> >> >>>> > >> > >> >>> > CPU MHz:
3192.064
> >> >>>> > >> > >> >>> > BogoMIPS:
6384.12
> >> >>>> > >> > >> >>> >
Virtualization: VT-x
> >> >>>> > >> > >> >>> > L1d cache:
32K
> >> >>>> > >> > >> >>> > L1i cache:
32K
> >> >>>> > >> > >> >>> > L2 cache:
256K
> >> >>>> > >> > >> >>> > L3 cache:
12288K
> >> >>>> > >> > >> >>> > NUMA node0
CPU(s): 0,2,4,6,8,10,12,14
> >> >>>> > >> > >> >>> > NUMA node1
CPU(s): 1,3,5,7,9,11,13,15
> >> >>>> > >> > >> >>> > Flags:
fpu vme de pse tsc msr pae
> mce cx8 apic sep
> >> >>>> > >> > >> >>> > mtrr pge mca
cmov pat pse36 clflush dts acpi mmx
> fxsr sse sse2 ss ht
> >> >>>> > >> > >> >>> > tm pbe
syscall nx pdpe1gb rdtscp lm constant_tsc
> arch_perfmon pebs bts
> >> >>>> > >> > >> >>> > rep_good nopl
xtopology nonstop_tsc aperfmperf
> eagerfpu pni pclmulqdq
> >> >>>> > >> > >> >>> > dtes64
monitor ds_cpl vmx smx est tm2 ssse3 cx16
> xtpr pdcm pcid dca
> >> >>>> > >> > >> >>> > sse4_1 sse4_2
popcnt aes lahf_lm ssbd ibrs ibpb
> stibp tpr_shadow vnmi
> >> >>>> > >> > >> >>> > flexpriority
ept vpid dtherm ida arat spec_ctrl
> intel_stibp flush_l1d
> >> >>>> > >> > >> >>> > [root@swm-02
~]#
> >> >>>> > >> > >> >>> >
_______________________________________________
> >> >>>> > >> > >> >>> > Users mailing
list -- users(a)ovirt.org
> >> >>>> > >> > >> >>> > To
unsubscribe send an email to
> users-leave(a)ovirt.org
> >> >>>> > >> > >> >>> > Privacy
Statement:
>
https://www.ovirt.org/site/privacy-policy/
> >> >>>> > >> > >> >>> > oVirt Code of
Conduct:
>
https://www.ovirt.org/community/about/community-guidelines/
> >> >>>> > >> > >> >>> > List
Archives:
>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/MTMZ5MF4CF2...
> >> >>>> > >> > >> >>
> >> >>>> > >> > >> >>
_______________________________________________
> >> >>>> > >> > >> >> Users mailing list --
users(a)ovirt.org
> >> >>>> > >> > >> >> To unsubscribe send an
email to
> users-leave(a)ovirt.org
> >> >>>> > >> > >> >> Privacy Statement:
>
https://www.ovirt.org/site/privacy-policy/
> >> >>>> > >> > >> >> oVirt Code of
Conduct:
>
https://www.ovirt.org/community/about/community-guidelines/
> >> >>>> > >> > >> >> List Archives:
>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/QBA7NYKAJNR...
>