On Fri, Aug 23, 2019 at 5:49 PM Curtis E. Combs Jr. <ej.albany(a)gmail.com>
wrote:
Sure! Right now, I only have a 500gb partition on each node shared
over
NFS, added as storage domains. This is on each node - so, currently 3.
How can the storage cause a node to drop out?
Thanks, I got it.
All three links go down on load, which causes NFS to fail.
Can you check in the switch port configuration if there is some kind of
Ethernet flow control enabled?
Can you try to modify the behavior by changing the settings of your host
interfaces, e.g.
ethtool -A em1 autoneg off rx off tx off
or
ethtool -A em1 autoneg on rx on tx on
?
On Fri, Aug 23, 2019, 11:46 AM Dominik Holler
<dholler(a)redhat.com> wrote:
>
>
> On Fri, Aug 23, 2019 at 5:41 PM Curtis E. Combs Jr. <ej.albany(a)gmail.com>
> wrote:
>
>> Also, if it helps, the hosts will sit there, quietly, for hours or
>> days before anything happens. They're up and working just fine. But
>> then, when I manually migrate a VM from one host to another, they
>> become completely inaccessible.
>>
>>
> Can you share some details about your storage?
> Maybe there is a feature used during live migration, which triggers the
> issue.
>
>
>
>> These are vanilla-as-possible CentOS7 nodes. Very basic ovirt install
>> and configuration.
>>
>> On Fri, Aug 23, 2019 at 11:33 AM Curtis E. Combs Jr.
>> <ej.albany(a)gmail.com> wrote:
>> >
>> > Hey Dominik,
>> >
>> > Thanks for helping. I really want to try to use ovirt.
>> >
>> > When these events happen, I cannot even SSH to the nodes due to the
>> > link being down. After a little while, the hosts come back...
>> >
>> > On Fri, Aug 23, 2019 at 11:30 AM Dominik Holler <dholler(a)redhat.com>
>> wrote:
>> > >
>> > > Is you storage connected via NFS?
>> > > Can you manually access the storage on the host?
>> > >
>> > >
>> > > On Fri, Aug 23, 2019 at 5:19 PM Curtis E. Combs Jr. <
>> ej.albany(a)gmail.com> wrote:
>> > >>
>> > >> Sorry to dead bump this, but I'm beginning to suspect that
maybe
>> it's
>> > >> not STP that's the problem.
>> > >>
>> > >> 2 of my hosts just went down when a few VMs tried to migrate.
>> > >>
>> > >> Do any of you have any idea what might be going on here? I
don't
>> even
>> > >> know where to start. I'm going to include the dmesg in case it
>> helps.
>> > >> This happens on both of the hosts whenever any migration attempts
>> to start.
>> > >>
>> > >>
>> > >>
>> > >>
>> > >>
>> > >>
>> > >>
>> > >>
>> > >>
>> > >> [68099.245833] bnx2 0000:01:00.0 em1: NIC Copper Link is Down
>> > >> [68099.246055] internal: port 1(em1) entered disabled state
>> > >> [68184.177343] ixgbe 0000:03:00.0 p1p1: NIC Link is Down
>> > >> [68184.177789] ovirtmgmt: port 1(p1p1) entered disabled state
>> > >> [68184.177856] ovirtmgmt: topology change detected, propagating
>> > >> [68277.078671] INFO: task qemu-kvm:8888 blocked for more than 120
>> seconds.
>> > >> [68277.078700] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
>> > >> disables this message.
>> > >> [68277.078723] qemu-kvm D ffff9db40c359040 0 8888
>> 1 0x000001a0
>> > >> [68277.078727] Call Trace:
>> > >> [68277.078738] [<ffffffff978fd2ac>] ?
avc_has_perm_flags+0xdc/0x1c0
>> > >> [68277.078743] [<ffffffff97d69f19>] schedule+0x29/0x70
>> > >> [68277.078746] [<ffffffff9785f3d9>]
inode_dio_wait+0xd9/0x100
>> > >> [68277.078751] [<ffffffff976c4010>] ?
wake_bit_function+0x40/0x40
>> > >> [68277.078765] [<ffffffffc09d6dd6>] nfs_getattr+0x1b6/0x250
[nfs]
>> > >> [68277.078768] [<ffffffff97848109>] vfs_getattr+0x49/0x80
>> > >> [68277.078769] [<ffffffff97848185>] vfs_fstat+0x45/0x80
>> > >> [68277.078771] [<ffffffff978486f4>] SYSC_newfstat+0x24/0x60
>> > >> [68277.078774] [<ffffffff97d76d21>] ?
>> system_call_after_swapgs+0xae/0x146
>> > >> [68277.078778] [<ffffffff97739f34>] ?
>> __audit_syscall_entry+0xb4/0x110
>> > >> [68277.078782] [<ffffffff9763aaeb>] ?
>> syscall_trace_enter+0x16b/0x220
>> > >> [68277.078784] [<ffffffff97848ace>] SyS_newfstat+0xe/0x10
>> > >> [68277.078786] [<ffffffff97d7706b>] tracesys+0xa3/0xc9
>> > >> [68397.072384] INFO: task qemu-kvm:8888 blocked for more than 120
>> seconds.
>> > >> [68397.072413] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
>> > >> disables this message.
>> > >> [68397.072436] qemu-kvm D ffff9db40c359040 0 8888
>> 1 0x000001a0
>> > >> [68397.072439] Call Trace:
>> > >> [68397.072453] [<ffffffff978fd2ac>] ?
avc_has_perm_flags+0xdc/0x1c0
>> > >> [68397.072458] [<ffffffff97d69f19>] schedule+0x29/0x70
>> > >> [68397.072462] [<ffffffff9785f3d9>]
inode_dio_wait+0xd9/0x100
>> > >> [68397.072467] [<ffffffff976c4010>] ?
wake_bit_function+0x40/0x40
>> > >> [68397.072480] [<ffffffffc09d6dd6>] nfs_getattr+0x1b6/0x250
[nfs]
>> > >> [68397.072485] [<ffffffff97848109>] vfs_getattr+0x49/0x80
>> > >> [68397.072486] [<ffffffff97848185>] vfs_fstat+0x45/0x80
>> > >> [68397.072488] [<ffffffff978486f4>] SYSC_newfstat+0x24/0x60
>> > >> [68397.072491] [<ffffffff97d76d21>] ?
>> system_call_after_swapgs+0xae/0x146
>> > >> [68397.072495] [<ffffffff97739f34>] ?
>> __audit_syscall_entry+0xb4/0x110
>> > >> [68397.072498] [<ffffffff9763aaeb>] ?
>> syscall_trace_enter+0x16b/0x220
>> > >> [68397.072500] [<ffffffff97848ace>] SyS_newfstat+0xe/0x10
>> > >> [68397.072502] [<ffffffff97d7706b>] tracesys+0xa3/0xc9
>> > >> [68401.573141] bnx2 0000:01:00.0 em1: NIC Copper Link is Up, 1000
>> Mbps
>> > >> full duplex
>> > >>
>> > >> [68401.573247] internal: port 1(em1) entered blocking state
>> > >> [68401.573255] internal: port 1(em1) entered listening state
>> > >> [68403.576985] internal: port 1(em1) entered learning state
>> > >> [68405.580907] internal: port 1(em1) entered forwarding state
>> > >> [68405.580916] internal: topology change detected, propagating
>> > >> [68469.565589] nfs: server
swm-01.hpc.moffitt.org not responding,
>> timed out
>> > >> [68469.565840] nfs: server
swm-01.hpc.moffitt.org not responding,
>> timed out
>> > >> [68487.193932] ixgbe 0000:03:00.0 p1p1: NIC Link is Up 10 Gbps,
Flow
>> > >> Control: RX/TX
>> > >> [68487.194105] ovirtmgmt: port 1(p1p1) entered blocking state
>> > >> [68487.194114] ovirtmgmt: port 1(p1p1) entered listening state
>> > >> [68489.196508] ovirtmgmt: port 1(p1p1) entered learning state
>> > >> [68491.200400] ovirtmgmt: port 1(p1p1) entered forwarding state
>> > >> [68491.200405] ovirtmgmt: topology change detected, sending tcn
bpdu
>> > >> [68493.672423] NFS: nfs4_reclaim_open_state: Lock reclaim failed!
>> > >> [68494.777996] NFSD: client 10.15.28.22 testing state ID with
>> > >> incorrect client ID
>> > >> [68494.778580] NFSD: client 10.15.28.22 testing state ID with
>> > >> incorrect client ID
>> > >>
>> > >>
>> > >> On Thu, Aug 22, 2019 at 2:53 PM Curtis E. Combs Jr. <
>> ej.albany(a)gmail.com> wrote:
>> > >> >
>> > >> > Thanks, I'm just going to revert back to bridges.
>> > >> >
>> > >> > On Thu, Aug 22, 2019 at 11:50 AM Dominik Holler <
>> dholler(a)redhat.com> wrote:
>> > >> > >
>> > >> > >
>> > >> > >
>> > >> > > On Thu, Aug 22, 2019 at 3:06 PM Curtis E. Combs Jr. <
>> ej.albany(a)gmail.com> wrote:
>> > >> > >>
>> > >> > >> Seems like the STP options are so common and
necessary that it
>> would
>> > >> > >> be a priority over seldom-used bridge_opts. I know
what STP is
>> and I'm
>> > >> > >> not even a networking guy - never even heard of half
of the
>> > >> > >> bridge_opts that have switches in the UI.
>> > >> > >>
>> > >> > >> Anyway. I wanted to try the openvswitches, so I
reinstalled
>> all of my
>> > >> > >> nodes and used "openvswitch (Technology
Preview)" as the
>> engine-setup
>> > >> > >> option for the first host. I made a new Cluster for
my nodes,
>> added
>> > >> > >> them all to the new cluster, created a new
"logical network"
>> for the
>> > >> > >> internal network and attached it to the internal
network ports.
>> > >> > >>
>> > >> > >> Now, when I go to create a new VM, I don't even
have either the
>> > >> > >> ovirtmgmt switch OR the internal switch as an option.
The
>> drop-down is
>> > >> > >> empy as if I don't have any vnic-profiles.
>> > >> > >>
>> > >> > >
>> > >> > > openvswitch clusters are limited to ovn networks.
>> > >> > > You can create one like described in
>> > >> > >
>>
https://www.ovirt.org/documentation/admin-guide/chap-External_Providers.h...
>> > >> > >
>> > >> > >
>> > >> > >>
>> > >> > >> On Thu, Aug 22, 2019 at 7:34 AM Tony Pearce
<tonyppe(a)gmail.com>
>> wrote:
>> > >> > >> >
>> > >> > >> > Hi Dominik, would you mind sharing the use case
for stp via
>> API Only? I am keen to know this.
>> > >> > >> > Thanks
>> > >> > >> >
>> > >> > >> >
>> > >> > >> > On Thu., 22 Aug. 2019, 19:24 Dominik Holler,
<
>> dholler(a)redhat.com> wrote:
>> > >> > >> >>
>> > >> > >> >>
>> > >> > >> >>
>> > >> > >> >> On Thu, Aug 22, 2019 at 1:08 PM Miguel
Duarte de Mora
>> Barroso <mdbarroso(a)redhat.com> wrote:
>> > >> > >> >>>
>> > >> > >> >>> On Sat, Aug 17, 2019 at 11:27 AM
<ej.albany(a)gmail.com>
>> wrote:
>> > >> > >> >>> >
>> > >> > >> >>> > Hello. I have been trying to figure
out an issue for a
>> very long time.
>> > >> > >> >>> > That issue relates to the ethernet
and 10gb fc links
>> that I have on my
>> > >> > >> >>> > cluster being disabled any time a
migration occurs.
>> > >> > >> >>> >
>> > >> > >> >>> > I believe this is because I need to
have STP turned on
>> in order to
>> > >> > >> >>> > participate with the switch.
However, there does not
>> seem to be any
>> > >> > >> >>> > way to tell oVirt to stop turning
it off! Very
>> frustrating.
>> > >> > >> >>> >
>> > >> > >> >>> > After entering a cronjob that
enables stp on all bridges
>> every 1
>> > >> > >> >>> > minute, the migration issue
disappears....
>> > >> > >> >>> >
>> > >> > >> >>> > Is there any way at all to do
without this cronjob and
>> set STP to be
>> > >> > >> >>> > ON without having to resort to such
a silly solution?
>> > >> > >> >>>
>> > >> > >> >>> Vdsm exposes a per bridge STP knob that
you can use for
>> this. By
>> > >> > >> >>> default it is set to false, which is
probably why you had
>> to use this
>> > >> > >> >>> shenanigan.
>> > >> > >> >>>
>> > >> > >> >>> You can, for instance:
>> > >> > >> >>>
>> > >> > >> >>> # show present state
>> > >> > >> >>> [vagrant@vdsm ~]$ ip a
>> > >> > >> >>> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu
65536 qdisc noqueue
>> state UNKNOWN
>> > >> > >> >>> group default qlen 1000
>> > >> > >> >>> link/loopback 00:00:00:00:00:00 brd
00:00:00:00:00:00
>> > >> > >> >>> inet 127.0.0.1/8 scope host lo
>> > >> > >> >>> valid_lft forever preferred_lft
forever
>> > >> > >> >>> 2: eth0:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>> pfifo_fast
>> > >> > >> >>> state UP group default qlen 1000
>> > >> > >> >>> link/ether 52:54:00:41:fb:37 brd
ff:ff:ff:ff:ff:ff
>> > >> > >> >>> 3: eth1:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>> pfifo_fast
>> > >> > >> >>> state UP group default qlen 1000
>> > >> > >> >>> link/ether 52:54:00:83:5b:6f brd
ff:ff:ff:ff:ff:ff
>> > >> > >> >>> inet 192.168.50.50/24 brd
192.168.50.255 scope global
>> noprefixroute eth1
>> > >> > >> >>> valid_lft forever preferred_lft
forever
>> > >> > >> >>> inet6 fe80::5054:ff:fe83:5b6f/64
scope link
>> > >> > >> >>> valid_lft forever preferred_lft
forever
>> > >> > >> >>> 19: ;vdsmdummy;:
<BROADCAST,MULTICAST> mtu 1500 qdisc noop
>> state DOWN
>> > >> > >> >>> group default qlen 1000
>> > >> > >> >>> link/ether 8e:5c:2e:87:fa:0b brd
ff:ff:ff:ff:ff:ff
>> > >> > >> >>>
>> > >> > >> >>> # show example bridge configuration -
you're looking for
>> the STP knob here.
>> > >> > >> >>> [root@vdsm ~]$ cat bridged_net_with_stp
>> > >> > >> >>> {
>> > >> > >> >>> "bondings": {},
>> > >> > >> >>> "networks": {
>> > >> > >> >>> "test-network": {
>> > >> > >> >>> "nic":
"eth0",
>> > >> > >> >>> "switch":
"legacy",
>> > >> > >> >>> "bridged": true,
>> > >> > >> >>> "stp": true
>> > >> > >> >>> }
>> > >> > >> >>> },
>> > >> > >> >>> "options": {
>> > >> > >> >>> "connectivityCheck":
false
>> > >> > >> >>> }
>> > >> > >> >>> }
>> > >> > >> >>>
>> > >> > >> >>> # issue setup networks command:
>> > >> > >> >>> [root@vdsm ~]$ vdsm-client -f
bridged_net_with_stp Host
>> setupNetworks
>> > >> > >> >>> {
>> > >> > >> >>> "code": 0,
>> > >> > >> >>> "message":
"Done"
>> > >> > >> >>> }
>> > >> > >> >>>
>> > >> > >> >>> # show bridges
>> > >> > >> >>> [root@vdsm ~]$ brctl show
>> > >> > >> >>> bridge name bridge id STP enabled
interfaces
>> > >> > >> >>> ;vdsmdummy; 8000.000000000000 no
>> > >> > >> >>> test-network 8000.52540041fb37 yes eth0
>> > >> > >> >>>
>> > >> > >> >>> # show final state
>> > >> > >> >>> [root@vdsm ~]$ ip a
>> > >> > >> >>> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu
65536 qdisc noqueue
>> state UNKNOWN
>> > >> > >> >>> group default qlen 1000
>> > >> > >> >>> link/loopback 00:00:00:00:00:00 brd
00:00:00:00:00:00
>> > >> > >> >>> inet 127.0.0.1/8 scope host lo
>> > >> > >> >>> valid_lft forever preferred_lft
forever
>> > >> > >> >>> 2: eth0:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>> pfifo_fast
>> > >> > >> >>> master test-network state UP group
default qlen 1000
>> > >> > >> >>> link/ether 52:54:00:41:fb:37 brd
ff:ff:ff:ff:ff:ff
>> > >> > >> >>> 3: eth1:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>> pfifo_fast
>> > >> > >> >>> state UP group default qlen 1000
>> > >> > >> >>> link/ether 52:54:00:83:5b:6f brd
ff:ff:ff:ff:ff:ff
>> > >> > >> >>> inet 192.168.50.50/24 brd
192.168.50.255 scope global
>> noprefixroute eth1
>> > >> > >> >>> valid_lft forever preferred_lft
forever
>> > >> > >> >>> inet6 fe80::5054:ff:fe83:5b6f/64
scope link
>> > >> > >> >>> valid_lft forever preferred_lft
forever
>> > >> > >> >>> 19: ;vdsmdummy;:
<BROADCAST,MULTICAST> mtu 1500 qdisc noop
>> state DOWN
>> > >> > >> >>> group default qlen 1000
>> > >> > >> >>> link/ether 8e:5c:2e:87:fa:0b brd
ff:ff:ff:ff:ff:ff
>> > >> > >> >>> 432: test-network:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu
>> 1500 qdisc
>> > >> > >> >>> noqueue state UP group default qlen
1000
>> > >> > >> >>> link/ether 52:54:00:41:fb:37 brd
ff:ff:ff:ff:ff:ff
>> > >> > >> >>>
>> > >> > >> >>> I don't think this STP parameter is
exposed via engine UI;
>> @Dominik
>> > >> > >> >>> Holler , could you confirm ? What are
our plans for it ?
>> > >> > >> >>>
>> > >> > >> >>
>> > >> > >> >> STP is only available via REST-API, see
>> > >> > >> >>
>>
http://ovirt.github.io/ovirt-engine-api-model/4.3/#types/network
>> > >> > >> >> please find an example how to enable STP in
>> > >> > >> >>
>>
https://gist.github.com/dominikholler/4e70c9ef9929d93b6807f56d43a70b95
>> > >> > >> >>
>> > >> > >> >> We have no plans to add STP to the web ui,
>> > >> > >> >> but new feature requests are always welcome
on
>> > >> > >> >>
>>
https://bugzilla.redhat.com/enter_bug.cgi?product=ovirt-engine
>> > >> > >> >>
>> > >> > >> >>
>> > >> > >> >>>
>> > >> > >> >>> >
>> > >> > >> >>> > Here are some details about my
systems, if you need it.
>> > >> > >> >>> >
>> > >> > >> >>> >
>> > >> > >> >>> > selinux is disabled.
>> > >> > >> >>> >
>> > >> > >> >>> >
>> > >> > >> >>> >
>> > >> > >> >>> >
>> > >> > >> >>> >
>> > >> > >> >>> >
>> > >> > >> >>> >
>> > >> > >> >>> >
>> > >> > >> >>> >
>> > >> > >> >>> > [root@swm-02 ~]# rpm -qa | grep
ovirt
>> > >> > >> >>> >
ovirt-imageio-common-1.5.1-0.el7.x86_64
>> > >> > >> >>> >
ovirt-release43-4.3.5.2-1.el7.noarch
>> > >> > >> >>> >
ovirt-imageio-daemon-1.5.1-0.el7.noarch
>> > >> > >> >>> >
ovirt-vmconsole-host-1.0.7-2.el7.noarch
>> > >> > >> >>> >
ovirt-hosted-engine-setup-2.3.11-1.el7.noarch
>> > >> > >> >>> >
ovirt-ansible-hosted-engine-setup-1.0.26-1.el7.noarch
>> > >> > >> >>> >
python2-ovirt-host-deploy-1.8.0-1.el7.noarch
>> > >> > >> >>> >
ovirt-ansible-engine-setup-1.1.9-1.el7.noarch
>> > >> > >> >>> >
python2-ovirt-setup-lib-1.2.0-1.el7.noarch
>> > >> > >> >>> >
cockpit-machines-ovirt-195.1-1.el7.noarch
>> > >> > >> >>> >
ovirt-hosted-engine-ha-2.3.3-1.el7.noarch
>> > >> > >> >>> > ovirt-vmconsole-1.0.7-2.el7.noarch
>> > >> > >> >>> >
cockpit-ovirt-dashboard-0.13.5-1.el7.noarch
>> > >> > >> >>> >
ovirt-provider-ovn-driver-1.2.22-1.el7.noarch
>> > >> > >> >>> >
ovirt-host-deploy-common-1.8.0-1.el7.noarch
>> > >> > >> >>> > ovirt-host-4.3.4-1.el7.x86_64
>> > >> > >> >>> >
python-ovirt-engine-sdk4-4.3.2-2.el7.x86_64
>> > >> > >> >>> >
ovirt-host-dependencies-4.3.4-1.el7.x86_64
>> > >> > >> >>> >
ovirt-ansible-repositories-1.1.5-1.el7.noarch
>> > >> > >> >>> > [root@swm-02 ~]# cat
/etc/redhat-release
>> > >> > >> >>> > CentOS Linux release 7.6.1810
(Core)
>> > >> > >> >>> > [root@swm-02 ~]# uname -r
>> > >> > >> >>> > 3.10.0-957.27.2.el7.x86_64
>> > >> > >> >>> > You have new mail in
/var/spool/mail/root
>> > >> > >> >>> > [root@swm-02 ~]# ip a
>> > >> > >> >>> > 1: lo: <LOOPBACK,UP,LOWER_UP>
mtu 65536 qdisc noqueue
>> state UNKNOWN
>> > >> > >> >>> > group default qlen 1000
>> > >> > >> >>> > link/loopback 00:00:00:00:00:00
brd 00:00:00:00:00:00
>> > >> > >> >>> > inet 127.0.0.1/8 scope host lo
>> > >> > >> >>> > valid_lft forever
preferred_lft forever
>> > >> > >> >>> > 2: em1:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
>> mq master
>> > >> > >> >>> > test state UP group default qlen
1000
>> > >> > >> >>> > link/ether d4:ae:52:8d:50:48
brd ff:ff:ff:ff:ff:ff
>> > >> > >> >>> > 3: em2: <BROADCAST,MULTICAST>
mtu 1500 qdisc noop state
>> DOWN group
>> > >> > >> >>> > default qlen 1000
>> > >> > >> >>> > link/ether d4:ae:52:8d:50:49
brd ff:ff:ff:ff:ff:ff
>> > >> > >> >>> > 4: p1p1:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500
>> qdisc mq master
>> > >> > >> >>> > ovirtmgmt state UP group default
qlen 1000
>> > >> > >> >>> > link/ether 90:e2:ba:1e:14:80
brd ff:ff:ff:ff:ff:ff
>> > >> > >> >>> > 5: p1p2:
<BROADCAST,MULTICAST> mtu 1500 qdisc noop state
>> DOWN group
>> > >> > >> >>> > default qlen 1000
>> > >> > >> >>> > link/ether 90:e2:ba:1e:14:81
brd ff:ff:ff:ff:ff:ff
>> > >> > >> >>> > 6: ovs-system:
<BROADCAST,MULTICAST> mtu 1500 qdisc noop
>> state DOWN
>> > >> > >> >>> > group default qlen 1000
>> > >> > >> >>> > link/ether a2:b8:d6:e8:b3:d8
brd ff:ff:ff:ff:ff:ff
>> > >> > >> >>> > 7: br-int:
<BROADCAST,MULTICAST> mtu 1500 qdisc noop
>> state DOWN group
>> > >> > >> >>> > default qlen 1000
>> > >> > >> >>> > link/ether 96:a0:c1:4a:45:4b
brd ff:ff:ff:ff:ff:ff
>> > >> > >> >>> > 25: test:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500
>> qdisc noqueue
>> > >> > >> >>> > state UP group default qlen 1000
>> > >> > >> >>> > link/ether d4:ae:52:8d:50:48
brd ff:ff:ff:ff:ff:ff
>> > >> > >> >>> > inet 10.15.11.21/24 brd
10.15.11.255 scope global
>> test
>> > >> > >> >>> > valid_lft forever
preferred_lft forever
>> > >> > >> >>> > 26: ovirtmgmt:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu
>> 1500 qdisc
>> > >> > >> >>> > noqueue state UP group default qlen
1000
>> > >> > >> >>> > link/ether 90:e2:ba:1e:14:80
brd ff:ff:ff:ff:ff:ff
>> > >> > >> >>> > inet 10.15.28.31/24 brd
10.15.28.255 scope global
>> ovirtmgmt
>> > >> > >> >>> > valid_lft forever
preferred_lft forever
>> > >> > >> >>> > 27: ;vdsmdummy;:
<BROADCAST,MULTICAST> mtu 1500 qdisc
>> noop state DOWN
>> > >> > >> >>> > group default qlen 1000
>> > >> > >> >>> > link/ether 62:e5:e5:07:99:eb
brd ff:ff:ff:ff:ff:ff
>> > >> > >> >>> > 29: vnet0:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500
>> qdisc mq master
>> > >> > >> >>> > ovirtmgmt state UNKNOWN group
default qlen 1000
>> > >> > >> >>> > link/ether fe:6f:9c:95:00:02
brd ff:ff:ff:ff:ff:ff
>> > >> > >> >>> > [root@swm-02 ~]# free -m
>> > >> > >> >>> > total used
free shared
>> buff/cache available
>> > >> > >> >>> > Mem: 64413 1873
61804 9
>> 735 62062
>> > >> > >> >>> > Swap: 16383 0
16383
>> > >> > >> >>> > [root@swm-02 ~]# free -h
>> > >> > >> >>> > total used
free shared
>> buff/cache available
>> > >> > >> >>> > Mem: 62G 1.8G
60G 9.5M
>> 735M 60G
>> > >> > >> >>> > Swap: 15G 0B
15G
>> > >> > >> >>> > [root@swm-02 ~]# ls
>> > >> > >> >>> > ls lsb_release
lshw
>> lslocks
>> > >> > >> >>> > lsmod lspci
lssubsys
>> > >> > >> >>> > lsusb.py
>> > >> > >> >>> > lsattr lscgroup
lsinitrd
>> lslogins
>> > >> > >> >>> > lsns
lss16toppm
>> lstopo-no-graphics
>> > >> > >> >>> > lsblk lscpu
lsipc
>> lsmem
>> > >> > >> >>> > lsof lsscsi
lsusb
>> > >> > >> >>> > [root@swm-02 ~]# lscpu
>> > >> > >> >>> > Architecture: x86_64
>> > >> > >> >>> > CPU op-mode(s): 32-bit,
64-bit
>> > >> > >> >>> > Byte Order: Little
Endian
>> > >> > >> >>> > CPU(s): 16
>> > >> > >> >>> > On-line CPU(s) list: 0-15
>> > >> > >> >>> > Thread(s) per core: 2
>> > >> > >> >>> > Core(s) per socket: 4
>> > >> > >> >>> > Socket(s): 2
>> > >> > >> >>> > NUMA node(s): 2
>> > >> > >> >>> > Vendor ID:
GenuineIntel
>> > >> > >> >>> > CPU family: 6
>> > >> > >> >>> > Model: 44
>> > >> > >> >>> > Model name: Intel(R)
Xeon(R) CPU
>> X5672 @ 3.20GHz
>> > >> > >> >>> > Stepping: 2
>> > >> > >> >>> > CPU MHz: 3192.064
>> > >> > >> >>> > BogoMIPS: 6384.12
>> > >> > >> >>> > Virtualization: VT-x
>> > >> > >> >>> > L1d cache: 32K
>> > >> > >> >>> > L1i cache: 32K
>> > >> > >> >>> > L2 cache: 256K
>> > >> > >> >>> > L3 cache: 12288K
>> > >> > >> >>> > NUMA node0 CPU(s):
0,2,4,6,8,10,12,14
>> > >> > >> >>> > NUMA node1 CPU(s):
1,3,5,7,9,11,13,15
>> > >> > >> >>> > Flags: fpu vme de
pse tsc msr pae mce
>> cx8 apic sep
>> > >> > >> >>> > mtrr pge mca cmov pat pse36 clflush
dts acpi mmx fxsr
>> sse sse2 ss ht
>> > >> > >> >>> > tm pbe syscall nx pdpe1gb rdtscp lm
constant_tsc
>> arch_perfmon pebs bts
>> > >> > >> >>> > rep_good nopl xtopology nonstop_tsc
aperfmperf eagerfpu
>> pni pclmulqdq
>> > >> > >> >>> > dtes64 monitor ds_cpl vmx smx est
tm2 ssse3 cx16 xtpr
>> pdcm pcid dca
>> > >> > >> >>> > sse4_1 sse4_2 popcnt aes lahf_lm
ssbd ibrs ibpb stibp
>> tpr_shadow vnmi
>> > >> > >> >>> > flexpriority ept vpid dtherm ida
arat spec_ctrl
>> intel_stibp flush_l1d
>> > >> > >> >>> > [root@swm-02 ~]#
>> > >> > >> >>> >
_______________________________________________
>> > >> > >> >>> > Users mailing list --
users(a)ovirt.org
>> > >> > >> >>> > To unsubscribe send an email to
users-leave(a)ovirt.org
>> > >> > >> >>> > Privacy Statement:
>>
https://www.ovirt.org/site/privacy-policy/
>> > >> > >> >>> > oVirt Code of Conduct:
>>
https://www.ovirt.org/community/about/community-guidelines/
>> > >> > >> >>> > List Archives:
>>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/MTMZ5MF4CF2...
>> > >> > >> >>
>> > >> > >> >>
_______________________________________________
>> > >> > >> >> Users mailing list -- users(a)ovirt.org
>> > >> > >> >> To unsubscribe send an email to
users-leave(a)ovirt.org
>> > >> > >> >> Privacy Statement:
>>
https://www.ovirt.org/site/privacy-policy/
>> > >> > >> >> oVirt Code of Conduct:
>>
https://www.ovirt.org/community/about/community-guidelines/
>> > >> > >> >> List Archives:
>>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/QBA7NYKAJNR...
>>
>