[ovirt-users] jumbo frames inside vm best practices?
Edward Haas
ehaas at redhat.com
Mon Feb 6 09:16:47 UTC 2017
Your described setup seems correct.
Please attempt to isolate the issue by trying to pass traffic between the
hosts, taking the VM/s out of the equation.
You may also consider connecting the hosts directly to each other, to make
sure this is not a switch problem.
Thanks,
Edy.
On Mon, Feb 6, 2017 at 1:50 AM, Gianluca Cecchi <gianluca.cecchi at gmail.com>
wrote:
> Hello,
> I'm testing an Oracle RAC with 2 Oracle Linux VMs inside a 4.0.6
> environment.
> They run on two different hosts
> I would like to configure RAC intracluster communication with jumbo frames.
> At VM level network adapter is eth1 (mapped to a vlan 95 at oVirt hosts
> side)
> At oVirt side I configured a vm enabled vlan with mtu=9000
> I verified that at hosts side I have
>
> vlan95: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 9000
> ether 00:1c:c4:ab:be:ba txqueuelen 1000 (Ethernet)
> RX packets 61706 bytes 3631426 (3.4 MiB)
> RX errors 0 dropped 0 overruns 0 frame 0
> TX packets 3 bytes 258 (258.0 B)
> TX errors 0 dropped 0 overruns 0 carrier 0 collisions 0
>
> And able to do a
> ping -M do -s 8972 ip
> from each host to the other one
> In VMs I configure the same MTU=9000 in ifcfg-eth1
>
> But actually inside VMs it works erratically: the same ping test is ok
> between the VMs but Oracle checks sometimes work and sometimes give error
> on communication.
> At initial cluster config, the second node fails to start the cluster.
> I tried 5-6 times and also tried then to set mtu=8000 inside the VMs,
> supposing some sort of inner overhead to consider (such as 2 times 28
> bytes) but nothing.
> As soon as I set MTU=1500 at VM side, the cluster is able to form without
> any problem.
> I can survive without jumbo frames in this particular case, because this
> is only a test, but the question remains about eventual best practices to
> put in place if I want to use jumbo frames.
>
> One thing I see is that at VM side I see many drops when interface mtu was
> 9000, such as
>
> eth1 Link encap:Ethernet HWaddr 00:1A:4A:17:01:57
> inet addr:192.168.10.32 Bcast:192.168.10.255 Mask:255.255.255.0
> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
> RX packets:93046 errors:0 dropped:54964 overruns:0 frame:0
> TX packets:26258 errors:0 dropped:0 overruns:0 carrier:0
> collisions:0 txqueuelen:1000
> RX bytes:25726242 (24.5 MiB) TX bytes:33573207 (32.0 MiB)
>
> at host side I see drops at bond0 level only:
>
> [root at ovmsrv05 ~]# brctl show
> bridge name bridge id STP enabled interfaces
> ;vdsmdummy; 8000.000000000000 no
> vlan100 8000.001cc446ef73 no bond1.100
> vlan65 8000.001cc446ef73 no bond1.65
> vnet0
> vnet1
> vlan95 8000.001cc4abbeba no bond0.95
> vnet2
>
> bond0: flags=5187<UP,BROADCAST,RUNNING,MASTER,MULTICAST> mtu 9000
> ether 00:1c:c4:ab:be:ba txqueuelen 1000 (Ethernet)
> RX packets 2855175 bytes 3126868334 <(312)%20686-8334> (2.9 GiB)
> RX errors 0 dropped 11686 overruns 0 frame 0
> TX packets 1012849 bytes 478702140 (456.5 MiB)
> TX errors 0 dropped 0 overruns 0 carrier 0 collisions 0
>
> bond0.95: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 9000
> ether 00:1c:c4:ab:be:ba txqueuelen 1000 (Ethernet)
> RX packets 100272 bytes 27125992 (25.8 MiB)
> RX errors 0 dropped 0 overruns 0 frame 0
> TX packets 42355 bytes 40833904 (38.9 MiB)
> TX errors 0 dropped 0 overruns 0 carrier 0 collisions 0
>
> vlan95: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 9000
> ether 00:1c:c4:ab:be:ba txqueuelen 1000 (Ethernet)
> RX packets 62576 bytes 3719175 (3.5 MiB)
> RX errors 0 dropped 0 overruns 0 frame 0
> TX packets 3 bytes 258 (258.0 B)
> TX errors 0 dropped 0 overruns 0 carrier 0 collisions 0
>
> vnet2: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 9000
> inet6 fe80::fc1a:4aff:fe17:157 prefixlen 64 scopeid 0x20<link>
> ether fe:1a:4a:17:01:57 txqueuelen 1000 (Ethernet)
> RX packets 21014 bytes 24139492 (23.0 MiB)
> RX errors 0 dropped 0 overruns 0 frame 0
> TX packets 85777 bytes 21089777 (20.1 MiB)
> TX errors 0 dropped 0 overruns 0 carrier 0 collisions 0
>
> [root at ovmsrv05 ~]# cat /proc/net/bonding/bond0
> Ethernet Channel Bonding Driver: v3.7.1 (April 27, 2011)
>
> Bonding Mode: fault-tolerance (active-backup)
> Primary Slave: None
> Currently Active Slave: enp3s0
> MII Status: up
> MII Polling Interval (ms): 100
> Up Delay (ms): 0
> Down Delay (ms): 0
>
> Slave Interface: enp3s0
> MII Status: up
> Speed: 1000 Mbps
> Duplex: full
> Link Failure Count: 0
> Permanent HW addr: 00:1c:c4:ab:be:ba
> Slave queue ID: 0
>
> Slave Interface: enp5s0
> MII Status: up
> Speed: 1000 Mbps
> Duplex: full
> Link Failure Count: 0
> Permanent HW addr: 00:1c:c4:ab:be:bc
> Slave queue ID: 0
>
>
> Any hint?
> Thanks in advance,
> Gianluca
>
>
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20170206/aae95064/attachment-0001.html>
More information about the Users
mailing list