
Hi guys, I've got a bit of a strange one, I'm setting up an internal ovirt system Centos 6.3 64bit dreyou repo... and I'm getting lots of packet loss on the guest I've installed, the packet loss doesn't happen on the physical hosts, only the VM gets it, when communicating from and to it. 1 node(Centos 6.3 64bit) vdsm-4.10.0-0.46.15.el6.x86_64 vdsm-cli-4.10.0-0.46.15.el6.noarch vdsm-xmlrpc-4.10.0-0.46.15.el6.noarch vdsm-python-4.10.0-0.46.15.el6.x86_64 The engine(also Centos 6.3 64bit(engine has local NFS storage which the node connects to) ovirt-engine-userportal-3.1.0-3.19.el6.noarch ovirt-engine-dbscripts-3.1.0-3.19.el6.noarch ovirt-engine-3.1.0-3.19.el6.noarch ovirt-engine-backend-3.1.0-3.19.el6.noarch ovirt-engine-notification-service-3.1.0-3.19.el6.noarch ovirt-image-uploader-3.1.0-16.el6.noarch ovirt-engine-genericapi-3.1.0-3.19.el6.noarch ovirt-iso-uploader-3.1.0-16.el6.noarch ovirt-engine-restapi-3.1.0-3.19.el6.noarch ovirt-engine-tools-common-3.1.0-3.19.el6.noarch ovirt-engine-sdk-3.2.0.8-1.el6.noarch ovirt-engine-webadmin-portal-3.1.0-3.19.el6.noarch ovirt-engine-cli-3.2.0.5-1.el6.noarch ovirt-log-collector-3.1.0-16.el6.noarch ovirt-engine-setup-3.1.0-3.19.el6.noarch ovirt-engine-jbossas711-1-0.x86_64 ovirt-engine-config-3.1.0-3.19.el6.noarch Both the node and engine have bonded interfaces all NICS are intel 82574L gigabit and the managed switch reflects gigabit on each of the ports. The ifcfg-bond0 is below... DEVICE=bond0 IPADDR=192.168.0.9 NETWORK=192.168.0.0 NETMASK=255.255.255.0 USERCTL=no BONDING_OPTS=mode=2 BOOTPROTO=none MTU=1500 ONBOOT=yes Then the ifcfg-eth0 and eth1 are almost identical... DEVICE=eth2 USERCTL=no ONBOOT=yes MASTER=bond0 SLAVE=yes MTU=1500 BOOTPROTO=none These are the network details on the guest, as you can see, there are no network errors showing on the guest at all, which is strange.... eth0 Link encap:Ethernet HWaddr 00:1A:4A:A8:00:00 inet addr:192.168.0.12 Bcast:192.168.0.255 Mask:255.255.255.0 inet6 addr: fe80::21a:4aff:fea8:0/64 Scope:Link UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 RX packets:5050 errors:0 dropped:0 overruns:0 frame:0 TX packets:255 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:1000 RX bytes:490762 (479.2 KiB) TX bytes:32516 (31.7 KiB) Ethernet controller: Red Hat, Inc Virtio network device Has anyone got any ideas? Have I set something up wrong? Any help or advice is greatly appreciated. Regards. Neil Wilson.

Hi Neil, I had a similar issue with my oVirt environment with some vms. The issue on my side was oVirt and RHEV environment in the same subnet and conflicting MAC addresses on some vms (as both use the same MAC range and I didn't change this with engine-config). So can you check if this MAC our your vm is in use by an other host/vm (maybe from a KVM installation)? -- Best regards René Koch Senior Solution Architect ============================================ ovido gmbh - "Das Linux Systemhaus" Brünner Straße 163, A-1210 Wien Phone: +43 720 / 530 670 Mobile: +43 660 / 512 21 31 E-Mail: r.koch@ovido.at ============================================ On Fri, 2013-03-08 at 11:27 +0200, Neil wrote:
Hi guys,
I've got a bit of a strange one, I'm setting up an internal ovirt system Centos 6.3 64bit dreyou repo...
and I'm getting lots of packet loss on the guest I've installed, the packet loss doesn't happen on the physical hosts, only the VM gets it, when communicating from and to it.
1 node(Centos 6.3 64bit) vdsm-4.10.0-0.46.15.el6.x86_64 vdsm-cli-4.10.0-0.46.15.el6.noarch vdsm-xmlrpc-4.10.0-0.46.15.el6.noarch vdsm-python-4.10.0-0.46.15.el6.x86_64
The engine(also Centos 6.3 64bit(engine has local NFS storage which the node connects to) ovirt-engine-userportal-3.1.0-3.19.el6.noarch ovirt-engine-dbscripts-3.1.0-3.19.el6.noarch ovirt-engine-3.1.0-3.19.el6.noarch ovirt-engine-backend-3.1.0-3.19.el6.noarch ovirt-engine-notification-service-3.1.0-3.19.el6.noarch ovirt-image-uploader-3.1.0-16.el6.noarch ovirt-engine-genericapi-3.1.0-3.19.el6.noarch ovirt-iso-uploader-3.1.0-16.el6.noarch ovirt-engine-restapi-3.1.0-3.19.el6.noarch ovirt-engine-tools-common-3.1.0-3.19.el6.noarch ovirt-engine-sdk-3.2.0.8-1.el6.noarch ovirt-engine-webadmin-portal-3.1.0-3.19.el6.noarch ovirt-engine-cli-3.2.0.5-1.el6.noarch ovirt-log-collector-3.1.0-16.el6.noarch ovirt-engine-setup-3.1.0-3.19.el6.noarch ovirt-engine-jbossas711-1-0.x86_64 ovirt-engine-config-3.1.0-3.19.el6.noarch
Both the node and engine have bonded interfaces all NICS are intel 82574L gigabit and the managed switch reflects gigabit on each of the ports.
The ifcfg-bond0 is below...
DEVICE=bond0 IPADDR=192.168.0.9 NETWORK=192.168.0.0 NETMASK=255.255.255.0 USERCTL=no BONDING_OPTS=mode=2 BOOTPROTO=none MTU=1500 ONBOOT=yes
Then the ifcfg-eth0 and eth1 are almost identical... DEVICE=eth2 USERCTL=no ONBOOT=yes MASTER=bond0 SLAVE=yes MTU=1500 BOOTPROTO=none
These are the network details on the guest, as you can see, there are no network errors showing on the guest at all, which is strange....
eth0 Link encap:Ethernet HWaddr 00:1A:4A:A8:00:00 inet addr:192.168.0.12 Bcast:192.168.0.255 Mask:255.255.255.0 inet6 addr: fe80::21a:4aff:fea8:0/64 Scope:Link UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 RX packets:5050 errors:0 dropped:0 overruns:0 frame:0 TX packets:255 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:1000 RX bytes:490762 (479.2 KiB) TX bytes:32516 (31.7 KiB)
Ethernet controller: Red Hat, Inc Virtio network device
Has anyone got any ideas? Have I set something up wrong?
Any help or advice is greatly appreciated.
Regards.
Neil Wilson. _______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users

Thanks Rene, I'll look into your suggestions. I don't think it's a conflicting MAC as there is only 1 guest, but will check it out. Would you advise running the engine and nodes on a separate network range to my existing network rather? Thanks. Regards. Neil Wilson. On Fri, Mar 8, 2013 at 1:52 PM, René Koch (ovido) <r.koch@ovido.at> wrote:
Hi Neil,
I had a similar issue with my oVirt environment with some vms. The issue on my side was oVirt and RHEV environment in the same subnet and conflicting MAC addresses on some vms (as both use the same MAC range and I didn't change this with engine-config).
So can you check if this MAC our your vm is in use by an other host/vm (maybe from a KVM installation)?
-- Best regards
René Koch Senior Solution Architect
============================================ ovido gmbh - "Das Linux Systemhaus" Brünner Straße 163, A-1210 Wien
Phone: +43 720 / 530 670 Mobile: +43 660 / 512 21 31 E-Mail: r.koch@ovido.at ============================================
On Fri, 2013-03-08 at 11:27 +0200, Neil wrote:
Hi guys,
I've got a bit of a strange one, I'm setting up an internal ovirt system Centos 6.3 64bit dreyou repo...
and I'm getting lots of packet loss on the guest I've installed, the packet loss doesn't happen on the physical hosts, only the VM gets it, when communicating from and to it.
1 node(Centos 6.3 64bit) vdsm-4.10.0-0.46.15.el6.x86_64 vdsm-cli-4.10.0-0.46.15.el6.noarch vdsm-xmlrpc-4.10.0-0.46.15.el6.noarch vdsm-python-4.10.0-0.46.15.el6.x86_64
The engine(also Centos 6.3 64bit(engine has local NFS storage which the node connects to) ovirt-engine-userportal-3.1.0-3.19.el6.noarch ovirt-engine-dbscripts-3.1.0-3.19.el6.noarch ovirt-engine-3.1.0-3.19.el6.noarch ovirt-engine-backend-3.1.0-3.19.el6.noarch ovirt-engine-notification-service-3.1.0-3.19.el6.noarch ovirt-image-uploader-3.1.0-16.el6.noarch ovirt-engine-genericapi-3.1.0-3.19.el6.noarch ovirt-iso-uploader-3.1.0-16.el6.noarch ovirt-engine-restapi-3.1.0-3.19.el6.noarch ovirt-engine-tools-common-3.1.0-3.19.el6.noarch ovirt-engine-sdk-3.2.0.8-1.el6.noarch ovirt-engine-webadmin-portal-3.1.0-3.19.el6.noarch ovirt-engine-cli-3.2.0.5-1.el6.noarch ovirt-log-collector-3.1.0-16.el6.noarch ovirt-engine-setup-3.1.0-3.19.el6.noarch ovirt-engine-jbossas711-1-0.x86_64 ovirt-engine-config-3.1.0-3.19.el6.noarch
Both the node and engine have bonded interfaces all NICS are intel 82574L gigabit and the managed switch reflects gigabit on each of the ports.
The ifcfg-bond0 is below...
DEVICE=bond0 IPADDR=192.168.0.9 NETWORK=192.168.0.0 NETMASK=255.255.255.0 USERCTL=no BONDING_OPTS=mode=2 BOOTPROTO=none MTU=1500 ONBOOT=yes
Then the ifcfg-eth0 and eth1 are almost identical... DEVICE=eth2 USERCTL=no ONBOOT=yes MASTER=bond0 SLAVE=yes MTU=1500 BOOTPROTO=none
These are the network details on the guest, as you can see, there are no network errors showing on the guest at all, which is strange....
eth0 Link encap:Ethernet HWaddr 00:1A:4A:A8:00:00 inet addr:192.168.0.12 Bcast:192.168.0.255 Mask:255.255.255.0 inet6 addr: fe80::21a:4aff:fea8:0/64 Scope:Link UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 RX packets:5050 errors:0 dropped:0 overruns:0 frame:0 TX packets:255 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:1000 RX bytes:490762 (479.2 KiB) TX bytes:32516 (31.7 KiB)
Ethernet controller: Red Hat, Inc Virtio network device
Has anyone got any ideas? Have I set something up wrong?
Any help or advice is greatly appreciated.
Regards.
Neil Wilson. _______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users

Hi, If you don't have another oVirt / RHEL KVM host in the same network than MACs want be an issue. So it's totally save to have oVirt in the same network as all other systems. I see in your email that you use bonding mode 2. This shouldn't cause issues with switches, but I think I had an issue with RHEV (bond 2) and a Cisco switch once on customer side - please don't ask me for details on what their network admins changed to make this setup work. Can you try bonding mode 0 (active-backup) and check if you still have packet loss? -- Best regards René Koch Senior Solution Architect ============================================ ovido gmbh - "Das Linux Systemhaus" Brünner Straße 163, A-1210 Wien Phone: +43 720 / 530 670 Mobile: +43 660 / 512 21 31 E-Mail: r.koch@ovido.at ============================================ On Fri, 2013-03-08 at 15:57 +0200, Neil wrote:
Thanks Rene, I'll look into your suggestions.
I don't think it's a conflicting MAC as there is only 1 guest, but will check it out.
Would you advise running the engine and nodes on a separate network range to my existing network rather?
Thanks.
Regards.
Neil Wilson.
On Fri, Mar 8, 2013 at 1:52 PM, René Koch (ovido) <r.koch@ovido.at> wrote:
Hi Neil,
I had a similar issue with my oVirt environment with some vms. The issue on my side was oVirt and RHEV environment in the same subnet and conflicting MAC addresses on some vms (as both use the same MAC range and I didn't change this with engine-config).
So can you check if this MAC our your vm is in use by an other host/vm (maybe from a KVM installation)?
-- Best regards
René Koch Senior Solution Architect
============================================ ovido gmbh - "Das Linux Systemhaus" Brünner Straße 163, A-1210 Wien
Phone: +43 720 / 530 670 Mobile: +43 660 / 512 21 31 E-Mail: r.koch@ovido.at ============================================
On Fri, 2013-03-08 at 11:27 +0200, Neil wrote:
Hi guys,
I've got a bit of a strange one, I'm setting up an internal ovirt system Centos 6.3 64bit dreyou repo...
and I'm getting lots of packet loss on the guest I've installed, the packet loss doesn't happen on the physical hosts, only the VM gets it, when communicating from and to it.
1 node(Centos 6.3 64bit) vdsm-4.10.0-0.46.15.el6.x86_64 vdsm-cli-4.10.0-0.46.15.el6.noarch vdsm-xmlrpc-4.10.0-0.46.15.el6.noarch vdsm-python-4.10.0-0.46.15.el6.x86_64
The engine(also Centos 6.3 64bit(engine has local NFS storage which the node connects to) ovirt-engine-userportal-3.1.0-3.19.el6.noarch ovirt-engine-dbscripts-3.1.0-3.19.el6.noarch ovirt-engine-3.1.0-3.19.el6.noarch ovirt-engine-backend-3.1.0-3.19.el6.noarch ovirt-engine-notification-service-3.1.0-3.19.el6.noarch ovirt-image-uploader-3.1.0-16.el6.noarch ovirt-engine-genericapi-3.1.0-3.19.el6.noarch ovirt-iso-uploader-3.1.0-16.el6.noarch ovirt-engine-restapi-3.1.0-3.19.el6.noarch ovirt-engine-tools-common-3.1.0-3.19.el6.noarch ovirt-engine-sdk-3.2.0.8-1.el6.noarch ovirt-engine-webadmin-portal-3.1.0-3.19.el6.noarch ovirt-engine-cli-3.2.0.5-1.el6.noarch ovirt-log-collector-3.1.0-16.el6.noarch ovirt-engine-setup-3.1.0-3.19.el6.noarch ovirt-engine-jbossas711-1-0.x86_64 ovirt-engine-config-3.1.0-3.19.el6.noarch
Both the node and engine have bonded interfaces all NICS are intel 82574L gigabit and the managed switch reflects gigabit on each of the ports.
The ifcfg-bond0 is below...
DEVICE=bond0 IPADDR=192.168.0.9 NETWORK=192.168.0.0 NETMASK=255.255.255.0 USERCTL=no BONDING_OPTS=mode=2 BOOTPROTO=none MTU=1500 ONBOOT=yes
Then the ifcfg-eth0 and eth1 are almost identical... DEVICE=eth2 USERCTL=no ONBOOT=yes MASTER=bond0 SLAVE=yes MTU=1500 BOOTPROTO=none
These are the network details on the guest, as you can see, there are no network errors showing on the guest at all, which is strange....
eth0 Link encap:Ethernet HWaddr 00:1A:4A:A8:00:00 inet addr:192.168.0.12 Bcast:192.168.0.255 Mask:255.255.255.0 inet6 addr: fe80::21a:4aff:fea8:0/64 Scope:Link UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 RX packets:5050 errors:0 dropped:0 overruns:0 frame:0 TX packets:255 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:1000 RX bytes:490762 (479.2 KiB) TX bytes:32516 (31.7 KiB)
Ethernet controller: Red Hat, Inc Virtio network device
Has anyone got any ideas? Have I set something up wrong?
Any help or advice is greatly appreciated.
Regards.
Neil Wilson. _______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
participants (2)
-
Neil
-
René Koch (ovido)