[Users] Packet loss to guests

Neil nwilson123 at gmail.com
Fri Mar 8 08:57:20 EST 2013


Thanks Rene, I'll look into your suggestions.

I don't think it's a conflicting MAC as there is only 1 guest, but
will check it out.

Would you advise running the engine and nodes on a separate network
range to my existing network rather?

Thanks.

Regards.

Neil Wilson.

On Fri, Mar 8, 2013 at 1:52 PM, René Koch (ovido) <r.koch at ovido.at> wrote:
> Hi Neil,
>
> I had a similar issue with my oVirt environment with some vms.
> The issue on my side was oVirt and RHEV environment in the same subnet
> and conflicting MAC addresses on some vms (as both use the same MAC
> range and I didn't change this with engine-config).
>
> So can you check if this MAC our your vm is in use by an other host/vm
> (maybe from a KVM installation)?
>
>
> --
> Best regards
>
> René Koch
> Senior Solution Architect
>
> ============================================
> ovido gmbh - "Das Linux Systemhaus"
> Brünner Straße 163, A-1210 Wien
>
> Phone:   +43 720 / 530 670
> Mobile:  +43 660 / 512 21 31
> E-Mail:  r.koch at ovido.at
> ============================================
>
>
> On Fri, 2013-03-08 at 11:27 +0200, Neil wrote:
>> Hi guys,
>>
>> I've got a bit of a strange one, I'm setting up an internal ovirt
>> system Centos 6.3 64bit dreyou repo...
>>
>> and I'm getting lots of packet loss on the guest I've installed, the
>> packet loss doesn't happen on the physical hosts, only the VM gets it,
>> when communicating from and to it.
>>
>> 1 node(Centos 6.3 64bit)
>> vdsm-4.10.0-0.46.15.el6.x86_64
>> vdsm-cli-4.10.0-0.46.15.el6.noarch
>> vdsm-xmlrpc-4.10.0-0.46.15.el6.noarch
>> vdsm-python-4.10.0-0.46.15.el6.x86_64
>>
>> The engine(also Centos 6.3 64bit(engine has local NFS storage which
>> the node connects to)
>> ovirt-engine-userportal-3.1.0-3.19.el6.noarch
>> ovirt-engine-dbscripts-3.1.0-3.19.el6.noarch
>> ovirt-engine-3.1.0-3.19.el6.noarch
>> ovirt-engine-backend-3.1.0-3.19.el6.noarch
>> ovirt-engine-notification-service-3.1.0-3.19.el6.noarch
>> ovirt-image-uploader-3.1.0-16.el6.noarch
>> ovirt-engine-genericapi-3.1.0-3.19.el6.noarch
>> ovirt-iso-uploader-3.1.0-16.el6.noarch
>> ovirt-engine-restapi-3.1.0-3.19.el6.noarch
>> ovirt-engine-tools-common-3.1.0-3.19.el6.noarch
>> ovirt-engine-sdk-3.2.0.8-1.el6.noarch
>> ovirt-engine-webadmin-portal-3.1.0-3.19.el6.noarch
>> ovirt-engine-cli-3.2.0.5-1.el6.noarch
>> ovirt-log-collector-3.1.0-16.el6.noarch
>> ovirt-engine-setup-3.1.0-3.19.el6.noarch
>> ovirt-engine-jbossas711-1-0.x86_64
>> ovirt-engine-config-3.1.0-3.19.el6.noarch
>>
>> Both the node and engine have bonded interfaces all NICS are intel
>> 82574L gigabit and the managed switch reflects gigabit on each of the
>> ports.
>>
>> The ifcfg-bond0 is below...
>>
>> DEVICE=bond0
>> IPADDR=192.168.0.9
>> NETWORK=192.168.0.0
>> NETMASK=255.255.255.0
>> USERCTL=no
>> BONDING_OPTS=mode=2
>> BOOTPROTO=none
>> MTU=1500
>> ONBOOT=yes
>>
>> Then the ifcfg-eth0 and eth1 are almost identical...
>> DEVICE=eth2
>> USERCTL=no
>> ONBOOT=yes
>> MASTER=bond0
>> SLAVE=yes
>> MTU=1500
>> BOOTPROTO=none
>>
>>
>> These are the network details on the guest, as you can see, there are
>> no network errors showing on the guest at all, which is strange....
>>
>> eth0 Link encap:Ethernet  HWaddr 00:1A:4A:A8:00:00
>>        inet addr:192.168.0.12  Bcast:192.168.0.255  Mask:255.255.255.0
>>        inet6 addr: fe80::21a:4aff:fea8:0/64 Scope:Link
>>        UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
>>        RX packets:5050 errors:0 dropped:0 overruns:0 frame:0
>>        TX packets:255 errors:0 dropped:0 overruns:0 carrier:0
>>        collisions:0 txqueuelen:1000
>>        RX bytes:490762 (479.2 KiB)  TX bytes:32516 (31.7 KiB)
>>
>> Ethernet controller: Red Hat, Inc Virtio network device
>>
>> Has anyone got any ideas? Have I set something up wrong?
>>
>> Any help or advice is greatly appreciated.
>>
>> Regards.
>>
>> Neil Wilson.
>> _______________________________________________
>> Users mailing list
>> Users at ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>


More information about the Users mailing list