Thanks Rene, I'll look into your suggestions.
I don't think it's a conflicting MAC as there is only 1 guest, but
will check it out.
Would you advise running the engine and nodes on a separate network
range to my existing network rather?
Thanks.
Regards.
Neil Wilson.
On Fri, Mar 8, 2013 at 1:52 PM, René Koch (ovido) <r.koch(a)ovido.at> wrote:
Hi Neil,
I had a similar issue with my oVirt environment with some vms.
The issue on my side was oVirt and RHEV environment in the same subnet
and conflicting MAC addresses on some vms (as both use the same MAC
range and I didn't change this with engine-config).
So can you check if this MAC our your vm is in use by an other host/vm
(maybe from a KVM installation)?
--
Best regards
René Koch
Senior Solution Architect
============================================
ovido gmbh - "Das Linux Systemhaus"
Brünner Straße 163, A-1210 Wien
Phone: +43 720 / 530 670
Mobile: +43 660 / 512 21 31
E-Mail: r.koch(a)ovido.at
============================================
On Fri, 2013-03-08 at 11:27 +0200, Neil wrote:
> Hi guys,
>
> I've got a bit of a strange one, I'm setting up an internal ovirt
> system Centos 6.3 64bit dreyou repo...
>
> and I'm getting lots of packet loss on the guest I've installed, the
> packet loss doesn't happen on the physical hosts, only the VM gets it,
> when communicating from and to it.
>
> 1 node(Centos 6.3 64bit)
> vdsm-4.10.0-0.46.15.el6.x86_64
> vdsm-cli-4.10.0-0.46.15.el6.noarch
> vdsm-xmlrpc-4.10.0-0.46.15.el6.noarch
> vdsm-python-4.10.0-0.46.15.el6.x86_64
>
> The engine(also Centos 6.3 64bit(engine has local NFS storage which
> the node connects to)
> ovirt-engine-userportal-3.1.0-3.19.el6.noarch
> ovirt-engine-dbscripts-3.1.0-3.19.el6.noarch
> ovirt-engine-3.1.0-3.19.el6.noarch
> ovirt-engine-backend-3.1.0-3.19.el6.noarch
> ovirt-engine-notification-service-3.1.0-3.19.el6.noarch
> ovirt-image-uploader-3.1.0-16.el6.noarch
> ovirt-engine-genericapi-3.1.0-3.19.el6.noarch
> ovirt-iso-uploader-3.1.0-16.el6.noarch
> ovirt-engine-restapi-3.1.0-3.19.el6.noarch
> ovirt-engine-tools-common-3.1.0-3.19.el6.noarch
> ovirt-engine-sdk-3.2.0.8-1.el6.noarch
> ovirt-engine-webadmin-portal-3.1.0-3.19.el6.noarch
> ovirt-engine-cli-3.2.0.5-1.el6.noarch
> ovirt-log-collector-3.1.0-16.el6.noarch
> ovirt-engine-setup-3.1.0-3.19.el6.noarch
> ovirt-engine-jbossas711-1-0.x86_64
> ovirt-engine-config-3.1.0-3.19.el6.noarch
>
> Both the node and engine have bonded interfaces all NICS are intel
> 82574L gigabit and the managed switch reflects gigabit on each of the
> ports.
>
> The ifcfg-bond0 is below...
>
> DEVICE=bond0
> IPADDR=192.168.0.9
> NETWORK=192.168.0.0
> NETMASK=255.255.255.0
> USERCTL=no
> BONDING_OPTS=mode=2
> BOOTPROTO=none
> MTU=1500
> ONBOOT=yes
>
> Then the ifcfg-eth0 and eth1 are almost identical...
> DEVICE=eth2
> USERCTL=no
> ONBOOT=yes
> MASTER=bond0
> SLAVE=yes
> MTU=1500
> BOOTPROTO=none
>
>
> These are the network details on the guest, as you can see, there are
> no network errors showing on the guest at all, which is strange....
>
> eth0 Link encap:Ethernet HWaddr 00:1A:4A:A8:00:00
> inet addr:192.168.0.12 Bcast:192.168.0.255 Mask:255.255.255.0
> inet6 addr: fe80::21a:4aff:fea8:0/64 Scope:Link
> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
> RX packets:5050 errors:0 dropped:0 overruns:0 frame:0
> TX packets:255 errors:0 dropped:0 overruns:0 carrier:0
> collisions:0 txqueuelen:1000
> RX bytes:490762 (479.2 KiB) TX bytes:32516 (31.7 KiB)
>
> Ethernet controller: Red Hat, Inc Virtio network device
>
> Has anyone got any ideas? Have I set something up wrong?
>
> Any help or advice is greatly appreciated.
>
> Regards.
>
> Neil Wilson.
> _______________________________________________
> Users mailing list
> Users(a)ovirt.org
>
http://lists.ovirt.org/mailman/listinfo/users