Hi Alex,
So if you're sending data out there is no or almost no data loss?
Have you tried to move the VM to the second host to see if it makes any
difference?
I actually tried to call you by phone but the call gets dropped once I
enter * and your ext. number, as the voice prompt says to.
Tried to enter the ext. number directly without *, still without success.
Thanks in advance,
On Thu, Jun 8, 2017 at 1:59 PM, Герасимов Александр <gerasimov.ay(a)eksmo.ru>
wrote:
Hi Lev.
I am create new virtual host, and run ping from internet.
Ping results 70% packet loss.
But, if i am running ping from virtual host to somebody host in the
internet and at the same time, i am pinging this virtual host from internet.
then ping results 0% packet loss or 10% packet loss.
28.05.2017 13:34, Lev Veyde пишет:
Hi Alex,
That is quite strange...
Does this happen on both hosts - have you tried to migrate the VM to the
second host and see if the issue still remains?
Thanks in advance,
On Fri, May 26, 2017 at 3:02 PM, Герасимов Александр <
gerasimov.ay(a)eksmo.ru> wrote:
> Hi Lev.
>
>
> On one of the VMs you only see 1 NIC instead of the 2?
>
> NO. both VM's sees two NIC, but on first VM ping with no error, and
> second VM ping with 75% error.
>
> OS version on hosts [root@node01 ~]# cat /etc/redhat-release
> CentOS Linux release 7.3.1611 (Core)
>
> OS veriosion on VM's [root@node03 ~]# cat /etc/redhat-release
> CentOS Linux release 7.3.1611 (Core)
>
>
>
> *first VM*
>
> 00:03.0 Ethernet controller: Red Hat, Inc Virtio network device
>
> 00:09.0 Ethernet controller: Red Hat, Inc Virtio network device
>
> [root@node03 ~]# ip l
> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN mode
> DEFAULT qlen 1
> link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
> state UP mode DEFAULT qlen 1000
> link/ether 00:1a:4a:16:01:51 brd ff:ff:ff:ff:ff:ff
> 3: eth1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
> state UP mode DEFAULT qlen 1000
> link/ether 00:1a:4a:16:01:55 brd ff:ff:ff:ff:ff:ff
>
> *second VM*
>
> 00:03.0 Ethernet controller: Realtek Semiconductor Co., Ltd.
> RTL-8100/8101L/8139 PCI Fast Ethernet Adapter (rev 20) - but i tested all
> version of NIC and no effect
>
> 00:0a.0 Ethernet controller: Red Hat, Inc Virtio network device
>
> [root@node04 ~]# ip link
> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN mode
> DEFAULT qlen 1
> link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
> state UP mode DEFAULT qlen 1000
> link/ether 00:1a:4a:16:01:53 brd ff:ff:ff:ff:ff:ff
> 3: ens3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
> state UP mode DEFAULT qlen 1000
> link/ether 00:1a:4a:16:01:52 brd ff:ff:ff:ff:ff:ff
>
> In logs no messages only like this :
>
> May 26 15:01:01 node04 systemd: Started Session 67263 of user root.
> May 26 15:01:01 node04 systemd: Starting Session 67263 of user root.
> May 26 15:01:01 node04 systemd: Created slice user-600.slice.
> May 26 15:01:01 node04 systemd: Starting user-600.slice.
> May 26 15:01:01 node04 systemd: Started Session 67262 of user bitrix.
> May 26 15:01:01 node04 systemd: Starting Session 67262 of user bitrix.
> May 26 15:01:01 node04 systemd: Removed slice user-600.slice.
> May 26 15:01:01 node04 systemd: Stopping user-600.slice.
>
>
> Hi Alexander,
>
> So if I understand it correctly, you have the following configuration:
> - 2 hosts, each having 2 NICs
> - 2 virtual machines, each have a connection to each one of the NICs
> available on the hosts
>
> On one of the VMs you only see 1 NIC instead of the 2?
>
> Are you sure that the VM is properly configured to have 2 NICs?
>
> What Linux distro and version you're using on the hosts and inside the
> VMs ?
>
> Can you please send us:
> - the logs from the VM, e.g. /var/log/messages
> - the output of lspci -v
> - the output of ip link
>
> Thanks in advance,
>
> 2017-05-18 12:19 GMT+03:00 Герасимов Александр <gerasimov.ay at eksmo.ru
> >:
>
> > Hi all.
> >
> > I have to servers with ovirt.
> >
> > And to identical virtual machines.
> >
> > Both servers are identical. But on second virtual server not working one
> > network interface. Ping have a problem. I tried to change network
> driver,
> > but has no effect.
> >
> > I don't understand that to do
> >
> >
> > ovirt version and package:
> >
> > rpm -qa|grep ovirt
> > ovirt-imageio-proxy-0.4.0-0.201608310602.gita9b573b.el7.centos.noarch
> > ovirt-engine-vmconsole-proxy-helper-4.0.5.5-1.el7.centos.noarch
> > ovirt-engine-restapi-4.0.5.5-1.el7.centos.noarch
> > ovirt-engine-extensions-api-impl-4.0.5.5-1.el7.centos.noarch
> > ovirt-imageio-daemon-0.4.0-1.el7.noarch
> > ovirt-engine-wildfly-10.1.0-1.el7.x86_64
> > ovirt-vmconsole-1.0.4-1.el7.centos.noarch
> > ovirt-engine-cli-3.6.9.2-1.el7.noarch
> > ovirt-engine-websocket-proxy-4.0.5.5-1.el7.centos.noarch
> > ovirt-engine-dashboard-1.0.5-1.el7.centos.noarch
> > ovirt-host-deploy-1.5.3-1.el7.centos.noarch
> > ovirt-engine-wildfly-overlay-10.0.0-1.el7.noarch
> > ovirt-engine-setup-base-4.0.5.5-1.el7.centos.noarch
> > ovirt-engine-dwh-setup-4.0.5-1.el7.centos.noarch
> > ovirt-engine-setup-plugin-websocket-proxy-4.0.5.5-1.el7.centos.noarch
> > ovirt-engine-setup-4.0.5.5-1.el7.centos.noarch
> > ovirt-engine-dbscripts-4.0.5.5-1.el7.centos.noarch
> > ovirt-engine-userportal-4.0.5.5-1.el7.centos.noarch
> > ovirt-imageio-common-0.4.0-1.el7.noarch
> > python-ovirt-engine-sdk4-4.0.2-1.el7.centos.x86_64
> > ovirt-vmconsole-host-1.0.4-1.el7.centos.noarch
> > ovirt-engine-dwh-4.0.5-1.el7.centos.noarch
> > ovirt-engine-tools-backup-4.0.5.5-1.el7.centos.noarch
> > ovirt-image-uploader-4.0.1-1.el7.centos.noarch
> > ovirt-engine-setup-plugin-ovirt-engine-4.0.5.5-1.el7.centos.noarch
> > ovirt-engine-tools-4.0.5.5-1.el7.centos.noarch
> > ovirt-engine-4.0.5.5-1.el7.centos.noarch
> > ovirt-release40-4.0.5-2.noarch
> > ovirt-host-deploy-java-1.5.3-1.el7.centos.noarch
> > ovirt-engine-setup-plugin-ovirt-engine-common-4.0.5.5-1.el7.
> centos.noarch
> > ovirt-iso-uploader-4.0.2-1.el7.centos.noarch
> > ovirt-engine-webadmin-portal-4.0.5.5-1.el7.centos.noarch
> > ovirt-setup-lib-1.0.2-1.el7.centos.noarch
> > ovirt-engine-sdk-python-3.6.9.1-1.el7.centos.noarch
> > ovirt-engine-lib-4.0.5.5-1.el7.centos.noarch
> > ovirt-imageio-proxy-setup-0.4.0-0.201608310602.gita9b573b.
> > el7.centos.noarch
> > ovirt-engine-setup-plugin-vmconsole-proxy-helper-4.0.5.
> > 5-1.el7.centos.noarch
> > ovirt-engine-backend-4.0.5.5-1.el7.centos.noarch
> > ovirt-engine-extension-aaa-jdbc-1.1.1-1.el7.noarch
> > ovirt-vmconsole-proxy-1.0.4-1.el7.centos.noarch
> >
> >
>
> --
> С уважением, базисный админстратор
> Гераcимов Александр
> тел. +7(495)4116886 <+7%20495%20411-68-86> доб. 5367
>
>
> _______________________________________________
> Users mailing list
> Users(a)ovirt.org
>
http://lists.ovirt.org/mailman/listinfo/users
>
>
--
Lev Veyde
Software Engineer, RHCE | RHCVA | MCITP
Red Hat Israel
<
https://www.redhat.com>
lev(a)redhat.com | lveyde(a)redhat.com
<
https://red.ht/sig>
TRIED. TESTED. TRUSTED. <
https://redhat.com/trusted>
--
С уважением, базисный админстратор
Гераcимов Александр
тел. +7(495)4116886 <+7%20495%20411-68-86> доб. 5367
TRIED. TESTED. TRUSTED. <