[ovirt-users] Network interface not working

Lev Veyde lveyde at redhat.com
Thu Jun 8 11:33:22 UTC 2017


Hi Alex,

So if you're sending data out there is no or almost no data loss?

Have you tried to move the VM to the second host to see if it makes any
difference?

I actually tried to call you by phone but the call gets dropped once I
enter * and your ext. number, as the voice prompt says to.
Tried to enter the ext. number directly without *, still without success.

Thanks in advance,

On Thu, Jun 8, 2017 at 1:59 PM, Герасимов Александр <gerasimov.ay at eksmo.ru>
wrote:

> Hi Lev.
>
> I am create new virtual host, and run ping from internet.
>
> Ping results 70% packet loss.
>
> But, if i am running ping from virtual host to somebody host in the
> internet and at the same time, i am pinging this virtual host from internet.
>
> then ping results 0% packet loss or 10% packet loss.
>
>
> 28.05.2017 13:34, Lev Veyde пишет:
>
> Hi Alex,
>
> That is quite strange...
>
> Does this happen on both hosts - have you tried to migrate the VM to the
> second host and see if the issue still remains?
>
> Thanks in advance,
>
>
> On Fri, May 26, 2017 at 3:02 PM, Герасимов Александр <
> gerasimov.ay at eksmo.ru> wrote:
>
>> Hi Lev.
>>
>>
>> On one of the VMs you only see 1 NIC instead of the 2?
>>
>> NO. both VM's sees two NIC, but on first VM ping with no error, and
>> second VM ping with 75% error.
>>
>> OS version on hosts [root at node01 ~]# cat /etc/redhat-release
>> CentOS Linux release 7.3.1611 (Core)
>>
>> OS veriosion on VM's [root at node03 ~]# cat /etc/redhat-release
>> CentOS Linux release 7.3.1611 (Core)
>>
>>
>>
>> *first VM*
>>
>> 00:03.0 Ethernet controller: Red Hat, Inc Virtio network device
>>
>> 00:09.0 Ethernet controller: Red Hat, Inc Virtio network device
>>
>> [root at node03 ~]# ip l
>> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN mode
>> DEFAULT qlen 1
>>     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
>> state UP mode DEFAULT qlen 1000
>>     link/ether 00:1a:4a:16:01:51 brd ff:ff:ff:ff:ff:ff
>> 3: eth1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
>> state UP mode DEFAULT qlen 1000
>>     link/ether 00:1a:4a:16:01:55 brd ff:ff:ff:ff:ff:ff
>>
>> *second VM*
>>
>> 00:03.0 Ethernet controller: Realtek Semiconductor Co., Ltd.
>> RTL-8100/8101L/8139 PCI Fast Ethernet Adapter (rev 20)  - but i tested all
>> version of NIC and no effect
>>
>> 00:0a.0 Ethernet controller: Red Hat, Inc Virtio network device
>>
>> [root at node04 ~]# ip link
>> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN mode
>> DEFAULT qlen 1
>>     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
>> state UP mode DEFAULT qlen 1000
>>     link/ether 00:1a:4a:16:01:53 brd ff:ff:ff:ff:ff:ff
>> 3: ens3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
>> state UP mode DEFAULT qlen 1000
>>     link/ether 00:1a:4a:16:01:52 brd ff:ff:ff:ff:ff:ff
>>
>> In logs no messages only like this :
>>
>> May 26 15:01:01 node04 systemd: Started Session 67263 of user root.
>> May 26 15:01:01 node04 systemd: Starting Session 67263 of user root.
>> May 26 15:01:01 node04 systemd: Created slice user-600.slice.
>> May 26 15:01:01 node04 systemd: Starting user-600.slice.
>> May 26 15:01:01 node04 systemd: Started Session 67262 of user bitrix.
>> May 26 15:01:01 node04 systemd: Starting Session 67262 of user bitrix.
>> May 26 15:01:01 node04 systemd: Removed slice user-600.slice.
>> May 26 15:01:01 node04 systemd: Stopping user-600.slice.
>>
>>
>> Hi Alexander,
>>
>> So if I understand it correctly, you have the following configuration:
>> - 2 hosts, each having 2 NICs
>> - 2 virtual machines, each have a connection to each one of the NICs
>> available on the hosts
>>
>> On one of the VMs you only see 1 NIC instead of the 2?
>>
>> Are you sure that the VM is properly configured to have 2 NICs?
>>
>> What Linux distro and version you're using on the hosts and inside the
>> VMs ?
>>
>> Can you please send us:
>> - the logs from the VM, e.g. /var/log/messages
>> - the output of lspci -v
>> - the output of ip link
>>
>> Thanks in advance,
>>
>> 2017-05-18 12:19 GMT+03:00 Герасимов Александр <gerasimov.ay at eksmo.ru
>> >:
>>
>> > Hi all.
>> >
>> > I have to servers with ovirt.
>> >
>> > And to identical virtual machines.
>> >
>> > Both servers are identical. But on second virtual server not working one
>> > network interface. Ping have a problem. I tried to change network
>> driver,
>> > but has no effect.
>> >
>> > I don't understand that to do
>> >
>> >
>> > ovirt version and package:
>> >
>> > rpm -qa|grep ovirt
>> > ovirt-imageio-proxy-0.4.0-0.201608310602.gita9b573b.el7.centos.noarch
>> > ovirt-engine-vmconsole-proxy-helper-4.0.5.5-1.el7.centos.noarch
>> > ovirt-engine-restapi-4.0.5.5-1.el7.centos.noarch
>> > ovirt-engine-extensions-api-impl-4.0.5.5-1.el7.centos.noarch
>> > ovirt-imageio-daemon-0.4.0-1.el7.noarch
>> > ovirt-engine-wildfly-10.1.0-1.el7.x86_64
>> > ovirt-vmconsole-1.0.4-1.el7.centos.noarch
>> > ovirt-engine-cli-3.6.9.2-1.el7.noarch
>> > ovirt-engine-websocket-proxy-4.0.5.5-1.el7.centos.noarch
>> > ovirt-engine-dashboard-1.0.5-1.el7.centos.noarch
>> > ovirt-host-deploy-1.5.3-1.el7.centos.noarch
>> > ovirt-engine-wildfly-overlay-10.0.0-1.el7.noarch
>> > ovirt-engine-setup-base-4.0.5.5-1.el7.centos.noarch
>> > ovirt-engine-dwh-setup-4.0.5-1.el7.centos.noarch
>> > ovirt-engine-setup-plugin-websocket-proxy-4.0.5.5-1.el7.centos.noarch
>> > ovirt-engine-setup-4.0.5.5-1.el7.centos.noarch
>> > ovirt-engine-dbscripts-4.0.5.5-1.el7.centos.noarch
>> > ovirt-engine-userportal-4.0.5.5-1.el7.centos.noarch
>> > ovirt-imageio-common-0.4.0-1.el7.noarch
>> > python-ovirt-engine-sdk4-4.0.2-1.el7.centos.x86_64
>> > ovirt-vmconsole-host-1.0.4-1.el7.centos.noarch
>> > ovirt-engine-dwh-4.0.5-1.el7.centos.noarch
>> > ovirt-engine-tools-backup-4.0.5.5-1.el7.centos.noarch
>> > ovirt-image-uploader-4.0.1-1.el7.centos.noarch
>> > ovirt-engine-setup-plugin-ovirt-engine-4.0.5.5-1.el7.centos.noarch
>> > ovirt-engine-tools-4.0.5.5-1.el7.centos.noarch
>> > ovirt-engine-4.0.5.5-1.el7.centos.noarch
>> > ovirt-release40-4.0.5-2.noarch
>> > ovirt-host-deploy-java-1.5.3-1.el7.centos.noarch
>> > ovirt-engine-setup-plugin-ovirt-engine-common-4.0.5.5-1.el7.
>> centos.noarch
>> > ovirt-iso-uploader-4.0.2-1.el7.centos.noarch
>> > ovirt-engine-webadmin-portal-4.0.5.5-1.el7.centos.noarch
>> > ovirt-setup-lib-1.0.2-1.el7.centos.noarch
>> > ovirt-engine-sdk-python-3.6.9.1-1.el7.centos.noarch
>> > ovirt-engine-lib-4.0.5.5-1.el7.centos.noarch
>> > ovirt-imageio-proxy-setup-0.4.0-0.201608310602.gita9b573b.
>> > el7.centos.noarch
>> > ovirt-engine-setup-plugin-vmconsole-proxy-helper-4.0.5.
>> > 5-1.el7.centos.noarch
>> > ovirt-engine-backend-4.0.5.5-1.el7.centos.noarch
>> > ovirt-engine-extension-aaa-jdbc-1.1.1-1.el7.noarch
>> > ovirt-vmconsole-proxy-1.0.4-1.el7.centos.noarch
>> >
>> >
>>
>> --
>> С уважением, базисный админстратор
>> Гераcимов Александр
>> тел. +7(495)4116886 <+7%20495%20411-68-86> доб. 5367
>>
>>
>> _______________________________________________
>> Users mailing list
>> Users at ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
>
>
> --
>
> Lev Veyde
>
> Software Engineer, RHCE | RHCVA | MCITP
>
> Red Hat Israel
>
> <https://www.redhat.com>
>
> lev at redhat.com | lveyde at redhat.com
> <https://red.ht/sig>
> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted>
>
>
> --
> С уважением, базисный админстратор
> Гераcимов Александр
> тел. +7(495)4116886 <+7%20495%20411-68-86> доб. 5367
>
>


-- 

Lev Veyde

Software Engineer, RHCE | RHCVA | MCITP

Red Hat Israel

<https://www.redhat.com>

lev at redhat.com | lveyde at redhat.com
<https://red.ht/sig>
TRIED. TESTED. TRUSTED. <https://redhat.com/trusted>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20170608/fc4bf6f7/attachment-0001.html>


More information about the Users mailing list