Hi Alex,

So if you're sending data out there is no or almost no data loss?

Have you tried to move the VM to the second host to see if it makes any
difference?

I actually tried to call you by phone but the call gets dropped once I
enter * and your ext. number, as the voice prompt says to.
Tried to enter the ext. number directly without *, still without success.

Thanks in advance,

On Thu, Jun 8, 2017 at 1:59 PM, Герасимов Александр <[email protected]>
wrote:

> Hi Lev.
>
> I am create new virtual host, and run ping from internet.
>
> Ping results 70% packet loss.
>
> But, if i am running ping from virtual host to somebody host in the
> internet and at the same time, i am pinging this virtual host from internet.
>
> then ping results 0% packet loss or 10% packet loss.
>
>
> 28.05.2017 13:34, Lev Veyde пишет:
>
> Hi Alex,
>
> That is quite strange...
>
> Does this happen on both hosts - have you tried to migrate the VM to the
> second host and see if the issue still remains?
>
> Thanks in advance,
>
>
> On Fri, May 26, 2017 at 3:02 PM, Герасимов Александр <
> [email protected]> wrote:
>
>> Hi Lev.
>>
>>
>> On one of the VMs you only see 1 NIC instead of the 2?
>>
>> NO. both VM's sees two NIC, but on first VM ping with no error, and
>> second VM ping with 75% error.
>>
>> OS version on hosts [root@node01 ~]# cat /etc/redhat-release
>> CentOS Linux release 7.3.1611 (Core)
>>
>> OS veriosion on VM's [root@node03 ~]# cat /etc/redhat-release
>> CentOS Linux release 7.3.1611 (Core)
>>
>>
>>
>> *first VM*
>>
>> 00:03.0 Ethernet controller: Red Hat, Inc Virtio network device
>>
>> 00:09.0 Ethernet controller: Red Hat, Inc Virtio network device
>>
>> [root@node03 ~]# ip l
>> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN mode
>> DEFAULT qlen 1
>>     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
>> state UP mode DEFAULT qlen 1000
>>     link/ether 00:1a:4a:16:01:51 brd ff:ff:ff:ff:ff:ff
>> 3: eth1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
>> state UP mode DEFAULT qlen 1000
>>     link/ether 00:1a:4a:16:01:55 brd ff:ff:ff:ff:ff:ff
>>
>> *second VM*
>>
>> 00:03.0 Ethernet controller: Realtek Semiconductor Co., Ltd.
>> RTL-8100/8101L/8139 PCI Fast Ethernet Adapter (rev 20)  - but i tested all
>> version of NIC and no effect
>>
>> 00:0a.0 Ethernet controller: Red Hat, Inc Virtio network device
>>
>> [root@node04 ~]# ip link
>> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN mode
>> DEFAULT qlen 1
>>     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
>> state UP mode DEFAULT qlen 1000
>>     link/ether 00:1a:4a:16:01:53 brd ff:ff:ff:ff:ff:ff
>> 3: ens3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
>> state UP mode DEFAULT qlen 1000
>>     link/ether 00:1a:4a:16:01:52 brd ff:ff:ff:ff:ff:ff
>>
>> In logs no messages only like this :
>>
>> May 26 15:01:01 node04 systemd: Started Session 67263 of user root.
>> May 26 15:01:01 node04 systemd: Starting Session 67263 of user root.
>> May 26 15:01:01 node04 systemd: Created slice user-600.slice.
>> May 26 15:01:01 node04 systemd: Starting user-600.slice.
>> May 26 15:01:01 node04 systemd: Started Session 67262 of user bitrix.
>> May 26 15:01:01 node04 systemd: Starting Session 67262 of user bitrix.
>> May 26 15:01:01 node04 systemd: Removed slice user-600.slice.
>> May 26 15:01:01 node04 systemd: Stopping user-600.slice.
>>
>>
>> Hi Alexander,
>>
>> So if I understand it correctly, you have the following configuration:
>> - 2 hosts, each having 2 NICs
>> - 2 virtual machines, each have a connection to each one of the NICs
>> available on the hosts
>>
>> On one of the VMs you only see 1 NIC instead of the 2?
>>
>> Are you sure that the VM is properly configured to have 2 NICs?
>>
>> What Linux distro and version you're using on the hosts and inside the
>> VMs ?
>>
>> Can you please send us:
>> - the logs from the VM, e.g. /var/log/messages
>> - the output of lspci -v
>> - the output of ip link
>>
>> Thanks in advance,
>>
>> 2017-05-18 12:19 GMT+03:00 Герасимов Александр <gerasimov.ay at eksmo.ru
>> >:
>>
>> > Hi all.
>> >
>> > I have to servers with ovirt.
>> >
>> > And to identical virtual machines.
>> >
>> > Both servers are identical. But on second virtual server not working one
>> > network interface. Ping have a problem. I tried to change network
>> driver,
>> > but has no effect.
>> >
>> > I don't understand that to do
>> >
>> >
>> > ovirt version and package:
>> >
>> > rpm -qa|grep ovirt
>> > ovirt-imageio-proxy-0.4.0-0.201608310602.gita9b573b.el7.centos.noarch
>> > ovirt-engine-vmconsole-proxy-helper-4.0.5.5-1.el7.centos.noarch
>> > ovirt-engine-restapi-4.0.5.5-1.el7.centos.noarch
>> > ovirt-engine-extensions-api-impl-4.0.5.5-1.el7.centos.noarch
>> > ovirt-imageio-daemon-0.4.0-1.el7.noarch
>> > ovirt-engine-wildfly-10.1.0-1.el7.x86_64
>> > ovirt-vmconsole-1.0.4-1.el7.centos.noarch
>> > ovirt-engine-cli-3.6.9.2-1.el7.noarch
>> > ovirt-engine-websocket-proxy-4.0.5.5-1.el7.centos.noarch
>> > ovirt-engine-dashboard-1.0.5-1.el7.centos.noarch
>> > ovirt-host-deploy-1.5.3-1.el7.centos.noarch
>> > ovirt-engine-wildfly-overlay-10.0.0-1.el7.noarch
>> > ovirt-engine-setup-base-4.0.5.5-1.el7.centos.noarch
>> > ovirt-engine-dwh-setup-4.0.5-1.el7.centos.noarch
>> > ovirt-engine-setup-plugin-websocket-proxy-4.0.5.5-1.el7.centos.noarch
>> > ovirt-engine-setup-4.0.5.5-1.el7.centos.noarch
>> > ovirt-engine-dbscripts-4.0.5.5-1.el7.centos.noarch
>> > ovirt-engine-userportal-4.0.5.5-1.el7.centos.noarch
>> > ovirt-imageio-common-0.4.0-1.el7.noarch
>> > python-ovirt-engine-sdk4-4.0.2-1.el7.centos.x86_64
>> > ovirt-vmconsole-host-1.0.4-1.el7.centos.noarch
>> > ovirt-engine-dwh-4.0.5-1.el7.centos.noarch
>> > ovirt-engine-tools-backup-4.0.5.5-1.el7.centos.noarch
>> > ovirt-image-uploader-4.0.1-1.el7.centos.noarch
>> > ovirt-engine-setup-plugin-ovirt-engine-4.0.5.5-1.el7.centos.noarch
>> > ovirt-engine-tools-4.0.5.5-1.el7.centos.noarch
>> > ovirt-engine-4.0.5.5-1.el7.centos.noarch
>> > ovirt-release40-4.0.5-2.noarch
>> > ovirt-host-deploy-java-1.5.3-1.el7.centos.noarch
>> > ovirt-engine-setup-plugin-ovirt-engine-common-4.0.5.5-1.el7.
>> centos.noarch
>> > ovirt-iso-uploader-4.0.2-1.el7.centos.noarch
>> > ovirt-engine-webadmin-portal-4.0.5.5-1.el7.centos.noarch
>> > ovirt-setup-lib-1.0.2-1.el7.centos.noarch
>> > ovirt-engine-sdk-python-3.6.9.1-1.el7.centos.noarch
>> > ovirt-engine-lib-4.0.5.5-1.el7.centos.noarch
>> > ovirt-imageio-proxy-setup-0.4.0-0.201608310602.gita9b573b.
>> > el7.centos.noarch
>> > ovirt-engine-setup-plugin-vmconsole-proxy-helper-4.0.5.
>> > 5-1.el7.centos.noarch
>> > ovirt-engine-backend-4.0.5.5-1.el7.centos.noarch
>> > ovirt-engine-extension-aaa-jdbc-1.1.1-1.el7.noarch
>> > ovirt-vmconsole-proxy-1.0.4-1.el7.centos.noarch
>> >
>> >
>>
>> --
>> С уважением, базисный админстратор
>> Гераcимов Александр
>> тел. +7(495)4116886 <+7%20495%20411-68-86> доб. 5367
>>
>>
>> _______________________________________________
>> Users mailing list
>> [email protected]
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
>
>
> --
>
> Lev Veyde
>
> Software Engineer, RHCE | RHCVA | MCITP
>
> Red Hat Israel
>
> <https://www.redhat.com>
>
> [email protected] | [email protected]
> <https://red.ht/sig>
> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted>
>
>
> --
> С уважением, базисный админстратор
> Гераcимов Александр
> тел. +7(495)4116886 <+7%20495%20411-68-86> доб. 5367
>
>


-- 

Lev Veyde

Software Engineer, RHCE | RHCVA | MCITP

Red Hat Israel

<https://www.redhat.com>

[email protected] | [email protected]
<https://red.ht/sig>
TRIED. TESTED. TRUSTED. <https://redhat.com/trusted>
_______________________________________________
Users mailing list
[email protected]
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to