On Mon, Nov 19, 2018 at 7:17 PM Gregory Rose <gvrose8...@gmail.com
<mailto:gvrose8...@gmail.com>> wrote:
Hi Siva,
One more request - I need to see the underlying network
configuration
of the hypervisor running the two VMs.
Are both VMs on the same machine? If so then just the network
configuration of the base machine
running the VMs, otherwise the network configuration of each base
machine running their perspective
VM.
This is turning into quite the investigation and I apologize that
it is
taking so long. Please bear with me
if you can and we'll see if we can't get this problem solved.
I've seen
some puzzling bugs before and
this one is turning out to be one of the best. Or worst....
depends on
your outlook. :)
Thanks for all your help so far!
- Greg
Hi Greg,
Both the VMs run on same hypervisor in my setup. Created VMs and
virtual networks using virsh commands. Virsh XMLs for networks look
like below
[user@hyp1 ] virsh net-dumpxml route1
<network connections='2'>
<name>route1</name>
<uuid>2c935aaf-ebde-5b76-a903-4fccb115ff75</uuid>
<forward mode='route'/>
<bridge name='testbr1' stp='on' delay='0'/>
<mac address='42:54:00:84:4e:04'/>
<ip address='20.20.0.1' netmask='255.255.255.0'>
<dhcp>
<range start='20.20.0.2' end='20.20.0.254'/>
</dhcp>
</ip>
</network>
[user@hyp1 ] network virsh net-dumpxml route2
<network connections='2'>
<name>route2</name>
<uuid>2c935baf-ebde-5b76-a903-4fccb115ff75</uuid>
<forward mode='route'/>
<bridge name='testbr2' stp='on' delay='0'/>
<mac address='42:54:10:84:4e:04'/>
<ip address='30.30.0.1' netmask='255.255.255.0'>
<dhcp>
<range start='30.30.0.2' end='30.30.0.254'/>
</dhcp>
</ip>
</network>
Each VM is connected to both the networks.
Some network configuration of the hypervisor.
[user@hyp-1] ip a
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN
group default qlen 1
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
inet 127.0.0.1/8 <http://127.0.0.1/8> scope host lo
valid_lft forever preferred_lft forever
inet6 ::1/128 scope host
valid_lft forever preferred_lft forever
2: enp5s0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1400 qdisc pfifo_fast
state UP group default qlen 1000
link/ether <mac> brd ff:ff:ff:ff:ff:ff
inet A.B.C.D/24 brd X.Y.Z.W scope global dynamic enp5s0
valid_lft 318349sec preferred_lft 318349sec
3: virbr0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
state UP group default qlen 1000
link/ether fe:54:00:0a:d3:70 brd ff:ff:ff:ff:ff:ff
inet 192.168.122.1/24 <http://192.168.122.1/24> brd
192.168.122.255 scope global virbr0
valid_lft forever preferred_lft forever
4: virbr0-nic: <BROADCAST,MULTICAST> mtu 1500 qdisc pfifo_fast state
DOWN group default qlen 1000
link/ether 52:54:00:94:4e:04 brd ff:ff:ff:ff:ff:ff
11: docker0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
state UP group default
link/ether 02:42:89:28:db:a5 brd ff:ff:ff:ff:ff:ff
inet 172.17.0.1/16 <http://172.17.0.1/16> scope global docker0
valid_lft forever preferred_lft forever
inet6 fe80::42:89ff:fe28:dba5/64 scope link
valid_lft forever preferred_lft forever
96: vboxnet0: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc
pfifo_fast state DOWN group default qlen 1000
link/ether 0a:00:27:00:00:00 brd ff:ff:ff:ff:ff:ff
inet 192.168.99.1/24 <http://192.168.99.1/24> brd 192.168.99.255
scope global vboxnet0
valid_lft forever preferred_lft forever
inet6 fe80::800:27ff:fe00:0/64 scope link
valid_lft forever preferred_lft forever
193: testbr0: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc
noqueue state DOWN group default qlen 1000
link/ether 00:00:00:00:00:00 brd ff:ff:ff:ff:ff:ff
inet 10.10.0.1/24 <http://10.10.0.1/24> brd 10.10.0.255 scope
global testbr0
valid_lft forever preferred_lft forever
194: testbr0-nic: <BROADCAST,MULTICAST> mtu 1500 qdisc pfifo_fast
state DOWN group default qlen 1000
link/ether 42:54:00:94:4e:04 brd ff:ff:ff:ff:ff:ff
227: testbr1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
state UP group default qlen 1000
link/ether fe:54:00:05:93:7c brd ff:ff:ff:ff:ff:ff
inet 20.20.0.1/24 <http://20.20.0.1/24> brd 20.20.0.255 scope
global testbr1
valid_lft forever preferred_lft forever
228: testbr1-nic: <BROADCAST,MULTICAST> mtu 1500 qdisc pfifo_fast
state DOWN group default qlen 1000
link/ether 42:54:00:84:4e:04 brd ff:ff:ff:ff:ff:ff
229: testbr2: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
state UP group default qlen 1000
link/ether fe:54:00:79:ef:92 brd ff:ff:ff:ff:ff:ff
inet 30.30.0.1/24 <http://30.30.0.1/24> brd 30.30.0.255 scope
global testbr2
valid_lft forever preferred_lft forever
230: testbr2-nic: <BROADCAST,MULTICAST> mtu 1500 qdisc pfifo_fast
state DOWN group default qlen 1000
link/ether 42:54:10:84:4e:04 brd ff:ff:ff:ff:ff:ff
231: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
pfifo_fast master virbr0 state UNKNOWN group default qlen 1000
link/ether fe:54:00:0a:d3:70 brd ff:ff:ff:ff:ff:ff
inet6 fe80::fc54:ff:fe0a:d370/64 scope link
valid_lft forever preferred_lft forever
232: vnet1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
pfifo_fast master testbr2 state UNKNOWN group default qlen 1000
link/ether fe:54:00:b8:05:be brd ff:ff:ff:ff:ff:ff
inet6 fe80::fc54:ff:feb8:5be/64 scope link
valid_lft forever preferred_lft forever
233: vnet2: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
pfifo_fast master testbr1 state UNKNOWN group default qlen 1000
link/ether fe:54:00:f0:64:37 brd ff:ff:ff:ff:ff:ff
inet6 fe80::fc54:ff:fef0:6437/64 scope link
valid_lft forever preferred_lft forever
234: vnet3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
pfifo_fast master virbr0 state UNKNOWN group default qlen 1000
link/ether fe:54:00:56:cb:89 brd ff:ff:ff:ff:ff:ff
inet6 fe80::fc54:ff:fe56:cb89/64 scope link
valid_lft forever preferred_lft forever
235: vnet4: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
pfifo_fast master testbr2 state UNKNOWN group default qlen 1000
link/ether fe:54:00:79:ef:92 brd ff:ff:ff:ff:ff:ff
inet6 fe80::fc54:ff:fe79:ef92/64 scope link
valid_lft forever preferred_lft forever
236: vnet5: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
pfifo_fast master testbr1 state UNKNOWN group default qlen 1000
link/ether fe:54:00:05:93:7c brd ff:ff:ff:ff:ff:ff
inet6 fe80::fc54:ff:fe05:937c/64 scope link
valid_lft forever preferred_lft forever
[user@hyp-1] ip route
default via A.B.C.D dev enp5s0 proto static metric 100
10.10.0.0/24 <http://10.10.0.0/24> dev testbr0 proto kernel scope
link src 10.10.0.1 linkdown
20.20.0.0/24 <http://20.20.0.0/24> dev testbr1 proto kernel scope
link src 20.20.0.1
30.30.0.0/24 <http://30.30.0.0/24> dev testbr2 proto kernel scope
link src 30.30.0.1
A.B.C.D via P.Q.R.S dev enp5s0 proto dhcp metric 100
X.Y.Z.W dev enp5s0 proto kernel scope link src A.B.C.D metric 100
172.17.0.0/16 <http://172.17.0.0/16> dev docker0 proto kernel scope
link src 172.17.0.1
192.168.99.0/24 <http://192.168.99.0/24> dev vboxnet0 proto kernel
scope link src 192.168.99.1 linkdown
192.168.122.0/24 <http://192.168.122.0/24> dev virbr0 proto kernel
scope link src 192.168.122.1
I am not completely sure what you meant by network configuration.
Kindly let me know if you are looking for something more specific.
There is one strange behavior that I observed on the VMs. I am able to
ping across networks if I use IP address but it does not work if I use
interface name directly.
[root@vm2 ~]# ip addr show eth1
3: eth1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
state UP qlen 1000
link/ether 52:54:00:79:ef:92 brd ff:ff:ff:ff:ff:ff
inet 30.30.0.193/24 <http://30.30.0.193/24> brd 30.30.0.255 scope
global dynamic eth1
valid_lft 2728sec preferred_lft 2728sec
inet6 fe80::5054:ff:fe79:ef92/64 scope link
valid_lft forever preferred_lft forever
[root@vm2 ~]# ping 20.20.0.183 -I eth1
PING 20.20.0.183 (20.20.0.183) from 30.30.0.193 eth1: 56(84) bytes of
data.
^C
--- 20.20.0.183 ping statistics ---
7 packets transmitted, 0 received, 100% packet loss, time 6000ms
[root@vm2 ~]# ping 20.20.0.183 -I 30.30.0.193
PING 20.20.0.183 (20.20.0.183) from 30.30.0.193 : 56(84) bytes of data.
64 bytes from 20.20.0.183 <http://20.20.0.183>: icmp_seq=1 ttl=64
time=0.766 ms
64 bytes from 20.20.0.183 <http://20.20.0.183>: icmp_seq=2 ttl=64
time=0.561 ms
64 bytes from 20.20.0.183 <http://20.20.0.183>: icmp_seq=3 ttl=64
time=0.605 ms
64 bytes from 20.20.0.183 <http://20.20.0.183>: icmp_seq=4 ttl=64
time=0.537 ms
64 bytes from 20.20.0.183 <http://20.20.0.183>: icmp_seq=5 ttl=64
time=0.607 ms
64 bytes from 20.20.0.183 <http://20.20.0.183>: icmp_seq=6 ttl=64
time=0.618 ms
64 bytes from 20.20.0.183 <http://20.20.0.183>: icmp_seq=7 ttl=64
time=0.624 ms
^C
--- 20.20.0.183 ping statistics ---
7 packets transmitted, 7 received, 0% packet loss, time 6000ms
rtt min/avg/max/mdev = 0.537/0.616/0.766/0.075 ms
[root@vm2 ~]#
I don't the reason behind this and I will need to understand this when
I get some time. Just letting you know if it might make any difference
in your setup.
Siva Teja.