Hi Andrei, Understand. To get all the obvious things out the way – have you allowed egress traffic on the two networks (you mention ACLs which we only use on VPCs and basic networks)?
Regards, Dag Sonstebo Cloud Architect ShapeBlue On 21/02/2018, 14:51, "Andrei Mikhailovsky" <and...@arhont.com.INVALID> wrote: Hi Dag, Please see my comments below: > Hi Andrei, > > You’re confusing the matter with your masking of public IP ranges. You said you > have “2 x Public IP ranges with /26 netmask” – but since you are masking them > out with X’s your email doesn’t make sense. If all the X’s are the same then a > .10 and a .20 IP address would be on the same /26 network. > > I will assume that you do in fact have 2 x 26-bit networks, e.g.: > > 192.168.0.0/26 – with default gateway 192.168.0.1 > 192.168.0.64/26 – with default gateway 192.168.0.65 > That is correct. I do have two separate /26 networks similar to what you've described above. However, one /26 is used for direct public IP service offering, where VRs are not involved in networking at all and the second /26 is used for the service offering where VRs are used to provide the networking function. > If your two guest networks have VRs on separate public IP ranges you will have > e.g. > > VR1: public IP 192.168.0.10 > VR2: public IP 192.168.0.70 > Nope, the guest vms with VRs that can't talk to each other are on the same /26 network. (in your example that would be on the same 192.168.0.0/26) > For a VM hosted behind VR1 to reach a service NAT’ed on VR2 you need to set up > routing and possibly firewalling on the data centre device which handles the > default gateway for the two networks – i.e. the top of rack switch or router > which hosts default gateways 192.168.0.1 and 192.168.0.65. The fact that you > can reach services on both networks from outside this range makes sense. > These has been set up and vms between separate /26 networks CAN talk to each other. The VMs on the same /26 network that doesn't use the VR service can also talk to each other. The problem with VMs on the same /26 that use VRs can't talk to each other using their public IP addresses. > So once you have fixed this you will have VM1 > VR1 > DC_SWITCH_OR_ROUTER > VR2 > > VM2. > > > Regards, > Dag Sonstebo > Cloud Architect > ShapeBlue > > On 21/02/2018, 12:27, "Andrei Mikhailovsky" <and...@arhont.com.INVALID> wrote: > > Hello > > Could someone help me to identify the routing issues that we have. The problem > is the traffic from different guest networks can not reach each other via the > public IPs. > > Here is my ACS setup: > ACS 4.9.3.0 (both management and agents) > KVM Hypervisor based on Ubuntu 16.04 > Ceph as primary storage. NFS as secondary storage > Advanced Networking with vlan separation > 2 x Public IP ranges with /26 netmask. > > > > Here is an example when routing DOES NOT work: > > Case 1 - Advanced Networking, vlan separation, VRs route all traffic and provide > all networking services (dhcp, fw, port forwarding, load balancing, etc) > > Guest Network 1: > > Public IP: XXX.XXX.XXX.10/26 > Private IP range: 10.1.1.0/24 > guest vm1 IP: 10.1.1.100/24 > > Guest Network 2: > Public IP: XXX.XXX.XXX.20/26 > Private IP range: 10.1.1.0/24 > guest vm2 IP: 10.1.1.200/24 > > > I've created ACLs on both guest networks to allow traffic from 0.0.0.0/0 on port > 80. I've created the port forwarding rules to forward port 80 from public > XXX.XXX.XXX.10 and XXX.XXX.XXX.XXX.20 onto 10.1.1.100 and 10.1.1.200 > respectively. > > This setup works perfectly well when I am initiating the connections from > outside of our CloudStack. However, vm2 can't reach vm1 on port 80 using the > public IP XXX.XXX.XXX.10 and vice versa, vm1 can't reach vm2 on public IP > XXX.XXX.XXX.20. > > > > > Here is an example when the routing DOES work: > > Case 2 - Advanced Networking, vlan separation, VRs are not used. Public IPs are > given directly to a guest vm > > Guest Network 1: > > guest vm1 Public IP: XXX.XXX.XXX.100/26 > > Guest Network 2: > > guest vm2 Public IP: XXX.XXX.XXX.110/26 > > In the Case 2, the guest vm has a public IP address directly assigned to its > network interface. VRs are not used for this networking. Each guest has a fw > rule to allow incoming traffic on port 80 from 0.0.0.0/0. Both vm1 and vm2 can > access each other on port 80. Also, vms from Case 1 above can access port 80 on > vms from Case 2, similarly, vms from Case 2 can access port 80 on vms from Case > 1. > > > > So, it seems that the rules on the VR in Case 1 do not allow traffic that > originates from other VRs within the same public network range. The trace route > shows the last hop being the VR's private IP address. How do I change that > behaviour and fix the networking issue? > > Thanks > > Andrei > > > > dag.sonst...@shapeblue.com > www.shapeblue.com > 53 Chandos Place, Covent Garden, London WC2N 4HSUK > @shapeblue dag.sonst...@shapeblue.com www.shapeblue.com 53 Chandos Place, Covent Garden, London WC2N 4HSUK @shapeblue