Doesn't solution like this works seamlessly for large VXLAN networks? https://vincent.bernat.ch/en/blog/2017-vxlan-bgp-evpn
вт, 23 окт. 2018 г., 8:34 Simon Weller <swel...@ena.com.invalid>: > Linux native VXLAN uses multicast and each host has to participate in > multicast in order to see the VXLAN networks. We haven't tried using PIM > across a L3 boundary with ACS, although it will probably work fine. > > Another option is to use a L3 VTEP, but right now there is no native > support for that in CloudStack's VXLAN implementation, although we've > thought about proposing it as feature. > > > ________________________________ > From: Wido den Hollander <w...@widodh.nl> > Sent: Tuesday, October 23, 2018 7:17 AM > To: dev@cloudstack.apache.org; Simon Weller > Subject: Re: VXLAN and KVm experiences > > > > On 10/23/18 1:51 PM, Simon Weller wrote: > > We've also been using VXLAN on KVM for all of our isolated VPC guest > networks for quite a long time now. As Andrija pointed out, make sure you > increase the max_igmp_memberships param and also put an ip address on each > interface host VXLAN interface in the same subnet for all hosts that will > share networking, or multicast won't work. > > > > Thanks! So you are saying that all hypervisors need to be in the same L2 > network or are you routing the multicast? > > My idea was that each POD would be an isolated Layer 3 domain and that a > VNI would span over the different Layer 3 networks. > > I don't like STP and other Layer 2 loop-prevention systems. > > Wido > > > > > - Si > > > > > > ________________________________ > > From: Wido den Hollander <w...@widodh.nl> > > Sent: Tuesday, October 23, 2018 5:21 AM > > To: dev@cloudstack.apache.org > > Subject: Re: VXLAN and KVm experiences > > > > > > > > On 10/23/18 11:21 AM, Andrija Panic wrote: > >> Hi Wido, > >> > >> I have "pioneered" this one in production for last 3 years (and > suffered a > >> nasty pain of silent drop of packages on kernel 3.X back in the days > >> because of being unaware of max_igmp_memberships kernel parameters, so I > >> have updated the manual long time ago). > >> > >> I never had any issues (beside above nasty one...) and it works very > well. > > > > That's what I want to hear! > > > >> To avoid above issue that I described - you should increase > >> max_igmp_memberships (/proc/sys/net/ipv4/igmp_max_memberships) - > otherwise > >> with more than 20 vxlan interfaces, some of them will stay in down state > >> and have a hard traffic drop (with proper message in agent.log) with > kernel > >>> 4.0 (or I silent, bitchy random packet drop on kernel 3.X...) - and > also > >> pay attention to MTU size as well - anyway everything is in the manual > (I > >> updated everything I though was missing) - so please check it. > >> > > > > Yes, the underlying network will all be 9000 bytes MTU. > > > >> Our example setup: > >> > >> We have i.e. bond.950 as the main VLAN which will carry all vxlan > "tunnels" > >> - so this is defined as KVM traffic label. In our case it didn't make > sense > >> to use bridge on top of this bond0.950 (as the traffic label) - you can > >> test it on your own - since this bridge is used only to extract child > >> bond0.950 interface name, then based on vxlan ID, ACS will provision > >> vxlan...@bond0.xxx and join this new vxlan interface to NEW bridge > created > >> (and then of course vNIC goes to this new bridge), so original bridge > (to > >> which bond0.xxx belonged) is not used for anything. > >> > > > > Clear, I indeed thought something like that would happen. > > > >> Here is sample from above for vxlan 867 used for tenant isolation: > >> > >> root@hostname:~# brctl show brvx-867 > >> > >> bridge name bridge id STP enabled interfaces > >> brvx-867 8000.2215cfce99ce no vnet6 > >> > >> vxlan867 > >> > >> root@hostname:~# ip -d link show vxlan867 > >> > >> 297: vxlan867: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 8142 qdisc noqueue > >> master brvx-867 state UNKNOWN mode DEFAULT group default qlen 1000 > >> link/ether 22:15:cf:ce:99:ce brd ff:ff:ff:ff:ff:ff promiscuity 1 > >> vxlan id 867 group 239.0.3.99 dev bond0.950 port 0 0 ttl 10 ageing > 300 > >> > >> root@ix1-c7-2:~# ifconfig bond0.950 | grep MTU > >> UP BROADCAST RUNNING MULTICAST MTU:8192 Metric:1 > >> > >> So note how the vxlan interface has by 50 bytes smaller MTU than the > >> bond0.950 parent interface (which could affects traffic inside VM) - so > >> jumbo frames are needed anyway on the parent interface (bond.950 in > example > >> above with minimum of 1550 MTU) > >> > > > > Yes, thanks! We will be using 1500 MTU inside the VMs, so all the > > networks underneath will be ~9k. > > > >> Ping me if more details needed, happy to help. > >> > > > > Awesome! We'll be doing a PoC rather soon. I'll come back with our > > experiences later. > > > > Wido > > > >> Cheers > >> Andrija > >> > >> On Tue, 23 Oct 2018 at 08:23, Wido den Hollander <w...@widodh.nl> > wrote: > >> > >>> Hi, > >>> > >>> I just wanted to know if there are people out there using KVM with > >>> Advanced Networking and using VXLAN for different networks. > >>> > >>> Our main goal would be to spawn a VM and based on the network the NIC > is > >>> in attach it to a different VXLAN bridge on the KVM host. > >>> > >>> It seems to me that this should work, but I just wanted to check and > see > >>> if people have experience with it. > >>> > >>> Wido > >>> > >> > >> > > >