Hello Wido and community, Just to have the basics in simple terms:
To use the VXLAN in Cloudstack - We need Management and Storage reacheable by plain L3, Correct ? For the KVM Host initially: Option 1) At KVM, Do I need a PLAIN L3 only initially, WITH NO Initial cloudbr1/VXLAN? Yes/ NO? Option 2) At KVM host, I need the Cloudbr1 with the initial VXLAN reacheable somehow by the management server ? Thank you very Much. Tata Y. > On Jan 14, 2025, at 3:30 AM, Wido den Hollander <w...@widodh.nl> wrote: > > > > Op 13/01/2025 om 17:06 schreef Chi vediamo: >> Hello Wido, Thank you so much for taking the time to respond.: >> I bonded the interfaces for capacity. > > I would just create additional BGP sessions and use ECMP balancing for more > capacity. > >> I was following your YouTube videos and users mail. >> Yes, VXLAN seems to be working now. The most dificult/radical part was the >> tunning of the parameres on Ubuntu 22. or should I use a different Linux >> system? >> Let me put on a logic it may clear for me; then, with VXLAN: >> Wido: Regarding your Youtube video: To me seems there is a VLAN on the TOR. >> But makes no sense based on what you mention here. What is the VLAN624 being >> used on your Youtube video @time 24:10 >> https://www.youtube.com/watch?v=X02bxtIC0u4 > > VLAN is a Cumulus naming. These are all VXLAN VNI. In this case I showed the > example how Cumulus would be the gateway for VNI 624 > >> Then, Management servers will be on the same vxlan as kvm cloudbr1 then this >> should be the management vxlan. Correct? >> Storage I did put the ceph on vxlan2 with BGP; then, should I removed from >> vxlan ior keep it on the management vxlan? which one should be the best >> practice.? > > Storage doesn't need to be in a VNI/VXLAN. It just needs to be reachable by > the hypervisor. Plain L3 routing, nothing special. Just make sure both can > ping eachother. > > This is really just basic routing. > >> What About Network planning with Cloudstack for NFVs and low latency: >> Final Question. Does Cloudstack supports Sriov ? Any documentation besides >> an old document that inidicates we should list the PCI interfaces ? IS there >> ANy documentation about this? > > > No idea :-) > >> I was thinnking SRIOV will align better with VXLAN/EVPN >> Or do you recommend to use DPDk ? >> If DPDK wihich one will you recommend, >> - OpenVswitch - if this is selected then I will have to create the Bridges >> with Openvswitch, and should work with VXLAN/EVPN , right? >> - Or tunsgteen(OpenSDN) - I know Contrail pretty much very well, Not sure >> if there is any documentation about OpenSDN - Formerly Tungteen - with >> Cloudstack and VXLAN/EVPN, >> - Seems VXLAN/EVPN and tungsteen are mutually exclusive, plus not supported >> on Ubuntu 22 as of yet, are both of this statements correct? >> Thank you so much for taking the time to respond. >> Tata Y. >>> On Jan 13, 2025, at 8:45 AM, Wido den Hollander <w...@widodh.nl> wrote: >>> >>> >>> >>> Op 11/01/2025 om 00:36 schreef Chi vediamo: >>>> Forgot to add Wido >>>> The Isolation should start, and should happen at the HOST running KVM, No >>>> VLANS needed at the Host, neither at the Leafs. >>>> But seems isolation is not happening! using cloudstack 4.20, maybe is a >>>> detail missing on my side. >>>> What Cloudstac documentation mean with "On the Spine router(s) the VNIs >>>> will terminate and they will act as IPv4/IPv6 gateways" >>>> DO "Need" vlans to be configured on the Leafs. I will understand if the VM >>>> isolation is based on VLANS that will be needed, but if the VM Isolation >>>> is going to be VXLAN >>>> what for I will need the VLANS at the leafs? >>> >>> No VLANs needed. Forget Guest and Storage traffic. The hypervisor will do >>> everything routed! It will talk BGP with the Top-of-Rack. >>> >>> You will reach your storage via routing without the need of VXLAN, have it >>> route through your network. >>> >>> No need for bonding on your hypervisor either, it just has two BGP uplink >>> sessions towards the ToR. >>> >>> Just connect your storage to your network and make it reachable via L3, >>> same goes for the management server. >>> >>> But before starting to use CloudStack with VXLAN, make sure you get VXLAN >>> working without CloudStack so you really understand how VXLAN works. >>> >>> Wido >>> >>>> based on this scenario from Guido presso: >>>> [spine1] [spine2] no ibgp between spines >>>> / / >>>> / / evpn/bgp unnumbered >>>> / / >>>> / / >>>> [leaf1]-----ibgp---[Leaf2] >>>> \ / >>>> \ / >>>> \ / >>>> \ / >>>> | bond1 bond2 | >>>> | [cloudbr1] | >>>> | vxlan1 | >>>> | vxlan2 | >>>> | Host 1 - KVM | >>>> | SRIOV | >>>> -------------------- >>>> the vxlan1 will handle the guest traffic >>>> the vxlan2 will handle main storage >>>> I have management over a regular vlan for now but will create a vxlan3 for >>>> it >>>> and vxlan2000 for Public traffic. >>>> vxlan1, vxlan2, vxlan3, and vxlan2000 are goint to be created manually or >>>> Cloudstack should/will create them when I choose the VXLAN for each one of >>>> the traffic names? >>>> Or should I use VXLAN for the Guest traffic only and use regular vlan >>>> isolation for the the remaining - management, storage and public ? >>>> I have 4 physical hosts, and i am using same ip addressing on vxlan 1 and >>>> vxlan 2 connect to different leafs and i am still able to ping between >>>> them. >>>> Thank you >>>> Tata Y. >>>>> On Jan 6, 2025, at 9:08 PM, Chi vediamo <tatay...@gmail.com> wrote: >>>>> >>>>> >>>>> Hello Community, >>>>> First: Thank you Wido, your answers on previous emails to the community >>>>> help a lot. I read the vincent.bernat document, but his example uses VLAN >>>>> mapping at the Switch level. >>>>> >>>>> I was thinking to use the LEAF-SPINE as a transport only , and the >>>>> seetings on the Host will take care of the Isolation. >>>>> >>>>> But is not working that way. Should I create the traditional VXLAN, >>>>> VLAN/VNI/VRF on the LEAF Switches to properly isolate it? >>>>> We are using SONIC NOS community version, nothing fancy. >>>>> >>>>> The BGP unnumbered evpn etc works fine. >>>>> >>>>> The output: >>>>> >>>>> vtysh -c 'show interface vxlan2' >>>>> VNI: 2 >>>>> Type: L2 >>>>> Tenant VRF: default >>>>> VxLAN interface: vxlan2 >>>>> VxLAN ifIndex: 14 >>>>> SVI interface: storage0 >>>>> SVI ifIndex: 13 >>>>> Local VTEP IP: 172.2.0.60 >>>>> Mcast group: 0.0.0.0 >>>>> Remote VTEPs for this VNI: >>>>> 172.2.0.59 flood: HER >>>>> 172.2.0.32 flood: HER >>>>> 172.2.0.30 flood: HER >>>>> 172.2.0.28 flood: HER >>>>> 172.2.0.26 flood: HER >>>>> >>>>> bridge fdb show dev vxlan2 >>>>> 8a:be:71:4c:e0:20 vlan 1 extern_learn master storage0 >>>>> 8a:be:71:4c:e0:20 extern_learn master storage0 >>>>> b2:33:bb:84:cc:38 vlan 1 extern_learn master storage0 >>>>> b2:33:bb:84:cc:38 extern_learn master storage0 >>>>> 86:07:90:2b:db:db vlan 1 extern_learn master storage0 >>>>> 86:07:90:2b:db:db extern_learn master storage0 >>>>> 4a:28:60:90:76:42 vlan 1 extern_learn master storage0 >>>>> 4a:28:60:90:76:42 extern_learn master storage0 >>>>> 22:d6:49:9f:08:07 vlan 1 extern_learn master storage0 >>>>> 22:d6:49:9f:08:07 extern_learn master storage0 >>>>> fe:4a:fb:63:9d:3a vlan 1 extern_learn master storage0 >>>>> fe:4a:fb:63:9d:3a extern_learn master storage0 >>>>> ee:78:b4:d8:3f:a0 vlan 1 master storage0 permanent >>>>> ee:78:b4:d8:3f:a0 master storage0 permanent >>>>> 00:00:00:00:00:00 dst 172.2.0.24 self permanent >>>>> 00:00:00:00:00:00 dst 172.2.0.26 self permanent >>>>> 00:00:00:00:00:00 dst 172.2.0.28 self permanent >>>>> 00:00:00:00:00:00 dst 172.2.0.30 self permanent >>>>> 00:00:00:00:00:00 dst 172.2.0.32 self permanent >>>>> 00:00:00:00:00:00 dst 172.2.0.59 self permanent >>>>> >>>>> fe:4a:fb:63:9d:3a dst 172.2.0.24 self extern_learn >>>>> >>>>> >>>>> vtysh -c 'show interface vxlan1' >>>>> >>>>> Interface vxlan1 is up, line protocol is up >>>>> Link ups: 1 last: 2025/01/06 23:53:01.17 >>>>> Link downs: 1 last: 2025/01/06 23:53:01.17 >>>>> vrf: default >>>>> index 14 metric 0 mtu 9050 speed 4294967295 >>>>> flags: <UP,BROADCAST,RUNNING,MULTICAST> >>>>> Type: Ethernet >>>>> HWaddr: ea:d3:68:02:7d:f7 >>>>> inet6 fe80::e8d3:68ff:fe02:7df7/64 >>>>> Interface Type Vxlan >>>>> Interface Slave Type None >>>>> VxLAN Id 100 VTEP IP: 10.23.13.14 Access VLAN Id 1 >>>>> >>>>> protodown: off >>>>> >>>>> vtysh -c 'show evpn vni 1' >>>>> VNI: 1 >>>>> Type: L2 >>>>> Tenant VRF: default >>>>> VxLAN interface: vxlan1 >>>>> VxLAN ifIndex: 14 >>>>> SVI interface: cloudbr1 >>>>> SVI ifIndex: 12 >>>>> Local VTEP IP: 10.23.13.14 >>>>> Mcast group: 0.0.0.0 >>>>> No remote VTEPs known for this VNI >>>>> Number of MACs (local and remote) known for this VNI: 0 >>>>> Number of ARPs (IPv4 and IPv6, local and remote) known for this VNI: 0 >>>>> Advertise-gw-macip: No >>>>> Advertise-svi-macip: No >>>>> >>>>> and I can ping the IPV6 that is routed using the FRR from :60 which is in >>>>> VXLAN 2 to :14 which is in VXLAN 1 >>>>> >>>>>> ping -I 20XX:5XX:56XX:fff0::2:60 20XX:5XX:56XX:fff0:0:2:13:14 >>>>>> PING 20XX:5XX:56XX:fff0:0:2:13:14(20XX:5XX:56XX:fff0:0:2:13:14) from >>>>>> 20XX:5XX:56XX:fff0::2:60 : 56 data bytes >>>>>> 64 bytes from 20XX:5XX:56XX:fff0:0:2:13:14: icmp_seq=1 ttl=61 time=0.293 >>>>>> ms >>>>>> 64 bytes from 20XX:5XX:56XX:fff0:0:2:13:14: icmp_seq=2 ttl=61 time=0.222 >>>>>> ms >>>>> >>>>> >>>>> Then, my questions are: >>>>> are you using at the Leaf Switches/routers a regular Mapping VLAN to VNI >>>>> VXLAN with VRF ? IF not, Can you share a FRR config of your Switches? >>>>> Or should I use an enterprise SONIC switch software ? >>>>> What other possibilities are with the modifyvxlan.sh That Wido states on >>>>> some user mails. >>>>> >>>>> >>>>> Thank you >>>>> >>>>> Tata Y. >>>>> >>>>> >>> >