I need some additional guidance.

I was following I think all emails I found:


https://www.mail-archive.com/users@cloudstack.apache.org/msg38549.html 

https://www.mail-archive.com/users@cloudstack.apache.org/msg36477.html 

https://www.mail-archive.com/users@cloudstack.apache.org/msg38144.html 

Based on those previous e-mails, I have very similar outputs as Wido show 
there, only the vni and IP addressing changes.

Does anybody have a tip, I am using Ubuntu and netplan with networkd


> On Jan 17, 2025, at 6:06 AM, Wido den Hollander <w...@widodh.nl> wrote:
> 
> 
> 
> Op 16/01/2025 om 13:04 schreef Chi vediamo:
>> Hello Wido and community,
>> Just to have the basics in simple terms:
>> *To use the VXLAN in Cloudstack -*
>> We need Management and Storage reacheable by plain L3, Correct ?
> 
> Correct. Although you can also use a VNI for reaching management. You need at 
> least one VNI for cloudbr1 where all the Hypervisors can reach eachother for 
> migration traffic and such.

What is the VNI on cloudbr1 used for ? Guest traffic?

> 
>> *For the KVM Host initially:*
>> *Option 1)* At KVM, Do I need a PLAIN L3 only initially, WITH*NO Initial 
>> cloudbr1/VXLAN*? Yes/ NO?
>> *Option 2)* At KVM host, I need the Cloudbr1 with the initial VXLAN 
>> reacheable somehow by the management server ?
> 
> Yes, you need that. As said above, you need cloudbr1 for traffic within the 
> cluster. We use a dedicated VNI which we configure in systemd-networkd on our 
> machines.

Do I need VRF's ?

For a test purpose I created additional vxlans and is using the script, a vpc 
on vxlan 200 can reach a vpc on vxlan 250 two different customers. those VPC 
can not have the same private IP addressing.

In that case I will need to create a bunch of VRFs on the Switches. and 
remodify the script from 
https://gist.github.com/wido/51cb9880d86f08f73766634d7f6df3f4, unless there is 
something new.


Tata Y.


> 
> Wido
> 
>> Thank you very Much.
>> Tata Y.
>>> On Jan 14, 2025, at 3:30 AM, Wido den Hollander <w...@widodh.nl> wrote:
>>> 
>>> 
>>> 
>>> Op 13/01/2025 om 17:06 schreef Chi vediamo:
>>>> Hello Wido, Thank you so much for taking the time to respond.:
>>>> I bonded the interfaces for capacity.
>>> 
>>> I would just create additional BGP sessions and use ECMP balancing for more 
>>> capacity.
>>> 
>>>> I was following your YouTube videos and users mail.
>>>> Yes, VXLAN seems to be working now. The most dificult/radical part was the 
>>>> tunning of the parameres on Ubuntu 22. or should I use a different Linux 
>>>> system?
>>>> Let me put on a logic it may clear for me; then, with VXLAN:
>>>> Wido: Regarding your Youtube video: To me seems there is a VLAN on the 
>>>> TOR. But makes no sense based on what you mention here. What is the 
>>>> VLAN624 being used on your Youtube video @time 24:10 https:// 
>>>> www.youtube.com/watch?v=X02bxtIC0u4
>>> 
>>> VLAN is a Cumulus naming. These are all VXLAN VNI. In this case I showed 
>>> the example how Cumulus would be the gateway for VNI 624
>>> 
>>>> Then, Management servers will be on the same vxlan as kvm cloudbr1 then 
>>>> this should be the management vxlan. Correct?
>>>> Storage I did put the ceph on vxlan2 with BGP; then, should I removed from 
>>>> vxlan ior keep it on the management vxlan? which one should be the best 
>>>> practice.?
>>> 
>>> Storage doesn't need to be in a VNI/VXLAN. It just needs to be reachable by 
>>> the hypervisor. Plain L3 routing, nothing special. Just make sure both can 
>>> ping eachother.
>>> 
>>> This is really just basic routing.
>>> 
>>>> What About Network planning with Cloudstack for NFVs and low latency:
>>>> Final Question. Does Cloudstack supports Sriov ? Any documentation besides 
>>>> an old document that inidicates we should list the PCI interfaces ? IS 
>>>> there ANy documentation about this?
>>> 
>>> 
>>> No idea :-)
>>> 
>>>> I was thinnking SRIOV will align better with VXLAN/EVPN
>>>> Or do you recommend to use DPDk ?
>>>> If DPDK wihich one will you recommend,
>>>>  - OpenVswitch - if this is selected then I will have to create the 
>>>> Bridges with Openvswitch, and should work with VXLAN/EVPN , right?
>>>>  - Or tunsgteen(OpenSDN) - I know Contrail pretty much very well, Not sure 
>>>> if there is any documentation about OpenSDN - Formerly Tungteen - with 
>>>> Cloudstack and VXLAN/EVPN,
>>>> -  Seems VXLAN/EVPN and tungsteen are mutually exclusive, plus not 
>>>> supported on Ubuntu 22 as of yet, are both of this statements correct?
>>>> Thank you so much for taking the time to respond.
>>>> Tata Y.
>>>>> On Jan 13, 2025, at 8:45 AM, Wido den Hollander <w...@widodh.nl> wrote:
>>>>> 
>>>>> 
>>>>> 
>>>>> Op 11/01/2025 om 00:36 schreef Chi vediamo:
>>>>>> Forgot to add Wido
>>>>>> The Isolation should start, and should happen at the HOST running KVM, 
>>>>>> No VLANS needed at the Host, neither at the Leafs.
>>>>>> But seems isolation is not happening! using cloudstack 4.20, maybe is a 
>>>>>> detail missing on my side.
>>>>>> What Cloudstac documentation mean with "On the Spine router(s) the VNIs 
>>>>>> will terminate and they will act as IPv4/IPv6 gateways"
>>>>>> DO "Need" vlans to be configured on the Leafs. I will understand if the 
>>>>>> VM isolation is based on VLANS that will be needed, but if the VM 
>>>>>> Isolation is going to be VXLAN
>>>>>> what for I will need the VLANS at the leafs?
>>>>> 
>>>>> No VLANs needed. Forget Guest and Storage traffic. The hypervisor will do 
>>>>> everything routed! It will talk BGP with the Top-of-Rack.
>>>>> 
>>>>> You will reach your storage via routing without the need of VXLAN, have 
>>>>> it route through your network.
>>>>> 
>>>>> No need for bonding on your hypervisor either, it just has two BGP uplink 
>>>>> sessions towards the ToR.
>>>>> 
>>>>> Just connect your storage to your network and make it reachable via L3, 
>>>>> same goes for the management server.
>>>>> 
>>>>> But before starting to use CloudStack with VXLAN, make sure you get VXLAN 
>>>>> working without CloudStack so you really understand how VXLAN works.
>>>>> 
>>>>> Wido
>>>>> 
>>>>>> based on this scenario from Guido presso:
>>>>>>            [spine1]           [spine2] no ibgp between spines
>>>>>>          /                          /
>>>>>>        /                           /   evpn/bgp unnumbered
>>>>>>      /                           /
>>>>>>    /                            /
>>>>>> [leaf1]-----ibgp---[Leaf2]
>>>>>>     \                           /
>>>>>>      \                         /
>>>>>>        \                      /
>>>>>>          \                  /
>>>>>>        | bond1     bond2 |
>>>>>>        |     [cloudbr1]      |
>>>>>>        |       vxlan1          |
>>>>>>        |       vxlan2          |
>>>>>>        |   Host 1 - KVM   |
>>>>>>        |        SRIOV         |
>>>>>>        --------------------
>>>>>> the vxlan1 will handle the guest traffic
>>>>>> the vxlan2 will handle main storage
>>>>>> I have management over a regular vlan for now but will create a vxlan3 
>>>>>> for it
>>>>>> and vxlan2000 for Public traffic.
>>>>>> vxlan1, vxlan2, vxlan3, and vxlan2000 are goint to be created manually 
>>>>>> or Cloudstack should/will create them when I choose the VXLAN for each 
>>>>>> one of the traffic names?
>>>>>> Or should I use VXLAN for the Guest traffic only and use regular vlan 
>>>>>> isolation for the the remaining - management, storage and public ?
>>>>>> I have 4 physical hosts, and i am using same ip addressing on vxlan 1 
>>>>>> and vxlan 2 connect to different leafs and i am still able to ping 
>>>>>> between them.
>>>>>> Thank you
>>>>>> Tata Y.
>>>>>>> On Jan 6, 2025, at 9:08 PM, Chi vediamo <tatay...@gmail.com> wrote:
>>>>>>> 
>>>>>>> 
>>>>>>> Hello Community,
>>>>>>> First: Thank you Wido, your answers on previous emails to the community 
>>>>>>> help a lot. I read the vincent.bernat document, but his example uses 
>>>>>>> VLAN mapping at the Switch level.
>>>>>>> 
>>>>>>> I was thinking to use the LEAF-SPINE as a transport only , and the 
>>>>>>> seetings on the Host will take care of the Isolation.
>>>>>>> 
>>>>>>> But is not working that way. Should I create the traditional VXLAN, 
>>>>>>> VLAN/VNI/VRF on the LEAF Switches to properly isolate it?
>>>>>>> We are using SONIC NOS community version, nothing fancy.
>>>>>>> 
>>>>>>> The BGP unnumbered evpn etc works fine.
>>>>>>> 
>>>>>>> The output:
>>>>>>> 
>>>>>>> vtysh -c 'show interface vxlan2'
>>>>>>> VNI: 2
>>>>>>>  Type: L2
>>>>>>>  Tenant VRF: default
>>>>>>>  VxLAN interface: vxlan2
>>>>>>>  VxLAN ifIndex: 14
>>>>>>>  SVI interface: storage0
>>>>>>>  SVI ifIndex: 13
>>>>>>>  Local VTEP IP: 172.2.0.60
>>>>>>>  Mcast group: 0.0.0.0
>>>>>>>  Remote VTEPs for this VNI:
>>>>>>>   172.2.0.59 flood: HER
>>>>>>>   172.2.0.32 flood: HER
>>>>>>>   172.2.0.30 flood: HER
>>>>>>>   172.2.0.28 flood: HER
>>>>>>>   172.2.0.26 flood: HER
>>>>>>> 
>>>>>>> bridge fdb show dev vxlan2
>>>>>>> 8a:be:71:4c:e0:20 vlan 1 extern_learn master storage0
>>>>>>> 8a:be:71:4c:e0:20 extern_learn master storage0
>>>>>>> b2:33:bb:84:cc:38 vlan 1 extern_learn master storage0
>>>>>>> b2:33:bb:84:cc:38 extern_learn master storage0
>>>>>>> 86:07:90:2b:db:db vlan 1 extern_learn master storage0
>>>>>>> 86:07:90:2b:db:db extern_learn master storage0
>>>>>>> 4a:28:60:90:76:42 vlan 1 extern_learn master storage0
>>>>>>> 4a:28:60:90:76:42 extern_learn master storage0
>>>>>>> 22:d6:49:9f:08:07 vlan 1 extern_learn master storage0
>>>>>>> 22:d6:49:9f:08:07 extern_learn master storage0
>>>>>>> fe:4a:fb:63:9d:3a vlan 1 extern_learn master storage0
>>>>>>> fe:4a:fb:63:9d:3a extern_learn master storage0
>>>>>>> ee:78:b4:d8:3f:a0 vlan 1 master storage0 permanent
>>>>>>> ee:78:b4:d8:3f:a0 master storage0 permanent
>>>>>>> 00:00:00:00:00:00 dst 172.2.0.24 self permanent
>>>>>>> 00:00:00:00:00:00 dst 172.2.0.26 self permanent
>>>>>>> 00:00:00:00:00:00 dst 172.2.0.28 self permanent
>>>>>>> 00:00:00:00:00:00 dst 172.2.0.30 self permanent
>>>>>>> 00:00:00:00:00:00 dst 172.2.0.32 self permanent
>>>>>>> 00:00:00:00:00:00 dst 172.2.0.59 self permanent
>>>>>>> 
>>>>>>> fe:4a:fb:63:9d:3a dst 172.2.0.24 self extern_learn
>>>>>>> 
>>>>>>> 
>>>>>>> vtysh -c 'show interface vxlan1'
>>>>>>> 
>>>>>>> Interface vxlan1 is up, line protocol is up
>>>>>>>   Link ups:       1    last: 2025/01/06 23:53:01.17
>>>>>>>   Link downs:     1    last: 2025/01/06 23:53:01.17
>>>>>>>   vrf: default
>>>>>>>   index 14 metric 0 mtu 9050 speed 4294967295
>>>>>>>   flags: <UP,BROADCAST,RUNNING,MULTICAST>
>>>>>>>   Type: Ethernet
>>>>>>>   HWaddr: ea:d3:68:02:7d:f7
>>>>>>>   inet6 fe80::e8d3:68ff:fe02:7df7/64
>>>>>>>   Interface Type Vxlan
>>>>>>>   Interface Slave Type None
>>>>>>>   VxLAN Id 100 VTEP IP: 10.23.13.14 Access VLAN Id 1
>>>>>>> 
>>>>>>>   protodown: off
>>>>>>> 
>>>>>>> vtysh -c 'show evpn vni 1'
>>>>>>> VNI: 1
>>>>>>>  Type: L2
>>>>>>>  Tenant VRF: default
>>>>>>>  VxLAN interface: vxlan1
>>>>>>>  VxLAN ifIndex: 14
>>>>>>>  SVI interface: cloudbr1
>>>>>>>  SVI ifIndex: 12
>>>>>>>  Local VTEP IP: 10.23.13.14
>>>>>>>  Mcast group: 0.0.0.0
>>>>>>>  No remote VTEPs known for this VNI
>>>>>>>  Number of MACs (local and remote) known for this VNI: 0
>>>>>>>  Number of ARPs (IPv4 and IPv6, local and remote) known for this VNI: 0
>>>>>>>  Advertise-gw-macip: No
>>>>>>>  Advertise-svi-macip: No
>>>>>>> 
>>>>>>> and I can ping the IPV6 that is routed using the FRR from :60 which is 
>>>>>>> in VXLAN 2 to :14 which is in VXLAN 1
>>>>>>> 
>>>>>>>> ping -I 20XX:5XX:56XX:fff0::2:60 20XX:5XX:56XX:fff0:0:2:13:14
>>>>>>>> PING 20XX:5XX:56XX:fff0:0:2:13:14(20XX:5XX:56XX:fff0:0:2:13:14) from 
>>>>>>>> 20XX:5XX:56XX:fff0::2:60 : 56 data bytes
>>>>>>>> 64 bytes from 20XX:5XX:56XX:fff0:0:2:13:14: icmp_seq=1 ttl=61 
>>>>>>>> time=0.293 ms
>>>>>>>> 64 bytes from 20XX:5XX:56XX:fff0:0:2:13:14: icmp_seq=2 ttl=61 
>>>>>>>> time=0.222 ms
>>>>>>> 
>>>>>>> 
>>>>>>> Then, my questions are:
>>>>>>> are you using at the Leaf Switches/routers a regular Mapping VLAN to 
>>>>>>> VNI VXLAN with VRF ? IF not, Can you share a FRR config of your 
>>>>>>> Switches?
>>>>>>> Or should I use an enterprise SONIC switch software ?
>>>>>>> What other possibilities are with the modifyvxlan.sh That Wido states 
>>>>>>> on some user mails.
>>>>>>> 
>>>>>>> 
>>>>>>> Thank you
>>>>>>> 
>>>>>>> Tata Y.
>>>>>>> 
>>>>>>> 
>>>>> 
>>> 
> 

Reply via email to