CoS will not work on the SD ports.

On 15 Nov 2018, at 04:51, Hugo Slabbert <[email protected]> wrote:

>> This was all while talking about a data center redesign that we are working 
>> on currently.  Replacing ToR VC EX4550’s connected LAG to ASR9K with new 
>> dual QFX5120 leaf to single MX960, dual MPC7E-MRATE
>> 
>> I think we will connect each QFX to each mpc7e card.  Is it best practice to 
>> not interconnect directly between the two QFX’s ? If so why not.
> 
> Glib answer: because then it's not spine & leaf anymore ;)
> 
> Less glib answer:
> 
> 1. it's not needed and is suboptimal
> 
> Going with a basic 3-stage (2 layer) spine & leaf, each leaf is connected to 
> each spine.  Connectivity between any two leafs is via any spine to which 
> they are both connected.  Suppose you have 2 spines, spine1 and spine2, and, 
> say, 10 leaf switches. If a given leaf loses its connection to spine1, it 
> would then just reach all other leafs via spine2.
> 
> If you add a connection between two spines, you do create an alternate path, 
> but it's also not an equal cost or optimal path.  If we're going simple least 
> hops / shortest path, provided leaf1's connection to spine1 is lost, in 
> theory leaf2 could reach leaf1 via:
> 
> leaf2 -> spine1 -> spine2 -> leaf1
> 
> ...but that would be a longer path than just going via the remaining:
> 
> leaf2 -> spine2 -> leaf2
> 
> ...path.  You could force it through the longer path, but why?
> 
> 2. What's your oversub?
> 
> The pitch on spine & leaf networks is generally their high bandwith, high 
> availability (lots of links), and low oversubscription ratios.  For the sake 
> of illustration let's go away from chassis gear for spines to a simpler 
> option like, say, 32x100G Tomahawk spines.  The spines there have capacity to 
> connect 32x leaf switches at line rate.  Whatever connections the leaf 
> switches have to the spines do not have any further oversub imposed within 
> the spine layer.
> 
> Now you interconnect your spines.  How many of those 32x 100G ports are you 
> going to dedicate to spine interconnect?  2 links?  If so, you've now dropped 
> the capacity for 2x more leafs in your fabric (and however many compute nodes 
> they were going to connect), and you're also only providing 200G interconnect 
> between spines for 3 Tbps of leaf connection capacity.  Even if you ignore 
> the less optimal path thing from above and try to intentionally force a 
> fallback path on spine:leaf link failure to traverse your spine xconnect, you 
> can impose up to 15:1 oversub in that scenario.
> 
> Or you could kill the oversub and carve out 16x of your 32x spine ports for 
> spine interconnects.  But now you've shrunk your fabric significantly (can 
> only support 16 leaf switches)...and you've done so unnecessarily because the 
> redundancy model is for leafs to use their uplinks through spines directly 
> rather than using inter-spine links.
> 
> 3. >2 spines
> 
> What if we leaf1 loses its connection to spine2 and leafx loses its 
> connection to spine1?  Have we not created a reachability problem?
> 
>     spine1     spine2
>    /               \
>  /                  \
> leaf1              leafx
> 
> Why, yes we have.  The design solution here is either >1 links between each 
> leaf & spine (cheating; blergh) or a greater number of spines.  What's your 
> redundancy factor?  Augment the above to 4x spines and you've significantly 
> shrunk your risk of creating connectivity islands.
> 
> But if you've designed for interconnecting your spines, what do you for 
> interconnecting 4x spines?  What about if you reach 6x spines?  Again: the 
> model is that resilience is achieved at the leaf:spine interconnectivity 
> rather than at the "top of the tree" as you would have in a standard 
> hierarchical, 3-tier-type setup.
> 
> -- 
> Hugo Slabbert       | email, xmpp/jabber: [email protected]
> pgp key: B178313E   | also on Signal
> 
>> On Tue 2018-Nov-06 12:38:22 -0600, Aaron1 <[email protected]> wrote:
>> 
>> This is a timely topic for me as I just got off a con-call yesterday with my 
>> Juniper SE and an SP specialist...
>> 
>> They also recommended EVPN as the way ahead in place of things like fusion.  
>> They even somewhat shy away from MC-lag
>> 
>> This was all while talking about a data center redesign that we are working 
>> on currently.  Replacing ToR VC EX4550’s connected LAG to ASR9K with new 
>> dual QFX5120 leaf to single MX960, dual MPC7E-MRATE
>> 
>> I think we will connect each QFX to each mpc7e card.  Is it best practice to 
>> not interconnect directly between the two QFX’s ? If so why not.
>> 
>> (please forgive, don’t mean to hijack thread, just some good topics going on 
>> here)
>> 
>> Aaron
> _______________________________________________
> juniper-nsp mailing list [email protected]
> https://puck.nether.net/mailman/listinfo/juniper-nsp
_______________________________________________
juniper-nsp mailing list [email protected]
https://puck.nether.net/mailman/listinfo/juniper-nsp

Reply via email to