Iftekhar,
I'm replying to this email because this email thread accounts for most
of the changes that I just made to draft-so-yong-rtgwg-cl-framework.
Since the call for support as a WG item is ongoing I will not submit
the updates until that call is completed.
Briefly the changes to the document are:
1. Author affiliation change for Ning, address change for Lucy.
@@ -61,23 +61,17 @@
@@ -107,12 +101,12 @@
2. Some clarification in "Flow Identification" section resulting
from your comments.
@@ -474,29 +470,35 @@
3. A brief summary added to the "Flow Identification" section.
@@ -474,29 +470,35 @@ (last addition in that set of diffs)
4. Wording changes suggested in this thread (see inline below).
@@ -648,17 +680,18 @@
5. Significant rewording in "Composite Link in Data Plane" section
to reduce redundancy and put the general description of traffic
distribution here and issues later, but referencing here.
@@ -770,24 +803,26 @@
@@ -795,34 +830,38 @@
6. Clarified wording about large networks and scalability.
@@ -833,12 +872,20 @@
7. Clarified wording about large routing changes.
@@ -851,15 +898,32 @@
8. Removed redundant description of issues in "Data Plane
Challenges" that were already covered in "Flow Identification"
section or "Composite Link in Data Plane" section, and moved
some text that was not covered in either place.
@@ -1391,59 +1455,14 @@
9. Added clarifications and cross reference to "Very Large LSP"
section.
@@ -1453,8 +1472,18 @@
10. Better defined what a "very large microflow" is.
@@ -1464,8 +1493,8 @@
11. Added some text to the acknowledgements.
@@ -2506,9 +2535,22 @@
Diffs will be sent in a separate email - too long to attach. The @@
notation above refers to specific diff chunks. Followup to the most
recent email is inline.
It was pointed out in another thread that some text on carrying IP and
LDP is in more than one place and similar enough to be redundant.
I'll get that in a next set of changes.
Best Regards,
Curtis
In message <[email protected]>
Curtis Villamizar writes:
>
> Iftekhar,
>
> Thank you for the detailed comments. Responses to individual
> comments/suggestions are inline.
>
> This is a long email message, so if I missed anything, please point
> out what I missed.
>
> Please let me know if you are in general agreement with the responses
> and if so I'll make specific proposals for replacement text or propose
> more detailed action items.
>
> Best Regards,
>
> Curtis
>
>
> In message
> <d7d7ab44c06a2440b716f1f1f5e70ae534638...@sv-exdb-prod1.infinera.com>
> Iftekhar Hussain writes:
>
> > Dear Authors,
> >
> > Please find below some comments on the
> > http://tools.ietf.org/html/draft-so-yong-rtgwg-cl-framework-05.
> >
> >
> > ----------------------------
> >
> > 2.1. Flow Identification
> >
> > "Operator may have other objectives such as ...composite link energy
> > saving, and etc. These new requirements are described in
> > [I-D.ietf-rtgwg-cl-requirement]"
> >
> > Comment:
> >
> > I don't recall any energy saving related requirement (or discussion)
> > in the [I-D.ietf-rtgwg-cl-requirement. Suggest removing the text
> > "energy saving etc."
>
> You are entirely correct that there are no energy savings mentioned in
> the requirements document. This wording came from one of the other
> authors but I think I understand the intent. There is a daily cycle
> (and less so a weekly cycle) to traffic levels. During the low
> traffic hours (generally late at night for local hours of night)
> traffic can be rebalanced to reduce the number of active component
> links. Where it is possible to depower interfaces, this could result
> in energy savings.
>
> I'm OK with removing this example, though it is a very good example
> and an goal that we should be pursuing. Another option is to add a
> "MAY" in the requirements docuemt somewhere that indicates that "Load
> balancing MAY be used during sustained low traffic periods to reduce
> the number of active component links for the purpose of power
> reduction".
>
> I'll wait for further comments on this before making a suggestion on
> how we will proceed.
No further comments on the mailing list on this. The requirements was
added to CL requirements.
> > 2.2. Composite Link in Control Plane
> >
> > "LDP follows the IGP, therefore failure to forward on the IGP path
> > will often result in loss of connectivity if the IGP adjacency is not
> > withdrawn when an LDP FEC is refused. This is a pathologic case that
> > can occur if LDP is carried natively and there is a high volume of LDP
> > traffic. This situation can be avoided by carrying LDP within RSVP-TE
> > LSP."
> >
> >
> > Comment:
> >
> > Is it the loss of connectivity referring to LDP control plane
> > connectivity only or LDP signaled LSP data plane or both?
>
> The bottom line is that admission control cannot be applied to LDP
> without loss of connectivity. Complete loss of connectivity for a
> subset of traffic, no matter how low its priority, is unacceptable.
>
> LDP does not support TE. There are two options. Either don't carry
> enough LDP traffic such that this situation can occur (this is the
> case when a small subset of traffic is high priority traffic carried
> within LDP, such as VOIP or enterprise VPN traffic mixed with a much
> larger volume of Internet traffic). The alternative is to carry the
> LDP traffic within RSVP-TE LSP when enough LDP traffic is aggregated
> such that TE is needed.
>
> The goal was to refer to this problem without going into an
> excessively long explanation. If you think it is too brief and
> therefore is unclear, then I'll reword it.
No further comments so I'm assuming that the wording was adequate as
is.
> > Comment:
> >
> > "Composite link capacity is aggregated capacity and MAY be larger than
> > individual component link capacity."
>
> Before the MAY insert "LSP capacity" and the sentence makes more
> sense. I'll reread this and see if that is what was intended.
Fixed in @@ -648,17 +680,18 @@ (#4 above)
> > Composite link aggregate capacity should always be larger than
> > individual component link capacity. Did I misunderstand?
>
> The statement as written is not useful. The sum of positive
> quantities is always greater than the quantities being summed. Thanks
> for pointing out this statement. I'll look at the context and figure
> out the intent.
Looked fine after the two word insertion mentioned above.
> > "...1. If no other information is available the largest microflow is
> > bound by one of the following:"
> >
> > Suggested text:
> >
> > If no other information is available the largest microflow is bound
> > (i.e., signaling extensions don't indicate a bound) by one of the
> > following:
>
> I'll reword this. The other information could be signaled or
> configured.
Added this in @@ -648,17 +680,18 @@ last change in the chunk.
> > Comment:
> >
> > Section 2.2 provides architectural guidelines e.g., " ...Available
> > capacity in other component links MUST be used to carry impacted
> > traffic. The available bandwidth after failure MUST be advertised
> > immediately to avoid looped crankback" and provides illustrative
> > examples e.g.,"... no microflow larger than 10 Gb/s will be present on
> > the RSVP-TE LSP that aggregate traffic across the core, even if the
> > core interfaces are 100 Gb/s interfaces."
>
> Be careful not to take the last sentence out of the context of the
> example (where no interface feeding the core is greater than 10 Gb/s).
No issue here.
> > In contrast, section 2.1 appears to be little bit too vague e.g., "
> > ... technique of grouping flows, such as hashing on the flow
> > identification criteria, becomes essential to reduce the stored state,
> > and is an essential scaling technique. Other means of grouping flows
> > may be possible"
> >
> > Suggestion:
> >
> > Add some examples which flow identification scheme to use for
> > composite links or add a reference to section 4.2 which discusses flow
> > identification trade-offs.
>
> The intent in Section 2.1 is to make three points. First, tracking
> every flow is not scalable. Second, IP src/dst address hashing has
> proven (in over two decades of experience) to be an excellent way of
> identifying groups of flows (for reasons briefly listed). Third, if
> you find a better way to identify groups of flows, then use it.
>
> We don't want to require the use of IP address hashing, but wanted to
> strongly encourage its use, given its long history of successful
> deployment.
>
> I'll look at rewording the section.
Above comments insired the changes in diff chunk @@ -474,29 +470,35 @@
> > 4.1.4. Requirements for Contained LSP
> >
> > Comment:
> >
> > Add reference to specific relevant to requirements in
> > I-D.ietf-rtgwg-cl-requirement] similar to section 4.1.2 and 4.1.3.
>
> OK. The sentence "[I-D.ietf-rtgwg-cl-requirement] calls for new LSP
> constraints." needs followup indicating which requirements are being
> referred to.
The existing text identifies the requirement topics. I don't want to
embed the requirement numbers here since the numbers can change.
> > 4.2. Data Plane Challenges
> >
> > Comment:
> >
> > Minor typo: "very course..." should be "very coarse..."
>
> Of coarse. How could I miss that? :-)
>
> I searched for other uses of the word "course" and found quite a few
> that should be "coarse". Thanks for the catch.
>
> > Comment:
> >
> > "In practice using the MPLS label stack alone has proven too course
> > to acheive a reasonably good load balance, due to bin-packing issues
> > and discrpencies between signaled bandwidth and actual traffic loads
> > on LSP."
> >
> > Suggest adding a reference to an IETF standard or best practices
> > document that discusses these issues.
>
> OK, if I can find something. This has been discussed on and off on
> the mailing lists for about a decade. There may be something in the
> original link bundling RFC or elsewhere.
That text is gone. The topic is covered in Flow Identification.
Another reference to draft-symmvo-rtgwg-cl-use-cases was added.
RFC4928 indicates that more than MPLS label stack is needed but
doesn't say why. There is a little more detail in the Entropy Label
draft (now in last call).
> > Comment:
> >
> > Sections 4.2, 2.1, 2.3, appear to be somewhat redundant. Suggest
> > trimming section 2.1 and absorbing that information in section 4.2.
>
> Thanks for pointing this out. I'll reread and try to reduce
> redundancy, using cross references where appropriate rather than
> repeating a point.
Redundant information in 4.2 was removed. Some information in 4.2 was
put into 2.3. This left 2.1 with the background on current flow
identification and 2.3 with background on current traffic
distribution.
> > 3. Architecture Tradeoffs
> >
> > Comment:
> >
> > "Composite Link is applicable to large networks, and therefore
> > scalability must be a major consideration."
> >
> > What is a typical definition of a large network? Suggestion: Add an
> > example (or a forward reference to section 3.3.1 which has an example)
>
> Any network that one of your customers wants to build would be a good
> example of large. :-) [But not a good definition of large.]
>
> By "large" we mean the type of networks that service providers or
> large content providers are building today.
>
> In the future we may find that enterprises are building large enough
> private networks to have certain scaling problems that had previously
> only been experienced by service providers and content providers in
> the past. A good example of that in the past is overuse of bridging.
> Some of the NSFNET regional networks in the late 1980s used bridging
> but experienced severe scaling issues very quickly. It wasn't until
> the early to mid 1990s that large enterprises began having similar
> problems.
>
> The point of the phrase "Composite Link is applicable to large
> networks" is that if you need parallel links because the single link
> capacity je jour is too small, then your network is large relative to
> most other networks. Considering the magnitude of "large" where
> 10Gb/s links are too small, the phrase "and therefore scalability must
> be a major consideration" may seem almost rhetorical to some with
> deployment experience with IP networks.
>
> The reason that the statement here is brief is that there are a number
> of IETF base documents dating back to the mid-1990s on the importance
> of scalability. These realization came from deployment experience.
> Despite this someone occasionally makes a naive comment about
> processors getting faster and memory getting larger, ignoring the fact
> that comparable grouth with order NlogN or N-squared growth in
> computation or memory requirements far outpaces Moore's Law
> improvements.
>
> I think at least one somewhat ancient RFC indicatea that scalability
> must be considered in every document in the IETF routing area. If I
> find it I'll cite it. It will probably have Brian Carpenter or
> Christian Huitman on the author list, making it easier to find. Maybe
> Fred Baker. I think it was an IAB or IESG RFC.
Best general IAB statement I found was "All designs must scale readily
to very many nodes per site and to many millions of sites." in RFC
1958 "Architectural Principles of the Internet" (1996). RFC 2791
"Scalable Routing Design Principles" by Jessica Yu (2000) is very good
but focused on BGP. RFC 3439 "Internet Architectural Guidelines" by
Randy Bush and David Meyer (2002) was an IAB document about scaling
but touched on a lot of topics without driving home any one point very
well (IMHO).
I didn't add any reference here. I did make changes saying a bit
about what a "large network" means.
> > 3.1. Scalability Motivations
> >
> > Comment:
> >
> > "....a large routing change to be accomplished more quickly,"
> >
> > What is a typical definition of a large routing change? Suggestion:
> > Add an example.
>
> Not entirely serious: Hurricane Katrina resulted in a number of "large
> routing changes". There was the collaspse of the Raratan River bridge
> in NJ taking out fibers on both sides of the bridge (thought to be
> redundant), an earthquake east of San Diego taking out three of four
> fiber paths, etc.
>
> This is again a relative term that has been used for quite some time.
> A customer circuit going down or a new customer coming online results
> in a tiny routing change. A metro fault may be handled entirely in
> the metro and have no effect at all on routing elsewhere in the
> provider network or global network. Major fiber outages generally
> result in large routing changes.
>
> In an MPLS context, a large routing change is one that affects a large
> number of LSP. A classic example is where one hop along one of two or
> three east-west paths across the continental US goes down (or comes
> up, though this can be made far less disruptive). Many LSP traverse
> the small number of US east-west paths and terminate in a large number
> of medium to large sized cities along either coast.
>
> Again, I'll look at clarifying, but in less words than this response.
I added changes giving some limited information about what a large
routing change entails in an MPLS network.
> > 7.2.5. Dynamic Multipath Balance
> >
> > Comment:
> >
> > "...uses a course granularity, the adjustments would have to be
> > equally course, in the worst case moving entire LSP"
> >
> > Minor typo "course" should be "coarse".
>
> s/course/coarse/ with selective replacement.
>
> > 7. Required Protocol Extensions and Mechanisms
> >
> > Comment/suggestion:
> >
> > Organizing section 7 as a matrix containing something like:
> > Requirement#, Existing Mechanisms, Gaps, Ongoing/new extensions...,
> > might be more clearer. AT a minimum, add a traceability to each
> > requirement in [I-D.ietf-rtgwg-cl-requirement] and the section number
> > in framework document that addresses each requirement (or set of
> > requirement).
>
> Section 7.2 groups the requirements but there is no where that lists
> every requirement in numeric order and points to where in the grouping
> it falls. It should be easy enough to add that.
>
> At the moment, potential issues with the requirements or in this
> document are listed in Section 7.3. This is a different approach,
> listing the omissions rather than cross referencing. Ideally we
> address all of the issues and drop this section.
>
> Section 7.4 lists the requirements by protocol or functionality
> affected. This as stated in the first paragraph is for the benefit of
> implementors. Subsections of 7.4 refers back to the groupings in
> subsections of 7.2. It would not be difficult to put in the reverse
> cross references (functional group refers to protocol change or
> functionality supporting it).
I didn't add this cross reference because the requirement numbers
change. There is no way to add a tag in the requirements document
that would allow a tag to number mapping if the numbers change.
Keeping it this way I only have to fix the requirement numbering in
Section 7.2 if the CL requirements document changes.
> > Is there a minimal set of requirements which must be met to form a
> > deployable (useful) composite link based solution?
>
> That's a great question.
>
> "Enough to be useful" is interpreted differently by different network
> operators deploying a network. For an equipment vendor the best
> approach is to plan to implement everything over time under the
> assumption that every feature will appear in a check box in someone's
> RFI/RFP/RFQ eventually, but ask current key customers which features
> to prioritize. That process may yield a different answer for
> different equipment vendors, depending on who their current key
> customers are.
>
> BTW- Our friends at Verizon made sure that almost every requirement is
> a MUST, so ask them to identify the requirements that they are not
> really serious about. Wear protective fireproof undergarments. :-)
Did you ever ask?
> > -----------------------
> >
> > Regards,
> > Iftekhar
> _______________________________________________
> rtgwg mailing list
> [email protected]
> https://www.ietf.org/mailman/listinfo/rtgwg
_______________________________________________
rtgwg mailing list
[email protected]
https://www.ietf.org/mailman/listinfo/rtgwg