Iftekhar,

Thank you for the detailed comments.  Responses to individual
comments/suggestions are inline.

This is a long email message, so if I missed anything, please point
out what I missed.

Please let me know if you are in general agreement with the responses
and if so I'll make specific proposals for replacement text or propose
more detailed action items.

Best Regards,

Curtis


In message <d7d7ab44c06a2440b716f1f1f5e70ae534638...@sv-exdb-prod1.infinera.com>
Iftekhar Hussain writes:

> Dear  Authors,
>
> Please find below some comments on the
> http://tools.ietf.org/html/draft-so-yong-rtgwg-cl-framework-05.
>
>
> ----------------------------
>
> 2.1. Flow Identification
>
> "Operator may have other objectives such as ...composite link energy
> saving, and etc. These new requirements are described in
> [I-D.ietf-rtgwg-cl-requirement]"
>
> Comment:
>
> I don't recall any energy saving related requirement (or discussion)
> in the [I-D.ietf-rtgwg-cl-requirement.  Suggest removing the text
> "energy saving etc."

You are entirely correct that there are no energy savings mentioned in
the requirements document.  This wording came from one of the other
authors but I think I understand the intent.  There is a daily cycle
(and less so a weekly cycle) to traffic levels.  During the low
traffic hours (generally late at night for local hours of night)
traffic can be rebalanced to reduce the number of active component
links.  Where it is possible to depower interfaces, this could result
in energy savings.

I'm OK with removing this example, though it is a very good example
and an goal that we should be pursuing.  Another option is to add a
"MAY" in the requirements docuemt somewhere that indicates that "Load
balancing MAY be used during sustained low traffic periods to reduce
the number of active component links for the purpose of power
reduction".

I'll wait for further comments on this before making a suggestion on
how we will proceed.


> 2.2. Composite Link in Control Plane
>
> "LDP follows the IGP, therefore failure to forward on  the IGP path
> will often result in loss of connectivity if the IGP adjacency is not
> withdrawn when an LDP FEC is refused.  This is a pathologic case that
> can occur if LDP is carried natively and there is a high volume of LDP
> traffic.  This situation can be avoided by carrying LDP within RSVP-TE
> LSP."
>
>
> Comment:
>
> Is it the loss of connectivity referring to LDP control plane
> connectivity only or LDP signaled LSP data plane or both?

The bottom line is that admission control cannot be applied to LDP
without loss of connectivity.  Complete loss of connectivity for a
subset of traffic, no matter how low its priority, is unacceptable.

LDP does not support TE.  There are two options.  Either don't carry
enough LDP traffic such that this situation can occur (this is the
case when a small subset of traffic is high priority traffic carried
within LDP, such as VOIP or enterprise VPN traffic mixed with a much
larger volume of Internet traffic).  The alternative is to carry the
LDP traffic within RSVP-TE LSP when enough LDP traffic is aggregated
such that TE is needed.

The goal was to refer to this problem without going into an
excessively long explanation.  If you think it is too brief and
therefore is unclear, then I'll reword it.

> Comment:
>
> "Composite link capacity is aggregated capacity and MAY be larger than
> individual component link capacity."

Before the MAY insert "LSP capacity" and the sentence makes more
sense.  I'll reread this and see if that is what was intended.

> Composite link aggregate capacity should always be larger than
> individual component link capacity. Did I misunderstand?

The statement as written is not useful.  The sum of positive
quantities is always greater than the quantities being summed.  Thanks
for pointing out this statement.  I'll look at the context and figure
out the intent.

>
> "...1. If no other information is available the largest microflow is
> bound by one of the following:"
>
> Suggested text:
>
> If no other information is available the largest microflow is bound
> (i.e., signaling extensions don't indicate a bound) by one of the
> following:

I'll reword this.  The other information could be signaled or
configured.

> Comment:
>
> Section 2.2 provides architectural guidelines e.g., "  ...Available
> capacity in other component links MUST be used to carry impacted
> traffic.  The available bandwidth after failure MUST be advertised
> immediately to avoid looped crankback" and provides illustrative
> examples e.g.,"... no microflow larger than 10 Gb/s will be present on
> the RSVP-TE LSP that aggregate traffic across the core, even if the
> core interfaces are 100 Gb/s interfaces."

Be careful not to take the last sentence out of the context of the
example (where no interface feeding the core is greater than 10 Gb/s).

> In contrast, section 2.1 appears to be little bit too vague e.g., "
> ... technique of grouping flows, such as hashing on the flow
> identification criteria, becomes essential to reduce the stored state,
> and is an essential scaling technique.  Other means of grouping flows
> may be possible"
>
> Suggestion:
>
> Add some examples which flow identification scheme to use for
> composite links or add a reference to section 4.2 which discusses flow
> identification trade-offs.

The intent in Section 2.1 is to make three points.  First, tracking
every flow is not scalable.  Second, IP src/dst address hashing has
proven (in over two decades of experience) to be an excellent way of
identifying groups of flows (for reasons briefly listed).  Third, if
you find a better way to identify groups of flows, then use it.

We don't want to require the use of IP address hashing, but wanted to
strongly encourage its use, given its long history of successful
deployment.

I'll look at rewording the section.

> 4.1.4. Requirements for Contained LSP
>
> Comment:
>
> Add reference to specific relevant to requirements in
> I-D.ietf-rtgwg-cl-requirement] similar to section 4.1.2 and 4.1.3.

OK.  The sentence "[I-D.ietf-rtgwg-cl-requirement] calls for new LSP
constraints." needs followup indicating which requirements are being
referred to.

> 4.2. Data Plane Challenges
>
> Comment:
>
> Minor typo: "very course..." should be "very coarse..."

Of coarse.  How could I miss that?  :-)

I searched for other uses of the word "course" and found quite a few
that should be "coarse".  Thanks for the catch.

> Comment:
>
> "In practice  using the MPLS label stack alone has proven too course
> to acheive a reasonably good load balance, due to bin-packing issues
> and discrpencies between signaled bandwidth and actual traffic loads
> on LSP."
>
> Suggest adding a reference to an IETF standard or best practices
> document that discusses these issues.

OK, if I can find something.  This has been discussed on and off on
the mailing lists for about a decade.  There may be something in the
original link bundling RFC or elsewhere.

> Comment:
>
> Sections 4.2, 2.1, 2.3, appear to be somewhat redundant. Suggest
> trimming section 2.1 and absorbing that information in section 4.2.

Thanks for pointing this out.  I'll reread and try to reduce
redundancy, using cross references where appropriate rather than
repeating a point.

> 3. Architecture Tradeoffs
>
> Comment:
>
> "Composite Link is applicable to large networks, and therefore
> scalability must be a major consideration."
>
> What is a typical definition of a large network?  Suggestion: Add an
> example (or a forward reference to section 3.3.1 which has an example)

Any network that one of your customers wants to build would be a good
example of large.  :-)  [But not a good definition of large.]

By "large" we mean the type of networks that service providers or
large content providers are building today.

In the future we may find that enterprises are building large enough
private networks to have certain scaling problems that had previously
only been experienced by service providers and content providers in
the past.  A good example of that in the past is overuse of bridging.
Some of the NSFNET regional networks in the late 1980s used bridging
but experienced severe scaling issues very quickly.  It wasn't until
the early to mid 1990s that large enterprises began having similar
problems.

The point of the phrase "Composite Link is applicable to large
networks" is that if you need parallel links because the single link
capacity je jour is too small, then your network is large relative to
most other networks.  Considering the magnitude of "large" where
10Gb/s links are too small, the phrase "and therefore scalability must
be a major consideration" may seem almost rhetorical to some with
deployment experience with IP networks.

The reason that the statement here is brief is that there are a number
of IETF base documents dating back to the mid-1990s on the importance
of scalability.  These realization came from deployment experience.
Despite this someone occasionally makes a naive comment about
processors getting faster and memory getting larger, ignoring the fact
that comparable grouth with order NlogN or N-squared growth in
computation or memory requirements far outpaces Moore's Law
improvements.

I think at least one somewhat ancient RFC indicatea that scalability
must be considered in every document in the IETF routing area.  If I
find it I'll cite it.  It will probably have Brian Carpenter or
Christian Huitman on the author list, making it easier to find.  Maybe
Fred Baker.  I think it was an IAB or IESG RFC.

> 3.1. Scalability Motivations
>
> Comment:
>
> "....a large routing change to be accomplished more quickly,"
>
> What is a typical definition of a large routing change?  Suggestion:
> Add an example.

Not entirely serious: Hurricane Katrina resulted in a number of "large
routing changes".  There was the collaspse of the Raratan River bridge
in NJ taking out fibers on both sides of the bridge (thought to be
redundant), an earthquake east of San Diego taking out three of four
fiber paths, etc.

This is again a relative term that has been used for quite some time.
A customer circuit going down or a new customer coming online results
in a tiny routing change.  A metro fault may be handled entirely in
the metro and have no effect at all on routing elsewhere in the
provider network or global network.  Major fiber outages generally
result in large routing changes.

In an MPLS context, a large routing change is one that affects a large
number of LSP.  A classic example is where one hop along one of two or
three east-west paths across the continental US goes down (or comes
up, though this can be made far less disruptive).  Many LSP traverse
the small number of US east-west paths and terminate in a large number
of medium to large sized cities along either coast.

Again, I'll look at clarifying, but in less words than this response.

> 7.2.5. Dynamic Multipath Balance
>
> Comment:
>
> "...uses a course granularity, the adjustments would have to be
> equally  course, in the worst case moving entire LSP"
>
> Minor typo "course" should be "coarse".

s/course/coarse/ with selective replacement.

> 7. Required Protocol Extensions and Mechanisms
>
> Comment/suggestion:
>
> Organizing section 7 as  a matrix containing something like:
> Requirement#, Existing Mechanisms, Gaps, Ongoing/new extensions...,
> might be more clearer. AT a minimum, add a traceability to each
> requirement in [I-D.ietf-rtgwg-cl-requirement] and the section number
> in framework document that addresses each requirement (or set of
> requirement).

Section 7.2 groups the requirements but there is no where that lists
every requirement in numeric order and points to where in the grouping
it falls.  It should be easy enough to add that.

At the moment, potential issues with the requirements or in this
document are listed in Section 7.3.  This is a different approach,
listing the omissions rather than cross referencing.  Ideally we
address all of the issues and drop this section.

Section 7.4 lists the requirements by protocol or functionality
affected.  This as stated in the first paragraph is for the benefit of
implementors.  Subsections of 7.4 refers back to the groupings in
subsections of 7.2.  It would not be difficult to put in the reverse
cross references (functional group refers to protocol change or
functionality supporting it).

> Is there a minimal set of requirements which must be met to form a
> deployable (useful) composite link based solution?

That's a great question.

"Enough to be useful" is interpreted differently by different network
operators deploying a network.  For an equipment vendor the best
approach is to plan to implement everything over time under the
assumption that every feature will appear in a check box in someone's
RFI/RFP/RFQ eventually, but ask current key customers which features
to prioritize.  That process may yield a different answer for
different equipment vendors, depending on who their current key
customers are.

BTW- Our friends at Verizon made sure that almost every requirement is
a MUST, so ask them to identify the requirements that they are not
really serious about.  Wear protective fireproof undergarments.  :-)

> -----------------------
>
> Regards,
> Iftekhar
_______________________________________________
rtgwg mailing list
[email protected]
https://www.ietf.org/mailman/listinfo/rtgwg

Reply via email to