Pedro,

A couple of high level responses ...

-- Different requirements

> The WG needs to support 2 (perhaps more) sets of requirements that are
> opposite in many respects. Or split into multiple WGs.

> My overall comment is that the attempt to build a problem statement for
> all types of DCs is the fundamental issue.

That may be.  One of my biggest reactions in reading the VPN4DC problem
statement draft was that I wished that it contained an applicability statement
expanding on (quoting from Section 3's first paragraph) "the assumption that
applications require IP connectivity but no Layer 2 direct adjacencies" as
opposed to the draft's current attempts to argue that a layer 2 service
that preserves such adjacencies is not needed.  As an aside - starting from
that quoted assumption, the absence of a need for such a service is effectively
a foregone conclusion ;-).

I have examples where MAC addressing and MAC adjacencies are required
(and where use of a single or very limited number of MAC addresses for
all virtual machines is unworkable), but they involve enterprise-centric
use cases and typically existing deployed-and-unmodifiable applications
and/or functionality that could be out-of-scope for a well-crafted
VPN4DC applicability statement (e.g., one that picks up some of the
examples from your email).

-- Section 2.7

> I don't understand your assertion that Section 2.7 appears to "dismiss"
> RFC 4364, when the latter part of your message appears to describe how
> RFC 4364 (BGP/MPLS VPNs) meets the requirements in that section.  What
> did I miss?
>
> quote:
  > "They are not
  >  necessarily seen as supporting the characteristics outlined in
  >  Section 
2.7<http://tools.ietf.org/html/draft-narten-nvo3-overlay-problem-statement-02#section-2.7>."

Ah, I'll need to check w/my co-authors, but I suggest just deleting that
sentence, as it seems to be adding a lot of heat and not much light to the
discussion.

> It is also unclear to me whether the requirements of section 2.7 do satisfy a
> realistic scenario. For instance, there are several proposals that are being
> supported by vendors in the "legacy DC" space that clearly do not scale up...
> but that may not be a problem for a DC in which the cluster size is 30 
> machines.

Right, and doing something about some of those proposals is one of the 
opportunities
that I see for nvo3 to contribute.  An obvious example is that flooding-based
data plane learning has scaling limitations, but one can preserve the interface
to the end system even while doing something smarter at the NVE.

Thanks,
--David

From: Pedro Roque Marques [mailto:[email protected]]
Sent: Tuesday, June 26, 2012 11:36 AM
To: Black, David
Cc: [email protected]
Subject: Re: [nvo3] call for adoption: 
draft-narten-nvo3-overlay-problem-statement-02


On Jun 26, 2012, at 7:25 AM, <[email protected]<mailto:[email protected]>> 
<[email protected]<mailto:[email protected]>> wrote:


Hi Pedro,

Could you provide some examples of the requirements differences between
the types of data centers?

Differences in aggregate bandwidth and locality of data access drive very 
different network designs.

To give you one example: large scale/large aggregate throughput DC networks 
typically do not support any form of broadcast / multicast. Solutions that rely 
on broadcast / IP multicast for learning are not very useful for this category 
of DCs.



I'm not sure that a data center taxonomy is
an important addition to the draft, but I'm definitely interested in
what may be missing in area of requirements.  Specific requirements
belong in requirements drafts, but I want to understand how you see
the differences affecting what the WG needs to support.

The WG needs to support 2 (perhaps more) sets of requirements that are opposite 
in many respects. Or split into multiple WGs.

Another example: "traditional / legacy" DCs may require the transport of 
unmodified ethernet frames; large scale/low cost DCs do not have any use for 
ethernet headers since applications are not allowed to send / receive non IP 
packets.

It is also unclear to me whether the requirements of section 2.7 do satisfy a 
realistic scenario. For instance, there are several proposals that are being 
supported by vendors in the "legacy DC" space that clearly do not scale up... 
but that may not be a problem for a DC in which the cluster size is 30 machines.



I don't understand your assertion that Section 2.7 appears to "dismiss"
RFC 4364, when the latter part of your message appears to describe how
RFC 4364 (BGP/MPLS VPNs) meets the requirements in that section.  What
did I miss?

quote:
  "They are not

   necessarily seen as supporting the characteristics outlined in

   Section 
2.7<http://tools.ietf.org/html/draft-narten-nvo3-overlay-problem-statement-02#section-2.7>."




I will admit that I don't expect to see BGP implemented/deployed in
hypervisor softswitches

That isn't what is being proposed in 
http://datatracker.ietf.org/doc/draft-marques-l3vpn-end-system/.



(although OSPF as the edge protocol for BGP/MPLS
VPNs seems more realistic for that implementation location), and hence
I'm interested in a standard protocol for NVEs to talk to an "oracle"
which could be a set of network nodes that use a routing protocol to
distribute information among themselves .  That possibility of a distributed
structure is not easy to infer from the current draft, so we probably
should add text to describe it ... and knock out the one parenthetical
remark that the "oracle" is directory-based.  I don't care how the
"oracle" is structured internally (a routing protocol is a fine way
to distribute information), although NVEs that don't directly participate
in the "oracle" probably view the "oracle" as some sort of directory on
which NVEs perform address lookups.

There are several reasons to have multiple and interoperable "oracles":
  - multiple oracles means multiple failure domains;
  - it means the possibility of operation across multiple administrative 
domains (or even multiple clusters).
  - and interoperability always ends up creating a bigger ecosystem which in 
the end benefits network operators.



Inter-VN traffic (what you refer to as inter-CUG traffic) is handled by
a straightforward application of IP routing to the inner IP headers; this
is similar to the well-understood application of IP routing to forward
traffic across VLANs.

That is where again the differences between different types of data-centers do 
play in. If for instance 90% of a VMs traffic happens to be between the Host OS 
and a network attached storage file system run as-a-Service (with the 
appropriate multi-tenent support) then the question of where are the routers 
becomes a very important issue. In a large scale data-center where the Host VM 
and the CPU that hosts the filesystem block can be randomly spread where is the 
router ?
Is every switch a router ? Does it have all the CUGs present ?
In some DC designs the problem to solve is the inter-CUG traffic. With L2 
headers being totally irrelevant.



  We should talk about VRFs as something other than
a limitation of current approaches - for VLANs, VRFs (separate instances
of routing) are definitely a feature, and I expect this to carry forward
to nvo3 VNs.  In addition, we need to make changes to address Dimitri's
comments about problems with the current VRF text.

My overall comment is that the attempt to build a problem statement for all 
types of DCs is the fundamental issue. There are DCs that have opposite design 
goals. This results in different server hardware, software, management 
processes, application architectures, storage solutions. An attempt to create a 
single unified problem statement is, imho, counter-productive.

Reasonable people, solving totally different problems with opposite goals will 
necessarily disagree.



Thanks,
--David

From: [email protected]<mailto:[email protected]> 
[mailto:[email protected]] On Behalf Of Pedro Roque Marques
Sent: Friday, June 22, 2012 10:57 AM
To: Benson Schliesser
Cc: Matthew (Matthew) Bocci; [email protected]<mailto:[email protected]>; 
[email protected]<mailto:[email protected]>
Subject: Re: [nvo3] call for adoption: 
draft-narten-nvo3-overlay-problem-statement-02

Benson,
I object to the document on the following points:

1) It doesn't appear to recognize that there are multiple type of data-center 
deployments with different requirements.
2) It appears to "dismiss" RFC4364 based on the requirements of section 2.7.
3) Does not discuss the requirements for inter-CUG traffic.

On the first point:
  - One can place data-center design approaches on a continuum where at one 
extreme one encounters data-centers that in general:
       o Have a small number of physical servers per cluster (orchestration 
system scope).
       o Rely on hardware resiliency.
       o Use a single VM per application tier.
       o The VM lifespan is very large.
       o Use a SAN or local disks for storage.
       o Use a oversubscribed network that is L2 based.
       o Tend to exchange traffic within a "VLAN"

    At the other end of the spectrum you have data-centers in which the design 
points are in general:
      o Very large number of physical servers.
      o Rely on software resiliency.
      o Use multiple VMs per app tier, often 10s or 100s.
      o The VM lifespan is small since VMs are created and deleted according to 
load requirements.
      o Use a distributed block storage and distributed NAS file-systems.
      o Use a non-oversubscribed network from the TOR up that is L3 only and 
typically does not support multicast.
      o Most of the traffic crosses "closed user group" boundaries.

Between these two extremes there is often non common elements: server hardware, 
storage, software architectures and network are completely different. The 
document in question seems to be focusing on solutions at the left end of the 
spectrum (first approach) and does not at all reflect the requirements of 
data-ceterns at the middle or at the right end of the spectrum (the latter 
approach).

On the second point:
  - Section 2.7 of the document states that the overlay technology must take 
into account:
    1. Highly distributed systems
    2. Many highly distributed virtual networks with sparse membership.
    3. Highly dynamic end systems.
    4. Work with existing [...] routers and switches.
    5. administered by a single administrative domain

L3VPN has proven to work in highly distributed systems, with many distributed 
virtual networks with space membership and with highly dynamic rechability 
information. It can be encapsulated as MPLS over GRE which implies IP packets 
and will work with any existing routers and switches. It can be administered by 
a single administrative domain.

Among the related work that is cited, i don't believe that trill, .1aq and arms 
can claim to work with existing routers and swathes. And none of the approaches 
has proven the same sort of route scaling as l3vpn.

thank you,
  Pedro.

On Jun 16, 2012, at 6:01 PM, Benson Schliesser wrote:



Dear NVO3 Participants -

This message begins a two week Call for Adoption of 
http://tools.ietf.org/html/draft-narten-nvo3-overlay-problem-statement-02 by 
the NVO3 working group, ending on 30-June-2012.

Please respond to the NVO3 mailing list with any statements of approval or 
disapproval, along with any additional comments that might explain your 
position. Also, if any NVO3 participant is aware of IPR associated with this 
draft, please inform the mailing list and/or the NVO3 chairs.

Thanks,
-Benson & Matthew

_______________________________________________
nvo3 mailing list
[email protected]<mailto:[email protected]>
https://www.ietf.org/mailman/listinfo/nvo3


_______________________________________________
nvo3 mailing list
[email protected]
https://www.ietf.org/mailman/listinfo/nvo3

Reply via email to