On Mon, Nov 4, 2019 at 6:36 PM Dumitru Ceara <[email protected]> wrote:
>
> ARP request and ND NS packets for router owned IPs were being
> flooded in the complete L2 domain (using the MC_FLOOD multicast group).
> However this creates a scaling issue in scenarios where aggregation
> logical switches are connected to more logical routers (~350). The
> logical pipelines of all routers would have to be executed before the
> packet is finally replied to by a single router, the owner of the IP
> address.
>
> This commit limits the broadcast domain by bypassing the L2 Lookup stage
> for ARP requests that will be replied by a single router. The packets
> are still flooded in the L2 domain but not on any of the other patch
> ports towards other routers connected to the switch. This restricted
> flooding is done by using a new multicast group (MC_ARP_ND_FLOOD).
>
> IPs that are owned by the routers and for which this fix applies are:
> - IP addresses configured on the router ports.
> - VIPs.
> - NAT IPs.
>
> This commit also fixes function get_router_load_balancer_ips() which
> was incorrectly returning a single address_family even though the
> IP set could contain a mix of IPv4 and IPv6 addresses.
>
> Reported-at: https://bugzilla.redhat.com/1756945
> Reported-by: Anil Venkata <[email protected]>
> Signed-off-by: Dumitru Ceara <[email protected]>
Hi Dumitru,
The patch LGTM. I have just a couple of minor comments.
Thanks
Numan
>
> ---
> v4: Rebase.
> v3: Properly deal with VXLAN traffic. Address review comments from
> Numan (add autotests). Fix function get_router_load_balancer_ips.
> Rebase -> deal with IPv6 NAT too.
> v2: Move ARP broadcast domain limiting to table S_SWITCH_IN_L2_LKUP to
> address localnet ports too.
> ---
> lib/mcast-group-index.h | 1 +
> northd/ovn-northd.8.xml | 16 +++
> northd/ovn-northd.c | 340
> ++++++++++++++++++++++++++++++++++++------------
> tests/ovn.at | 244 +++++++++++++++++++++++++++++++++-
> 4 files changed, 517 insertions(+), 84 deletions(-)
>
> diff --git a/lib/mcast-group-index.h b/lib/mcast-group-index.h
> index ba995ba..06bd8b3 100644
> --- a/lib/mcast-group-index.h
> +++ b/lib/mcast-group-index.h
> @@ -27,6 +27,7 @@ enum ovn_mcast_tunnel_keys {
>
> OVN_MCAST_FLOOD_TUNNEL_KEY = OVN_MIN_MULTICAST,
> OVN_MCAST_UNKNOWN_TUNNEL_KEY,
> + OVN_MCAST_ARP_ND_TUNNEL_KEY,
> OVN_MCAST_MROUTER_FLOOD_TUNNEL_KEY,
> OVN_MCAST_MROUTER_STATIC_TUNNEL_KEY,
> OVN_MCAST_STATIC_TUNNEL_KEY,
> diff --git a/northd/ovn-northd.8.xml b/northd/ovn-northd.8.xml
> index 0a33dcd..6fbb3ab 100644
> --- a/northd/ovn-northd.8.xml
> +++ b/northd/ovn-northd.8.xml
> @@ -1005,6 +1005,22 @@ output;
> </li>
>
> <li>
> + Priority-80 flows for each port connected to a logical router
> + matching self originated GARP/ARP request/ND packets. These packets
> + are flooded to the <code>MC_FLOOD</code> which contains all logical
> + ports.
> + </li>
> +
> + <li>
> + Priority-75 flows for each IP address/VIP/NAT address owned by a
> + router port connected to the switch. These flows match ARP requests
> + and ND packets for the specific IP addresses. Matched packets are
> + forwarded in the L3 domain only to the router that owns the IP
> + address and flooded in the L2 domain on all ports except patch
> + ports connected to logical routers.
> + </li>
> +
> + <li>
> A priority-70 flow that outputs all packets with an Ethernet
> broadcast
> or multicast <code>eth.dst</code> to the <code>MC_FLOOD</code>
> multicast group.
> diff --git a/northd/ovn-northd.c b/northd/ovn-northd.c
> index c23c270..b13b432 100644
> --- a/northd/ovn-northd.c
> +++ b/northd/ovn-northd.c
> @@ -210,6 +210,8 @@ enum ovn_stage {
> #define REGBIT_LOOKUP_NEIGHBOR_RESULT "reg9[4]"
> #define REGBIT_SKIP_LOOKUP_NEIGHBOR "reg9[5]"
>
> +#define REGBIT_NOT_VXLAN "flags[1] == 0"
> +
> /* Returns an "enum ovn_stage" built from the arguments. */
> static enum ovn_stage
> ovn_stage_build(enum ovn_datapath_type dp_type, enum ovn_pipeline pipeline,
> @@ -1202,6 +1204,34 @@ ovn_port_allocate_key(struct ovn_datapath *od)
> 1, (1u << 15) - 1, &od->port_key_hint);
> }
>
> +/* Returns true if the logical switch port 'enabled' column is empty or
> + * set to true. Otherwise, returns false. */
> +static bool
> +lsp_is_enabled(const struct nbrec_logical_switch_port *lsp)
> +{
> + return !lsp->n_enabled || *lsp->enabled;
> +}
> +
> +/* Returns true only if the logical switch port 'up' column is set to true.
> + * Otherwise, if the column is not set or set to false, returns false. */
> +static bool
> +lsp_is_up(const struct nbrec_logical_switch_port *lsp)
> +{
> + return lsp->n_up && *lsp->up;
> +}
> +
> +static bool
> +lsp_is_external(const struct nbrec_logical_switch_port *nbsp)
> +{
> + return !strcmp(nbsp->type, "external");
> +}
> +
> +static bool
> +lrport_is_enabled(const struct nbrec_logical_router_port *lrport)
> +{
> + return !lrport->enabled || *lrport->enabled;
> +}
> +
> static char *
> chassis_redirect_name(const char *port_name)
> {
> @@ -2184,7 +2214,7 @@ ip_address_and_port_from_lb_key(const char *key, char
> **ip_address,
>
> static void
> get_router_load_balancer_ips(const struct ovn_datapath *od,
> - struct sset *all_ips, int *addr_family)
> + struct sset *all_ips_v4, struct sset
> *all_ips_v6)
> {
> if (!od->nbr) {
> return;
> @@ -2199,13 +2229,21 @@ get_router_load_balancer_ips(const struct
> ovn_datapath *od,
> /* node->key contains IP:port or just IP. */
> char *ip_address = NULL;
> uint16_t port;
> + int addr_family;
>
> ip_address_and_port_from_lb_key(node->key, &ip_address, &port,
> - addr_family);
> + &addr_family);
> if (!ip_address) {
> continue;
> }
>
> + struct sset *all_ips;
> + if (addr_family == AF_INET) {
> + all_ips = all_ips_v4;
> + } else {
> + all_ips = all_ips_v6;
> + }
> +
> if (!sset_contains(all_ips, ip_address)) {
> sset_add(all_ips, ip_address);
> }
> @@ -2299,17 +2337,22 @@ get_nat_addresses(const struct ovn_port *op, size_t
> *n)
> }
> }
>
> - /* A set to hold all load-balancer vips. */
> - struct sset all_ips = SSET_INITIALIZER(&all_ips);
> - int addr_family;
> - get_router_load_balancer_ips(op->od, &all_ips, &addr_family);
> + /* Two sets to hold all load-balancer vips. */
> + struct sset all_ips_v4 = SSET_INITIALIZER(&all_ips_v4);
> + struct sset all_ips_v6 = SSET_INITIALIZER(&all_ips_v6);
> + get_router_load_balancer_ips(op->od, &all_ips_v4, &all_ips_v6);
>
> const char *ip_address;
> - SSET_FOR_EACH (ip_address, &all_ips) {
> + SSET_FOR_EACH (ip_address, &all_ips_v4) {
> ds_put_format(&c_addresses, " %s", ip_address);
> central_ip_address = true;
> }
> - sset_destroy(&all_ips);
> + SSET_FOR_EACH (ip_address, &all_ips_v6) {
> + ds_put_format(&c_addresses, " %s", ip_address);
> + central_ip_address = true;
> + }
> + sset_destroy(&all_ips_v4);
> + sset_destroy(&all_ips_v6);
>
> if (central_ip_address) {
> /* Gratuitous ARP for centralized NAT rules on distributed gateway
> @@ -3036,6 +3079,10 @@ static const struct multicast_group mc_static =
> static const struct multicast_group mc_unknown =
> { MC_UNKNOWN, OVN_MCAST_UNKNOWN_TUNNEL_KEY };
>
> +#define MC_ARP_ND "_MC_arp_nd"
> +static const struct multicast_group mc_arp_nd =
> + { MC_ARP_ND, OVN_MCAST_ARP_ND_TUNNEL_KEY };
> +
> static bool
> multicast_group_equal(const struct multicast_group *a,
> const struct multicast_group *b)
> @@ -3737,28 +3784,6 @@ build_port_security_ip(enum ovn_pipeline pipeline,
> struct ovn_port *op,
>
> }
>
> -/* Returns true if the logical switch port 'enabled' column is empty or
> - * set to true. Otherwise, returns false. */
> -static bool
> -lsp_is_enabled(const struct nbrec_logical_switch_port *lsp)
> -{
> - return !lsp->n_enabled || *lsp->enabled;
> -}
> -
> -/* Returns true only if the logical switch port 'up' column is set to true.
> - * Otherwise, if the column is not set or set to false, returns false. */
> -static bool
> -lsp_is_up(const struct nbrec_logical_switch_port *lsp)
> -{
> - return lsp->n_up && *lsp->up;
> -}
> -
> -static bool
> -lsp_is_external(const struct nbrec_logical_switch_port *nbsp)
> -{
> - return !strcmp(nbsp->type, "external");
> -}
> -
> static bool
> build_dhcpv4_action(struct ovn_port *op, ovs_be32 offer_ip,
> struct ds *options_action, struct ds *response_action,
> @@ -5161,6 +5186,143 @@ build_lrouter_groups(struct hmap *ports, struct
> ovs_list *lr_list)
> }
> }
>
> +/*
> + * Ingress table 17: Flows that forward ARP/ND requests only to the routers
> + * that own the addresses. Packets are still flooded in the switching domain
> + * as regular broadcast.
> + */
> +static void
> +build_lswitch_rport_arp_flow(const char *target_address, int addr_family,
> + struct ovn_port *patch_op,
> + struct ovn_datapath *od,
> + uint32_t priority,
> + struct hmap *lflows)
> +{
> + struct ds match = DS_EMPTY_INITIALIZER;
> + struct ds actions = DS_EMPTY_INITIALIZER;
> +
> + if (addr_family == AF_INET) {
> + ds_put_format(&match, "arp.tpa == %s && arp.op == 1",
> target_address);
> + } else {
> + ds_put_format(&match, "nd.target == %s && nd_ns", target_address);
> + }
> +
> + /* Packets received from VXLAN tunnels have already been through the
> + * router pipeline so we should skip them. Normally this is done by the
> + * multicast_group implementation (VXLAN packets skip table 32 which
> + * delivers to patch ports) but we're bypassing multicast_groups.
> + */
> + ds_put_format(&match, " && "REGBIT_NOT_VXLAN);
> +
> + /* Send a clone of the packet to the router pipeline and flood the
> + * original in the broadcast domain (skipping router ports). */
> + ds_put_format(&actions,
> + "clone { outport = %s; output; }; "
> + "outport = \""MC_ARP_ND"\"; output;",
> + patch_op->json_key);
> + ovn_lflow_add(lflows, od, S_SWITCH_IN_L2_LKUP, priority,
> + ds_cstr(&match), ds_cstr(&actions));
> +
> + ds_destroy(&match);
> + ds_destroy(&actions);
> +}
> +
> +/*
> + * Ingress table 17: Flows that forward ARP/ND requests only to the routers
> + * that own the addresses.
> + * Priorities:
> + * - 80: self originated GARPs that need to follow regular processing.
> + * - 75: ARP requests to router owned IPs (interface IP/LB/NAT).
> + */
> +static void
> +build_lswitch_rport_arp_responders(struct ovn_port *op,
> + struct ovn_datapath *sw_od,
> + struct ovn_port *sw_op,
> + struct hmap *lflows)
> +{
> + if (!op || !op->nbrp) {
> + return;
> + }
> +
> + if (!lrport_is_enabled(op->nbrp)) {
> + return;
> + }
> +
> + struct ds match = DS_EMPTY_INITIALIZER;
> +
> + /* Self originated (G)ARP requests/ND need to be flooded as usual.
> + * Priority: 40.
You mean priority 80 here right ?
I think you need to udpate the comments to change from 40 to 80
and 30 to 75.
> + */
> + ds_put_format(&match, "inport == %s && (arp.op == 1 || nd_ns)",
> + sw_op->json_key);
> + ovn_lflow_add(lflows, sw_od, S_SWITCH_IN_L2_LKUP, 80,
> + ds_cstr(&match),
> + "outport = \""MC_FLOOD"\"; output;");
> +
> + ds_destroy(&match);
> +
> + /* Forward ARP requests for IPs configured on the router only to this
> + * router port.
> + * Priority: 30.
Same here.
> + */
> + for (int i = 0; i < op->lrp_networks.n_ipv4_addrs; i++) {
> + build_lswitch_rport_arp_flow(op->lrp_networks.ipv4_addrs[i].addr_s,
> + AF_INET, sw_op, sw_od, 75, lflows);
> + }
> + for (int i = 0; i < op->lrp_networks.n_ipv6_addrs; i++) {
> + build_lswitch_rport_arp_flow(op->lrp_networks.ipv6_addrs[i].addr_s,
> + AF_INET6, sw_op, sw_od, 75, lflows);
> + }
> +
> + /* Forward ARP requests to load-balancer VIPs configured on the router
> + * only to this router port.
> + * Priority: 30.
> + */
> + struct sset all_ips_v4 = SSET_INITIALIZER(&all_ips_v4);
> + struct sset all_ips_v6 = SSET_INITIALIZER(&all_ips_v6);
> + const char *ip_address;
> +
> + get_router_load_balancer_ips(op->od, &all_ips_v4, &all_ips_v6);
> +
> + SSET_FOR_EACH (ip_address, &all_ips_v4) {
> + build_lswitch_rport_arp_flow(ip_address, AF_INET, sw_op, sw_od,
> + 75, lflows);
> + }
> + SSET_FOR_EACH (ip_address, &all_ips_v6) {
> + build_lswitch_rport_arp_flow(ip_address, AF_INET6, sw_op, sw_od,
> + 75, lflows);
> + }
> + sset_destroy(&all_ips_v4);
> + sset_destroy(&all_ips_v6);
> +
> + /* Forward ARP requests to NAT addresses configured on the router
> + * only to this router port.
> + * Priority: 30.
> + */
> + for (int i = 0; i < op->od->nbr->n_nat; i++) {
> + const struct nbrec_nat *nat = op->od->nbr->nat[i];
> +
> + if (!strcmp(nat->type, "snat")) {
> + continue;
> + }
> +
> + ovs_be32 ip;
> + ovs_be32 mask;
> + struct in6_addr ipv6;
> + struct in6_addr mask_v6;
> +
> + if (ip_parse_masked(nat->external_ip, &ip, &mask)) {
> + if (!ipv6_parse_masked(nat->external_ip, &ipv6, &mask_v6)) {
> + build_lswitch_rport_arp_flow(nat->external_ip, AF_INET6,
> sw_op,
> + sw_od, 75, lflows);
> + }
> + } else {
> + build_lswitch_rport_arp_flow(nat->external_ip, AF_INET, sw_op,
> + sw_od, 75, lflows);
> + }
> + }
> +}
> +
> static void
> build_lswitch_flows(struct hmap *datapaths, struct hmap *ports,
> struct hmap *port_groups, struct hmap *lflows,
> @@ -5748,6 +5910,15 @@ build_lswitch_flows(struct hmap *datapaths, struct
> hmap *ports,
> continue;
> }
>
> + /* For ports connected to logical routers add flows to bypass the
> + * broadcast flooding of ARP/ND requests in table 17. We direct the
> + * requests only to the router port that owns the IP address.
> + */
> + if (!strcmp(op->nbsp->type, "router")) {
> + build_lswitch_rport_arp_responders(op->peer, op->od, op,
> + lflows);
> + }
> +
> for (size_t i = 0; i < op->nbsp->n_addresses; i++) {
> /* Addresses are owned by the logical port.
> * Ethernet address followed by zero or more IPv4
> @@ -5879,12 +6050,6 @@ build_lswitch_flows(struct hmap *datapaths, struct
> hmap *ports,
> ds_destroy(&actions);
> }
>
> -static bool
> -lrport_is_enabled(const struct nbrec_logical_router_port *lrport)
> -{
> - return !lrport->enabled || *lrport->enabled;
> -}
> -
> /* Returns a string of the IP address of the router port 'op' that
> * overlaps with 'ip_s". If one is not found, returns NULL.
> *
> @@ -6904,61 +7069,66 @@ build_lrouter_flows(struct hmap *datapaths, struct
> hmap *ports,
> }
>
> /* A set to hold all load-balancer vips that need ARP responses. */
> - struct sset all_ips = SSET_INITIALIZER(&all_ips);
> - int addr_family;
> - get_router_load_balancer_ips(op->od, &all_ips, &addr_family);
> + struct sset all_ips_v4 = SSET_INITIALIZER(&all_ips_v4);
> + struct sset all_ips_v6 = SSET_INITIALIZER(&all_ips_v6);
> + get_router_load_balancer_ips(op->od, &all_ips_v4, &all_ips_v6);
>
> const char *ip_address;
> - SSET_FOR_EACH(ip_address, &all_ips) {
> + SSET_FOR_EACH (ip_address, &all_ips_v4) {
> ds_clear(&match);
> - if (addr_family == AF_INET) {
> - ds_put_format(&match,
> - "inport == %s && arp.tpa == %s && arp.op == 1",
> - op->json_key, ip_address);
> - } else {
> - ds_put_format(&match,
> - "inport == %s && nd_ns && nd.target == %s",
> - op->json_key, ip_address);
> - }
> + ds_put_format(&match,
> + "inport == %s && arp.tpa == %s && arp.op == 1",
> + op->json_key, ip_address);
>
> ds_clear(&actions);
> - if (addr_family == AF_INET) {
> - ds_put_format(&actions,
> - "eth.dst = eth.src; "
> - "eth.src = %s; "
> - "arp.op = 2; /* ARP reply */ "
> - "arp.tha = arp.sha; "
> - "arp.sha = %s; "
> - "arp.tpa = arp.spa; "
> - "arp.spa = %s; "
> - "outport = %s; "
> - "flags.loopback = 1; "
> - "output;",
> - op->lrp_networks.ea_s,
> - op->lrp_networks.ea_s,
> - ip_address,
> - op->json_key);
> - } else {
> - ds_put_format(&actions,
> - "nd_na { "
> - "eth.src = %s; "
> - "ip6.src = %s; "
> - "nd.target = %s; "
> - "nd.tll = %s; "
> - "outport = inport; "
> - "flags.loopback = 1; "
> - "output; "
> - "};",
> - op->lrp_networks.ea_s,
> - ip_address,
> - ip_address,
> - op->lrp_networks.ea_s);
> - }
> + ds_put_format(&actions,
> + "eth.dst = eth.src; "
> + "eth.src = %s; "
> + "arp.op = 2; /* ARP reply */ "
> + "arp.tha = arp.sha; "
> + "arp.sha = %s; "
> + "arp.tpa = arp.spa; "
> + "arp.spa = %s; "
> + "outport = %s; "
> + "flags.loopback = 1; "
> + "output;",
> + op->lrp_networks.ea_s,
> + op->lrp_networks.ea_s,
> + ip_address,
> + op->json_key);
> +
> ovn_lflow_add(lflows, op->od, S_ROUTER_IN_IP_INPUT, 90,
> ds_cstr(&match), ds_cstr(&actions));
> }
>
> - sset_destroy(&all_ips);
> + SSET_FOR_EACH (ip_address, &all_ips_v6) {
> + ds_clear(&match);
> + ds_put_format(&match,
> + "inport == %s && nd_ns && nd.target == %s",
> + op->json_key, ip_address);
> +
> + ds_clear(&actions);
> + ds_put_format(&actions,
> + "nd_na { "
> + "eth.src = %s; "
> + "ip6.src = %s; "
> + "nd.target = %s; "
> + "nd.tll = %s; "
> + "outport = inport; "
> + "flags.loopback = 1; "
> + "output; "
> + "};",
> + op->lrp_networks.ea_s,
> + ip_address,
> + ip_address,
> + op->lrp_networks.ea_s);
> +
> + ovn_lflow_add(lflows, op->od, S_ROUTER_IN_IP_INPUT, 90,
> + ds_cstr(&match), ds_cstr(&actions));
> + }
> +
> + sset_destroy(&all_ips_v4);
> + sset_destroy(&all_ips_v6);
>
> /* A gateway router can have 2 SNAT IP addresses to force DNATed and
> * LBed traffic respectively to be SNATed. In addition, there can be
> @@ -9392,6 +9562,12 @@ build_mcast_groups(struct northd_context *ctx,
> } else if (op->nbsp && lsp_is_enabled(op->nbsp)) {
> ovn_multicast_add(mcast_groups, &mc_flood, op);
>
> + /* Add all non-router ports to the ARP ND L2 broadcast flood
> + * domain entry. */
> + if (strcmp(op->nbsp->type, "router")) {
> + ovn_multicast_add(mcast_groups, &mc_arp_nd, op);
> + }
> +
> /* If this port is connected to a multicast router then add it
> * to the MC_MROUTER_FLOOD group.
> */
> diff --git a/tests/ovn.at b/tests/ovn.at
> index 410f4b5..c56bdce 100644
> --- a/tests/ovn.at
> +++ b/tests/ovn.at
> @@ -9595,7 +9595,7 @@ ovn-nbctl --wait=hv --timeout=3 sync
> # Check that there is a logical flow in logical switch foo's pipeline
> # to set the outport to rp-foo (which is expected).
> OVS_WAIT_UNTIL([test 1 = `ovn-sbctl dump-flows foo | grep ls_in_l2_lkup | \
> -grep rp-foo | grep -v is_chassis_resident | wc -l`])
> +grep rp-foo | grep -v is_chassis_resident | grep priority=50 -c`])
>
> # Set the option 'reside-on-redirect-chassis' for foo
> ovn-nbctl set logical_router_port foo options:reside-on-redirect-chassis=true
> @@ -9603,7 +9603,7 @@ ovn-nbctl set logical_router_port foo
> options:reside-on-redirect-chassis=true
> # to set the outport to rp-foo with the condition is_chassis_redirect.
> ovn-sbctl dump-flows foo
> OVS_WAIT_UNTIL([test 1 = `ovn-sbctl dump-flows foo | grep ls_in_l2_lkup | \
> -grep rp-foo | grep is_chassis_resident | wc -l`])
> +grep rp-foo | grep is_chassis_resident | grep priority=50 -c`])
>
> echo "---------NB dump-----"
> ovn-nbctl show
> @@ -16676,3 +16676,243 @@ as hv4 ovs-appctl fdb/show br-phys
> OVN_CLEANUP([hv1],[hv2],[hv3],[hv4])
>
> AT_CLEANUP
> +
> +AT_SETUP([ovn -- ARP/ND request broadcast limiting])
> +AT_SKIP_IF([test $HAVE_PYTHON = no])
> +ovn_start
> +
> +ip_to_hex() {
> + printf "%02x%02x%02x%02x" "$@"
> +}
> +
> +send_arp_request() {
> + local hv=$1 inport=$2 eth_src=$3 spa=$4 tpa=$5
> + local eth_dst=ffffffffffff
> + local eth_type=0806
> + local eth=${eth_dst}${eth_src}${eth_type}
> +
> + local arp=0001080006040001${eth_src}${spa}${eth_dst}${tpa}
> +
> + local request=${eth}${arp}
> + as hv$hv ovs-appctl netdev-dummy/receive hv${hv}-vif$inport $request
> +}
> +
> +send_nd_ns() {
> + local hv=$1 inport=$2 eth_src=$3 spa=$4 tpa=$5 cksum=$6
> +
> + local eth_dst=ffffffffffff
> + local eth_type=86dd
> + local eth=${eth_dst}${eth_src}${eth_type}
> +
> + local ip_vhlen=60000000
> + local ip_plen=0020
> + local ip_next=3a
> + local ip_ttl=ff
> + local ip=${ip_vhlen}${ip_plen}${ip_next}${ip_ttl}${spa}${tpa}
> +
> + # Neighbor Solicitation
> + local icmp6_type=87
> + local icmp6_code=00
> + local icmp6_rsvd=00000000
> + # ICMPv6 source lla option
> + local icmp6_opt=01
> + local icmp6_optlen=01
> + local
> icmp6=${icmp6_type}${icmp6_code}${cksum}${icmp6_rsvd}${tpa}${icmp6_opt}${icmp6_optlen}${eth_src}
> +
> + local request=${eth}${ip}${icmp6}
> +
> + as hv$hv ovs-appctl netdev-dummy/receive hv${hv}-vif$inport $request
> +}
> +
> +src_mac=000000000001
> +
> +net_add n1
> +sim_add hv1
> +as hv1
> +ovs-vsctl add-br br-phys
> +ovn_attach n1 br-phys 192.168.0.1
> +
> +ovs-vsctl -- add-port br-int hv1-vif1 -- \
> + set interface hv1-vif1 external-ids:iface-id=sw-agg-ext \
> + options:tx_pcap=hv1/vif1-tx.pcap \
> + options:rxq_pcap=hv1/vif1-rx.pcap \
> + ofport-request=1
> +
> +# One Aggregation Switch connected to two Logical networks (routers).
> +ovn-nbctl ls-add sw-agg
> +ovn-nbctl lsp-add sw-agg sw-agg-ext \
> + -- lsp-set-addresses sw-agg-ext 00:00:00:00:00:01
> +
> +ovn-nbctl lsp-add sw-agg sw-rtr1 \
> + -- lsp-set-type sw-rtr1 router \
> + -- lsp-set-addresses sw-rtr1 00:00:00:00:01:00 \
> + -- lsp-set-options sw-rtr1 router-port=rtr1-sw
> +ovn-nbctl lsp-add sw-agg sw-rtr2 \
> + -- lsp-set-type sw-rtr2 router \
> + -- lsp-set-addresses sw-rtr2 00:00:00:00:02:00 \
> + -- lsp-set-options sw-rtr2 router-port=rtr2-sw
> +
> +# Configure L3 interface IPv4 & IPv6 on both routers
> +ovn-nbctl lr-add rtr1
> +ovn-nbctl lrp-add rtr1 rtr1-sw 00:00:00:00:01:00 10.0.0.1/24 10::1/64
> +
> +ovn-nbctl lr-add rtr2
> +ovn-nbctl lrp-add rtr2 rtr2-sw 00:00:00:00:02:00 10.0.0.2/24 10::2/64
> +
> +OVN_POPULATE_ARP
> +ovn-nbctl --wait=hv sync
> +
> +sw_dp_uuid=$(ovn-sbctl --bare --columns _uuid list datapath_binding sw-agg)
> +sw_dp_key=$(ovn-sbctl --bare --columns tunnel_key list datapath_binding
> sw-agg)
> +
> +r1_tnl_key=$(ovn-sbctl --bare --columns tunnel_key list port_binding sw-rtr1)
> +r2_tnl_key=$(ovn-sbctl --bare --columns tunnel_key list port_binding sw-rtr2)
> +
> +mc_key=$(ovn-sbctl --bare --columns tunnel_key find multicast_group
> datapath=${sw_dp_uuid} name="_MC_flood")
> +mc_key=$(printf "%04x" $mc_key)
> +
> +match_sw_metadata="metadata=0x${sw_dp_key}"
> +
> +# Inject ARP request for first router owned IP address.
> +send_arp_request 1 1 ${src_mac} $(ip_to_hex 10 0 0 254) $(ip_to_hex 10 0 0 1)
> +
> +# Verify that the ARP request is sent only to rtr1.
> +match_arp_req1="${match_sw_metadata}.*arp_tpa=10.0.0.1,arp_op=1"
> +match_arp_req2="${match_sw_metadata}.*arp_tpa=10.0.0.2,arp_op=1"
> +match_send_rtr1="clone(load:0x${r1_tnl_key}->NXM_NX_REG15"
> +match_send_rtr2="clone(load:0x${r2_tnl_key}->NXM_NX_REG15"
> +
> +as hv1
> +pkts_to_rtr1=$(ovs-ofctl dump-flows br-int | grep -E "${match_arp_req1}" | \
> + grep "${match_send_rtr1}" | grep n_packets=1 -c)
> +pkts_to_rtr2=$(ovs-ofctl dump-flows br-int | grep -E "${match_arp_req2}" | \
> + grep "${match_send_rtr2}" | grep n_packets=1 -c)
> +pkts_flooded=$(ovs-ofctl dump-flows br-int | grep -E "${match_sw_metadata}"
> | \
> + grep ${mc_key} | grep -v n_packets=0 -c)
> +
> +AT_CHECK([test "1" = "${pkts_to_rtr1}"])
> +AT_CHECK([test "0" = "${pkts_to_rtr2}"])
> +AT_CHECK([test "0" = "${pkts_flooded}"])
> +
> +# Inject ND_NS for ofirst router owned IP address.
> +src_ipv6=00100000000000000000000000000254
> +dst_ipv6=00100000000000000000000000000001
> +send_nd_ns 1 1 ${src_mac} ${src_ipv6} ${dst_ipv6} 751d
> +
> +# Verify that the ND_NS is sent only to rtr1.
> +match_nd_ns1="${match_sw_metadata}.*icmp_type=135.*nd_target=10::1"
> +match_nd_ns2="${match_sw_metadata}.*icmp_type=135.*nd_target=10::2"
> +as hv1
> +pkts_to_rtr1=$(ovs-ofctl dump-flows br-int | grep -E "${match_nd_ns1}" | \
> + grep "${match_send_rtr1}" | grep n_packets=1 -c)
> +pkts_to_rtr2=$(ovs-ofctl dump-flows br-int | grep -E "${match_nd_ns1}" | \
> + grep "${match_send_rtr2}" | grep n_packets=1 -c)
> +pkts_flooded=$(ovs-ofctl dump-flows br-int | grep -E "${match_sw_metadata}"
> | \
> + grep ${mc_key} | grep -v n_packets=0 -c)
> +
> +AT_CHECK([test "1" = "${pkts_to_rtr1}"])
> +AT_CHECK([test "0" = "${pkts_to_rtr2}"])
> +AT_CHECK([test "0" = "${pkts_flooded}"])
> +
> +# Configure load balancing on both routers.
> +ovn-nbctl lb-add lb1-v4 10.0.0.11 42.42.42.1
> +ovn-nbctl lb-add lb1-v6 10::11 42::1
> +ovn-nbctl lr-lb-add rtr1 lb1-v4
> +ovn-nbctl lr-lb-add rtr1 lb1-v6
> +
> +ovn-nbctl lb-add lb2-v4 10.0.0.22 42.42.42.2
> +ovn-nbctl lb-add lb2-v6 10::22 42::2
> +ovn-nbctl lr-lb-add rtr2 lb2-v4
> +ovn-nbctl lr-lb-add rtr2 lb2-v6
> +ovn-nbctl --wait=hv sync
> +
> +# Inject ARP request for first router owned VIP address.
> +send_arp_request 1 1 ${src_mac} $(ip_to_hex 10 0 0 254) $(ip_to_hex 10 0 0
> 11)
> +
> +# Verify that the ARP request is sent only to rtr1.
> +match_arp_req1="${match_sw_metadata}.*arp_tpa=10.0.0.11,arp_op=1"
> +match_arp_req2="${match_sw_metadata}.*arp_tpa=10.0.0.22,arp_op=1"
> +match_send_rtr1="clone(load:0x${r1_tnl_key}->NXM_NX_REG15"
> +match_send_rtr2="clone(load:0x${r2_tnl_key}->NXM_NX_REG15"
> +
> +as hv1
> +pkts_to_rtr1=$(ovs-ofctl dump-flows br-int | grep -E "${match_arp_req1}" | \
> + grep "${match_send_rtr1}" | grep n_packets=1 -c)
When we run tests with -j10 for instance, this could fail in the below AT_CHECK,
if ovn-northd/ovn-controlle is little slow to catch up.
Can you please use OVS_WAIT_UNTIL for the first check instead of AT_CHECK
whenever you call send_arp_request/send_nd_ns, to be sure that we
don't fail the tests
because of timing issues ?
> +pkts_to_rtr2=$(ovs-ofctl dump-flows br-int | grep -E "${match_arp_req2}" | \
> + grep "${match_send_rtr2}" | grep n_packets=1 -c)
> +pkts_flooded=$(ovs-ofctl dump-flows br-int | grep -E "${match_sw_metadata}"
> | \
> + grep ${mc_key} | grep -v n_packets=0 -c)
> +
> +AT_CHECK([test "1" = "${pkts_to_rtr1}"])
> +AT_CHECK([test "0" = "${pkts_to_rtr2}"])
> +AT_CHECK([test "0" = "${pkts_flooded}"])
> +
> +# Inject ND_NS for first router owned VIP address.
> +src_ipv6=00100000000000000000000000000254
> +dst_ipv6=00100000000000000000000000000011
> +send_nd_ns 1 1 ${src_mac} ${src_ipv6} ${dst_ipv6} 751d
> +
> +# Verify that the ND_NS is sent only to rtr1.
> +match_nd_ns1="${match_sw_metadata}.*icmp_type=135.*nd_target=10::11"
> +match_nd_ns2="${match_sw_metadata}.*icmp_type=135.*nd_target=10::22"
> +as hv1
> +pkts_to_rtr1=$(ovs-ofctl dump-flows br-int | grep -E "${match_nd_ns1}" | \
> + grep "${match_send_rtr1}" | grep n_packets=1 -c)
> +pkts_to_rtr2=$(ovs-ofctl dump-flows br-int | grep -E "${match_nd_ns1}" | \
> + grep "${match_send_rtr2}" | grep n_packets=1 -c)
> +pkts_flooded=$(ovs-ofctl dump-flows br-int | grep -E "${match_sw_metadata}"
> | \
> + grep ${mc_key} | grep -v n_packets=0 -c)
> +
> +AT_CHECK([test "1" = "${pkts_to_rtr1}"])
> +AT_CHECK([test "0" = "${pkts_to_rtr2}"])
> +AT_CHECK([test "0" = "${pkts_flooded}"])
> +
> +# Configure NAT on both routers
> +ovn-nbctl lr-nat-add rtr1 dnat_and_snat 10.0.0.111 42.42.42.1
> +ovn-nbctl lr-nat-add rtr1 dnat_and_snat 10::111 42::1
> +ovn-nbctl lr-nat-add rtr2 dnat_and_snat 10.0.0.222 42.42.42.2
> +ovn-nbctl lr-nat-add rtr2 dnat_and_snat 10::222 42::2
> +
> +# Inject ARP request for first router owned NAT address.
> +send_arp_request 1 1 ${src_mac} $(ip_to_hex 10 0 0 254) $(ip_to_hex 10 0 0
> 111)
> +
> +# Verify that the ARP request is sent only to rtr1.
> +match_arp_req1="${match_sw_metadata}.*arp_tpa=10.0.0.111,arp_op=1"
> +match_arp_req2="${match_sw_metadata}.*arp_tpa=10.0.0.222,arp_op=1"
> +match_send_rtr1="clone(load:0x${r1_tnl_key}->NXM_NX_REG15"
> +match_send_rtr2="clone(load:0x${r2_tnl_key}->NXM_NX_REG15"
> +
> +as hv1
> +pkts_to_rtr1=$(ovs-ofctl dump-flows br-int | grep -E "${match_arp_req1}" | \
> + grep "${match_send_rtr1}" | grep n_packets=1 -c)
> +pkts_to_rtr2=$(ovs-ofctl dump-flows br-int | grep -E "${match_arp_req2}" | \
> + grep "${match_send_rtr2}" | grep n_packets=1 -c)
> +pkts_flooded=$(ovs-ofctl dump-flows br-int | grep -E "${match_sw_metadata}"
> | \
> + grep ${mc_key} | grep -v n_packets=0 -c)
> +
> +AT_CHECK([test "1" = "${pkts_to_rtr1}"])
> +AT_CHECK([test "0" = "${pkts_to_rtr2}"])
> +AT_CHECK([test "0" = "${pkts_flooded}"])
> +
> +# Inject ND_NS for first router owned IP address.
> +src_ipv6=00100000000000000000000000000254
> +dst_ipv6=00100000000000000000000000000111
> +send_nd_ns 1 1 ${src_mac} ${src_ipv6} ${dst_ipv6} 751d
> +
> +# Verify that the ND_NS is sent only to rtr1.
> +match_nd_ns1="${match_sw_metadata}.*icmp_type=135.*nd_target=10::111"
> +match_nd_ns2="${match_sw_metadata}.*icmp_type=135.*nd_target=10::222"
> +as hv1
> +pkts_to_rtr1=$(ovs-ofctl dump-flows br-int | grep -E "${match_nd_ns1}" | \
> + grep "${match_send_rtr1}" | grep n_packets=1 -c)
> +pkts_to_rtr2=$(ovs-ofctl dump-flows br-int | grep -E "${match_nd_ns1}" | \
> + grep "${match_send_rtr2}" | grep n_packets=1 -c)
> +pkts_flooded=$(ovs-ofctl dump-flows br-int | grep -E "${match_sw_metadata}"
> | \
> + grep ${mc_key} | grep -v n_packets=0 -c)
> +
> +AT_CHECK([test "1" = "${pkts_to_rtr1}"])
> +AT_CHECK([test "0" = "${pkts_to_rtr2}"])
> +AT_CHECK([test "0" = "${pkts_flooded}"])
> +
> +OVN_CLEANUP([hv1])
> +AT_CLEANUP
> --
> 1.8.3.1
>
> _______________________________________________
> dev mailing list
> [email protected]
> https://mail.openvswitch.org/mailman/listinfo/ovs-dev
_______________________________________________
dev mailing list
[email protected]
https://mail.openvswitch.org/mailman/listinfo/ovs-dev