On Thu, Jun 2, 2022 at 10:09 AM Mark Michelson <[email protected]> wrote:
>
> On 6/2/22 11:22, Dumitru Ceara wrote:
> > Change the way ovn-controller decides whether it should match on
> > ct_mark.natted or ct_label.natted for hairpin load balancer traffic.
> > Until now this was done solely based on the northd-reported internal
> > version.
> >
> > However, to cover the case when OVN central components are not the last
> > ones to be updated, ovn-northd now explicitly informs ovn-controller
> > whether it should use ct_mark or ct_label.natted via a new option in the
> > OVN_Southbound.Load_Balancer record: hairpin_use_ct_mark.
> >
> > Signed-off-by: Dumitru Ceara <[email protected]>
> > ---
> > controller/lflow.c | 32 ++++++++++++--------------------
> > controller/lflow.h | 1 -
> > controller/ovn-controller.c | 9 ---------
> > lib/lb.c | 3 +++
> > lib/lb.h | 4 ++++
> > northd/northd.c | 8 ++++++--
> > ovn-sb.xml | 8 ++++++++
> > tests/ovn-northd.at | 2 +-
> > 8 files changed, 34 insertions(+), 33 deletions(-)
> >
> > diff --git a/controller/lflow.c b/controller/lflow.c
> > index 7a3419305..0d9184285 100644
> > --- a/controller/lflow.c
> > +++ b/controller/lflow.c
> > @@ -1942,7 +1942,6 @@ add_lb_vip_hairpin_flows(struct ovn_controller_lb
*lb,
> > struct ovn_lb_vip *lb_vip,
> > struct ovn_lb_backend *lb_backend,
> > uint8_t lb_proto,
> > - bool check_ct_label_for_lb_hairpin,
> > struct ovn_desired_flow_table *flow_table)
> > {
> > uint64_t stub[1024 / 8];
> > @@ -2033,18 +2032,19 @@ add_lb_vip_hairpin_flows(struct
ovn_controller_lb *lb,
> > * - packets must have ip.src == ip.dst at this point.
> > * - the destination protocol and port must be of a valid backend
that
> > * has the same IP as ip.dst.
> > + *
> > + * During upgrades logical flow might still use the old way of
storing
> > + * ct.natted in ct_label. For backwards compatibility, only use
ct_mark
> > + * if ovn-northd notified ovn-controller to do that.
> > */
> > - uint32_t lb_ct_mark = OVN_CT_NATTED;
> > - match_set_ct_mark_masked(&hairpin_match, lb_ct_mark, lb_ct_mark);
> > -
> > - ofctrl_add_flow(flow_table, OFTABLE_CHK_LB_HAIRPIN, 100,
> > - lb->slb->header_.uuid.parts[0], &hairpin_match,
> > - &ofpacts, &lb->slb->header_.uuid);
> > + if (lb->hairpin_use_ct_mark) {
> > + uint32_t lb_ct_mark = OVN_CT_NATTED;
> > + match_set_ct_mark_masked(&hairpin_match, lb_ct_mark,
lb_ct_mark);
> >
> > - /* The below flow is identical to the above except that it checks
> > - * ct_label.natted instead of ct_mark.natted, for backward
compatibility
> > - * during the upgrade from a previous version that uses ct_label.
*/
> > - if (check_ct_label_for_lb_hairpin) {
> > + ofctrl_add_flow(flow_table, OFTABLE_CHK_LB_HAIRPIN, 100,
> > + lb->slb->header_.uuid.parts[0], &hairpin_match,
> > + &ofpacts, &lb->slb->header_.uuid);
> > + } else {
> > match_set_ct_mark_masked(&hairpin_match, 0, 0);
> > ovs_u128 lb_ct_label = {
> > .u64.lo = OVN_CT_NATTED,
> > @@ -2328,7 +2328,6 @@ add_lb_ct_snat_hairpin_flows(struct
ovn_controller_lb *lb,
> > static void
> > consider_lb_hairpin_flows(const struct sbrec_load_balancer *sbrec_lb,
> > const struct hmap *local_datapaths,
> > - bool check_ct_label_for_lb_hairpin,
> > struct ovn_desired_flow_table *flow_table,
> > struct simap *ids)
> > {
> > @@ -2368,7 +2367,6 @@ consider_lb_hairpin_flows(const struct
sbrec_load_balancer *sbrec_lb,
> > struct ovn_lb_backend *lb_backend = &lb_vip->backends[j];
> >
> > add_lb_vip_hairpin_flows(lb, lb_vip, lb_backend, lb_proto,
> > - check_ct_label_for_lb_hairpin,
> > flow_table);
> > }
> > }
> > @@ -2383,7 +2381,6 @@ consider_lb_hairpin_flows(const struct
sbrec_load_balancer *sbrec_lb,
> > static void
> > add_lb_hairpin_flows(const struct sbrec_load_balancer_table *lb_table,
> > const struct hmap *local_datapaths,
> > - bool check_ct_label_for_lb_hairpin,
> > struct ovn_desired_flow_table *flow_table,
> > struct simap *ids,
> > struct id_pool *pool)
> > @@ -2406,9 +2403,7 @@ add_lb_hairpin_flows(const struct
sbrec_load_balancer_table *lb_table,
> > ovs_assert(id_pool_alloc_id(pool, &id));
> > simap_put(ids, lb->name, id);
> > }
> > - consider_lb_hairpin_flows(lb, local_datapaths,
> > - check_ct_label_for_lb_hairpin,
> > - flow_table, ids);
> > + consider_lb_hairpin_flows(lb, local_datapaths, flow_table,
ids);
> > }
> > }
> >
> > @@ -2545,7 +2540,6 @@ lflow_run(struct lflow_ctx_in *l_ctx_in, struct
lflow_ctx_out *l_ctx_out)
> > l_ctx_in->local_datapaths,
> > l_ctx_out->flow_table);
> > add_lb_hairpin_flows(l_ctx_in->lb_table,
l_ctx_in->local_datapaths,
> > - l_ctx_in->check_ct_label_for_lb_hairpin,
> > l_ctx_out->flow_table,
> > l_ctx_out->hairpin_lb_ids,
> > l_ctx_out->hairpin_id_pool);
> > @@ -2709,7 +2703,6 @@ lflow_add_flows_for_datapath(const struct
sbrec_datapath_binding *dp,
> > * associated. */
> > for (size_t i = 0; i < n_dp_lbs; i++) {
> > consider_lb_hairpin_flows(dp_lbs[i],
l_ctx_in->local_datapaths,
> > -
l_ctx_in->check_ct_label_for_lb_hairpin,
> > l_ctx_out->flow_table,
> > l_ctx_out->hairpin_lb_ids);
> > }
> > @@ -2840,7 +2833,6 @@ lflow_handle_changed_lbs(struct lflow_ctx_in
*l_ctx_in,
> > VLOG_DBG("Add load balancer hairpin flows for "UUID_FMT,
> > UUID_ARGS(&lb->header_.uuid));
> > consider_lb_hairpin_flows(lb, l_ctx_in->local_datapaths,
> > -
l_ctx_in->check_ct_label_for_lb_hairpin,
> > l_ctx_out->flow_table,
> > l_ctx_out->hairpin_lb_ids);
> > }
> > diff --git a/controller/lflow.h b/controller/lflow.h
> > index ad9449d3a..2aa896a75 100644
> > --- a/controller/lflow.h
> > +++ b/controller/lflow.h
> > @@ -160,7 +160,6 @@ struct lflow_ctx_in {
> > const struct sset *related_lport_ids;
> > const struct shash *binding_lports;
> > const struct hmap *chassis_tunnels;
> > - bool check_ct_label_for_lb_hairpin;
> > };
> >
> > struct lflow_ctx_out {
> > diff --git a/controller/ovn-controller.c b/controller/ovn-controller.c
> > index b597c0e37..b9801c42d 100644
> > --- a/controller/ovn-controller.c
> > +++ b/controller/ovn-controller.c
> > @@ -486,13 +486,6 @@ get_ovs_chassis_id(const struct
ovsrec_open_vswitch_table *ovs_table)
> > return chassis_id;
> > }
> >
> > -static bool
> > -get_check_ct_label_for_lb_hairpin(const char *northd_internal_ver)
> > -{
> > - unsigned int minor =
ovn_parse_internal_version_minor(northd_internal_ver);
> > - return (minor <= 3);
> > -}
> > -
> > static void
> > update_ssl_config(const struct ovsrec_ssl_table *ssl_table)
> > {
> > @@ -2529,8 +2522,6 @@ init_lflow_ctx(struct engine_node *node,
> > l_ctx_in->related_lport_ids = &rt_data->related_lports.lport_ids;
> > l_ctx_in->binding_lports = &rt_data->lbinding_data.lports;
> > l_ctx_in->chassis_tunnels = &non_vif_data->chassis_tunnels;
> > - l_ctx_in->check_ct_label_for_lb_hairpin =
> > - get_check_ct_label_for_lb_hairpin(n_ver->ver);
> >
> > l_ctx_out->flow_table = &fo->flow_table;
> > l_ctx_out->group_table = &fo->group_table;
> > diff --git a/lib/lb.c b/lib/lb.c
> > index 7b0ed1abe..6a48175a7 100644
> > --- a/lib/lb.c
> > +++ b/lib/lb.c
> > @@ -304,6 +304,9 @@ ovn_controller_lb_create(const struct
sbrec_load_balancer *sbrec_lb)
> > lb->hairpin_orig_tuple = smap_get_bool(&sbrec_lb->options,
> > "hairpin_orig_tuple",
> > false);
> > + lb->hairpin_use_ct_mark = smap_get_bool(&sbrec_lb->options,
> > + "hairpin_use_ct_mark",
> > + false);
>
> For the backports to 22.03 and 21.12, I understand why this would
> default to false. However, for "new" versions of OVN (main and 22.06),
> why don't we default to true? I know this would cause trouble for the
> case of upgrading the central by a major version before updating the
> hosts, but I thought the point of these patches was to make sure that
> minor release upgrades would not break.
I agree with Mark here. Otherwise, we would need to keep the option
explicitly forever, which looks quite noisy.
In addition, I wonder if it is better to just have a global NB option
rather than per LB option? A global option looks more clear/obvious and
sufficient.
Thanks,
Han
>
> > ovn_lb_get_hairpin_snat_ip(&sbrec_lb->header_.uuid,
&sbrec_lb->options,
> > &lb->hairpin_snat_ips);
> > return lb;
> > diff --git a/lib/lb.h b/lib/lb.h
> > index 832ed31fb..d703c6bf5 100644
> > --- a/lib/lb.h
> > +++ b/lib/lb.h
> > @@ -101,6 +101,10 @@ struct ovn_controller_lb {
> > bool hairpin_orig_tuple; /* True if ovn-northd stores the original
> > * destination tuple in registers.
> > */
> > + bool hairpin_use_ct_mark; /* True if ovn-northd uses ct_mark for
> > + * load balancer sessions. False if it
uses
> > + * ct_label.
> > + */
> >
> > struct lport_addresses hairpin_snat_ips; /* IP (v4 and/or v6) to
be used
> > * as source for
hairpinned
> > diff --git a/northd/northd.c b/northd/northd.c
> > index 450e05ad6..6e2f2a880 100644
> > --- a/northd/northd.c
> > +++ b/northd/northd.c
> > @@ -4146,7 +4146,8 @@ build_lb_port_related_data(struct hmap
*datapaths, struct hmap *ports,
> > */
> > static void
> > sync_lbs(struct northd_input *input_data, struct ovsdb_idl_txn
*ovnsb_txn,
> > - struct hmap *datapaths, struct hmap *lbs)
> > + struct hmap *datapaths, struct hmap *lbs,
> > + const struct chassis_features *features)
> > {
> > struct ovn_northd_lb *lb;
> >
> > @@ -4193,6 +4194,8 @@ sync_lbs(struct northd_input *input_data, struct
ovsdb_idl_txn *ovnsb_txn,
> > struct smap options;
> > smap_clone(&options, &lb->nlb->options);
> > smap_replace(&options, "hairpin_orig_tuple", "true");
> > + smap_replace(&options, "hairpin_use_ct_mark",
> > + features->ct_lb_mark ? "true" : "false");
> >
> > struct sbrec_datapath_binding **lb_dps =
> > xmalloc(lb->n_nb_ls * sizeof *lb_dps);
> > @@ -15344,7 +15347,8 @@ ovnnb_db_run(struct northd_input *input_data,
> > ovn_update_ipv6_options(&data->ports);
> > ovn_update_ipv6_prefix(&data->ports);
> >
> > - sync_lbs(input_data, ovnsb_txn, &data->datapaths, &data->lbs);
> > + sync_lbs(input_data, ovnsb_txn, &data->datapaths, &data->lbs,
> > + &data->features);
> > sync_address_sets(input_data, ovnsb_txn, &data->datapaths);
> > sync_port_groups(input_data, ovnsb_txn, &data->port_groups);
> > sync_meters(input_data, ovnsb_txn, &data->meter_groups);
> > diff --git a/ovn-sb.xml b/ovn-sb.xml
> > index 2dc0d5bea..b8b8074ec 100644
> > --- a/ovn-sb.xml
> > +++ b/ovn-sb.xml
> > @@ -4566,6 +4566,14 @@ tcp.flags = RST;
> > of the load balanced packets are stored in registers
> > <code>reg1, reg2, xxreg1</code>.
> > </column>
> > + <column name="options" key="hairpin_use_ct_mark"
> > + type='{"type": "boolean"}'>
> > + This value is automatically set to <code>true</code> by
> > + <code>ovn-northd</code> when action <code>ct_lb_mark</code> is
used
> > + for new load balancer sessions. <code>ovn-controller</code>
then knows
> > + that it should check <code>ct_mark.natted</code> to detect load
balanced
> > + traffic.
> > + </column>
> > </group>
> >
> > <group title="Common Columns">
> > diff --git a/tests/ovn-northd.at b/tests/ovn-northd.at
> > index 7bb6d33ac..4bb815c7b 100644
> > --- a/tests/ovn-northd.at
> > +++ b/tests/ovn-northd.at
> > @@ -2563,7 +2563,7 @@ check_column "" sb:datapath_binding
load_balancers external_ids:name=sw1
> > echo
> > echo "__file__:__line__: Set hairpin_snat_ip on lb1 and check that SB
DB is updated."
> > check ovn-nbctl --wait=sb set Load_Balancer lb1
options:hairpin_snat_ip="42.42.42.42 4242::4242"
> > -check_column "$lb1_uuid" sb:load_balancer _uuid name=lb1
options='{hairpin_orig_tuple="true", hairpin_snat_ip="42.42.42.42
4242::4242"}'
> > +check_column "$lb1_uuid" sb:load_balancer _uuid name=lb1
options='{hairpin_orig_tuple="true", hairpin_snat_ip="42.42.42.42
4242::4242", hairpin_use_ct_mark="true"}'
> >
> > echo
> > echo "__file__:__line__: Delete load balancers lb1 and lbg1 and check
that datapath sw1's load_balancers is still empty."
> >
>
_______________________________________________
dev mailing list
[email protected]
https://mail.openvswitch.org/mailman/listinfo/ovs-dev