On 28 Jan 2022, at 17:04, Adrian Moreno wrote:

> In order to be able to reuse the core extraction logic, split the command
> in two parts. The core extraction logic is moved to python/build while
> the command that writes the different files out of the extracted field
> info is kept in build-aux.
>
> Signed-off-by: Adrian Moreno <amore...@redhat.com>

Guess you missed one previous comment, rest looks good.

> ---
>  build-aux/extract-ofp-fields       | 706 ++++++++---------------------
>  python/automake.mk                 |   1 +
>  python/build/extract_ofp_fields.py | 386 ++++++++++++++++
>  3 files changed, 579 insertions(+), 514 deletions(-)
>  create mode 100644 python/build/extract_ofp_fields.py
>
> diff --git a/build-aux/extract-ofp-fields b/build-aux/extract-ofp-fields
> index 8766995d9..efec59c25 100755
> --- a/build-aux/extract-ofp-fields
> +++ b/build-aux/extract-ofp-fields
> @@ -3,85 +3,23 @@
>  import getopt
>  import sys
>  import os.path
> -import re
>  import xml.dom.minidom
>  import build.nroff
>
> -line = ""
> -
> -# Maps from user-friendly version number to its protocol encoding.
> -VERSION = {"1.0": 0x01,
> -           "1.1": 0x02,
> -           "1.2": 0x03,
> -           "1.3": 0x04,
> -           "1.4": 0x05,
> -           "1.5": 0x06}
> -VERSION_REVERSE = dict((v,k) for k, v in VERSION.items())
> -
> -TYPES = {"u8":       (1,   False),
> -         "be16":     (2,   False),
> -         "be32":     (4,   False),
> -         "MAC":      (6,   False),
> -         "be64":     (8,   False),
> -         "be128":    (16,  False),
> -         "tunnelMD": (124, True)}
> -
> -FORMATTING = {"decimal":            ("MFS_DECIMAL",      1,   8),
> -              "hexadecimal":        ("MFS_HEXADECIMAL",  1, 127),
> -              "ct state":           ("MFS_CT_STATE",     4,   4),
> -              "Ethernet":           ("MFS_ETHERNET",     6,   6),
> -              "IPv4":               ("MFS_IPV4",         4,   4),
> -              "IPv6":               ("MFS_IPV6",        16,  16),
> -              "OpenFlow 1.0 port":  ("MFS_OFP_PORT",     2,   2),
> -              "OpenFlow 1.1+ port": ("MFS_OFP_PORT_OXM", 4,   4),
> -              "frag":               ("MFS_FRAG",         1,   1),
> -              "tunnel flags":       ("MFS_TNL_FLAGS",    2,   2),
> -              "TCP flags":          ("MFS_TCP_FLAGS",    2,   2),
> -              "packet type":        ("MFS_PACKET_TYPE",  4,   4)}
> -
> -PREREQS = {"none": "MFP_NONE",
> -           "Ethernet": "MFP_ETHERNET",
> -           "ARP": "MFP_ARP",
> -           "VLAN VID": "MFP_VLAN_VID",
> -           "IPv4": "MFP_IPV4",
> -           "IPv6": "MFP_IPV6",
> -           "IPv4/IPv6": "MFP_IP_ANY",
> -           "NSH": "MFP_NSH",
> -           "CT": "MFP_CT_VALID",
> -           "MPLS": "MFP_MPLS",
> -           "TCP": "MFP_TCP",
> -           "UDP": "MFP_UDP",
> -           "SCTP": "MFP_SCTP",
> -           "ICMPv4": "MFP_ICMPV4",
> -           "ICMPv6": "MFP_ICMPV6",
> -           "ND": "MFP_ND",
> -           "ND solicit": "MFP_ND_SOLICIT",
> -           "ND advert": "MFP_ND_ADVERT"}
> -
> -# Maps a name prefix into an (experimenter ID, class) pair, so:
> -#
> -#      - Standard OXM classes are written as (0, <oxm_class>)
> -#
> -#      - Experimenter OXM classes are written as (<oxm_vender>, 0xffff)
> -#
> -# If a name matches more than one prefix, the longest one is used.
> -OXM_CLASSES = {"NXM_OF_":        (0,          0x0000, 'extension'),
> -               "NXM_NX_":        (0,          0x0001, 'extension'),
> -               "NXOXM_NSH_":     (0x005ad650, 0xffff, 'extension'),
> -               "OXM_OF_":        (0,          0x8000, 'standard'),
> -               "OXM_OF_PKT_REG": (0,          0x8001, 'standard'),
> -               "ONFOXM_ET_":     (0x4f4e4600, 0xffff, 'standard'),
> -               "ERICOXM_OF_":    (0,          0x1000, 'extension'),
> -
> -               # This is the experimenter OXM class for Nicira, which is the
> -               # one that OVS would be using instead of NXM_OF_ and NXM_NX_
> -               # if OVS didn't have those grandfathered in.  It is currently
> -               # used only to test support for experimenter OXM, since there
> -               # are barely any real uses of experimenter OXM in the wild.
> -               "NXOXM_ET_":      (0x00002320, 0xffff, 'extension')}
> +from build.extract_ofp_fields import (
> +    extract_ofp_fields,
> +    PREREQS,
> +    OXM_CLASSES,
> +    VERSION,
> +    fatal,
> +    n_errors,
> +)
> +
> +VERSION_REVERSE = dict((v, k) for k, v in VERSION.items())
> +
>
>  def oxm_name_to_class(name):
> -    prefix = ''
> +    prefix = ""
>      class_ = None
>      for p, c in OXM_CLASSES.items():
>          if name.startswith(p) and len(p) > len(prefix):
> @@ -92,267 +30,76 @@ def oxm_name_to_class(name):
>
>  def is_standard_oxm(name):
>      oxm_vendor, oxm_class, oxm_class_type = oxm_name_to_class(name)
> -    return oxm_class_type == 'standard'
> -
> -
> -def decode_version_range(range):
> -    if range in VERSION:
> -        return (VERSION[range], VERSION[range])
> -    elif range.endswith('+'):
> -        return (VERSION[range[:-1]], max(VERSION.values()))
> -    else:
> -        a, b = re.match(r'^([^-]+)-([^-]+)$', range).groups()
> -        return (VERSION[a], VERSION[b])
> -
> -
> -def get_line():
> -    global line
> -    global line_number
> -    line = input_file.readline()
> -    line_number += 1
> -    if line == "":
> -        fatal("unexpected end of input")
> -
> -
> -n_errors = 0
> -
> -
> -def error(msg):
> -    global n_errors
> -    sys.stderr.write("%s:%d: %s\n" % (file_name, line_number, msg))
> -    n_errors += 1
> -
> -
> -def fatal(msg):
> -    error(msg)
> -    sys.exit(1)
> +    return oxm_class_type == "standard"
>
>
>  def usage():
>      argv0 = os.path.basename(sys.argv[0])
> -    print('''\
> +    print(
> +        """\
>  %(argv0)s, for extracting OpenFlow field properties from meta-flow.h
>  usage: %(argv0)s INPUT [--meta-flow | --nx-match]
>    where INPUT points to lib/meta-flow.h in the source directory.
>  Depending on the option given, the output written to stdout is intended to be
>  saved either as lib/meta-flow.inc or lib/nx-match.inc for the respective C
>  file to #include.\
> -''' % {"argv0": argv0})
> +"""
> +        % {"argv0": argv0}
> +    )
>      sys.exit(0)
>
>
> -def make_sizeof(s):
> -    m = re.match(r'(.*) up to (.*)', s)
> -    if m:
> -        struct, member = m.groups()
> -        return "offsetof(%s, %s)" % (struct, member)
> -    else:
> -        return "sizeof(%s)" % s
> -
> -
> -def parse_oxms(s, prefix, n_bytes):
> -    if s == 'none':
> -        return ()
> -
> -    return tuple(parse_oxm(s2.strip(), prefix, n_bytes) for s2 in 
> s.split(','))
> -
> -
> -match_types = dict()
> -
> -
> -def parse_oxm(s, prefix, n_bytes):
> -    global match_types
> -
> -    m = re.match('([A-Z0-9_]+)\(([0-9]+)\) since(?: OF(1\.[0-9]+) and)? 
> v([12]\.[0-9]+)$', s)
> -    if not m:
> -        fatal("%s: syntax error parsing %s" % (s, prefix))
> -
> -    name, oxm_type, of_version, ovs_version = m.groups()
> -
> -    class_ = oxm_name_to_class(name)
> -    if class_ is None:
> -        fatal("unknown OXM class for %s" % name)
> -    oxm_vendor, oxm_class, oxm_class_type = class_
> -
> -    if class_ in match_types:
> -        if oxm_type in match_types[class_]:
> -            fatal("duplicate match type for %s (conflicts with %s)" %
> -                  (name, match_types[class_][oxm_type]))
> -    else:
> -        match_types[class_] = dict()
> -    match_types[class_][oxm_type] = name
> -
> -    # Normally the oxm_length is the size of the field, but for experimenter
> -    # OXMs oxm_length also includes the 4-byte experimenter ID.
> -    oxm_length = n_bytes
> -    if oxm_class == 0xffff:
> -        oxm_length += 4
> -
> -    header = (oxm_vendor, oxm_class, int(oxm_type), oxm_length)
> -
> -    if of_version:
> -        if oxm_class_type == 'extension':
> -            fatal("%s: OXM extension can't have OpenFlow version" % name)
> -        if of_version not in VERSION:
> -            fatal("%s: unknown OpenFlow version %s" % (name, of_version))
> -        of_version_nr = VERSION[of_version]
> -        if of_version_nr < VERSION['1.2']:
> -            fatal("%s: claimed version %s predates OXM" % (name, of_version))
> -    else:
> -        if oxm_class_type == 'standard':
> -            fatal("%s: missing OpenFlow version number" % name)
> -        of_version_nr = 0
> -
> -    return (header, name, of_version_nr, ovs_version)
> -
> -
> -def parse_field(mff, comment):
> -    f = {'mff': mff}
> -
> -    # First line of comment is the field name.
> -    m = re.match(r'"([^"]+)"(?:\s+\(aka "([^"]+)"\))?(?:\s+\(.*\))?\.', 
> comment[0])
> -    if not m:
> -        fatal("%s lacks field name" % mff)
> -    f['name'], f['extra_name'] = m.groups()
> -
> -    # Find the last blank line the comment.  The field definitions
> -    # start after that.
> -    blank = None
> -    for i in range(len(comment)):
> -        if not comment[i]:
> -            blank = i
> -    if not blank:
> -        fatal("%s: missing blank line in comment" % mff)
> -
> -    d = {}
> -    for key in ("Type", "Maskable", "Formatting", "Prerequisites",
> -                "Access", "Prefix lookup member",
> -                "OXM", "NXM", "OF1.0", "OF1.1"):
> -        d[key] = None
> -    for fline in comment[blank + 1:]:
> -        m = re.match(r'([^:]+):\s+(.*)\.$', fline)
> -        if not m:
> -            fatal("%s: syntax error parsing key-value pair as part of %s"
> -                  % (fline, mff))
> -        key, value = m.groups()
> -        if key not in d:
> -            fatal("%s: unknown key" % key)
> -        elif key == 'Code point':
> -            d[key] += [value]
> -        elif d[key] is not None:
> -            fatal("%s: duplicate key" % key)
> -        d[key] = value
> -    for key, value in d.items():
> -        if not value and key not in ("OF1.0", "OF1.1",
> -                                     "Prefix lookup member", "Notes"):
> -            fatal("%s: missing %s" % (mff, key))
> -
> -    m = re.match(r'([a-zA-Z0-9]+)(?: \(low ([0-9]+) bits\))?$', d['Type'])
> -    if not m:
> -        fatal("%s: syntax error in type" % mff)
> -    type_ = m.group(1)
> -    if type_ not in TYPES:
> -        fatal("%s: unknown type %s" % (mff, d['Type']))
> -
> -    f['n_bytes'] = TYPES[type_][0]
> -    if m.group(2):
> -        f['n_bits'] = int(m.group(2))
> -        if f['n_bits'] > f['n_bytes'] * 8:
> -            fatal("%s: more bits (%d) than field size (%d)"
> -                  % (mff, f['n_bits'], 8 * f['n_bytes']))
> -    else:
> -        f['n_bits'] = 8 * f['n_bytes']
> -    f['variable'] = TYPES[type_][1]
> -
> -    if d['Maskable'] == 'no':
> -        f['mask'] = 'MFM_NONE'
> -    elif d['Maskable'] == 'bitwise':
> -        f['mask'] = 'MFM_FULLY'
> -    else:
> -        fatal("%s: unknown maskable %s" % (mff, d['Maskable']))
> -
> -    fmt = FORMATTING.get(d['Formatting'])
> -    if not fmt:
> -        fatal("%s: unknown format %s" % (mff, d['Formatting']))
> -    f['formatting'] = d['Formatting']
> -    if f['n_bytes'] < fmt[1] or f['n_bytes'] > fmt[2]:
> -        fatal("%s: %d-byte field can't be formatted as %s"
> -              % (mff, f['n_bytes'], d['Formatting']))
> -    f['string'] = fmt[0]
> -
> -    f['prereqs'] = d['Prerequisites']
> -    if f['prereqs'] not in PREREQS:
> -        fatal("%s: unknown prerequisites %s" % (mff, d['Prerequisites']))
> -
> -    if d['Access'] == 'read-only':
> -        f['writable'] = False
> -    elif d['Access'] == 'read/write':
> -        f['writable'] = True
> -    else:
> -        fatal("%s: unknown access %s" % (mff, d['Access']))
> -
> -    f['OF1.0'] = d['OF1.0']
> -    if not d['OF1.0'] in (None, 'exact match', 'CIDR mask'):
> -        fatal("%s: unknown OF1.0 match type %s" % (mff, d['OF1.0']))
> -
> -    f['OF1.1'] = d['OF1.1']
> -    if not d['OF1.1'] in (None, 'exact match', 'bitwise mask'):
> -        fatal("%s: unknown OF1.1 match type %s" % (mff, d['OF1.1']))
> -
> -    f['OXM'] = (parse_oxms(d['OXM'], 'OXM', f['n_bytes']) +
> -                parse_oxms(d['NXM'], 'NXM', f['n_bytes']))
> -
> -    f['prefix'] = d["Prefix lookup member"]
> -
> -    return f
> -
> -
>  def protocols_to_c(protocols):
> -    if protocols == set(['of10', 'of11', 'oxm']):
> -        return 'OFPUTIL_P_ANY'
> -    elif protocols == set(['of11', 'oxm']):
> -        return 'OFPUTIL_P_NXM_OF11_UP'
> -    elif protocols == set(['oxm']):
> -        return 'OFPUTIL_P_NXM_OXM_ANY'
> +    if protocols == set(["of10", "of11", "oxm"]):
> +        return "OFPUTIL_P_ANY"
> +    elif protocols == set(["of11", "oxm"]):
> +        return "OFPUTIL_P_NXM_OF11_UP"
> +    elif protocols == set(["oxm"]):
> +        return "OFPUTIL_P_NXM_OXM_ANY"
>      elif protocols == set([]):
> -        return 'OFPUTIL_P_NONE'
> +        return "OFPUTIL_P_NONE"
>      else:
>          assert False
>
>
>  def autogen_c_comment():
>      return [
> -"/* Generated automatically; do not modify!    -*- buffer-read-only: t -*- 
> */",
> -""]
> +        "/* Generated automatically; do not modify!    "
> +        "-*- buffer-read-only: t -*- */",
> +        "",
> +    ]
> +
>
>  def make_meta_flow(meta_flow_h):
>      fields = extract_ofp_fields(meta_flow_h)
>      output = autogen_c_comment()
>      for f in fields:
>          output += ["{"]
> -        output += ["    %s," % f['mff']]
> -        if f['extra_name']:
> -            output += ["    \"%s\", \"%s\"," % (f['name'], f['extra_name'])]
> +        output += ["    %s," % f["mff"]]
> +        if f["extra_name"]:
> +            output += ['    "%s", "%s",' % (f["name"], f["extra_name"])]
>          else:
> -            output += ["    \"%s\", NULL," % f['name']]
> +            output += ['    "%s", NULL,' % f["name"]]
>
> -        if f['variable']:
> -            variable = 'true'
> +        if f["variable"]:
> +            variable = "true"
>          else:
> -            variable = 'false'
> -        output += ["    %d, %d, %s," % (f['n_bytes'], f['n_bits'], variable)]
> +            variable = "false"
> +        output += ["    %d, %d, %s," % (f["n_bytes"], f["n_bits"], variable)]
>
> -        if f['writable']:
> -            rw = 'true'
> +        if f["writable"]:
> +            rw = "true"
>          else:
> -            rw = 'false'
> -        output += ["    %s, %s, %s, %s, false,"
> -                   % (f['mask'], f['string'], PREREQS[f['prereqs']], rw)]
> -
> -        oxm = f['OXM']
> -        of10 = f['OF1.0']
> -        of11 = f['OF1.1']
> -        if f['mff'] in ('MFF_DL_VLAN', 'MFF_DL_VLAN_PCP'):
> +            rw = "false"
> +        output += [
> +            "    %s, %s, %s, %s, false,"
> +            % (f["mask"], f["string"], PREREQS[f["prereqs"]], rw)
> +        ]
> +
> +        oxm = f["OXM"]
> +        of10 = f["OF1.0"]
> +        of11 = f["OF1.1"]
> +        if f["mff"] in ("MFF_DL_VLAN", "MFF_DL_VLAN_PCP"):
>              # MFF_DL_VLAN and MFF_DL_VLAN_PCP don't exactly correspond to
>              # OF1.1, nor do they have NXM or OXM assignments, but their
>              # meanings can be expressed in every protocol, which is the goal 
> of
> @@ -367,25 +114,25 @@ def make_meta_flow(meta_flow_h):
>              if oxm:
>                  protocols |= set(["oxm"])
>
> -        if f['mask'] == 'MFM_FULLY':
> +        if f["mask"] == "MFM_FULLY":
>              cidr_protocols = protocols.copy()
>              bitwise_protocols = protocols.copy()
>
> -            if of10 == 'exact match':
> -                bitwise_protocols -= set(['of10'])
> -                cidr_protocols -= set(['of10'])
> -            elif of10 == 'CIDR mask':
> -                bitwise_protocols -= set(['of10'])
> +            if of10 == "exact match":
> +                bitwise_protocols -= set(["of10"])
> +                cidr_protocols -= set(["of10"])
> +            elif of10 == "CIDR mask":
> +                bitwise_protocols -= set(["of10"])
>              else:
>                  assert of10 is None
>
> -            if of11 == 'exact match':
> -                bitwise_protocols -= set(['of11'])
> -                cidr_protocols -= set(['of11'])
> +            if of11 == "exact match":
> +                bitwise_protocols -= set(["of11"])
> +                cidr_protocols -= set(["of11"])
>              else:
> -                assert of11 in (None, 'bitwise mask')
> +                assert of11 in (None, "bitwise mask")
>          else:
> -            assert f['mask'] == 'MFM_NONE'
> +            assert f["mask"] == "MFM_NONE"
>              cidr_protocols = set([])
>              bitwise_protocols = set([])
>
> @@ -393,8 +140,8 @@ def make_meta_flow(meta_flow_h):
>          output += ["    %s," % protocols_to_c(cidr_protocols)]
>          output += ["    %s," % protocols_to_c(bitwise_protocols)]
>
> -        if f['prefix']:
> -            output += ["    FLOW_U32OFS(%s)," % f['prefix']]
> +        if f["prefix"]:
> +            output += ["    FLOW_U32OFS(%s)," % f["prefix"]]
>          else:
>              output += ["    -1, /* not usable for prefix lookup */"]
>
> @@ -409,147 +156,37 @@ def make_nx_match(meta_flow_h):
>      print("static struct nxm_field_index all_nxm_fields[] = {")
>      for f in fields:
>          # Sort by OpenFlow version number (nx-match.c depends on this).
> -        for oxm in sorted(f['OXM'], key=lambda x: x[2]):
> -            header = ("NXM_HEADER(0x%x,0x%x,%s,0,%d)" % oxm[0])
> -            print("""{ .nf = { %s, %d, "%s", %s } },""" % (
> -                header, oxm[2], oxm[1], f['mff']))
> +        for oxm in sorted(f["OXM"], key=lambda x: x[2]):
> +            header = "NXM_HEADER(0x%x,0x%x,%s,0,%d)" % oxm[0]
> +            print(
> +                """{ .nf = { %s, %d, "%s", %s } },"""
> +                % (header, oxm[2], oxm[1], f["mff"])
> +            )
>      print("};")
>      for oline in output:
>          print(oline)
>
>
> -def extract_ofp_fields(fn):
> -    global file_name
> -    global input_file
> -    global line_number
> -    global line
> -
> -    file_name = fn
> -    input_file = open(file_name)
> -    line_number = 0
> -
> -    fields = []
> -
> -    while True:
> -        get_line()
> -        if re.match('enum.*mf_field_id', line):
> -            break
> -
> -    while True:
> -        get_line()
> -        first_line_number = line_number
> -        here = '%s:%d' % (file_name, line_number)
> -        if (line.startswith('/*')
> -            or line.startswith(' *')
> -            or line.startswith('#')
> -            or not line
> -            or line.isspace()):
> -            continue
> -        elif re.match('}', line) or re.match('\s+MFF_N_IDS', line):
> -            break
> -
> -        # Parse the comment preceding an MFF_ constant into 'comment',
> -        # one line to an array element.
> -        line = line.strip()
> -        if not line.startswith('/*'):
> -            fatal("unexpected syntax between fields")
> -        line = line[1:]
> -        comment = []
> -        end = False
> -        while not end:
> -            line = line.strip()
> -            if line.startswith('*/'):
> -                get_line()
> -                break
> -            if not line.startswith('*'):
> -                fatal("unexpected syntax within field")
> -
> -            line = line[1:]
> -            if line.startswith(' '):
> -                line = line[1:]
> -            if line.startswith(' ') and comment:
> -                continuation = True
> -                line = line.lstrip()
> -            else:
> -                continuation = False
> -
> -            if line.endswith('*/'):
> -                line = line[:-2].rstrip()
> -                end = True
> -            else:
> -                end = False
> -
> -            if continuation:
> -                comment[-1] += " " + line
> -            else:
> -                comment += [line]
> -            get_line()
> -
> -        # Drop blank lines at each end of comment.
> -        while comment and not comment[0]:
> -            comment = comment[1:]
> -        while comment and not comment[-1]:
> -            comment = comment[:-1]
> -
> -        # Parse the MFF_ constant(s).
> -        mffs = []
> -        while True:
> -            m = re.match('\s+(MFF_[A-Z0-9_]+),?\s?$', line)
> -            if not m:
> -                break
> -            mffs += [m.group(1)]
> -            get_line()
> -        if not mffs:
> -            fatal("unexpected syntax looking for MFF_ constants")
> -
> -        if len(mffs) > 1 or '<N>' in comment[0]:
> -            for mff in mffs:
> -                # Extract trailing integer.
> -                m = re.match('.*[^0-9]([0-9]+)$', mff)
> -                if not m:
> -                    fatal("%s lacks numeric suffix in register group" % mff)
> -                n = m.group(1)
> -
> -                # Search-and-replace <N> within the comment,
> -                # and drop lines that have <x> for x != n.
> -                instance = []
> -                for x in comment:
> -                    y = x.replace('<N>', n)
> -                    if re.search('<[0-9]+>', y):
> -                        if ('<%s>' % n) not in y:
> -                            continue
> -                        y = re.sub('<[0-9]+>', '', y)
> -                    instance += [y.strip()]
> -                fields += [parse_field(mff, instance)]
> -        else:
> -            fields += [parse_field(mffs[0], comment)]
> -        continue
> -
> -    input_file.close()
> -
> -    if n_errors:
> -        sys.exit(1)
> -
> -    return fields
> -
>  ## ------------------------ ##
>  ## Documentation Generation ##
>  ## ------------------------ ##
>
> +
>  def field_to_xml(field_node, f, body, summary):
>      f["used"] = True
>
>      # Summary.
> -    if field_node.hasAttribute('internal'):
> +    if field_node.hasAttribute("internal"):
>          return
>
>      min_of_version = None
>      min_ovs_version = None
> -    for header, name, of_version_nr, ovs_version_s in f['OXM']:
> -        if (is_standard_oxm(name)
> -            and (min_ovs_version is None or of_version_nr < min_of_version)):
> +    for header, name, of_version_nr, ovs_version_s in f["OXM"]:
> +        if is_standard_oxm(name) and (
> +            min_ovs_version is None or of_version_nr < min_of_version
> +        ):
>              min_of_version = of_version_nr
> -        ovs_version = [int(x) for x in ovs_version_s.split('.')]
> +        ovs_version = [int(x) for x in ovs_version_s.split(".")]
>          if min_ovs_version is None or ovs_version < min_ovs_version:
>              min_ovs_version = ovs_version
>      summary += ["\\fB%s\\fR" % f["name"]]
> @@ -565,124 +202,152 @@ def field_to_xml(field_node, f, body, summary):
>      if min_of_version is not None:
>          support += ["OF %s+" % VERSION_REVERSE[min_of_version]]
>      if min_ovs_version is not None:
> -        support += ["OVS %s+" % '.'.join([str(x) for x in min_ovs_version])]
> -    summary += ' and '.join(support)
> +        support += ["OVS %s+" % ".".join([str(x) for x in min_ovs_version])]
> +    summary += " and ".join(support)
>      summary += ["\n"]
>
>      # Full description.
> -    if field_node.hasAttribute('hidden'):
> +    if field_node.hasAttribute("hidden"):
>          return
>
> -    title = field_node.attributes['title'].nodeValue
> +    title = field_node.attributes["title"].nodeValue
>
> -    body += [""".PP
> +    body += [
> +        """.PP
>  \\fB%s Field\\fR
>  .TS
>  tab(;);
>  l lx.
> -""" % title]
> +"""
> +        % title
> +    ]
>
>      body += ["Name:;\\fB%s\\fR" % f["name"]]
>      if f["extra_name"]:
>          body += [" (aka \\fB%s\\fR)" % f["extra_name"]]
> -    body += ['\n']
> +    body += ["\n"]
>
>      body += ["Width:;"]
>      if f["n_bits"] != 8 * f["n_bytes"]:
> -        body += ["%d bits (only the least-significant %d bits "
> -                 "may be nonzero)" % (f["n_bytes"] * 8, f["n_bits"])]
> +        body += [
> +            "%d bits (only the least-significant %d bits "
> +            "may be nonzero)" % (f["n_bytes"] * 8, f["n_bits"])
> +        ]
>      elif f["n_bits"] <= 128:
>          body += ["%d bits" % f["n_bits"]]
>      else:
>          body += ["%d bits (%d bytes)" % (f["n_bits"], f["n_bits"] / 8)]
> -    body += ['\n']
> +    body += ["\n"]
>
>      body += ["Format:;%s\n" % f["formatting"]]
>
> -    masks = {"MFM_NONE": "not maskable",
> -             "MFM_FULLY": "arbitrary bitwise masks"}
> +    masks = {
> +        "MFM_NONE": "not maskable",
> +        "MFM_FULLY": "arbitrary bitwise masks",
> +    }
>      body += ["Masking:;%s\n" % masks[f["mask"]]]
>      body += ["Prerequisites:;%s\n" % f["prereqs"]]
>
> -    access = {True: "read/write",
> -              False: "read-only"}[f["writable"]]
> +    access = {True: "read/write", False: "read-only"}[f["writable"]]
>      body += ["Access:;%s\n" % access]
>
> -    of10 = {None: "not supported",
> -            "exact match": "yes (exact match only)",
> -            "CIDR mask": "yes (CIDR match only)"}
> +    of10 = {
> +        None: "not supported",
> +        "exact match": "yes (exact match only)",
> +        "CIDR mask": "yes (CIDR match only)",
> +    }
>      body += ["OpenFlow 1.0:;%s\n" % of10[f["OF1.0"]]]
>
> -    of11 = {None: "not supported",
> -            "exact match": "yes (exact match only)",
> -            "bitwise mask": "yes"}
> +    of11 = {
> +        None: "not supported",
> +        "exact match": "yes (exact match only)",
> +        "bitwise mask": "yes",
> +    }
>      body += ["OpenFlow 1.1:;%s\n" % of11[f["OF1.1"]]]
>
>      oxms = []
> -    for header, name, of_version_nr, ovs_version in [x for x in 
> sorted(f['OXM'], key=lambda x: x[2]) if is_standard_oxm(x[1])]:
> +    for header, name, of_version_nr, ovs_version in [
> +        x
> +        for x in sorted(f["OXM"], key=lambda x: x[2])
> +        if is_standard_oxm(x[1])
> +    ]:
>          of_version = VERSION_REVERSE[of_version_nr]
> -        oxms += [r"\fB%s\fR (%d) since OpenFlow %s and Open vSwitch %s" % 
> (name, header[2], of_version, ovs_version)]
> +        oxms += [
> +            r"\fB%s\fR (%d) since OpenFlow %s and Open vSwitch %s"
> +            % (name, header[2], of_version, ovs_version)
> +        ]
>      if not oxms:
> -        oxms = ['none']
> -    body += ['OXM:;T{\n%s\nT}\n' % r'\[char59] '.join(oxms)]
> +        oxms = ["none"]
> +    body += ["OXM:;T{\n%s\nT}\n" % r"\[char59] ".join(oxms)]
>
>      nxms = []
> -    for header, name, of_version_nr, ovs_version in [x for x in 
> sorted(f['OXM'], key=lambda x: x[2]) if not is_standard_oxm(x[1])]:
> -        nxms += [r"\fB%s\fR (%d) since Open vSwitch %s" % (name, header[2], 
> ovs_version)]
> +    for header, name, of_version_nr, ovs_version in [
> +        x
> +        for x in sorted(f["OXM"], key=lambda x: x[2])
> +        if not is_standard_oxm(x[1])
> +    ]:
> +        nxms += [
> +            r"\fB%s\fR (%d) since Open vSwitch %s"
> +            % (name, header[2], ovs_version)
> +        ]
>      if not nxms:
> -        nxms = ['none']
> -    body += ['NXM:;T{\n%s\nT}\n' % r'\[char59] '.join(nxms)]
> +        nxms = ["none"]
> +    body += ["NXM:;T{\n%s\nT}\n" % r"\[char59] ".join(nxms)]
>
>      body += [".TE\n"]
>
> -    body += ['.PP\n']
> +    body += [".PP\n"]
>      body += [build.nroff.block_xml_to_nroff(field_node.childNodes)]
>
> +
>  def group_xml_to_nroff(group_node, fields):
> -    title = group_node.attributes['title'].nodeValue
> +    title = group_node.attributes["title"].nodeValue
>
>      summary = []
>      body = []
>      for node in group_node.childNodes:
> -        if node.nodeType == node.ELEMENT_NODE and node.tagName == 'field':
> -            id_ = node.attributes['id'].nodeValue
> +        if node.nodeType == node.ELEMENT_NODE and node.tagName == "field":
> +            id_ = node.attributes["id"].nodeValue
>              field_to_xml(node, fields[id_], body, summary)
>          else:
>              body += [build.nroff.block_xml_to_nroff([node])]
>
>      content = [
> -        '.bp\n',
> -        '.SH \"%s\"\n' % build.nroff.text_to_nroff(title.upper() + " 
> FIELDS"),
> +        ".bp\n",
> +        '.SH "%s"\n' % build.nroff.text_to_nroff(title.upper() + " FIELDS"),
>          '.SS "Summary:"\n',
> -        '.TS\n',
> -        'tab(;);\n',
> -        'l l l l l l l.\n',
> -        'Name;Bytes;Mask;RW?;Prereqs;NXM/OXM Support\n',
> -        '\_;\_;\_;\_;\_;\_\n']
> +        ".TS\n",
> +        "tab(;);\n",
> +        "l l l l l l l.\n",
> +        "Name;Bytes;Mask;RW?;Prereqs;NXM/OXM Support\n",
> +        "\_;\_;\_;\_;\_;\_\n",
> +    ]
>      content += summary
> -    content += ['.TE\n']
> +    content += [".TE\n"]
>      content += body
> -    return ''.join(content)
> +    return "".join(content)
> +
>
>  def make_oxm_classes_xml(document):
> -    s = '''tab(;);
> +    s = """tab(;);
>  l l l.
>  Prefix;Vendor;Class
>  \_;\_;\_
> -'''
> +"""
>      for key in sorted(OXM_CLASSES, key=OXM_CLASSES.get):
>          vendor, class_, class_type = OXM_CLASSES.get(key)
> -        s += r"\fB%s\fR;" % key.rstrip('_')
> +        s += r"\fB%s\fR;" % key.rstrip("_")
>          if vendor:
>              s += r"\fL0x%08x\fR;" % vendor
>          else:
>              s += "(none);"
>          s += r"\fL0x%04x\fR;" % class_
>          s += "\n"
> -    e = document.createElement('tbl')
> +    e = document.createElement("tbl")
>      e.appendChild(document.createTextNode(s))
>      return e
>
> +
>  def recursively_replace(node, name, replacement):
>      for child in node.childNodes:
>          if child.nodeType == node.ELEMENT_NODE:
> @@ -691,11 +356,12 @@ def recursively_replace(node, name, replacement):
>              else:
>                  recursively_replace(child, name, replacement)
>
> +
>  def make_ovs_fields(meta_flow_h, meta_flow_xml):
>      fields = extract_ofp_fields(meta_flow_h)
>      fields_map = {}
>      for f in fields:
> -        fields_map[f['mff']] = f
> +        fields_map[f["mff"]] = f
>
>      document = xml.dom.minidom.parse(meta_flow_xml)
>      doc = document.documentElement
> @@ -704,7 +370,8 @@ def make_ovs_fields(meta_flow_h, meta_flow_xml):
>      if version == None:
>          version = "UNKNOWN"
>
> -    print('''\
> +    print(
> +        """\
>  '\\" tp
>  .\\" -*- mode: troff; coding: utf-8 -*-
>  .TH "ovs\-fields" 7 "%s" "Open vSwitch" "Open vSwitch Manual"
> @@ -740,11 +407,13 @@ def make_ovs_fields(meta_flow_h, meta_flow_xml):
>  ovs\-fields \- protocol header fields in OpenFlow and Open vSwitch
>  .
>  .PP
> -''' % version)
> +"""
> +        % version
> +    )
>
> -    recursively_replace(doc, 'oxm_classes', make_oxm_classes_xml(document))
> +    recursively_replace(doc, "oxm_classes", make_oxm_classes_xml(document))
>
> -    s = ''
> +    s = ""
>      for node in doc.childNodes:
>          if node.nodeType == node.ELEMENT_NODE and node.tagName == "group":
>              s += group_xml_to_nroff(node, fields_map)
> @@ -757,9 +426,10 @@ ovs\-fields \- protocol header fields in OpenFlow and 
> Open vSwitch
>
>      for f in fields:
>          if "used" not in f:
> -            fatal("%s: field not documented "
> -                  "(please add documentation in lib/meta-flow.xml)"
> -                  % f["mff"])
> +            fatal(
> +                "%s: field not documented "
> +                "(please add documentation in lib/meta-flow.xml)" % f["mff"]
> +            )
>      if n_errors:
>          sys.exit(1)
>
> @@ -769,26 +439,27 @@ ovs\-fields \- protocol header fields in OpenFlow and 
> Open vSwitch
>
>          # Life is easier with nroff if we don't try to feed it Unicode.
>          # Fortunately, we only use a few characters outside the ASCII range.
> -        oline = oline.replace(u'\u2208', r'\[mo]')
> -        oline = oline.replace(u'\u2260', r'\[!=]')
> -        oline = oline.replace(u'\u2264', r'\[<=]')
> -        oline = oline.replace(u'\u2265', r'\[>=]')
> -        oline = oline.replace(u'\u00d7', r'\[mu]')
> +        oline = oline.replace(u"\u2208", r"\[mo]")
> +        oline = oline.replace(u"\u2260", r"\[!=]")
> +        oline = oline.replace(u"\u2264", r"\[<=]")
> +        oline = oline.replace(u"\u2265", r"\[>=]")
> +        oline = oline.replace(u"\u00d7", r"\[mu]")
>          if len(oline):
>              output += [oline]
>
>      # nroff tends to ignore .bp requests if they come after .PP requests,
>      # so remove .PPs that precede .bp.
>      for i in range(len(output)):
> -        if output[i] == '.bp':
> +        if output[i] == ".bp":
>              j = i - 1
> -            while j >= 0 and output[j] == '.PP':
> +            while j >= 0 and output[j] == ".PP":
>                  output[j] = None
>                  j -= 1
>      for i in range(len(output)):
>          if output[i] is not None:
>              print(output[i])
> -
> +
> +
>  ## ------------ ##
>  ## Main Program ##
>  ## ------------ ##
> @@ -796,8 +467,9 @@ ovs\-fields \- protocol header fields in OpenFlow and 
> Open vSwitch
>  if __name__ == "__main__":
>      argv0 = sys.argv[0]
>      try:
> -        options, args = getopt.gnu_getopt(sys.argv[1:], 'h',
> -                                          ['help', 'ovs-version='])
> +        options, args = getopt.gnu_getopt(
> +            sys.argv[1:], "h", ["help", "ovs-version="]
> +        )
>      except getopt.GetoptError as geo:
>          sys.stderr.write("%s: %s\n" % (argv0, geo.msg))
>          sys.exit(1)
> @@ -805,32 +477,38 @@ if __name__ == "__main__":
>      global version
>      version = None
>      for key, value in options:
> -        if key in ['-h', '--help']:
> +        if key in ["-h", "--help"]:
>              usage()
> -        elif key == '--ovs-version':
> +        elif key == "--ovs-version":
>              version = value
>          else:
>              sys.exit(0)
>
>      if not args:
> -        sys.stderr.write("%s: missing command argument "
> -                         "(use --help for help)\n" % argv0)
> +        sys.stderr.write(
> +            "%s: missing command argument " "(use --help for help)\n" % argv0
> +        )
>          sys.exit(1)
>
> -    commands = {"meta-flow": (make_meta_flow, 1),
> -                "nx-match": (make_nx_match, 1),
> -                "ovs-fields": (make_ovs_fields, 2)}
> +    commands = {
> +        "meta-flow": (make_meta_flow, 1),
> +        "nx-match": (make_nx_match, 1),
> +        "ovs-fields": (make_ovs_fields, 2),
> +    }
>
>      if not args[0] in commands:
> -        sys.stderr.write("%s: unknown command \"%s\" "
> -                         "(use --help for help)\n" % (argv0, args[0]))
> +        sys.stderr.write(
> +            '%s: unknown command "%s" '
> +            "(use --help for help)\n" % (argv0, args[0])
> +        )
>          sys.exit(1)
>
>      func, n_args = commands[args[0]]
>      if len(args) - 1 != n_args:
> -        sys.stderr.write("%s: \"%s\" requires %d arguments but %d "
> -                         "provided\n"
> -                         % (argv0, args[0], n_args, len(args) - 1))
> +        sys.stderr.write(
> +            '%s: "%s" requires %d arguments but %d '
> +            "provided\n" % (argv0, args[0], n_args, len(args) - 1)
> +        )
>          sys.exit(1)
>
>      func(*args[1:])
> diff --git a/python/automake.mk b/python/automake.mk
> index 73438d615..54c2321a9 100644
> --- a/python/automake.mk
> +++ b/python/automake.mk
> @@ -51,6 +51,7 @@ ovs_pyfiles = \
>  # so they are not installed.
>  EXTRA_DIST += \
>       python/build/__init__.py \
> +     python/build/extract_ofp_fields.py \
>       python/build/nroff.py \
>       python/build/soutil.py
>
> diff --git a/python/build/extract_ofp_fields.py 
> b/python/build/extract_ofp_fields.py


As suggested in my previous review, I think we should add it to FLAKE8_PYFILES, 
so errors flake errors get caught:


diff --git a/python/automake.mk b/python/automake.mk
index fac32ff2b..a5151338d 100644
--- a/python/automake.mk
+++ b/python/automake.mk
@@ -86,11 +86,12 @@ PYCOV_CLEAN_FILES += $(PYFILES:.py=.py,cover)

 FLAKE8_PYFILES += \
        $(filter-out python/ovs/compat/% python/ovs/dirs.py,$(PYFILES)) \
-       python/setup.py \
        python/build/__init__.py \
+       python/build/extract_ofp_fields.py
        python/build/flow-parse-deps.py \
-       python/build/nroff.py \
-       python/ovs/dirs.py.template
+       python/build/soutil.py \
+       python/ovs/dirs.py.template \
+       python/setup.py \


Guess the soutil.py should also be added!


With this I get the following errors:

python/build/extract_ofp_fields.py:1:1: F401 'getopt' imported but unused
python/build/extract_ofp_fields.py:3:1: F401 'os.path' imported but unused
python/build/extract_ofp_fields.py:5:1: F401 'xml.dom.minidom' imported but 
unused
python/build/extract_ofp_fields.py:6:1: F401 'build.nroff' imported but unused
python/build/extract_ofp_fields.py:17:26: E231 missing whitespace after ','
python/build/extract_ofp_fields.py:19:15: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:19:25: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:20:17: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:20:25: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:21:17: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:21:25: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:22:16: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:22:25: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:23:17: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:23:25: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:24:18: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:24:26: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:27:25: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:27:52: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:27:60: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:28:29: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:28:56: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:29:26: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:29:53: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:29:60: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:30:26: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:30:53: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:30:60: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:31:22: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:31:49: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:31:60: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:32:22: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:32:49: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:32:60: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:33:35: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:33:53: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:33:60: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:34:60: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:35:22: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:35:49: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:35:60: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:36:30: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:36:54: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:36:60: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:37:27: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:37:54: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:37:60: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:38:29: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:38:56: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:38:60: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:66:26: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:66:37: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:67:26: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:67:37: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:68:29: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:69:26: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:69:37: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:70:37: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:71:29: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:72:30: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:72:37: E241 multiple spaces after ','
python/build/extract_ofp_fields.py:79:28: E241 multiple spaces after ':'
python/build/extract_ofp_fields.py:81:1: E302 expected 2 blank lines, found 1
python/build/extract_ofp_fields.py:118:1: E302 expected 2 blank lines, found 1
python/build/extract_ofp_fields.py:131:31: W605 invalid escape sequence '\('
python/build/extract_ofp_fields.py:131:41: W605 invalid escape sequence '\)'
python/build/extract_ofp_fields.py:131:57: W605 invalid escape sequence '\.'
python/build/extract_ofp_fields.py:131:79: W605 invalid escape sequence '\.'
python/build/extract_ofp_fields.py:131:80: E501 line too long (93 > 79 
characters)
python/build/extract_ofp_fields.py:178:80: E501 line too long (83 > 79 
characters)
python/build/extract_ofp_fields.py:274:1: E302 expected 2 blank lines, found 1
python/build/extract_ofp_fields.py:293:9: F841 local variable 
'first_line_number' is assigned to but never used
python/build/extract_ofp_fields.py:294:9: F841 local variable 'here' is 
assigned to but never used
python/build/extract_ofp_fields.py:301:47: W605 invalid escape sequence '\s'
python/build/extract_ofp_fields.py:350:27: W605 invalid escape sequence '\s'
python/build/extract_ofp_fields.py:350:48: W605 invalid escape sequence '\s'
make[2]: [Makefile:6758: flake8-check] Error 1 (ignored)

> new file mode 100644
> index 000000000..f6938b6dd
> --- /dev/null
> +++ b/python/build/extract_ofp_fields.py
> @@ -0,0 +1,386 @@
> +import getopt
> +import sys
> +import os.path
> +import re
> +import xml.dom.minidom
> +import build.nroff
> +
> +line = ""
> +
> +# Maps from user-friendly version number to its protocol encoding.
> +VERSION = {"1.0": 0x01,
> +           "1.1": 0x02,
> +           "1.2": 0x03,
> +           "1.3": 0x04,
> +           "1.4": 0x05,
> +           "1.5": 0x06}
> +VERSION_REVERSE = dict((v,k) for k, v in VERSION.items())
> +
> +TYPES = {"u8":       (1,   False),
> +         "be16":     (2,   False),
> +         "be32":     (4,   False),
> +         "MAC":      (6,   False),
> +         "be64":     (8,   False),
> +         "be128":    (16,  False),
> +         "tunnelMD": (124, True)}
> +
> +FORMATTING = {"decimal":            ("MFS_DECIMAL",      1,   8),
> +              "hexadecimal":        ("MFS_HEXADECIMAL",  1, 127),
> +              "ct state":           ("MFS_CT_STATE",     4,   4),
> +              "Ethernet":           ("MFS_ETHERNET",     6,   6),
> +              "IPv4":               ("MFS_IPV4",         4,   4),
> +              "IPv6":               ("MFS_IPV6",        16,  16),
> +              "OpenFlow 1.0 port":  ("MFS_OFP_PORT",     2,   2),
> +              "OpenFlow 1.1+ port": ("MFS_OFP_PORT_OXM", 4,   4),
> +              "frag":               ("MFS_FRAG",         1,   1),
> +              "tunnel flags":       ("MFS_TNL_FLAGS",    2,   2),
> +              "TCP flags":          ("MFS_TCP_FLAGS",    2,   2),
> +              "packet type":        ("MFS_PACKET_TYPE",  4,   4)}
> +
> +PREREQS = {"none": "MFP_NONE",
> +           "Ethernet": "MFP_ETHERNET",
> +           "ARP": "MFP_ARP",
> +           "VLAN VID": "MFP_VLAN_VID",
> +           "IPv4": "MFP_IPV4",
> +           "IPv6": "MFP_IPV6",
> +           "IPv4/IPv6": "MFP_IP_ANY",
> +           "NSH": "MFP_NSH",
> +           "CT": "MFP_CT_VALID",
> +           "MPLS": "MFP_MPLS",
> +           "TCP": "MFP_TCP",
> +           "UDP": "MFP_UDP",
> +           "SCTP": "MFP_SCTP",
> +           "ICMPv4": "MFP_ICMPV4",
> +           "ICMPv6": "MFP_ICMPV6",
> +           "ND": "MFP_ND",
> +           "ND solicit": "MFP_ND_SOLICIT",
> +           "ND advert": "MFP_ND_ADVERT"}
> +
> +# Maps a name prefix into an (experimenter ID, class) pair, so:
> +#
> +#      - Standard OXM classes are written as (0, <oxm_class>)
> +#
> +#      - Experimenter OXM classes are written as (<oxm_vender>, 0xffff)
> +#
> +# If a name matches more than one prefix, the longest one is used.
> +OXM_CLASSES = {"NXM_OF_":        (0,          0x0000, 'extension'),
> +               "NXM_NX_":        (0,          0x0001, 'extension'),
> +               "NXOXM_NSH_":     (0x005ad650, 0xffff, 'extension'),
> +               "OXM_OF_":        (0,          0x8000, 'standard'),
> +               "OXM_OF_PKT_REG": (0,          0x8001, 'standard'),
> +               "ONFOXM_ET_":     (0x4f4e4600, 0xffff, 'standard'),
> +               "ERICOXM_OF_":    (0,          0x1000, 'extension'),
> +
> +               # This is the experimenter OXM class for Nicira, which is the
> +               # one that OVS would be using instead of NXM_OF_ and NXM_NX_
> +               # if OVS didn't have those grandfathered in.  It is currently
> +               # used only to test support for experimenter OXM, since there
> +               # are barely any real uses of experimenter OXM in the wild.
> +               "NXOXM_ET_":      (0x00002320, 0xffff, 'extension')}
> +
> +def oxm_name_to_class(name):
> +    prefix = ''
> +    class_ = None
> +    for p, c in OXM_CLASSES.items():
> +        if name.startswith(p) and len(p) > len(prefix):
> +            prefix = p
> +            class_ = c
> +    return class_
> +
> +
> +def is_standard_oxm(name):
> +    oxm_vendor, oxm_class, oxm_class_type = oxm_name_to_class(name)
> +    return oxm_class_type == 'standard'
> +
> +
> +def get_line():
> +    global line
> +    global line_number
> +    line = input_file.readline()
> +    line_number += 1
> +    if line == "":
> +        fatal("unexpected end of input")
> +
> +
> +n_errors = 0
> +
> +
> +def error(msg):
> +    global n_errors
> +    sys.stderr.write("%s:%d: %s\n" % (file_name, line_number, msg))
> +    n_errors += 1
> +
> +
> +def fatal(msg):
> +    error(msg)
> +    sys.exit(1)
> +
> +def parse_oxms(s, prefix, n_bytes):
> +    if s == 'none':
> +        return ()
> +
> +    return tuple(parse_oxm(s2.strip(), prefix, n_bytes) for s2 in 
> s.split(','))
> +
> +
> +match_types = dict()
> +
> +
> +def parse_oxm(s, prefix, n_bytes):
> +    global match_types
> +
> +    m = re.match('([A-Z0-9_]+)\(([0-9]+)\) since(?: OF(1\.[0-9]+) and)? 
> v([12]\.[0-9]+)$', s)
> +    if not m:
> +        fatal("%s: syntax error parsing %s" % (s, prefix))
> +
> +    name, oxm_type, of_version, ovs_version = m.groups()
> +
> +    class_ = oxm_name_to_class(name)
> +    if class_ is None:
> +        fatal("unknown OXM class for %s" % name)
> +    oxm_vendor, oxm_class, oxm_class_type = class_
> +
> +    if class_ in match_types:
> +        if oxm_type in match_types[class_]:
> +            fatal("duplicate match type for %s (conflicts with %s)" %
> +                  (name, match_types[class_][oxm_type]))
> +    else:
> +        match_types[class_] = dict()
> +    match_types[class_][oxm_type] = name
> +
> +    # Normally the oxm_length is the size of the field, but for experimenter
> +    # OXMs oxm_length also includes the 4-byte experimenter ID.
> +    oxm_length = n_bytes
> +    if oxm_class == 0xffff:
> +        oxm_length += 4
> +
> +    header = (oxm_vendor, oxm_class, int(oxm_type), oxm_length)
> +
> +    if of_version:
> +        if oxm_class_type == 'extension':
> +            fatal("%s: OXM extension can't have OpenFlow version" % name)
> +        if of_version not in VERSION:
> +            fatal("%s: unknown OpenFlow version %s" % (name, of_version))
> +        of_version_nr = VERSION[of_version]
> +        if of_version_nr < VERSION['1.2']:
> +            fatal("%s: claimed version %s predates OXM" % (name, of_version))
> +    else:
> +        if oxm_class_type == 'standard':
> +            fatal("%s: missing OpenFlow version number" % name)
> +        of_version_nr = 0
> +
> +    return (header, name, of_version_nr, ovs_version)
> +
> +
> +def parse_field(mff, comment):
> +    f = {'mff': mff}
> +
> +    # First line of comment is the field name.
> +    m = re.match(r'"([^"]+)"(?:\s+\(aka "([^"]+)"\))?(?:\s+\(.*\))?\.', 
> comment[0])
> +    if not m:
> +        fatal("%s lacks field name" % mff)
> +    f['name'], f['extra_name'] = m.groups()
> +
> +    # Find the last blank line the comment.  The field definitions
> +    # start after that.
> +    blank = None
> +    for i in range(len(comment)):
> +        if not comment[i]:
> +            blank = i
> +    if not blank:
> +        fatal("%s: missing blank line in comment" % mff)
> +
> +    d = {}
> +    for key in ("Type", "Maskable", "Formatting", "Prerequisites",
> +                "Access", "Prefix lookup member",
> +                "OXM", "NXM", "OF1.0", "OF1.1"):
> +        d[key] = None
> +    for fline in comment[blank + 1:]:
> +        m = re.match(r'([^:]+):\s+(.*)\.$', fline)
> +        if not m:
> +            fatal("%s: syntax error parsing key-value pair as part of %s"
> +                  % (fline, mff))
> +        key, value = m.groups()
> +        if key not in d:
> +            fatal("%s: unknown key" % key)
> +        elif key == 'Code point':
> +            d[key] += [value]
> +        elif d[key] is not None:
> +            fatal("%s: duplicate key" % key)
> +        d[key] = value
> +    for key, value in d.items():
> +        if not value and key not in ("OF1.0", "OF1.1",
> +                                     "Prefix lookup member", "Notes"):
> +            fatal("%s: missing %s" % (mff, key))
> +
> +    m = re.match(r'([a-zA-Z0-9]+)(?: \(low ([0-9]+) bits\))?$', d['Type'])
> +    if not m:
> +        fatal("%s: syntax error in type" % mff)
> +    type_ = m.group(1)
> +    if type_ not in TYPES:
> +        fatal("%s: unknown type %s" % (mff, d['Type']))
> +
> +    f['n_bytes'] = TYPES[type_][0]
> +    if m.group(2):
> +        f['n_bits'] = int(m.group(2))
> +        if f['n_bits'] > f['n_bytes'] * 8:
> +            fatal("%s: more bits (%d) than field size (%d)"
> +                  % (mff, f['n_bits'], 8 * f['n_bytes']))
> +    else:
> +        f['n_bits'] = 8 * f['n_bytes']
> +    f['variable'] = TYPES[type_][1]
> +
> +    if d['Maskable'] == 'no':
> +        f['mask'] = 'MFM_NONE'
> +    elif d['Maskable'] == 'bitwise':
> +        f['mask'] = 'MFM_FULLY'
> +    else:
> +        fatal("%s: unknown maskable %s" % (mff, d['Maskable']))
> +
> +    fmt = FORMATTING.get(d['Formatting'])
> +    if not fmt:
> +        fatal("%s: unknown format %s" % (mff, d['Formatting']))
> +    f['formatting'] = d['Formatting']
> +    if f['n_bytes'] < fmt[1] or f['n_bytes'] > fmt[2]:
> +        fatal("%s: %d-byte field can't be formatted as %s"
> +              % (mff, f['n_bytes'], d['Formatting']))
> +    f['string'] = fmt[0]
> +
> +    f['prereqs'] = d['Prerequisites']
> +    if f['prereqs'] not in PREREQS:
> +        fatal("%s: unknown prerequisites %s" % (mff, d['Prerequisites']))
> +
> +    if d['Access'] == 'read-only':
> +        f['writable'] = False
> +    elif d['Access'] == 'read/write':
> +        f['writable'] = True
> +    else:
> +        fatal("%s: unknown access %s" % (mff, d['Access']))
> +
> +    f['OF1.0'] = d['OF1.0']
> +    if not d['OF1.0'] in (None, 'exact match', 'CIDR mask'):
> +        fatal("%s: unknown OF1.0 match type %s" % (mff, d['OF1.0']))
> +
> +    f['OF1.1'] = d['OF1.1']
> +    if not d['OF1.1'] in (None, 'exact match', 'bitwise mask'):
> +        fatal("%s: unknown OF1.1 match type %s" % (mff, d['OF1.1']))
> +
> +    f['OXM'] = (parse_oxms(d['OXM'], 'OXM', f['n_bytes']) +
> +                parse_oxms(d['NXM'], 'NXM', f['n_bytes']))
> +
> +    f['prefix'] = d["Prefix lookup member"]
> +
> +    return f
> +
> +def extract_ofp_fields(fn):
> +    global file_name
> +    global input_file
> +    global line_number
> +    global line
> +
> +    file_name = fn
> +    input_file = open(file_name)
> +    line_number = 0
> +
> +    fields = []
> +
> +    while True:
> +        get_line()
> +        if re.match('enum.*mf_field_id', line):
> +            break
> +
> +    while True:
> +        get_line()
> +        first_line_number = line_number
> +        here = '%s:%d' % (file_name, line_number)
> +        if (line.startswith('/*')
> +            or line.startswith(' *')
> +            or line.startswith('#')
> +            or not line
> +            or line.isspace()):
> +            continue
> +        elif re.match('}', line) or re.match('\s+MFF_N_IDS', line):
> +            break
> +
> +        # Parse the comment preceding an MFF_ constant into 'comment',
> +        # one line to an array element.
> +        line = line.strip()
> +        if not line.startswith('/*'):
> +            fatal("unexpected syntax between fields")
> +        line = line[1:]
> +        comment = []
> +        end = False
> +        while not end:
> +            line = line.strip()
> +            if line.startswith('*/'):
> +                get_line()
> +                break
> +            if not line.startswith('*'):
> +                fatal("unexpected syntax within field")
> +
> +            line = line[1:]
> +            if line.startswith(' '):
> +                line = line[1:]
> +            if line.startswith(' ') and comment:
> +                continuation = True
> +                line = line.lstrip()
> +            else:
> +                continuation = False
> +
> +            if line.endswith('*/'):
> +                line = line[:-2].rstrip()
> +                end = True
> +            else:
> +                end = False
> +
> +            if continuation:
> +                comment[-1] += " " + line
> +            else:
> +                comment += [line]
> +            get_line()
> +
> +        # Drop blank lines at each end of comment.
> +        while comment and not comment[0]:
> +            comment = comment[1:]
> +        while comment and not comment[-1]:
> +            comment = comment[:-1]
> +
> +        # Parse the MFF_ constant(s).
> +        mffs = []
> +        while True:
> +            m = re.match('\s+(MFF_[A-Z0-9_]+),?\s?$', line)
> +            if not m:
> +                break
> +            mffs += [m.group(1)]
> +            get_line()
> +        if not mffs:
> +            fatal("unexpected syntax looking for MFF_ constants")
> +
> +        if len(mffs) > 1 or '<N>' in comment[0]:
> +            for mff in mffs:
> +                # Extract trailing integer.
> +                m = re.match('.*[^0-9]([0-9]+)$', mff)
> +                if not m:
> +                    fatal("%s lacks numeric suffix in register group" % mff)
> +                n = m.group(1)
> +
> +                # Search-and-replace <N> within the comment,
> +                # and drop lines that have <x> for x != n.
> +                instance = []
> +                for x in comment:
> +                    y = x.replace('<N>', n)
> +                    if re.search('<[0-9]+>', y):
> +                        if ('<%s>' % n) not in y:
> +                            continue
> +                        y = re.sub('<[0-9]+>', '', y)
> +                    instance += [y.strip()]
> +                fields += [parse_field(mff, instance)]
> +        else:
> +            fields += [parse_field(mffs[0], comment)]
> +        continue
> +
> +    input_file.close()
> +
> +    if n_errors:
> +        sys.exit(1)
> +
> +    return fields
> -- 
> 2.34.1

_______________________________________________
dev mailing list
d...@openvswitch.org
https://mail.openvswitch.org/mailman/listinfo/ovs-dev

Reply via email to