diff mbox series

[ovs-dev,v3,04/18] build-aux: split extract-ofp-fields

Message ID 20220311152128.3988946-5-amorenoz@redhat.com
State Changes Requested
Headers show
Series python: add flow parsing library | expand

Checks

Context Check Description
ovsrobot/apply-robot success apply and check: success
ovsrobot/intel-ovs-compilation success test: success
ovsrobot/github-robot-_Build_and_Test success github build: passed

Commit Message

Adrian Moreno March 11, 2022, 3:21 p.m. UTC
In order to be able to reuse the core extraction logic, split the command
in two parts. The core extraction logic is moved to python/build while
the command that writes the different files out of the extracted field
info is kept in build-aux.

Signed-off-by: Adrian Moreno <amorenoz@redhat.com>
---
 build-aux/extract-ofp-fields       | 706 ++++++++---------------------
 python/automake.mk                 |   7 +-
 python/build/extract_ofp_fields.py | 421 +++++++++++++++++
 3 files changed, 618 insertions(+), 516 deletions(-)
 create mode 100644 python/build/extract_ofp_fields.py

Comments

Eelco Chaudron March 18, 2022, 10:21 a.m. UTC | #1
On 11 Mar 2022, at 16:21, Adrian Moreno wrote:

> In order to be able to reuse the core extraction logic, split the command
> in two parts. The core extraction logic is moved to python/build while
> the command that writes the different files out of the extracted field
> info is kept in build-aux.
>
> Signed-off-by: Adrian Moreno <amorenoz@redhat.com>

Changes look good to me!

Acked-by: Eelco Chaudron <echaudro@redhat.com>
diff mbox series

Patch

diff --git a/build-aux/extract-ofp-fields b/build-aux/extract-ofp-fields
index 8766995d9..efec59c25 100755
--- a/build-aux/extract-ofp-fields
+++ b/build-aux/extract-ofp-fields
@@ -3,85 +3,23 @@ 
 import getopt
 import sys
 import os.path
-import re
 import xml.dom.minidom
 import build.nroff
 
-line = ""
-
-# Maps from user-friendly version number to its protocol encoding.
-VERSION = {"1.0": 0x01,
-           "1.1": 0x02,
-           "1.2": 0x03,
-           "1.3": 0x04,
-           "1.4": 0x05,
-           "1.5": 0x06}
-VERSION_REVERSE = dict((v,k) for k, v in VERSION.items())
-
-TYPES = {"u8":       (1,   False),
-         "be16":     (2,   False),
-         "be32":     (4,   False),
-         "MAC":      (6,   False),
-         "be64":     (8,   False),
-         "be128":    (16,  False),
-         "tunnelMD": (124, True)}
-
-FORMATTING = {"decimal":            ("MFS_DECIMAL",      1,   8),
-              "hexadecimal":        ("MFS_HEXADECIMAL",  1, 127),
-              "ct state":           ("MFS_CT_STATE",     4,   4),
-              "Ethernet":           ("MFS_ETHERNET",     6,   6),
-              "IPv4":               ("MFS_IPV4",         4,   4),
-              "IPv6":               ("MFS_IPV6",        16,  16),
-              "OpenFlow 1.0 port":  ("MFS_OFP_PORT",     2,   2),
-              "OpenFlow 1.1+ port": ("MFS_OFP_PORT_OXM", 4,   4),
-              "frag":               ("MFS_FRAG",         1,   1),
-              "tunnel flags":       ("MFS_TNL_FLAGS",    2,   2),
-              "TCP flags":          ("MFS_TCP_FLAGS",    2,   2),
-              "packet type":        ("MFS_PACKET_TYPE",  4,   4)}
-
-PREREQS = {"none": "MFP_NONE",
-           "Ethernet": "MFP_ETHERNET",
-           "ARP": "MFP_ARP",
-           "VLAN VID": "MFP_VLAN_VID",
-           "IPv4": "MFP_IPV4",
-           "IPv6": "MFP_IPV6",
-           "IPv4/IPv6": "MFP_IP_ANY",
-           "NSH": "MFP_NSH",
-           "CT": "MFP_CT_VALID",
-           "MPLS": "MFP_MPLS",
-           "TCP": "MFP_TCP",
-           "UDP": "MFP_UDP",
-           "SCTP": "MFP_SCTP",
-           "ICMPv4": "MFP_ICMPV4",
-           "ICMPv6": "MFP_ICMPV6",
-           "ND": "MFP_ND",
-           "ND solicit": "MFP_ND_SOLICIT",
-           "ND advert": "MFP_ND_ADVERT"}
-
-# Maps a name prefix into an (experimenter ID, class) pair, so:
-#
-#      - Standard OXM classes are written as (0, <oxm_class>)
-#
-#      - Experimenter OXM classes are written as (<oxm_vender>, 0xffff)
-#
-# If a name matches more than one prefix, the longest one is used.
-OXM_CLASSES = {"NXM_OF_":        (0,          0x0000, 'extension'),
-               "NXM_NX_":        (0,          0x0001, 'extension'),
-               "NXOXM_NSH_":     (0x005ad650, 0xffff, 'extension'),
-               "OXM_OF_":        (0,          0x8000, 'standard'),
-               "OXM_OF_PKT_REG": (0,          0x8001, 'standard'),
-               "ONFOXM_ET_":     (0x4f4e4600, 0xffff, 'standard'),
-               "ERICOXM_OF_":    (0,          0x1000, 'extension'),
-
-               # This is the experimenter OXM class for Nicira, which is the
-               # one that OVS would be using instead of NXM_OF_ and NXM_NX_
-               # if OVS didn't have those grandfathered in.  It is currently
-               # used only to test support for experimenter OXM, since there
-               # are barely any real uses of experimenter OXM in the wild.
-               "NXOXM_ET_":      (0x00002320, 0xffff, 'extension')}
+from build.extract_ofp_fields import (
+    extract_ofp_fields,
+    PREREQS,
+    OXM_CLASSES,
+    VERSION,
+    fatal,
+    n_errors,
+)
+
+VERSION_REVERSE = dict((v, k) for k, v in VERSION.items())
+
 
 def oxm_name_to_class(name):
-    prefix = ''
+    prefix = ""
     class_ = None
     for p, c in OXM_CLASSES.items():
         if name.startswith(p) and len(p) > len(prefix):
@@ -92,267 +30,76 @@  def oxm_name_to_class(name):
 
 def is_standard_oxm(name):
     oxm_vendor, oxm_class, oxm_class_type = oxm_name_to_class(name)
-    return oxm_class_type == 'standard'
-
-
-def decode_version_range(range):
-    if range in VERSION:
-        return (VERSION[range], VERSION[range])
-    elif range.endswith('+'):
-        return (VERSION[range[:-1]], max(VERSION.values()))
-    else:
-        a, b = re.match(r'^([^-]+)-([^-]+)$', range).groups()
-        return (VERSION[a], VERSION[b])
-
-
-def get_line():
-    global line
-    global line_number
-    line = input_file.readline()
-    line_number += 1
-    if line == "":
-        fatal("unexpected end of input")
-
-
-n_errors = 0
-
-
-def error(msg):
-    global n_errors
-    sys.stderr.write("%s:%d: %s\n" % (file_name, line_number, msg))
-    n_errors += 1
-
-
-def fatal(msg):
-    error(msg)
-    sys.exit(1)
+    return oxm_class_type == "standard"
 
 
 def usage():
     argv0 = os.path.basename(sys.argv[0])
-    print('''\
+    print(
+        """\
 %(argv0)s, for extracting OpenFlow field properties from meta-flow.h
 usage: %(argv0)s INPUT [--meta-flow | --nx-match]
   where INPUT points to lib/meta-flow.h in the source directory.
 Depending on the option given, the output written to stdout is intended to be
 saved either as lib/meta-flow.inc or lib/nx-match.inc for the respective C
 file to #include.\
-''' % {"argv0": argv0})
+"""
+        % {"argv0": argv0}
+    )
     sys.exit(0)
 
 
-def make_sizeof(s):
-    m = re.match(r'(.*) up to (.*)', s)
-    if m:
-        struct, member = m.groups()
-        return "offsetof(%s, %s)" % (struct, member)
-    else:
-        return "sizeof(%s)" % s
-
-
-def parse_oxms(s, prefix, n_bytes):
-    if s == 'none':
-        return ()
-
-    return tuple(parse_oxm(s2.strip(), prefix, n_bytes) for s2 in s.split(','))
-
-
-match_types = dict()
-
-
-def parse_oxm(s, prefix, n_bytes):
-    global match_types
-
-    m = re.match('([A-Z0-9_]+)\(([0-9]+)\) since(?: OF(1\.[0-9]+) and)? v([12]\.[0-9]+)$', s)
-    if not m:
-        fatal("%s: syntax error parsing %s" % (s, prefix))
-
-    name, oxm_type, of_version, ovs_version = m.groups()
-
-    class_ = oxm_name_to_class(name)
-    if class_ is None:
-        fatal("unknown OXM class for %s" % name)
-    oxm_vendor, oxm_class, oxm_class_type = class_
-
-    if class_ in match_types:
-        if oxm_type in match_types[class_]:
-            fatal("duplicate match type for %s (conflicts with %s)" %
-                  (name, match_types[class_][oxm_type]))
-    else:
-        match_types[class_] = dict()
-    match_types[class_][oxm_type] = name
-
-    # Normally the oxm_length is the size of the field, but for experimenter
-    # OXMs oxm_length also includes the 4-byte experimenter ID.
-    oxm_length = n_bytes
-    if oxm_class == 0xffff:
-        oxm_length += 4
-
-    header = (oxm_vendor, oxm_class, int(oxm_type), oxm_length)
-
-    if of_version:
-        if oxm_class_type == 'extension':
-            fatal("%s: OXM extension can't have OpenFlow version" % name)
-        if of_version not in VERSION:
-            fatal("%s: unknown OpenFlow version %s" % (name, of_version))
-        of_version_nr = VERSION[of_version]
-        if of_version_nr < VERSION['1.2']:
-            fatal("%s: claimed version %s predates OXM" % (name, of_version))
-    else:
-        if oxm_class_type == 'standard':
-            fatal("%s: missing OpenFlow version number" % name)
-        of_version_nr = 0
-
-    return (header, name, of_version_nr, ovs_version)
-
-
-def parse_field(mff, comment):
-    f = {'mff': mff}
-
-    # First line of comment is the field name.
-    m = re.match(r'"([^"]+)"(?:\s+\(aka "([^"]+)"\))?(?:\s+\(.*\))?\.', comment[0])
-    if not m:
-        fatal("%s lacks field name" % mff)
-    f['name'], f['extra_name'] = m.groups()
-
-    # Find the last blank line the comment.  The field definitions
-    # start after that.
-    blank = None
-    for i in range(len(comment)):
-        if not comment[i]:
-            blank = i
-    if not blank:
-        fatal("%s: missing blank line in comment" % mff)
-
-    d = {}
-    for key in ("Type", "Maskable", "Formatting", "Prerequisites",
-                "Access", "Prefix lookup member",
-                "OXM", "NXM", "OF1.0", "OF1.1"):
-        d[key] = None
-    for fline in comment[blank + 1:]:
-        m = re.match(r'([^:]+):\s+(.*)\.$', fline)
-        if not m:
-            fatal("%s: syntax error parsing key-value pair as part of %s"
-                  % (fline, mff))
-        key, value = m.groups()
-        if key not in d:
-            fatal("%s: unknown key" % key)
-        elif key == 'Code point':
-            d[key] += [value]
-        elif d[key] is not None:
-            fatal("%s: duplicate key" % key)
-        d[key] = value
-    for key, value in d.items():
-        if not value and key not in ("OF1.0", "OF1.1",
-                                     "Prefix lookup member", "Notes"):
-            fatal("%s: missing %s" % (mff, key))
-
-    m = re.match(r'([a-zA-Z0-9]+)(?: \(low ([0-9]+) bits\))?$', d['Type'])
-    if not m:
-        fatal("%s: syntax error in type" % mff)
-    type_ = m.group(1)
-    if type_ not in TYPES:
-        fatal("%s: unknown type %s" % (mff, d['Type']))
-
-    f['n_bytes'] = TYPES[type_][0]
-    if m.group(2):
-        f['n_bits'] = int(m.group(2))
-        if f['n_bits'] > f['n_bytes'] * 8:
-            fatal("%s: more bits (%d) than field size (%d)"
-                  % (mff, f['n_bits'], 8 * f['n_bytes']))
-    else:
-        f['n_bits'] = 8 * f['n_bytes']
-    f['variable'] = TYPES[type_][1]
-
-    if d['Maskable'] == 'no':
-        f['mask'] = 'MFM_NONE'
-    elif d['Maskable'] == 'bitwise':
-        f['mask'] = 'MFM_FULLY'
-    else:
-        fatal("%s: unknown maskable %s" % (mff, d['Maskable']))
-
-    fmt = FORMATTING.get(d['Formatting'])
-    if not fmt:
-        fatal("%s: unknown format %s" % (mff, d['Formatting']))
-    f['formatting'] = d['Formatting']
-    if f['n_bytes'] < fmt[1] or f['n_bytes'] > fmt[2]:
-        fatal("%s: %d-byte field can't be formatted as %s"
-              % (mff, f['n_bytes'], d['Formatting']))
-    f['string'] = fmt[0]
-
-    f['prereqs'] = d['Prerequisites']
-    if f['prereqs'] not in PREREQS:
-        fatal("%s: unknown prerequisites %s" % (mff, d['Prerequisites']))
-
-    if d['Access'] == 'read-only':
-        f['writable'] = False
-    elif d['Access'] == 'read/write':
-        f['writable'] = True
-    else:
-        fatal("%s: unknown access %s" % (mff, d['Access']))
-
-    f['OF1.0'] = d['OF1.0']
-    if not d['OF1.0'] in (None, 'exact match', 'CIDR mask'):
-        fatal("%s: unknown OF1.0 match type %s" % (mff, d['OF1.0']))
-
-    f['OF1.1'] = d['OF1.1']
-    if not d['OF1.1'] in (None, 'exact match', 'bitwise mask'):
-        fatal("%s: unknown OF1.1 match type %s" % (mff, d['OF1.1']))
-
-    f['OXM'] = (parse_oxms(d['OXM'], 'OXM', f['n_bytes']) +
-                parse_oxms(d['NXM'], 'NXM', f['n_bytes']))
-
-    f['prefix'] = d["Prefix lookup member"]
-
-    return f
-
-
 def protocols_to_c(protocols):
-    if protocols == set(['of10', 'of11', 'oxm']):
-        return 'OFPUTIL_P_ANY'
-    elif protocols == set(['of11', 'oxm']):
-        return 'OFPUTIL_P_NXM_OF11_UP'
-    elif protocols == set(['oxm']):
-        return 'OFPUTIL_P_NXM_OXM_ANY'
+    if protocols == set(["of10", "of11", "oxm"]):
+        return "OFPUTIL_P_ANY"
+    elif protocols == set(["of11", "oxm"]):
+        return "OFPUTIL_P_NXM_OF11_UP"
+    elif protocols == set(["oxm"]):
+        return "OFPUTIL_P_NXM_OXM_ANY"
     elif protocols == set([]):
-        return 'OFPUTIL_P_NONE'
+        return "OFPUTIL_P_NONE"
     else:
         assert False
 
 
 def autogen_c_comment():
     return [
-"/* Generated automatically; do not modify!    -*- buffer-read-only: t -*- */",
-""]
+        "/* Generated automatically; do not modify!    "
+        "-*- buffer-read-only: t -*- */",
+        "",
+    ]
+
 
 def make_meta_flow(meta_flow_h):
     fields = extract_ofp_fields(meta_flow_h)
     output = autogen_c_comment()
     for f in fields:
         output += ["{"]
-        output += ["    %s," % f['mff']]
-        if f['extra_name']:
-            output += ["    \"%s\", \"%s\"," % (f['name'], f['extra_name'])]
+        output += ["    %s," % f["mff"]]
+        if f["extra_name"]:
+            output += ['    "%s", "%s",' % (f["name"], f["extra_name"])]
         else:
-            output += ["    \"%s\", NULL," % f['name']]
+            output += ['    "%s", NULL,' % f["name"]]
 
-        if f['variable']:
-            variable = 'true'
+        if f["variable"]:
+            variable = "true"
         else:
-            variable = 'false'
-        output += ["    %d, %d, %s," % (f['n_bytes'], f['n_bits'], variable)]
+            variable = "false"
+        output += ["    %d, %d, %s," % (f["n_bytes"], f["n_bits"], variable)]
 
-        if f['writable']:
-            rw = 'true'
+        if f["writable"]:
+            rw = "true"
         else:
-            rw = 'false'
-        output += ["    %s, %s, %s, %s, false,"
-                   % (f['mask'], f['string'], PREREQS[f['prereqs']], rw)]
-
-        oxm = f['OXM']
-        of10 = f['OF1.0']
-        of11 = f['OF1.1']
-        if f['mff'] in ('MFF_DL_VLAN', 'MFF_DL_VLAN_PCP'):
+            rw = "false"
+        output += [
+            "    %s, %s, %s, %s, false,"
+            % (f["mask"], f["string"], PREREQS[f["prereqs"]], rw)
+        ]
+
+        oxm = f["OXM"]
+        of10 = f["OF1.0"]
+        of11 = f["OF1.1"]
+        if f["mff"] in ("MFF_DL_VLAN", "MFF_DL_VLAN_PCP"):
             # MFF_DL_VLAN and MFF_DL_VLAN_PCP don't exactly correspond to
             # OF1.1, nor do they have NXM or OXM assignments, but their
             # meanings can be expressed in every protocol, which is the goal of
@@ -367,25 +114,25 @@  def make_meta_flow(meta_flow_h):
             if oxm:
                 protocols |= set(["oxm"])
 
-        if f['mask'] == 'MFM_FULLY':
+        if f["mask"] == "MFM_FULLY":
             cidr_protocols = protocols.copy()
             bitwise_protocols = protocols.copy()
 
-            if of10 == 'exact match':
-                bitwise_protocols -= set(['of10'])
-                cidr_protocols -= set(['of10'])
-            elif of10 == 'CIDR mask':
-                bitwise_protocols -= set(['of10'])
+            if of10 == "exact match":
+                bitwise_protocols -= set(["of10"])
+                cidr_protocols -= set(["of10"])
+            elif of10 == "CIDR mask":
+                bitwise_protocols -= set(["of10"])
             else:
                 assert of10 is None
 
-            if of11 == 'exact match':
-                bitwise_protocols -= set(['of11'])
-                cidr_protocols -= set(['of11'])
+            if of11 == "exact match":
+                bitwise_protocols -= set(["of11"])
+                cidr_protocols -= set(["of11"])
             else:
-                assert of11 in (None, 'bitwise mask')
+                assert of11 in (None, "bitwise mask")
         else:
-            assert f['mask'] == 'MFM_NONE'
+            assert f["mask"] == "MFM_NONE"
             cidr_protocols = set([])
             bitwise_protocols = set([])
 
@@ -393,8 +140,8 @@  def make_meta_flow(meta_flow_h):
         output += ["    %s," % protocols_to_c(cidr_protocols)]
         output += ["    %s," % protocols_to_c(bitwise_protocols)]
 
-        if f['prefix']:
-            output += ["    FLOW_U32OFS(%s)," % f['prefix']]
+        if f["prefix"]:
+            output += ["    FLOW_U32OFS(%s)," % f["prefix"]]
         else:
             output += ["    -1, /* not usable for prefix lookup */"]
 
@@ -409,147 +156,37 @@  def make_nx_match(meta_flow_h):
     print("static struct nxm_field_index all_nxm_fields[] = {")
     for f in fields:
         # Sort by OpenFlow version number (nx-match.c depends on this).
-        for oxm in sorted(f['OXM'], key=lambda x: x[2]):
-            header = ("NXM_HEADER(0x%x,0x%x,%s,0,%d)" % oxm[0])
-            print("""{ .nf = { %s, %d, "%s", %s } },""" % (
-                header, oxm[2], oxm[1], f['mff']))
+        for oxm in sorted(f["OXM"], key=lambda x: x[2]):
+            header = "NXM_HEADER(0x%x,0x%x,%s,0,%d)" % oxm[0]
+            print(
+                """{ .nf = { %s, %d, "%s", %s } },"""
+                % (header, oxm[2], oxm[1], f["mff"])
+            )
     print("};")
     for oline in output:
         print(oline)
 
 
-def extract_ofp_fields(fn):
-    global file_name
-    global input_file
-    global line_number
-    global line
-
-    file_name = fn
-    input_file = open(file_name)
-    line_number = 0
-
-    fields = []
-
-    while True:
-        get_line()
-        if re.match('enum.*mf_field_id', line):
-            break
-
-    while True:
-        get_line()
-        first_line_number = line_number
-        here = '%s:%d' % (file_name, line_number)
-        if (line.startswith('/*')
-            or line.startswith(' *')
-            or line.startswith('#')
-            or not line
-            or line.isspace()):
-            continue
-        elif re.match('}', line) or re.match('\s+MFF_N_IDS', line):
-            break
-
-        # Parse the comment preceding an MFF_ constant into 'comment',
-        # one line to an array element.
-        line = line.strip()
-        if not line.startswith('/*'):
-            fatal("unexpected syntax between fields")
-        line = line[1:]
-        comment = []
-        end = False
-        while not end:
-            line = line.strip()
-            if line.startswith('*/'):
-                get_line()
-                break
-            if not line.startswith('*'):
-                fatal("unexpected syntax within field")
-
-            line = line[1:]
-            if line.startswith(' '):
-                line = line[1:]
-            if line.startswith(' ') and comment:
-                continuation = True
-                line = line.lstrip()
-            else:
-                continuation = False
-
-            if line.endswith('*/'):
-                line = line[:-2].rstrip()
-                end = True
-            else:
-                end = False
-
-            if continuation:
-                comment[-1] += " " + line
-            else:
-                comment += [line]
-            get_line()
-
-        # Drop blank lines at each end of comment.
-        while comment and not comment[0]:
-            comment = comment[1:]
-        while comment and not comment[-1]:
-            comment = comment[:-1]
-
-        # Parse the MFF_ constant(s).
-        mffs = []
-        while True:
-            m = re.match('\s+(MFF_[A-Z0-9_]+),?\s?$', line)
-            if not m:
-                break
-            mffs += [m.group(1)]
-            get_line()
-        if not mffs:
-            fatal("unexpected syntax looking for MFF_ constants")
-
-        if len(mffs) > 1 or '<N>' in comment[0]:
-            for mff in mffs:
-                # Extract trailing integer.
-                m = re.match('.*[^0-9]([0-9]+)$', mff)
-                if not m:
-                    fatal("%s lacks numeric suffix in register group" % mff)
-                n = m.group(1)
-
-                # Search-and-replace <N> within the comment,
-                # and drop lines that have <x> for x != n.
-                instance = []
-                for x in comment:
-                    y = x.replace('<N>', n)
-                    if re.search('<[0-9]+>', y):
-                        if ('<%s>' % n) not in y:
-                            continue
-                        y = re.sub('<[0-9]+>', '', y)
-                    instance += [y.strip()]
-                fields += [parse_field(mff, instance)]
-        else:
-            fields += [parse_field(mffs[0], comment)]
-        continue
-
-    input_file.close()
-
-    if n_errors:
-        sys.exit(1)
-
-    return fields
-
 ## ------------------------ ##
 ## Documentation Generation ##
 ## ------------------------ ##
 
+
 def field_to_xml(field_node, f, body, summary):
     f["used"] = True
 
     # Summary.
-    if field_node.hasAttribute('internal'):
+    if field_node.hasAttribute("internal"):
         return
 
     min_of_version = None
     min_ovs_version = None
-    for header, name, of_version_nr, ovs_version_s in f['OXM']:
-        if (is_standard_oxm(name)
-            and (min_ovs_version is None or of_version_nr < min_of_version)):
+    for header, name, of_version_nr, ovs_version_s in f["OXM"]:
+        if is_standard_oxm(name) and (
+            min_ovs_version is None or of_version_nr < min_of_version
+        ):
             min_of_version = of_version_nr
-        ovs_version = [int(x) for x in ovs_version_s.split('.')]
+        ovs_version = [int(x) for x in ovs_version_s.split(".")]
         if min_ovs_version is None or ovs_version < min_ovs_version:
             min_ovs_version = ovs_version
     summary += ["\\fB%s\\fR" % f["name"]]
@@ -565,124 +202,152 @@  def field_to_xml(field_node, f, body, summary):
     if min_of_version is not None:
         support += ["OF %s+" % VERSION_REVERSE[min_of_version]]
     if min_ovs_version is not None:
-        support += ["OVS %s+" % '.'.join([str(x) for x in min_ovs_version])]
-    summary += ' and '.join(support)
+        support += ["OVS %s+" % ".".join([str(x) for x in min_ovs_version])]
+    summary += " and ".join(support)
     summary += ["\n"]
 
     # Full description.
-    if field_node.hasAttribute('hidden'):
+    if field_node.hasAttribute("hidden"):
         return
 
-    title = field_node.attributes['title'].nodeValue
+    title = field_node.attributes["title"].nodeValue
 
-    body += [""".PP
+    body += [
+        """.PP
 \\fB%s Field\\fR
 .TS
 tab(;);
 l lx.
-""" % title]
+"""
+        % title
+    ]
 
     body += ["Name:;\\fB%s\\fR" % f["name"]]
     if f["extra_name"]:
         body += [" (aka \\fB%s\\fR)" % f["extra_name"]]
-    body += ['\n']
+    body += ["\n"]
 
     body += ["Width:;"]
     if f["n_bits"] != 8 * f["n_bytes"]:
-        body += ["%d bits (only the least-significant %d bits "
-                 "may be nonzero)" % (f["n_bytes"] * 8, f["n_bits"])]
+        body += [
+            "%d bits (only the least-significant %d bits "
+            "may be nonzero)" % (f["n_bytes"] * 8, f["n_bits"])
+        ]
     elif f["n_bits"] <= 128:
         body += ["%d bits" % f["n_bits"]]
     else:
         body += ["%d bits (%d bytes)" % (f["n_bits"], f["n_bits"] / 8)]
-    body += ['\n']
+    body += ["\n"]
 
     body += ["Format:;%s\n" % f["formatting"]]
 
-    masks = {"MFM_NONE": "not maskable",
-             "MFM_FULLY": "arbitrary bitwise masks"}
+    masks = {
+        "MFM_NONE": "not maskable",
+        "MFM_FULLY": "arbitrary bitwise masks",
+    }
     body += ["Masking:;%s\n" % masks[f["mask"]]]
     body += ["Prerequisites:;%s\n" % f["prereqs"]]
 
-    access = {True: "read/write",
-              False: "read-only"}[f["writable"]]
+    access = {True: "read/write", False: "read-only"}[f["writable"]]
     body += ["Access:;%s\n" % access]
 
-    of10 = {None: "not supported",
-            "exact match": "yes (exact match only)",
-            "CIDR mask": "yes (CIDR match only)"}
+    of10 = {
+        None: "not supported",
+        "exact match": "yes (exact match only)",
+        "CIDR mask": "yes (CIDR match only)",
+    }
     body += ["OpenFlow 1.0:;%s\n" % of10[f["OF1.0"]]]
 
-    of11 = {None: "not supported",
-            "exact match": "yes (exact match only)",
-            "bitwise mask": "yes"}
+    of11 = {
+        None: "not supported",
+        "exact match": "yes (exact match only)",
+        "bitwise mask": "yes",
+    }
     body += ["OpenFlow 1.1:;%s\n" % of11[f["OF1.1"]]]
 
     oxms = []
-    for header, name, of_version_nr, ovs_version in [x for x in sorted(f['OXM'], key=lambda x: x[2]) if is_standard_oxm(x[1])]:
+    for header, name, of_version_nr, ovs_version in [
+        x
+        for x in sorted(f["OXM"], key=lambda x: x[2])
+        if is_standard_oxm(x[1])
+    ]:
         of_version = VERSION_REVERSE[of_version_nr]
-        oxms += [r"\fB%s\fR (%d) since OpenFlow %s and Open vSwitch %s" % (name, header[2], of_version, ovs_version)]
+        oxms += [
+            r"\fB%s\fR (%d) since OpenFlow %s and Open vSwitch %s"
+            % (name, header[2], of_version, ovs_version)
+        ]
     if not oxms:
-        oxms = ['none']
-    body += ['OXM:;T{\n%s\nT}\n' % r'\[char59] '.join(oxms)]
+        oxms = ["none"]
+    body += ["OXM:;T{\n%s\nT}\n" % r"\[char59] ".join(oxms)]
 
     nxms = []
-    for header, name, of_version_nr, ovs_version in [x for x in sorted(f['OXM'], key=lambda x: x[2]) if not is_standard_oxm(x[1])]:
-        nxms += [r"\fB%s\fR (%d) since Open vSwitch %s" % (name, header[2], ovs_version)]
+    for header, name, of_version_nr, ovs_version in [
+        x
+        for x in sorted(f["OXM"], key=lambda x: x[2])
+        if not is_standard_oxm(x[1])
+    ]:
+        nxms += [
+            r"\fB%s\fR (%d) since Open vSwitch %s"
+            % (name, header[2], ovs_version)
+        ]
     if not nxms:
-        nxms = ['none']
-    body += ['NXM:;T{\n%s\nT}\n' % r'\[char59] '.join(nxms)]
+        nxms = ["none"]
+    body += ["NXM:;T{\n%s\nT}\n" % r"\[char59] ".join(nxms)]
 
     body += [".TE\n"]
 
-    body += ['.PP\n']
+    body += [".PP\n"]
     body += [build.nroff.block_xml_to_nroff(field_node.childNodes)]
 
+
 def group_xml_to_nroff(group_node, fields):
-    title = group_node.attributes['title'].nodeValue
+    title = group_node.attributes["title"].nodeValue
 
     summary = []
     body = []
     for node in group_node.childNodes:
-        if node.nodeType == node.ELEMENT_NODE and node.tagName == 'field':
-            id_ = node.attributes['id'].nodeValue
+        if node.nodeType == node.ELEMENT_NODE and node.tagName == "field":
+            id_ = node.attributes["id"].nodeValue
             field_to_xml(node, fields[id_], body, summary)
         else:
             body += [build.nroff.block_xml_to_nroff([node])]
 
     content = [
-        '.bp\n',
-        '.SH \"%s\"\n' % build.nroff.text_to_nroff(title.upper() + " FIELDS"),
+        ".bp\n",
+        '.SH "%s"\n' % build.nroff.text_to_nroff(title.upper() + " FIELDS"),
         '.SS "Summary:"\n',
-        '.TS\n',
-        'tab(;);\n',
-        'l l l l l l l.\n',
-        'Name;Bytes;Mask;RW?;Prereqs;NXM/OXM Support\n',
-        '\_;\_;\_;\_;\_;\_\n']
+        ".TS\n",
+        "tab(;);\n",
+        "l l l l l l l.\n",
+        "Name;Bytes;Mask;RW?;Prereqs;NXM/OXM Support\n",
+        "\_;\_;\_;\_;\_;\_\n",
+    ]
     content += summary
-    content += ['.TE\n']
+    content += [".TE\n"]
     content += body
-    return ''.join(content)
+    return "".join(content)
+
 
 def make_oxm_classes_xml(document):
-    s = '''tab(;);
+    s = """tab(;);
 l l l.
 Prefix;Vendor;Class
 \_;\_;\_
-'''
+"""
     for key in sorted(OXM_CLASSES, key=OXM_CLASSES.get):
         vendor, class_, class_type = OXM_CLASSES.get(key)
-        s += r"\fB%s\fR;" % key.rstrip('_')
+        s += r"\fB%s\fR;" % key.rstrip("_")
         if vendor:
             s += r"\fL0x%08x\fR;" % vendor
         else:
             s += "(none);"
         s += r"\fL0x%04x\fR;" % class_
         s += "\n"
-    e = document.createElement('tbl')
+    e = document.createElement("tbl")
     e.appendChild(document.createTextNode(s))
     return e
 
+
 def recursively_replace(node, name, replacement):
     for child in node.childNodes:
         if child.nodeType == node.ELEMENT_NODE:
@@ -691,11 +356,12 @@  def recursively_replace(node, name, replacement):
             else:
                 recursively_replace(child, name, replacement)
 
+
 def make_ovs_fields(meta_flow_h, meta_flow_xml):
     fields = extract_ofp_fields(meta_flow_h)
     fields_map = {}
     for f in fields:
-        fields_map[f['mff']] = f
+        fields_map[f["mff"]] = f
 
     document = xml.dom.minidom.parse(meta_flow_xml)
     doc = document.documentElement
@@ -704,7 +370,8 @@  def make_ovs_fields(meta_flow_h, meta_flow_xml):
     if version == None:
         version = "UNKNOWN"
 
-    print('''\
+    print(
+        """\
 '\\" tp
 .\\" -*- mode: troff; coding: utf-8 -*-
 .TH "ovs\-fields" 7 "%s" "Open vSwitch" "Open vSwitch Manual"
@@ -740,11 +407,13 @@  def make_ovs_fields(meta_flow_h, meta_flow_xml):
 ovs\-fields \- protocol header fields in OpenFlow and Open vSwitch
 .
 .PP
-''' % version)
+"""
+        % version
+    )
 
-    recursively_replace(doc, 'oxm_classes', make_oxm_classes_xml(document))
+    recursively_replace(doc, "oxm_classes", make_oxm_classes_xml(document))
 
-    s = ''
+    s = ""
     for node in doc.childNodes:
         if node.nodeType == node.ELEMENT_NODE and node.tagName == "group":
             s += group_xml_to_nroff(node, fields_map)
@@ -757,9 +426,10 @@  ovs\-fields \- protocol header fields in OpenFlow and Open vSwitch
 
     for f in fields:
         if "used" not in f:
-            fatal("%s: field not documented "
-                  "(please add documentation in lib/meta-flow.xml)"
-                  % f["mff"])
+            fatal(
+                "%s: field not documented "
+                "(please add documentation in lib/meta-flow.xml)" % f["mff"]
+            )
     if n_errors:
         sys.exit(1)
 
@@ -769,26 +439,27 @@  ovs\-fields \- protocol header fields in OpenFlow and Open vSwitch
 
         # Life is easier with nroff if we don't try to feed it Unicode.
         # Fortunately, we only use a few characters outside the ASCII range.
-        oline = oline.replace(u'\u2208', r'\[mo]')
-        oline = oline.replace(u'\u2260', r'\[!=]')
-        oline = oline.replace(u'\u2264', r'\[<=]')
-        oline = oline.replace(u'\u2265', r'\[>=]')
-        oline = oline.replace(u'\u00d7', r'\[mu]')
+        oline = oline.replace(u"\u2208", r"\[mo]")
+        oline = oline.replace(u"\u2260", r"\[!=]")
+        oline = oline.replace(u"\u2264", r"\[<=]")
+        oline = oline.replace(u"\u2265", r"\[>=]")
+        oline = oline.replace(u"\u00d7", r"\[mu]")
         if len(oline):
             output += [oline]
 
     # nroff tends to ignore .bp requests if they come after .PP requests,
     # so remove .PPs that precede .bp.
     for i in range(len(output)):
-        if output[i] == '.bp':
+        if output[i] == ".bp":
             j = i - 1
-            while j >= 0 and output[j] == '.PP':
+            while j >= 0 and output[j] == ".PP":
                 output[j] = None
                 j -= 1
     for i in range(len(output)):
         if output[i] is not None:
             print(output[i])
-
+
+
 ## ------------ ##
 ## Main Program ##
 ## ------------ ##
@@ -796,8 +467,9 @@  ovs\-fields \- protocol header fields in OpenFlow and Open vSwitch
 if __name__ == "__main__":
     argv0 = sys.argv[0]
     try:
-        options, args = getopt.gnu_getopt(sys.argv[1:], 'h',
-                                          ['help', 'ovs-version='])
+        options, args = getopt.gnu_getopt(
+            sys.argv[1:], "h", ["help", "ovs-version="]
+        )
     except getopt.GetoptError as geo:
         sys.stderr.write("%s: %s\n" % (argv0, geo.msg))
         sys.exit(1)
@@ -805,32 +477,38 @@  if __name__ == "__main__":
     global version
     version = None
     for key, value in options:
-        if key in ['-h', '--help']:
+        if key in ["-h", "--help"]:
             usage()
-        elif key == '--ovs-version':
+        elif key == "--ovs-version":
             version = value
         else:
             sys.exit(0)
 
     if not args:
-        sys.stderr.write("%s: missing command argument "
-                         "(use --help for help)\n" % argv0)
+        sys.stderr.write(
+            "%s: missing command argument " "(use --help for help)\n" % argv0
+        )
         sys.exit(1)
 
-    commands = {"meta-flow": (make_meta_flow, 1),
-                "nx-match": (make_nx_match, 1),
-                "ovs-fields": (make_ovs_fields, 2)}
+    commands = {
+        "meta-flow": (make_meta_flow, 1),
+        "nx-match": (make_nx_match, 1),
+        "ovs-fields": (make_ovs_fields, 2),
+    }
 
     if not args[0] in commands:
-        sys.stderr.write("%s: unknown command \"%s\" "
-                         "(use --help for help)\n" % (argv0, args[0]))
+        sys.stderr.write(
+            '%s: unknown command "%s" '
+            "(use --help for help)\n" % (argv0, args[0])
+        )
         sys.exit(1)
 
     func, n_args = commands[args[0]]
     if len(args) - 1 != n_args:
-        sys.stderr.write("%s: \"%s\" requires %d arguments but %d "
-                         "provided\n"
-                         % (argv0, args[0], n_args, len(args) - 1))
+        sys.stderr.write(
+            '%s: "%s" requires %d arguments but %d '
+            "provided\n" % (argv0, args[0], n_args, len(args) - 1)
+        )
         sys.exit(1)
 
     func(*args[1:])
diff --git a/python/automake.mk b/python/automake.mk
index 73438d615..d7d33928a 100644
--- a/python/automake.mk
+++ b/python/automake.mk
@@ -51,6 +51,7 @@  ovs_pyfiles = \
 # so they are not installed.
 EXTRA_DIST += \
 	python/build/__init__.py \
+	python/build/extract_ofp_fields.py \
 	python/build/nroff.py \
 	python/build/soutil.py
 
@@ -69,10 +70,12 @@  PYCOV_CLEAN_FILES += $(PYFILES:.py=.py,cover)
 
 FLAKE8_PYFILES += \
 	$(filter-out python/ovs/compat/% python/ovs/dirs.py,$(PYFILES)) \
-	python/setup.py \
 	python/build/__init__.py \
+	python/build/extract_ofp_fields.py \
 	python/build/nroff.py \
-	python/ovs/dirs.py.template
+	python/build/soutil.py \
+	python/ovs/dirs.py.template \
+	python/setup.py
 
 nobase_pkgdata_DATA = $(ovs_pyfiles) $(ovstest_pyfiles)
 ovs-install-data-local:
diff --git a/python/build/extract_ofp_fields.py b/python/build/extract_ofp_fields.py
new file mode 100644
index 000000000..3fe62634a
--- /dev/null
+++ b/python/build/extract_ofp_fields.py
@@ -0,0 +1,421 @@ 
+import sys
+import re
+
+line = ""
+
+# Maps from user-friendly version number to its protocol encoding.
+VERSION = {
+    "1.0": 0x01,
+    "1.1": 0x02,
+    "1.2": 0x03,
+    "1.3": 0x04,
+    "1.4": 0x05,
+    "1.5": 0x06,
+}
+VERSION_REVERSE = dict((v, k) for k, v in VERSION.items())
+
+TYPES = {
+    "u8": (1, False),
+    "be16": (2, False),
+    "be32": (4, False),
+    "MAC": (6, False),
+    "be64": (8, False),
+    "be128": (16, False),
+    "tunnelMD": (124, True),
+}
+
+FORMATTING = {
+    "decimal": ("MFS_DECIMAL", 1, 8),
+    "hexadecimal": ("MFS_HEXADECIMAL", 1, 127),
+    "ct state": ("MFS_CT_STATE", 4, 4),
+    "Ethernet": ("MFS_ETHERNET", 6, 6),
+    "IPv4": ("MFS_IPV4", 4, 4),
+    "IPv6": ("MFS_IPV6", 16, 16),
+    "OpenFlow 1.0 port": ("MFS_OFP_PORT", 2, 2),
+    "OpenFlow 1.1+ port": ("MFS_OFP_PORT_OXM", 4, 4),
+    "frag": ("MFS_FRAG", 1, 1),
+    "tunnel flags": ("MFS_TNL_FLAGS", 2, 2),
+    "TCP flags": ("MFS_TCP_FLAGS", 2, 2),
+    "packet type": ("MFS_PACKET_TYPE", 4, 4),
+}
+
+PREREQS = {
+    "none": "MFP_NONE",
+    "Ethernet": "MFP_ETHERNET",
+    "ARP": "MFP_ARP",
+    "VLAN VID": "MFP_VLAN_VID",
+    "IPv4": "MFP_IPV4",
+    "IPv6": "MFP_IPV6",
+    "IPv4/IPv6": "MFP_IP_ANY",
+    "NSH": "MFP_NSH",
+    "CT": "MFP_CT_VALID",
+    "MPLS": "MFP_MPLS",
+    "TCP": "MFP_TCP",
+    "UDP": "MFP_UDP",
+    "SCTP": "MFP_SCTP",
+    "ICMPv4": "MFP_ICMPV4",
+    "ICMPv6": "MFP_ICMPV6",
+    "ND": "MFP_ND",
+    "ND solicit": "MFP_ND_SOLICIT",
+    "ND advert": "MFP_ND_ADVERT",
+}
+
+# Maps a name prefix into an (experimenter ID, class) pair, so:
+#
+#      - Standard OXM classes are written as (0, <oxm_class>)
+#
+#      - Experimenter OXM classes are written as (<oxm_vender>, 0xffff)
+#
+# If a name matches more than one prefix, the longest one is used.
+OXM_CLASSES = {
+    "NXM_OF_": (0, 0x0000, "extension"),
+    "NXM_NX_": (0, 0x0001, "extension"),
+    "NXOXM_NSH_": (0x005AD650, 0xFFFF, "extension"),
+    "OXM_OF_": (0, 0x8000, "standard"),
+    "OXM_OF_PKT_REG": (0, 0x8001, "standard"),
+    "ONFOXM_ET_": (0x4F4E4600, 0xFFFF, "standard"),
+    "ERICOXM_OF_": (0, 0x1000, "extension"),
+    # This is the experimenter OXM class for Nicira, which is the
+    # one that OVS would be using instead of NXM_OF_ and NXM_NX_
+    # if OVS didn't have those grandfathered in.  It is currently
+    # used only to test support for experimenter OXM, since there
+    # are barely any real uses of experimenter OXM in the wild.
+    "NXOXM_ET_": (0x00002320, 0xFFFF, "extension"),
+}
+
+
+def oxm_name_to_class(name):
+    prefix = ""
+    class_ = None
+    for p, c in OXM_CLASSES.items():
+        if name.startswith(p) and len(p) > len(prefix):
+            prefix = p
+            class_ = c
+    return class_
+
+
+def is_standard_oxm(name):
+    oxm_vendor, oxm_class, oxm_class_type = oxm_name_to_class(name)
+    return oxm_class_type == "standard"
+
+
+def get_line():
+    global line
+    global line_number
+    line = input_file.readline()
+    line_number += 1
+    if line == "":
+        fatal("unexpected end of input")
+
+
+n_errors = 0
+
+
+def error(msg):
+    global n_errors
+    sys.stderr.write("%s:%d: %s\n" % (file_name, line_number, msg))
+    n_errors += 1
+
+
+def fatal(msg):
+    error(msg)
+    sys.exit(1)
+
+
+def parse_oxms(s, prefix, n_bytes):
+    if s == "none":
+        return ()
+
+    return tuple(parse_oxm(s2.strip(), prefix, n_bytes) for s2 in s.split(","))
+
+
+match_types = dict()
+
+
+def parse_oxm(s, prefix, n_bytes):
+    global match_types
+
+    m = re.match(
+        r"([A-Z0-9_]+)\(([0-9]+)\) since(?: OF(1\.[0-9]+) and)? v([12]\.[0-9]+)$",  # noqa: E501
+        s,
+    )
+    if not m:
+        fatal("%s: syntax error parsing %s" % (s, prefix))
+
+    name, oxm_type, of_version, ovs_version = m.groups()
+
+    class_ = oxm_name_to_class(name)
+    if class_ is None:
+        fatal("unknown OXM class for %s" % name)
+    oxm_vendor, oxm_class, oxm_class_type = class_
+
+    if class_ in match_types:
+        if oxm_type in match_types[class_]:
+            fatal(
+                "duplicate match type for %s (conflicts with %s)"
+                % (name, match_types[class_][oxm_type])
+            )
+    else:
+        match_types[class_] = dict()
+    match_types[class_][oxm_type] = name
+
+    # Normally the oxm_length is the size of the field, but for experimenter
+    # OXMs oxm_length also includes the 4-byte experimenter ID.
+    oxm_length = n_bytes
+    if oxm_class == 0xFFFF:
+        oxm_length += 4
+
+    header = (oxm_vendor, oxm_class, int(oxm_type), oxm_length)
+
+    if of_version:
+        if oxm_class_type == "extension":
+            fatal("%s: OXM extension can't have OpenFlow version" % name)
+        if of_version not in VERSION:
+            fatal("%s: unknown OpenFlow version %s" % (name, of_version))
+        of_version_nr = VERSION[of_version]
+        if of_version_nr < VERSION["1.2"]:
+            fatal("%s: claimed version %s predates OXM" % (name, of_version))
+    else:
+        if oxm_class_type == "standard":
+            fatal("%s: missing OpenFlow version number" % name)
+        of_version_nr = 0
+
+    return (header, name, of_version_nr, ovs_version)
+
+
+def parse_field(mff, comment):
+    f = {"mff": mff}
+
+    # First line of comment is the field name.
+    m = re.match(
+        r'"([^"]+)"(?:\s+\(aka "([^"]+)"\))?(?:\s+\(.*\))?\.', comment[0]
+    )
+    if not m:
+        fatal("%s lacks field name" % mff)
+    f["name"], f["extra_name"] = m.groups()
+
+    # Find the last blank line the comment.  The field definitions
+    # start after that.
+    blank = None
+    for i in range(len(comment)):
+        if not comment[i]:
+            blank = i
+    if not blank:
+        fatal("%s: missing blank line in comment" % mff)
+
+    d = {}
+    for key in (
+        "Type",
+        "Maskable",
+        "Formatting",
+        "Prerequisites",
+        "Access",
+        "Prefix lookup member",
+        "OXM",
+        "NXM",
+        "OF1.0",
+        "OF1.1",
+    ):
+        d[key] = None
+    for fline in comment[blank + 1 :]:
+        m = re.match(r"([^:]+):\s+(.*)\.$", fline)
+        if not m:
+            fatal(
+                "%s: syntax error parsing key-value pair as part of %s"
+                % (fline, mff)
+            )
+        key, value = m.groups()
+        if key not in d:
+            fatal("%s: unknown key" % key)
+        elif key == "Code point":
+            d[key] += [value]
+        elif d[key] is not None:
+            fatal("%s: duplicate key" % key)
+        d[key] = value
+    for key, value in d.items():
+        if not value and key not in (
+            "OF1.0",
+            "OF1.1",
+            "Prefix lookup member",
+            "Notes",
+        ):
+            fatal("%s: missing %s" % (mff, key))
+
+    m = re.match(r"([a-zA-Z0-9]+)(?: \(low ([0-9]+) bits\))?$", d["Type"])
+    if not m:
+        fatal("%s: syntax error in type" % mff)
+    type_ = m.group(1)
+    if type_ not in TYPES:
+        fatal("%s: unknown type %s" % (mff, d["Type"]))
+
+    f["n_bytes"] = TYPES[type_][0]
+    if m.group(2):
+        f["n_bits"] = int(m.group(2))
+        if f["n_bits"] > f["n_bytes"] * 8:
+            fatal(
+                "%s: more bits (%d) than field size (%d)"
+                % (mff, f["n_bits"], 8 * f["n_bytes"])
+            )
+    else:
+        f["n_bits"] = 8 * f["n_bytes"]
+    f["variable"] = TYPES[type_][1]
+
+    if d["Maskable"] == "no":
+        f["mask"] = "MFM_NONE"
+    elif d["Maskable"] == "bitwise":
+        f["mask"] = "MFM_FULLY"
+    else:
+        fatal("%s: unknown maskable %s" % (mff, d["Maskable"]))
+
+    fmt = FORMATTING.get(d["Formatting"])
+    if not fmt:
+        fatal("%s: unknown format %s" % (mff, d["Formatting"]))
+    f["formatting"] = d["Formatting"]
+    if f["n_bytes"] < fmt[1] or f["n_bytes"] > fmt[2]:
+        fatal(
+            "%s: %d-byte field can't be formatted as %s"
+            % (mff, f["n_bytes"], d["Formatting"])
+        )
+    f["string"] = fmt[0]
+
+    f["prereqs"] = d["Prerequisites"]
+    if f["prereqs"] not in PREREQS:
+        fatal("%s: unknown prerequisites %s" % (mff, d["Prerequisites"]))
+
+    if d["Access"] == "read-only":
+        f["writable"] = False
+    elif d["Access"] == "read/write":
+        f["writable"] = True
+    else:
+        fatal("%s: unknown access %s" % (mff, d["Access"]))
+
+    f["OF1.0"] = d["OF1.0"]
+    if not d["OF1.0"] in (None, "exact match", "CIDR mask"):
+        fatal("%s: unknown OF1.0 match type %s" % (mff, d["OF1.0"]))
+
+    f["OF1.1"] = d["OF1.1"]
+    if not d["OF1.1"] in (None, "exact match", "bitwise mask"):
+        fatal("%s: unknown OF1.1 match type %s" % (mff, d["OF1.1"]))
+
+    f["OXM"] = parse_oxms(d["OXM"], "OXM", f["n_bytes"]) + parse_oxms(
+        d["NXM"], "NXM", f["n_bytes"]
+    )
+
+    f["prefix"] = d["Prefix lookup member"]
+
+    return f
+
+
+def extract_ofp_fields(fn):
+    global file_name
+    global input_file
+    global line_number
+    global line
+
+    file_name = fn
+    input_file = open(file_name)
+    line_number = 0
+
+    fields = []
+
+    while True:
+        get_line()
+        if re.match("enum.*mf_field_id", line):
+            break
+
+    while True:
+        get_line()
+        if (
+            line.startswith("/*")
+            or line.startswith(" *")
+            or line.startswith("#")
+            or not line
+            or line.isspace()
+        ):
+            continue
+        elif re.match(r"}", line) or re.match(r"\s+MFF_N_IDS", line):
+            break
+
+        # Parse the comment preceding an MFF_ constant into 'comment',
+        # one line to an array element.
+        line = line.strip()
+        if not line.startswith("/*"):
+            fatal("unexpected syntax between fields")
+        line = line[1:]
+        comment = []
+        end = False
+        while not end:
+            line = line.strip()
+            if line.startswith("*/"):
+                get_line()
+                break
+            if not line.startswith("*"):
+                fatal("unexpected syntax within field")
+
+            line = line[1:]
+            if line.startswith(" "):
+                line = line[1:]
+            if line.startswith(" ") and comment:
+                continuation = True
+                line = line.lstrip()
+            else:
+                continuation = False
+
+            if line.endswith("*/"):
+                line = line[:-2].rstrip()
+                end = True
+            else:
+                end = False
+
+            if continuation:
+                comment[-1] += " " + line
+            else:
+                comment += [line]
+            get_line()
+
+        # Drop blank lines at each end of comment.
+        while comment and not comment[0]:
+            comment = comment[1:]
+        while comment and not comment[-1]:
+            comment = comment[:-1]
+
+        # Parse the MFF_ constant(s).
+        mffs = []
+        while True:
+            m = re.match(r"\s+(MFF_[A-Z0-9_]+),?\s?$", line)
+            if not m:
+                break
+            mffs += [m.group(1)]
+            get_line()
+        if not mffs:
+            fatal("unexpected syntax looking for MFF_ constants")
+
+        if len(mffs) > 1 or "<N>" in comment[0]:
+            for mff in mffs:
+                # Extract trailing integer.
+                m = re.match(".*[^0-9]([0-9]+)$", mff)
+                if not m:
+                    fatal("%s lacks numeric suffix in register group" % mff)
+                n = m.group(1)
+
+                # Search-and-replace <N> within the comment,
+                # and drop lines that have <x> for x != n.
+                instance = []
+                for x in comment:
+                    y = x.replace("<N>", n)
+                    if re.search("<[0-9]+>", y):
+                        if ("<%s>" % n) not in y:
+                            continue
+                        y = re.sub("<[0-9]+>", "", y)
+                    instance += [y.strip()]
+                fields += [parse_field(mff, instance)]
+        else:
+            fields += [parse_field(mffs[0], comment)]
+        continue
+
+    input_file.close()
+
+    if n_errors:
+        sys.exit(1)
+
+    return fields