[ovs-dev] [PATCH v2 17/21] ovn: Rename Pipeline table to Rule table.

Ben Pfaff blp at nicira.com
Fri Jul 31 18:19:03 UTC 2015


I have a lot of misgivings about all of the possible names.

Thanks.

On Thu, Jul 30, 2015 at 04:33:23PM -0700, Justin Pettit wrote:
> I think Pipeline is more descriptive about what it actually is.  I also find it confusing since we use the term "rule" in the classifier.  I think Flow (or Logical_Flow) would be clearer than Rule, since we really are talking about flows, and people may look for a distinction that isn't there.  That, and the fact that we use "rule" for a different purpose in other parts of the tree, I think will make it more confusing.
> 
> All that said, I haven't looked ahead at the other patches yet, so maybe this is the right choice.  I'll defer to you.
> 
> Acked-by: Justin Pettit <jpettit at nicira.com>
> 
> --Justin
> 
> 
> > On Jul 28, 2015, at 8:44 AM, Ben Pfaff <blp at nicira.com> wrote:
> > 
> > The OVN pipeline is being split into two phases, which are most naturally
> > called "pipelines".  I kept getting very confused trying to call them
> > anything else, and in the end it seems to make more sense to just rename
> > the Pipeline table.
> > 
> > It would be even better to call this table Flow or Logical_Flow, but I
> > am worried that we already have far too many uses of the word "flow".
> > "Rule" is slightly less overloaded in OVS.
> > 
> > Signed-off-by: Ben Pfaff <blp at nicira.com>
> > ---
> > ovn/TODO                              |   2 +-
> > ovn/controller/automake.mk            |   6 +-
> > ovn/controller/ovn-controller.c       |   8 +-
> > ovn/controller/physical.c             |   2 +-
> > ovn/controller/{pipeline.c => rule.c} |  50 +++++-----
> > ovn/controller/{pipeline.h => rule.h} |  18 ++--
> > ovn/lib/actions.c                     |   4 +-
> > ovn/northd/ovn-northd.c               | 182 +++++++++++++++++-----------------
> > ovn/ovn-architecture.7.xml            |  20 ++--
> > ovn/ovn-nb.xml                        |   4 +-
> > ovn/ovn-sb.ovsschema                  |   2 +-
> > ovn/ovn-sb.xml                        |   6 +-
> > 12 files changed, 152 insertions(+), 152 deletions(-)
> > rename ovn/controller/{pipeline.c => rule.c} (89%)
> > rename ovn/controller/{pipeline.h => rule.h} (79%)
> > 
> > diff --git a/ovn/TODO b/ovn/TODO
> > index 07d66da..19c95ca 100644
> > --- a/ovn/TODO
> > +++ b/ovn/TODO
> > @@ -48,7 +48,7 @@
> >     Currently, clients monitor the entire contents of a table.  It
> >     might make sense to allow clients to monitor only rows that
> >     satisfy specific criteria, e.g. to allow an ovn-controller to
> > -    receive only Pipeline rows for logical networks on its hypervisor.
> > +    receive only Rule rows for logical networks on its hypervisor.
> > 
> > *** Reducing redundant data and code within ovsdb-server.
> > 
> > diff --git a/ovn/controller/automake.mk b/ovn/controller/automake.mk
> > index 9ed6bec..55134a3 100644
> > --- a/ovn/controller/automake.mk
> > +++ b/ovn/controller/automake.mk
> > @@ -10,10 +10,10 @@ ovn_controller_ovn_controller_SOURCES = \
> > 	ovn/controller/ofctrl.h \
> > 	ovn/controller/ovn-controller.c \
> > 	ovn/controller/ovn-controller.h \
> > -	ovn/controller/pipeline.c \
> > -	ovn/controller/pipeline.h \
> > 	ovn/controller/physical.c \
> > -	ovn/controller/physical.h
> > +	ovn/controller/physical.h \
> > +	ovn/controller/rule.c \
> > +	ovn/controller/rule.h
> > ovn_controller_ovn_controller_LDADD = ovn/lib/libovn.la lib/libopenvswitch.la
> > man_MANS += ovn/controller/ovn-controller.8
> > EXTRA_DIST += ovn/controller/ovn-controller.8.xml
> > diff --git a/ovn/controller/ovn-controller.c b/ovn/controller/ovn-controller.c
> > index 12515c3..cfd6eb9 100644
> > --- a/ovn/controller/ovn-controller.c
> > +++ b/ovn/controller/ovn-controller.c
> > @@ -44,7 +44,7 @@
> > #include "chassis.h"
> > #include "encaps.h"
> > #include "physical.h"
> > -#include "pipeline.h"
> > +#include "rule.h"
> > 
> > VLOG_DEFINE_THIS_MODULE(main);
> > 
> > @@ -224,7 +224,7 @@ main(int argc, char *argv[])
> >     sbrec_init();
> > 
> >     ofctrl_init();
> > -    pipeline_init();
> > +    rule_init();
> > 
> >     /* Connect to OVS OVSDB instance.  We do not monitor all tables by
> >      * default, so modules must register their interest explicitly.  */
> > @@ -266,7 +266,7 @@ main(int argc, char *argv[])
> > 
> >         if (br_int) {
> >             struct hmap flow_table = HMAP_INITIALIZER(&flow_table);
> > -            pipeline_run(&ctx, &flow_table);
> > +            rule_run(&ctx, &flow_table);
> >             if (chassis_id) {
> >                 physical_run(&ctx, br_int, chassis_id, &flow_table);
> >             }
> > @@ -318,7 +318,7 @@ main(int argc, char *argv[])
> >     }
> > 
> >     unixctl_server_destroy(unixctl);
> > -    pipeline_destroy();
> > +    rule_destroy();
> >     ofctrl_destroy();
> > 
> >     idl_loop_destroy(&ovs_idl_loop);
> > diff --git a/ovn/controller/physical.c b/ovn/controller/physical.c
> > index 55d6107..2dc96ab 100644
> > --- a/ovn/controller/physical.c
> > +++ b/ovn/controller/physical.c
> > @@ -21,7 +21,7 @@
> > #include "ofpbuf.h"
> > #include "ovn-controller.h"
> > #include "ovn/lib/ovn-sb-idl.h"
> > -#include "pipeline.h"
> > +#include "rule.h"
> > #include "simap.h"
> > #include "vswitch-idl.h"
> > 
> > diff --git a/ovn/controller/pipeline.c b/ovn/controller/rule.c
> > similarity index 89%
> > rename from ovn/controller/pipeline.c
> > rename to ovn/controller/rule.c
> > index 1927ce4..0f5971b 100644
> > --- a/ovn/controller/pipeline.c
> > +++ b/ovn/controller/rule.c
> > @@ -14,7 +14,7 @@
> >  */
> > 
> > #include <config.h>
> > -#include "pipeline.h"
> > +#include "rule.h"
> > #include "dynamic-string.h"
> > #include "ofctrl.h"
> > #include "ofp-actions.h"
> > @@ -26,11 +26,11 @@
> > #include "ovn/lib/ovn-sb-idl.h"
> > #include "simap.h"
> > 
> > -VLOG_DEFINE_THIS_MODULE(pipeline);
> > +VLOG_DEFINE_THIS_MODULE(rule);
> > 
> > /* Symbol table. */
> > 
> > -/* Contains "struct expr_symbol"s for fields supported by OVN pipeline. */
> > +/* Contains "struct expr_symbol"s for fields supported by OVN rules. */
> > static struct shash symtab;
> > 
> > static void
> > @@ -244,31 +244,31 @@ ldp_destroy(void)
> > }
> > 
> > void
> > -pipeline_init(void)
> > +rule_init(void)
> > {
> >     symtab_init();
> > }
> > 
> > -/* Translates logical flows in the Pipeline table in the OVN_SB database
> > - * into OpenFlow flows, adding the OpenFlow flows to 'flow_table'.
> > +/* Translates logical flows in the Rule table in the OVN_SB database into
> > + * OpenFlow flows, adding the OpenFlow flows to 'flow_table'.
> >  *
> > - * We put the Pipeline flows into OpenFlow tables 16 through 47 (inclusive). */
> > + * We put the Rule flows into OpenFlow tables 16 through 47 (inclusive). */
> > void
> > -pipeline_run(struct controller_ctx *ctx, struct hmap *flow_table)
> > +rule_run(struct controller_ctx *ctx, struct hmap *flow_table)
> > {
> >     struct hmap flows = HMAP_INITIALIZER(&flows);
> >     uint32_t conj_id_ofs = 1;
> > 
> >     ldp_run(ctx);
> > 
> > -    const struct sbrec_pipeline *pipeline;
> > -    SBREC_PIPELINE_FOR_EACH (pipeline, ctx->ovnsb_idl) {
> > -        /* Find the "struct logical_datapath" asssociated with this Pipeline
> > -         * row.  If there's no such struct, that must be because no logical
> > -         * ports are bound to that logical datapath, so there's no point in
> > -         * maintaining any flows for it anyway, so skip it. */
> > +    const struct sbrec_rule *rule;
> > +    SBREC_RULE_FOR_EACH (rule, ctx->ovnsb_idl) {
> > +        /* Find the "struct logical_datapath" asssociated with this Rule row.
> > +         * If there's no such struct, that must be because no logical ports are
> > +         * bound to that logical datapath, so there's no point in maintaining
> > +         * any flows for it anyway, so skip it. */
> >         const struct logical_datapath *ldp;
> > -        ldp = ldp_lookup(&pipeline->logical_datapath);
> > +        ldp = ldp_lookup(&rule->logical_datapath);
> >         if (!ldp) {
> >             continue;
> >         }
> > @@ -281,13 +281,13 @@ pipeline_run(struct controller_ctx *ctx, struct hmap *flow_table)
> >         char *error;
> > 
> >         ofpbuf_use_stub(&ofpacts, ofpacts_stub, sizeof ofpacts_stub);
> > -        next_table_id = pipeline->table_id < 31 ? pipeline->table_id + 17 : 0;
> > -        error = actions_parse_string(pipeline->actions, &symtab, &ldp->ports,
> > +        next_table_id = rule->table_id < 31 ? rule->table_id + 17 : 0;
> > +        error = actions_parse_string(rule->actions, &symtab, &ldp->ports,
> >                                      next_table_id, &ofpacts, &prereqs);
> >         if (error) {
> >             static struct vlog_rate_limit rl = VLOG_RATE_LIMIT_INIT(1, 1);
> >             VLOG_WARN_RL(&rl, "error parsing actions \"%s\": %s",
> > -                         pipeline->actions, error);
> > +                         rule->actions, error);
> >             free(error);
> >             continue;
> >         }
> > @@ -296,7 +296,7 @@ pipeline_run(struct controller_ctx *ctx, struct hmap *flow_table)
> >         struct hmap matches;
> >         struct expr *expr;
> > 
> > -        expr = expr_parse_string(pipeline->match, &symtab, &error);
> > +        expr = expr_parse_string(rule->match, &symtab, &error);
> >         if (!error) {
> >             if (prereqs) {
> >                 expr = expr_combine(EXPR_T_AND, expr, prereqs);
> > @@ -307,7 +307,7 @@ pipeline_run(struct controller_ctx *ctx, struct hmap *flow_table)
> >         if (error) {
> >             static struct vlog_rate_limit rl = VLOG_RATE_LIMIT_INIT(1, 1);
> >             VLOG_WARN_RL(&rl, "error parsing match \"%s\": %s",
> > -                         pipeline->match, error);
> > +                         rule->match, error);
> >             expr_destroy(prereqs);
> >             ofpbuf_uninit(&ofpacts);
> >             free(error);
> > @@ -327,8 +327,8 @@ pipeline_run(struct controller_ctx *ctx, struct hmap *flow_table)
> >                 m->match.flow.conj_id += conj_id_ofs;
> >             }
> >             if (!m->n) {
> > -                ofctrl_add_flow(flow_table, pipeline->table_id + 16,
> > -                                pipeline->priority, &m->match, &ofpacts);
> > +                ofctrl_add_flow(flow_table, rule->table_id + 16,
> > +                                rule->priority, &m->match, &ofpacts);
> >             } else {
> >                 uint64_t conj_stubs[64 / 8];
> >                 struct ofpbuf conj;
> > @@ -343,8 +343,8 @@ pipeline_run(struct controller_ctx *ctx, struct hmap *flow_table)
> >                     dst->clause = src->clause;
> >                     dst->n_clauses = src->n_clauses;
> >                 }
> > -                ofctrl_add_flow(flow_table, pipeline->table_id + 16,
> > -                                pipeline->priority, &m->match, &conj);
> > +                ofctrl_add_flow(flow_table, rule->table_id + 16,
> > +                                rule->priority, &m->match, &conj);
> >                 ofpbuf_uninit(&conj);
> >             }
> >         }
> > @@ -357,7 +357,7 @@ pipeline_run(struct controller_ctx *ctx, struct hmap *flow_table)
> > }
> > 
> > void
> > -pipeline_destroy(void)
> > +rule_destroy(void)
> > {
> >     expr_symtab_destroy(&symtab);
> >     ldp_destroy();
> > diff --git a/ovn/controller/pipeline.h b/ovn/controller/rule.h
> > similarity index 79%
> > rename from ovn/controller/pipeline.h
> > rename to ovn/controller/rule.h
> > index 7d33341..3998994 100644
> > --- a/ovn/controller/pipeline.h
> > +++ b/ovn/controller/rule.h
> > @@ -14,13 +14,13 @@
> >  */
> > 
> > 
> > -#ifndef OVN_PIPELINE_H
> > -#define OVN_PIPELINE_H 1
> > +#ifndef OVN_RULE_H
> > +#define OVN_RULE_H 1
> > 
> > -/* Pipeline table translation to OpenFlow
> > - * ======================================
> > +/* Rule table translation to OpenFlow
> > + * ==================================
> >  *
> > - * The Pipeline table obtained from the OVN_Southbound database works in terms
> > + * The Rule table obtained from the OVN_Southbound database works in terms
> >  * of logical entities, that is, logical flows among logical datapaths and
> >  * logical ports.  This code translates these logical flows into OpenFlow flows
> >  * that, again, work in terms of logical entities implemented through OpenFlow
> > @@ -41,10 +41,10 @@ struct uuid;
> > #define MFF_LOG_INPORT  MFF_REG6 /* Logical input port. */
> > #define MFF_LOG_OUTPORT MFF_REG7 /* Logical output port. */
> > 
> > -void pipeline_init(void);
> > -void pipeline_run(struct controller_ctx *, struct hmap *flow_table);
> > -void pipeline_destroy(void);
> > +void rule_init(void);
> > +void rule_run(struct controller_ctx *, struct hmap *flow_table);
> > +void rule_destroy(void);
> > 
> > uint32_t ldp_to_integer(const struct uuid *logical_datapath);
> > 
> > -#endif /* ovn/pipeline.h */
> > +#endif /* ovn/rule.h */
> > diff --git a/ovn/lib/actions.c b/ovn/lib/actions.c
> > index 28be688..0aabdcf 100644
> > --- a/ovn/lib/actions.c
> > +++ b/ovn/lib/actions.c
> > @@ -176,8 +176,8 @@ parse_actions(struct action_context *ctx)
> > }
> > 
> > /* Parses OVN actions, in the format described for the "actions" column in the
> > - * Pipeline table in ovn-sb(5), and appends the parsed versions of the actions
> > - * to 'ofpacts' as "struct ofpact"s.
> > + * Rule table in ovn-sb(5), and appends the parsed versions of the actions to
> > + * 'ofpacts' as "struct ofpact"s.
> >  *
> >  * 'symtab' provides a table of "struct expr_symbol"s to support (as one would
> >  * provide to expr_parse()).
> > diff --git a/ovn/northd/ovn-northd.c b/ovn/northd/ovn-northd.c
> > index 2ad727c..eac5546 100644
> > --- a/ovn/northd/ovn-northd.c
> > +++ b/ovn/northd/ovn-northd.c
> > @@ -120,35 +120,35 @@ macs_equal(char **binding_macs_, size_t b_n_macs,
> >     return (i == b_n_macs) ? true : false;
> > }
> > 
> > -/* Pipeline generation.
> > +/* Rule generation.
> >  *
> > - * This code generates the Pipeline table in the southbound database, as a
> > + * This code generates the Rule table in the southbound database, as a
> >  * function of most of the northbound database.
> >  */
> > 
> > -/* Enough context to add a Pipeline row, using pipeline_add(). */
> > -struct pipeline_ctx {
> > +/* Enough context to add a Rule row, using rule_add(). */
> > +struct rule_ctx {
> >     /* From northd_context. */
> >     struct ovsdb_idl *ovnsb_idl;
> >     struct ovsdb_idl_txn *ovnsb_txn;
> > 
> > -    /* Contains "struct pipeline_hash_node"s.  Used to figure out what existing
> > -     * Pipeline rows should be deleted: we index all of the Pipeline rows into
> > -     * this data structure, then as existing rows are generated we remove them.
> > -     * After generating all the rows, any remaining in 'pipeline_hmap' must be
> > +    /* Contains "struct rule_hash_node"s.  Used to figure out what existing
> > +     * Rule rows should be deleted: we index all of the Rule rows into this
> > +     * data structure, then as existing rows are generated we remove them.
> > +     * After generating all the rows, any remaining in 'rule_hmap' must be
> >      * deleted from the database. */
> > -    struct hmap pipeline_hmap;
> > +    struct hmap rule_hmap;
> > };
> > 
> > -/* A row in the Pipeline table, indexed by its full contents, */
> > -struct pipeline_hash_node {
> > +/* A row in the Rule table, indexed by its full contents, */
> > +struct rule_hash_node {
> >     struct hmap_node node;
> > -    const struct sbrec_pipeline *pipeline;
> > +    const struct sbrec_rule *rule;
> > };
> > 
> > static size_t
> > -pipeline_hash(const struct uuid *logical_datapath, uint8_t table_id,
> > -              uint16_t priority, const char *match, const char *actions)
> > +rule_hash(const struct uuid *logical_datapath, uint8_t table_id,
> > +          uint16_t priority, const char *match, const char *actions)
> > {
> >     size_t hash = uuid_hash(logical_datapath);
> >     hash = hash_2words((table_id << 16) | priority, hash);
> > @@ -157,52 +157,52 @@ pipeline_hash(const struct uuid *logical_datapath, uint8_t table_id,
> > }
> > 
> > static size_t
> > -pipeline_hash_rec(const struct sbrec_pipeline *pipeline)
> > +rule_hash_rec(const struct sbrec_rule *rule)
> > {
> > -    return pipeline_hash(&pipeline->logical_datapath, pipeline->table_id,
> > -                         pipeline->priority, pipeline->match,
> > -                         pipeline->actions);
> > +    return rule_hash(&rule->logical_datapath, rule->table_id,
> > +                         rule->priority, rule->match,
> > +                         rule->actions);
> > }
> > 
> > -/* Adds a row with the specified contents to the Pipeline table. */
> > +/* Adds a row with the specified contents to the Rule table. */
> > static void
> > -pipeline_add(struct pipeline_ctx *ctx,
> > -             const struct nbrec_logical_switch *logical_datapath,
> > -             uint8_t table_id,
> > -             uint16_t priority,
> > -             const char *match,
> > -             const char *actions)
> > +rule_add(struct rule_ctx *ctx,
> > +         const struct nbrec_logical_switch *logical_datapath,
> > +         uint8_t table_id,
> > +         uint16_t priority,
> > +         const char *match,
> > +         const char *actions)
> > {
> > -    struct pipeline_hash_node *hash_node;
> > +    struct rule_hash_node *hash_node;
> > 
> > -    /* Check whether such a row already exists in the Pipeline table.  If so,
> > -     * remove it from 'ctx->pipeline_hmap' and we're done. */
> > +    /* Check whether such a row already exists in the Rule table.  If so,
> > +     * remove it from 'ctx->rule_hmap' and we're done. */
> >     HMAP_FOR_EACH_WITH_HASH (hash_node, node,
> > -                             pipeline_hash(&logical_datapath->header_.uuid,
> > -                                           table_id, priority, match, actions),
> > -                             &ctx->pipeline_hmap) {
> > -        const struct sbrec_pipeline *pipeline = hash_node->pipeline;
> > -        if (uuid_equals(&pipeline->logical_datapath,
> > +                             rule_hash(&logical_datapath->header_.uuid,
> > +                                       table_id, priority, match, actions),
> > +                             &ctx->rule_hmap) {
> > +        const struct sbrec_rule *rule = hash_node->rule;
> > +        if (uuid_equals(&rule->logical_datapath,
> >                         &logical_datapath->header_.uuid)
> > -            && pipeline->table_id == table_id
> > -            && pipeline->priority == priority
> > -            && !strcmp(pipeline->match, match)
> > -            && !strcmp(pipeline->actions, actions)) {
> > -            hmap_remove(&ctx->pipeline_hmap, &hash_node->node);
> > +            && rule->table_id == table_id
> > +            && rule->priority == priority
> > +            && !strcmp(rule->match, match)
> > +            && !strcmp(rule->actions, actions)) {
> > +            hmap_remove(&ctx->rule_hmap, &hash_node->node);
> >             free(hash_node);
> >             return;
> >         }
> >     }
> > 
> > -    /* No such Pipeline row.  Add one. */
> > -    const struct sbrec_pipeline *pipeline;
> > -    pipeline = sbrec_pipeline_insert(ctx->ovnsb_txn);
> > -    sbrec_pipeline_set_logical_datapath(pipeline,
> > +    /* No such Rule row.  Add one. */
> > +    const struct sbrec_rule *rule;
> > +    rule = sbrec_rule_insert(ctx->ovnsb_txn);
> > +    sbrec_rule_set_logical_datapath(rule,
> >                                         logical_datapath->header_.uuid);
> > -    sbrec_pipeline_set_table_id(pipeline, table_id);
> > -    sbrec_pipeline_set_priority(pipeline, priority);
> > -    sbrec_pipeline_set_match(pipeline, match);
> > -    sbrec_pipeline_set_actions(pipeline, actions);
> > +    sbrec_rule_set_table_id(rule, table_id);
> > +    sbrec_rule_set_priority(rule, priority);
> > +    sbrec_rule_set_match(rule, match);
> > +    sbrec_rule_set_actions(rule, actions);
> > }
> > 
> > /* Appends port security constraints on L2 address field 'eth_addr_field'
> > @@ -241,43 +241,43 @@ lport_is_enabled(const struct nbrec_logical_port *lport)
> >     return !lport->enabled || *lport->enabled;
> > }
> > 
> > -/* Updates the Pipeline table in the OVN_SB database, constructing its contents
> > +/* Updates the Rule table in the OVN_SB database, constructing its contents
> >  * based on the OVN_NB database. */
> > static void
> > -build_pipeline(struct northd_context *ctx)
> > +build_rule(struct northd_context *ctx)
> > {
> > -    struct pipeline_ctx pc = {
> > +    struct rule_ctx pc = {
> >         .ovnsb_idl = ctx->ovnsb_idl,
> >         .ovnsb_txn = ctx->ovnsb_txn,
> > -        .pipeline_hmap = HMAP_INITIALIZER(&pc.pipeline_hmap)
> > +        .rule_hmap = HMAP_INITIALIZER(&pc.rule_hmap)
> >     };
> > 
> > -    /* Add all the Pipeline entries currently in the southbound database to
> > -     * 'pc.pipeline_hmap'.  We remove entries that we generate from the hmap,
> > +    /* Add all the Rule entries currently in the southbound database to
> > +     * 'pc.rule_hmap'.  We remove entries that we generate from the hmap,
> >      * thus by the time we're done only entries that need to be removed
> >      * remain. */
> > -    const struct sbrec_pipeline *pipeline;
> > -    SBREC_PIPELINE_FOR_EACH (pipeline, ctx->ovnsb_idl) {
> > -        struct pipeline_hash_node *hash_node = xzalloc(sizeof *hash_node);
> > -        hash_node->pipeline = pipeline;
> > -        hmap_insert(&pc.pipeline_hmap, &hash_node->node,
> > -                    pipeline_hash_rec(pipeline));
> > +    const struct sbrec_rule *rule;
> > +    SBREC_RULE_FOR_EACH (rule, ctx->ovnsb_idl) {
> > +        struct rule_hash_node *hash_node = xzalloc(sizeof *hash_node);
> > +        hash_node->rule = rule;
> > +        hmap_insert(&pc.rule_hmap, &hash_node->node,
> > +                    rule_hash_rec(rule));
> >     }
> > 
> >     /* Table 0: Admission control framework. */
> >     const struct nbrec_logical_switch *lswitch;
> >     NBREC_LOGICAL_SWITCH_FOR_EACH (lswitch, ctx->ovnnb_idl) {
> >         /* Logical VLANs not supported. */
> > -        pipeline_add(&pc, lswitch, 0, 100, "vlan.present", "drop;");
> > +        rule_add(&pc, lswitch, 0, 100, "vlan.present", "drop;");
> > 
> >         /* Broadcast/multicast source address is invalid. */
> > -        pipeline_add(&pc, lswitch, 0, 100, "eth.src[40]", "drop;");
> > +        rule_add(&pc, lswitch, 0, 100, "eth.src[40]", "drop;");
> > 
> >         /* Port security flows have priority 50 (see below) and will continue
> >          * to the next table if packet source is acceptable. */
> > 
> >         /* Otherwise drop the packet. */
> > -        pipeline_add(&pc, lswitch, 0, 0, "1", "drop;");
> > +        rule_add(&pc, lswitch, 0, 0, "1", "drop;");
> >     }
> > 
> >     /* Table 0: Ingress port security. */
> > @@ -290,8 +290,8 @@ build_pipeline(struct northd_context *ctx)
> >             build_port_security("eth.src",
> >                                 lport->port_security, lport->n_port_security,
> >                                 &match);
> > -            pipeline_add(&pc, lswitch, 0, 50, ds_cstr(&match),
> > -                         lport_is_enabled(lport) ? "next;" : "drop;");
> > +            rule_add(&pc, lswitch, 0, 50, ds_cstr(&match),
> > +                     lport_is_enabled(lport) ? "next;" : "drop;");
> >             ds_destroy(&match);
> >         }
> >     }
> > @@ -329,8 +329,8 @@ build_pipeline(struct northd_context *ctx)
> >                     ds_put_cstr(&unicast, "outport = ");
> >                     json_string_escape(lport->name, &unicast);
> >                     ds_put_cstr(&unicast, "; next;");
> > -                    pipeline_add(&pc, lswitch, 1, 50,
> > -                                 ds_cstr(&match), ds_cstr(&unicast));
> > +                    rule_add(&pc, lswitch, 1, 50,
> > +                             ds_cstr(&match), ds_cstr(&unicast));
> >                     ds_destroy(&unicast);
> >                     ds_destroy(&match);
> >                 } else if (!strcmp(s, "unknown")) {
> > @@ -347,12 +347,12 @@ build_pipeline(struct northd_context *ctx)
> >         }
> > 
> >         ds_chomp(&bcast, ' ');
> > -        pipeline_add(&pc, lswitch, 1, 100, "eth.dst[40]", ds_cstr(&bcast));
> > +        rule_add(&pc, lswitch, 1, 100, "eth.dst[40]", ds_cstr(&bcast));
> >         ds_destroy(&bcast);
> > 
> >         if (unknown.length) {
> >             ds_chomp(&unknown, ' ');
> > -            pipeline_add(&pc, lswitch, 1, 0, "1", ds_cstr(&unknown));
> > +            rule_add(&pc, lswitch, 1, 0, "1", ds_cstr(&unknown));
> >         }
> >         ds_destroy(&unknown);
> >     }
> > @@ -363,19 +363,19 @@ build_pipeline(struct northd_context *ctx)
> >             const struct nbrec_acl *acl = lswitch->acls[i];
> > 
> >             NBREC_ACL_FOR_EACH (acl, ctx->ovnnb_idl) {
> > -                pipeline_add(&pc, lswitch, 2, acl->priority, acl->match,
> > -                             (!strcmp(acl->action, "allow") ||
> > -                              !strcmp(acl->action, "allow-related")
> > -                              ? "next;" : "drop;"));
> > +                rule_add(&pc, lswitch, 2, acl->priority, acl->match,
> > +                         (!strcmp(acl->action, "allow") ||
> > +                          !strcmp(acl->action, "allow-related")
> > +                          ? "next;" : "drop;"));
> >             }
> >         }
> > 
> > -        pipeline_add(&pc, lswitch, 2, 0, "1", "next;");
> > +        rule_add(&pc, lswitch, 2, 0, "1", "next;");
> >     }
> > 
> >     /* Table 3: Egress port security. */
> >     NBREC_LOGICAL_SWITCH_FOR_EACH (lswitch, ctx->ovnnb_idl) {
> > -        pipeline_add(&pc, lswitch, 3, 100, "eth.dst[40]", "output;");
> > +        rule_add(&pc, lswitch, 3, 100, "eth.dst[40]", "output;");
> > 
> >         for (size_t i = 0; i < lswitch->n_ports; i++) {
> >             const struct nbrec_logical_port *lport = lswitch->ports[i];
> > @@ -388,21 +388,21 @@ build_pipeline(struct northd_context *ctx)
> >                                 lport->port_security, lport->n_port_security,
> >                                 &match);
> > 
> > -            pipeline_add(&pc, lswitch, 3, 50, ds_cstr(&match),
> > +            rule_add(&pc, lswitch, 3, 50, ds_cstr(&match),
> >                          lport_is_enabled(lport) ? "output;" : "drop;");
> > 
> >             ds_destroy(&match);
> >         }
> >     }
> > 
> > -    /* Delete any existing Pipeline rows that were not re-generated.  */
> > -    struct pipeline_hash_node *hash_node, *next_hash_node;
> > -    HMAP_FOR_EACH_SAFE (hash_node, next_hash_node, node, &pc.pipeline_hmap) {
> > -        hmap_remove(&pc.pipeline_hmap, &hash_node->node);
> > -        sbrec_pipeline_delete(hash_node->pipeline);
> > +    /* Delete any existing Rule rows that were not re-generated.  */
> > +    struct rule_hash_node *hash_node, *next_hash_node;
> > +    HMAP_FOR_EACH_SAFE (hash_node, next_hash_node, node, &pc.rule_hmap) {
> > +        hmap_remove(&pc.rule_hmap, &hash_node->node);
> > +        sbrec_rule_delete(hash_node->rule);
> >         free(hash_node);
> >     }
> > -    hmap_destroy(&pc.pipeline_hmap);
> > +    hmap_destroy(&pc.rule_hmap);
> > }
> > 
> > static bool
> > @@ -610,7 +610,7 @@ ovnnb_db_changed(struct northd_context *ctx)
> >     VLOG_DBG("ovn-nb db contents have changed.");
> > 
> >     set_port_bindings(ctx);
> > -    build_pipeline(ctx);
> > +    build_rule(ctx);
> > }
> > 
> > /*
> > @@ -804,16 +804,16 @@ main(int argc, char *argv[])
> >     ovsdb_idl_add_column(ovnsb_idl, &sbrec_port_binding_col_parent_port);
> >     ovsdb_idl_add_column(ovnsb_idl, &sbrec_port_binding_col_logical_datapath);
> >     ovsdb_idl_add_column(ovnsb_idl, &sbrec_port_binding_col_tunnel_key);
> > -    ovsdb_idl_add_column(ovnsb_idl, &sbrec_pipeline_col_logical_datapath);
> > -    ovsdb_idl_omit_alert(ovnsb_idl, &sbrec_pipeline_col_logical_datapath);
> > -    ovsdb_idl_add_column(ovnsb_idl, &sbrec_pipeline_col_table_id);
> > -    ovsdb_idl_omit_alert(ovnsb_idl, &sbrec_pipeline_col_table_id);
> > -    ovsdb_idl_add_column(ovnsb_idl, &sbrec_pipeline_col_priority);
> > -    ovsdb_idl_omit_alert(ovnsb_idl, &sbrec_pipeline_col_priority);
> > -    ovsdb_idl_add_column(ovnsb_idl, &sbrec_pipeline_col_match);
> > -    ovsdb_idl_omit_alert(ovnsb_idl, &sbrec_pipeline_col_match);
> > -    ovsdb_idl_add_column(ovnsb_idl, &sbrec_pipeline_col_actions);
> > -    ovsdb_idl_omit_alert(ovnsb_idl, &sbrec_pipeline_col_actions);
> > +    ovsdb_idl_add_column(ovnsb_idl, &sbrec_rule_col_logical_datapath);
> > +    ovsdb_idl_omit_alert(ovnsb_idl, &sbrec_rule_col_logical_datapath);
> > +    ovsdb_idl_add_column(ovnsb_idl, &sbrec_rule_col_table_id);
> > +    ovsdb_idl_omit_alert(ovnsb_idl, &sbrec_rule_col_table_id);
> > +    ovsdb_idl_add_column(ovnsb_idl, &sbrec_rule_col_priority);
> > +    ovsdb_idl_omit_alert(ovnsb_idl, &sbrec_rule_col_priority);
> > +    ovsdb_idl_add_column(ovnsb_idl, &sbrec_rule_col_match);
> > +    ovsdb_idl_omit_alert(ovnsb_idl, &sbrec_rule_col_match);
> > +    ovsdb_idl_add_column(ovnsb_idl, &sbrec_rule_col_actions);
> > +    ovsdb_idl_omit_alert(ovnsb_idl, &sbrec_rule_col_actions);
> > 
> >     /*
> >      * The loop here just runs the IDL in a loop waiting for the seqno to
> > diff --git a/ovn/ovn-architecture.7.xml b/ovn/ovn-architecture.7.xml
> > index 5d95e26..0334d82 100644
> > --- a/ovn/ovn-architecture.7.xml
> > +++ b/ovn/ovn-architecture.7.xml
> > @@ -346,7 +346,7 @@
> >       <code>ovn-northd</code> receives the OVN Northbound database update.
> >       In turn, it makes the corresponding updates to the OVN Southbound
> >       database, by adding rows to the OVN Southbound database
> > -      <code>Pipeline</code> table to reflect the new port, e.g. add a
> > +      <code>Rule</code> table to reflect the new port, e.g. add a
> >       flow to recognize that packets destined to the new port's MAC
> >       address should be delivered to it, and update the flow that
> >       delivers broadcast and multicast packets to include the new port.
> > @@ -357,7 +357,7 @@
> > 
> >     <li>
> >       On every hypervisor, <code>ovn-controller</code> receives the
> > -      <code>Pipeline</code> table updates that <code>ovn-northd</code> made
> > +      <code>Rule</code> table updates that <code>ovn-northd</code> made
> >       in the previous step.  As long as the VM that owns the VIF is powered off,
> >       <code>ovn-controller</code> cannot do much; it cannot, for example,
> >       arrange to send packets to or receive packets from the VIF, because the
> > @@ -404,7 +404,7 @@
> >       <code>Binding</code> table.  This provides <code>ovn-controller</code>
> >       the physical location of the logical port, so each instance updates the
> >       OpenFlow tables of its switch (based on logical datapath flows in the OVN
> > -      DB <code>Pipeline</code> table) so that packets to and from the VIF can
> > +      DB <code>Rule</code> table) so that packets to and from the VIF can
> >       be properly handled via tunnels.
> >     </li>
> > 
> > @@ -444,13 +444,13 @@
> >       <code>ovn-northd</code> receives the OVN Northbound update and in turn
> >       updates the OVN Southbound database accordingly, by removing or
> >       updating the rows from the OVN Southbound database
> > -      <code>Pipeline</code> table and <code>Binding</code> table that
> > +      <code>Rule</code> table and <code>Binding</code> table that
> >       were related to the now-destroyed VIF.
> >     </li>
> > 
> >     <li>
> >       On every hypervisor, <code>ovn-controller</code> receives the
> > -      <code>Pipeline</code> table updates that <code>ovn-northd</code> made
> > +      <code>Rule</code> table updates that <code>ovn-northd</code> made
> >       in the previous step.  <code>ovn-controller</code> updates OpenFlow tables
> >       to reflect the update, although there may not be much to do, since the VIF
> >       had already become unreachable when it was removed from the
> > @@ -541,7 +541,7 @@
> >       <code>ovn-northd</code> receives the OVN Northbound database update.
> >       In turn, it makes the corresponding updates to the OVN Southbound
> >       database, by adding rows to the OVN Southbound database's
> > -      <code>Pipeline</code> table to reflect the new port and also by
> > +      <code>Rule</code> table to reflect the new port and also by
> >       creating a new row in the <code>Binding</code> table and
> >       populating all its columns except the column that identifies the
> >       <code>chassis</code>.
> > @@ -582,16 +582,16 @@
> >       <code>ovn-northd</code> receives the OVN Northbound update and in turn
> >       updates the OVN Southbound database accordingly, by removing or
> >       updating the rows from the OVN Southbound database
> > -      <code>Pipeline</code> table that were related to the now-destroyed
> > +      <code>Rule</code> table that were related to the now-destroyed
> >       CIF.  It also deletes the row in the <code>Binding</code> table
> >       for that CIF.
> >     </li>
> > 
> >     <li>
> >       On every hypervisor, <code>ovn-controller</code> receives the
> > -      <code>Pipeline</code> table updates that <code>ovn-northd</code> made
> > -      in the previous step.  <code>ovn-controller</code> updates OpenFlow tables
> > -      to reflect the update.
> > +      <code>Rule</code> table updates that <code>ovn-northd</code> made in the
> > +      previous step.  <code>ovn-controller</code> updates OpenFlow tables to
> > +      reflect the update.
> >     </li>
> >   </ol>
> > 
> > diff --git a/ovn/ovn-nb.xml b/ovn/ovn-nb.xml
> > index 032e23d..d953fa5 100644
> > --- a/ovn/ovn-nb.xml
> > +++ b/ovn/ovn-nb.xml
> > @@ -202,9 +202,9 @@
> > 
> >     <column name="match">
> >       The packets that the ACL should match, in the same expression
> > -      language used for the <ref column="match" table="Pipeline"
> > +      language used for the <ref column="match" table="Rule"
> >       db="OVN_Southbound"/> column in the OVN Southbound database's <ref
> > -      table="Pipeline" db="OVN_Southbound"/> table.  Match
> > +      table="Rule" db="OVN_Southbound"/> table.  Match
> >       <code>inport</code> and <code>outport</code> against names of
> >       logical ports within <ref column="lswitch"/> to implement ingress
> >       and egress ACLs, respectively.  In logical switches connected to
> > diff --git a/ovn/ovn-sb.ovsschema b/ovn/ovn-sb.ovsschema
> > index 4a2df47..add908b 100644
> > --- a/ovn/ovn-sb.ovsschema
> > +++ b/ovn/ovn-sb.ovsschema
> > @@ -32,7 +32,7 @@
> >                                               "value": {"type": "string"},
> >                                               "min": 0,
> >                                               "max": "unlimited"}}}},
> > -        "Pipeline": {
> > +        "Rule": {
> >             "columns": {
> >                 "logical_datapath": {"type": "uuid"},
> >                 "table_id": {"type": {"key": {"type": "integer",
> > diff --git a/ovn/ovn-sb.xml b/ovn/ovn-sb.xml
> > index 13e5145..2f2a55e 100644
> > --- a/ovn/ovn-sb.xml
> > +++ b/ovn/ovn-sb.xml
> > @@ -74,7 +74,7 @@
> >   </p>
> > 
> >   <p>
> > -    The <ref table="Pipeline"/> table is currently the only LN table.
> > +    The <ref table="Rule"/> table is currently the only LN table.
> >   </p>
> > 
> >   <h3>Bindings data</h3>
> > @@ -198,7 +198,7 @@
> >     </column>
> >   </table>
> > 
> > -  <table name="Pipeline" title="Logical Network Pipeline">
> > +  <table name="Rule" title="Logical Network Rule">
> >     <p>
> >       Each row in this table represents one logical flow.  The cloud management
> >       system, via its OVN integration, populates this table with logical flows
> > @@ -663,7 +663,7 @@
> >     <column name="logical_datapath">
> >       The logical datapath to which the logical port belongs.  A logical
> >       datapath implements a logical pipeline via logical flows in the <ref
> > -      table="Pipeline"/> table.  (No table represents a logical datapath.)
> > +      table="Rule"/> table.  (No table represents a logical datapath.)
> >     </column>
> > 
> >     <column name="logical_port">
> > -- 
> > 2.1.3
> > 
> > _______________________________________________
> > dev mailing list
> > dev at openvswitch.org
> > http://openvswitch.org/mailman/listinfo/dev
> 



More information about the dev mailing list