[ovs-dev] [PATCH V3 10/19] dpif-netdev: Read hw stats during flow_dump_next() call

Eli Britstein elibr at mellanox.com
Sun Dec 8 13:22:55 UTC 2019


From: Ophir Munk <ophirmu at mellanox.com>

Use netdev flow get API in order to update the statistics of fully
offloaded flows.

Co-authored-by: Eli Britstein <elibr at mellanox.com>
Signed-off-by: Ophir Munk <ophirmu at mellanox.com>
Reviewed-by: Oz Shlomo <ozsh at mellanox.com>
Signed-off-by: Eli Britstein <elibr at mellanox.com>
---
 lib/dpif-netdev.c | 47 +++++++++++++++++++++++++++++++++++++++++++++--
 1 file changed, 45 insertions(+), 2 deletions(-)

diff --git a/lib/dpif-netdev.c b/lib/dpif-netdev.c
index 1e5493615..37e7e5c38 100644
--- a/lib/dpif-netdev.c
+++ b/lib/dpif-netdev.c
@@ -527,6 +527,7 @@ struct dp_netdev_flow {
 
     bool dead;
     uint32_t mark;               /* Unique flow mark assigned to a flow */
+    bool actions_offloaded;      /* true if flow is fully offloaded. */
 
     /* Statistics. */
     struct dp_netdev_flow_stats stats;
@@ -2409,6 +2410,7 @@ dp_netdev_flow_offload_put(struct dp_flow_offload_item *offload)
         }
     }
     info.flow_mark = mark;
+    info.actions_offloaded = &flow->actions_offloaded;
 
     port = netdev_ports_get(in_port, pmd->dp->dpif->dpif_class);
     if (!port || netdev_vport_is_vport_class(port->netdev_class)) {
@@ -3071,8 +3073,8 @@ dp_netdev_flow_to_dpif_flow(const struct dp_netdev_flow *netdev_flow,
     flow->pmd_id = netdev_flow->pmd_id;
     get_dpif_flow_stats(netdev_flow, &flow->stats);
 
-    flow->attrs.offloaded = false;
-    flow->attrs.dp_layer = "ovs";
+    flow->attrs.offloaded = netdev_flow->actions_offloaded;
+    flow->attrs.dp_layer = flow->attrs.offloaded ? "in_hw" : "ovs";
 }
 
 static int
@@ -3242,6 +3244,7 @@ dp_netdev_flow_add(struct dp_netdev_pmd_thread *pmd,
     flow->dead = false;
     flow->batch = NULL;
     flow->mark = INVALID_FLOW_MARK;
+    flow->actions_offloaded = false;
     *CONST_CAST(unsigned *, &flow->pmd_id) = pmd->core_id;
     *CONST_CAST(struct flow *, &flow->flow) = match->flow;
     *CONST_CAST(ovs_u128 *, &flow->ufid) = *ufid;
@@ -3596,6 +3599,42 @@ dpif_netdev_flow_dump_thread_destroy(struct dpif_flow_dump_thread *thread_)
     free(thread);
 }
 
+static int
+dpif_netdev_offload_used(struct dp_netdev_flow *netdev_flow,
+                         struct dp_netdev_pmd_thread *pmd)
+{
+    struct dpif_flow_stats stats;
+    struct netdev *netdev;
+    struct match match;
+    struct nlattr *actions;
+    struct dpif_flow_attrs attrs;
+    struct ofpbuf wbuffer;
+
+    ovs_u128 ufid;
+    int ret = 0;
+
+    netdev = netdev_ports_get(netdev_flow->flow.in_port.odp_port,
+                              pmd->dp->class);
+    if (!netdev) {
+        return -1;
+    }
+    /* get offloaded stats */
+    ufid = netdev_flow->mega_ufid;
+    ret = netdev_flow_get(netdev, &match, &actions, &ufid, &stats, &attrs,
+                          &wbuffer);
+    netdev_close(netdev);
+    if (ret) {
+        return -1;
+    }
+    if (stats.n_packets) {
+        atomic_store_relaxed(&netdev_flow->stats.used, pmd->ctx.now / 1000);
+        non_atomic_ullong_add(&netdev_flow->stats.packet_count, stats.n_packets);
+        non_atomic_ullong_add(&netdev_flow->stats.byte_count, stats.n_bytes);
+    }
+
+    return 0;
+}
+
 static int
 dpif_netdev_flow_dump_next(struct dpif_flow_dump_thread *thread_,
                            struct dpif_flow *flows, int max_flows)
@@ -3636,6 +3675,10 @@ dpif_netdev_flow_dump_next(struct dpif_flow_dump_thread *thread_,
                 netdev_flows[n_flows] = CONTAINER_OF(node,
                                                      struct dp_netdev_flow,
                                                      node);
+                /* Read hardware stats in case of hardware offload */
+                if (netdev_flows[n_flows]->actions_offloaded) {
+                    dpif_netdev_offload_used(netdev_flows[n_flows], pmd);
+                }
             }
             /* When finishing dumping the current pmd thread, moves to
              * the next. */
-- 
2.14.5



More information about the dev mailing list