[ovs-dev] [PATCH] netdev-dpdk: Add a statistic on vhost tx contention.
David Marchand
david.marchand at redhat.com
Sun Jul 7 16:13:20 UTC 2019
Add a statistic to help diagnose contention on the vhost txqs.
This is usually seen as dropped packets on the physical ports for rates
that are usually handled fine by OVS.
Signed-off-by: David Marchand <david.marchand at redhat.com>
---
This patch applies on top of Kevin "vhost-retry" series [1].
1: https://patchwork.ozlabs.org/project/openvswitch/list/?submitter=70482
---
Documentation/topics/dpdk/vhost-user.rst | 25 +++++++++++++++++++++++++
lib/netdev-dpdk.c | 20 ++++++++++++++++++--
2 files changed, 43 insertions(+), 2 deletions(-)
diff --git a/Documentation/topics/dpdk/vhost-user.rst b/Documentation/topics/dpdk/vhost-user.rst
index 724aa62..e276b21 100644
--- a/Documentation/topics/dpdk/vhost-user.rst
+++ b/Documentation/topics/dpdk/vhost-user.rst
@@ -553,6 +553,31 @@ shown with::
$ ovs-vsctl get Interface dpdkvhostclient0 statistics:tx_retries
+vhost tx contention
+-------------------
+
+Contrary to physical ports where the number of Transmit queues is decided at
+the configuration of the port and does not change once set up, the number of
+Transmit queues for a vhost-user or vhost-user-client interface is negociated
+with the guest when its driver enables a virtio interface and OVS has no
+control over it.
+
+Each PMD threads needs a Transmit queue to send packet on such a port.
+To accomodate with the cases where the number of PMD threads differs from the
+number of Transmit queues enabled by a Virtual Machine, the Transmit queues are
+distributed among the PMD threads and their accesses are protected by a
+spinlock.
+
+When contention occurs (because two PMD threads are mapped to the same Transmit
+queue), it can be hard to understand this phenomenon.
+It usually will be seen as OVS taking more cycles to handle the packet and can
+be seen as packets getting dropped on the receive side (rx_dropped statistic
+on the physical ports).
+
+A statistic for the port is available to catch those cases::
+
+ ovs-vsctl get Interface dpdkvhostclient0 statistics:tx_contentions
+
vhost-user Dequeue Zero Copy (experimental)
-------------------------------------------
diff --git a/lib/netdev-dpdk.c b/lib/netdev-dpdk.c
index 0e70e44..844fb3c 100644
--- a/lib/netdev-dpdk.c
+++ b/lib/netdev-dpdk.c
@@ -139,10 +139,11 @@ BUILD_ASSERT_DECL((MAX_NB_MBUF / ROUND_DOWN_POW2(MAX_NB_MBUF / MIN_NB_MBUF))
#define XSTAT_RX_JABBER_ERRORS "rx_jabber_errors"
/* Size of vHost custom stats. */
-#define VHOST_CUSTOM_STATS_SIZE 1
+#define VHOST_CUSTOM_STATS_SIZE 2
/* Names of vHost custom stats. */
#define VHOST_STAT_TX_RETRIES "tx_retries"
+#define VHOST_STAT_TX_CONTENTIONS "tx_contentions"
#define SOCKET0 0
@@ -340,6 +341,8 @@ struct dpdk_tx_queue {
* pmd threads (see 'concurrent_txq'). */
int map; /* Mapping of configured vhost-user queues
* to enabled by guest. */
+ uint64_t tx_contentions; /* Number of times a pmd has been waiting
+ * for another to xmit on this queue. */
};
/* dpdk has no way to remove dpdk ring ethernet devices
@@ -2387,7 +2390,10 @@ __netdev_dpdk_vhost_send(struct netdev *netdev, int qid,
goto out;
}
- rte_spinlock_lock(&dev->tx_q[qid].tx_lock);
+ if (unlikely(!rte_spinlock_trylock(&dev->tx_q[qid].tx_lock))) {
+ rte_spinlock_lock(&dev->tx_q[qid].tx_lock);
+ atomic_count_inc64(&dev->tx_q[qid].tx_contentions);
+ }
cnt = netdev_dpdk_filter_packet_len(dev, cur_pkts, cnt);
/* Check has QoS has been configured for the netdev */
@@ -2878,6 +2884,16 @@ netdev_dpdk_vhost_get_custom_stats(const struct netdev *netdev,
custom_stats->counters[0].value = dev->tx_retries;
rte_spinlock_unlock(&dev->stats_lock);
+ ovs_strlcpy(custom_stats->counters[1].name, VHOST_STAT_TX_CONTENTIONS,
+ NETDEV_CUSTOM_STATS_NAME_SIZE);
+ custom_stats->counters[1].value = 0;
+ for (unsigned int i = 0; i < netdev->n_txq; i++) {
+ uint64_t tx_contentions;
+
+ atomic_read_relaxed(&dev->tx_q[i].tx_contentions, &tx_contentions);
+ custom_stats->counters[1].value += tx_contentions;
+ }
+
ovs_mutex_unlock(&dev->mutex);
return 0;
--
1.8.3.1
More information about the dev
mailing list