[ovs-discuss] unable to get all packets from the physical interface to OVS bridge
ian.stokes at intel.com
Wed Sep 25 15:42:42 UTC 2019
On 9/18/2019 5:22 AM, Anish M wrote:
> Im trying to forward/mirror VM packets between two ovs-dpdk compute
> nodes. With the help of
> http://docs.openvswitch.org/en/latest/howto/userspace-tunneling/ link, i
> could able to setup additional NIC for forwarding the packet from source
> ovs-dpdk compute node to the destination compute node. I could able to
> see all the vxlan forwarded packets at the destination compute node's
> additional NIC. But the same is not visible at the ovs bridge where is
> attached that additional NIC.
Hi, can you please clarify the sentence above? Is it that you can see
packets arrive at NIC but the packets are not seen routing through the
> At both the compute nodes, i have same type of 2 port 10G NIC. ens1f0 &
> ens1f0 is acting as DPDK port and im using it inside the openstack for
> DPDK VMs at both compute nodes.
Just to clarify, ens1f0 is a dpdk port type on each node and ens1f1 is
not? I.e. is ens1f1 a netdev linux device?
> In order to mirror DPDK VM traffic from one compute node to another
> compute node, i followed the above userspace-tunnelling link and able to
> forward VM traffic from source compute node towards the
> ens1f1(172.28.41.101) of the destination compute node.
Can you list the mirroring commands you used to set this up?
> ovs-vsctl --may-exist add-br br-phy \
> -- set Bridge br-phy datapath_type=netdev \
> -- br-set-external-id br-phy bridge-id br-phy \
> -- set bridge br-phy fail-mode=standalone \
> ovs-vsctl --timeout 10 add-port br-phy ens1f1
> ip addr add 172.28.41.101/24 dev br-phy
> ip link set br-phy up
> ip addr flush dev ens1f1 2>/dev/null
> ip link set ens1f1 up
> Even though i receiving all the mirrored vxlan packets at ens1f1 port
> (checked using tcpdump), the same number of packets are not available
> inside the ovs br-phy bridge (only ~10% of the mirrored traffic is
> available inside the br-phy)
Just to be aware, low volumes of traffic are fine for mirroring but at
high volumes it would be expected that you would not see the same amount
of traffic mirrored due to the overhead associated with mirroring. In
this case are you sending high volumes of traffic? Have you tested with
smaller bursts of traffic?
> [root at overcloud-hpcomputeovsdpdk-0 ~]# ovs-ofctl dump-flows br-phy
> cookie=0x3435, duration=46778.969s, table=0, n_packets=30807,
> n_bytes=3889444, priority=0 actions=NORMAL
> [root at overcloud-hpcomputeovsdpdk-0 ~]# ovs-ofctl dump-ports br-phy
> OFPST_PORT reply (xid=0x2): 3 ports
> port LOCAL: rx pkts=77, bytes=5710, drop=0, errs=0, frame=0, over=0,
> tx pkts=30107, bytes=3604094, drop=28077, errs=0, coll=0
> port ens1f1: rx pkts=4385104, bytes=596066321, drop=0, errs=0,
> frame=0, over=0, crc=0
> tx pkts=33606, bytes=4021526, drop=0, errs=0, coll=0
> port "patch-tap-bint": rx pkts=30056, bytes=3591859, drop=?, errs=?,
> frame=?, over=?, crc=?
> tx pkts=739, bytes=294432, drop=?, errs=?, coll=?
> In ens1f1, i could see lot of rx pkts, but in the br-phy flow im seeing
> only few packets.
> Please provide any advice how i can mirror/forward packets between two
> OVS-DPDK compute nodes.
A diagram of your setup would be useful to help debug/understand the use
case. I'm slightly confused with the mix of DPDK/non-DPDK ports you have
vs what you want to achieve with mirroring. If you could provide more
info on this as well as the expected flow of a packet as it is
mirrored/forwarded it would be helpful.
> Best Regards,
> discuss mailing list
> discuss at openvswitch.org
More information about the discuss