[ovs-discuss] [bug] a new problem of ovs-dpdk
Chandran, Sugesh
sugesh.chandran at intel.com
Fri Mar 11 17:13:23 UTC 2016
Are you using multiqueue for this test setup??
Can you please try with Qemu >=2.4 version instead of 2.3?
Regards
_Sugesh
From: discuss [mailto:discuss-bounces at openvswitch.org] On Behalf Of ???
Sent: Friday, March 11, 2016 9:54 AM
To: discuss at openvswitch.org
Subject: [ovs-discuss] [bug] a new problem of ovs-dpdk
Hi all.
These days ,I tested the vm by using ovs-dpdk ,and found a new problem as fellows:
My test environment :
Host:
Linux version 3.10.0-229.14.1.el7.x86_64 (builder at kbuilder.dev.centos.org<mailto:builder at kbuilder.dev.centos.org>) (gcc version 4.8.3 20140911 (Red Hat 4.8.3-9) (GCC) ) #1 SMP Tue Sep 15 15:05:51 UTC 2015
Dpdk: version 2.2
Ovs: version 2.5
QEMU version 2.3.1,
Guest : Linux version 3.10.0-229.el7.x86_64 (builder at kbuilder.dev.centos.org<mailto:builder at kbuilder.dev.centos.org>) (gcc version 4.8.2 20140120 (Red Hat 4.8.2-16) (GCC) ) #1 SMP Fri Mar 6 11:36:42 UTC 2015
Ovs:
1 S root 61984 1 0 80 0 - 11923 poll_s 14:29 ? 00:00:31 ovsdb-server -v --remote=punix:/usr/local/var/run/openvswitch/db.sock --remote=db:Open_vSwitch,Open_vSwitch,manager_options --private-key=db:Open_vSwitch,SSL,private_key --certificate=db:Open_vSwitch,SSL,certificate --bootstrap-ca-cert=db:Open_vSwitch,SSL,ca_cert --pidfile --detach
5 S root 61998 1 99 80 0 - 2745735 poll_s 14:29 ? 12:25:52 /usr/local/sbin/ovs-vswitchd --dpdk -c 0x1 -n 4 --socket-mem 400 -- unix:/usr/local/var/run/openvswitch/db.sock --pidfile --detach
Ovs-port cfg:
Bridge "br1"
Port "dpdk0"
Interface "dpdk0"
type: dpdk
Port "br1"
Interface "br1"
type: internal
Bridge "br0"
Port "br0"
Interface "br0"
type: internal
Port "vxlan-1"
Interface "vxlan-1"
type: vxlan
options: {remote_ip="7.0.0.2"}
Port "vhost-user-0"
Interface "vhost-user-0"
type: dpdkvhostuser
The numa huagepages and vcpu configes in xml :
<memoryBacking>
<hugepages>
<page size='2048' unit='KiB'/>
</hugepages>
</memoryBacking>
<vcpu placement='static'>4</vcpu>
<os>
<type arch='x86_64' machine='pc-i440fx-2.2'>hvm</type>
<boot dev='hd'/>
</os>
<features>
<acpi/>
</features>
<cpu>
<numa>
<cell id='0' cpus='0-3' memory='4000000' unit='KiB' memAccess='shared'/>
</numa>
Test steps:
Step 1: create a vm for port vhost-user-0
Step 2: create other 15 vms likes step 1
Step 3 :destroy the 15 vms created by step 2
Step 4 repeat step 2 and step 3
Then sometimes I find I can’t reach the vm created by step 1, every time, this occurred in step 2
The logs of guest as fellows:
localhost kernel: virtio_net virtio0: output.0:id 222 is not a head!
localhost kernel: net eth0: Unexpected TXQ (0) queue failure: -5
localhost kernel: net eth0: Unexpected TXQ (0) queue failure: -5
localhost kernel: net eth0: Unexpected TXQ (0) queue failure: -5
then reboot the vm it can be recovered.
It is obvious that this is network problem. In my opinion, this the problems of the hugepages. But indeed , the hugepages has been mapped when the ovs-dpdk started by “/usr/local/sbin/ovs-vswitchd --dpdk -c 0x1 -n 4 --socket-mem 400 -- unix:/usr/local/var/run/openvswitch/db.sock --pidfile --detach”I am not sure. How the hugepages recovery when the vm is destroyed. If somebody met the same problem with me , I am do not know if this is a bug of ovs-dpdk,or dpdk???
Thanks
Eric wang
Wang Huaxia ---JD.COM
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://openvswitch.org/pipermail/ovs-discuss/attachments/20160311/b9e6a257/attachment-0002.html>
More information about the discuss
mailing list