[ovs-discuss] Sending UDP traffic in openflow network
Shan Hu
shan.hu at utdallas.edu
Fri Oct 28 23:27:30 UTC 2011
Hi Ben,
I just tried the Queue tables, what i used is linux-htb which seems to be also a token buket approach,so when i sent UDP traffic,t here is still packet lost. Then i tried to use linux-hfsc(i just simply follow the configuration cookbook of linux-htb, and change linux-htb to linux-hfsc),there is still packet lost, maybe im wrong with the configuration of linux-hfsc since the rate limiting doesnt work, but is there a way to stop packet lost in UDP traffic?
Shan
----- Original Message -----
From: "Shan Hu" <shan.hu at utdallas.edu>
To: "Ben Pfaff" <blp at nicira.com>
Cc: discuss at openvswitch.org
Sent: Wednesday, October 26, 2011 10:55:20 PM
Subject: Re: [ovs-discuss] Sending UDP traffic in openflow network
Thank you for the reply, Ben.
Yes im using policing, but even when i set policing_rate and burst to 0,there is still packet lost.
I will try to use the Queue tables anyway.
But also i have question for this Queue tables QoS,will bandwidth of one queue be
reserved all the time? By that i mean,say i reserve 20Mbps for queue1,and there is another queue2 i reserved
90Mbps for, however,the total bandwidth of this link is only 100Mbps. So when i push 20Mbps and 90Mbps
data to two queues,respectively, will they compromise to each other?
regards,
Shan
----- Original Message -----
From: "Ben Pfaff" <blp at nicira.com>
To: "Shan Hu" <shan.hu at utdallas.edu>
Cc: discuss at openvswitch.org
Sent: Wednesday, October 26, 2011 10:24:13 PM
Subject: Re: [ovs-discuss] Sending UDP traffic in openflow network
On Wed, Oct 26, 2011 at 10:22:18PM -0500, Shan Hu wrote:
> Im trying to test the QoS Rate-Limiting of Kernel vSwitch, i use iperf
> as my measurement tool. Everything is working fine with TCP part,that
> is, after i limit rate of one port to, say 50Mbps, the rate is limited
> to 50Mbps correctly and packets are tranferred 100%. But when i turn
> to UDP part,i ran into problems.I have to limit bandwidth to at most
> 4Mbps in order to tranfer 100% packets.And if i use bandwidth more
> than 4Mbps, the lost packets increase, the worst packet lost
> percentage is almost 99%.
Are you using policing? The documentation says a lot about problems
with policing:
Ingress Policing:
These settings control ingress policing for packets received on this
interface. On a physical interface, this limits the rate at which
traffic is allowed into the system from the outside; on a virtual
interface (one connected to a virtual machine), this limits the rate at
which the VM is able to transmit.
Policing is a simple form of quality-of-service that simply drops pack-
ets received in excess of the configured rate. Due to its simplicity,
policing is usually less accurate and less effective than egress QoS
(which is configured using the QoS and Queue tables).
Policing is currently implemented only on Linux. The Linux implementa-
tion uses a simple ``token bucket'' approach:
o The size of the bucket corresponds to ingress_polic-
ing_burst. Initially the bucket is full.
o Whenever a packet is received, its size (converted to
tokens) is compared to the number of tokens currently in
the bucket. If the required number of tokens are avail-
able, they are removed and the packet is forwarded. Oth-
erwise, the packet is dropped.
o Whenever it is not full, the bucket is refilled with
tokens at the rate specified by ingress_policing_rate.
Policing interacts badly with some network protocols, and especially
with fragmented IP packets. Suppose that there is enough network
activity to keep the bucket nearly empty all the time. Then this token
bucket algorithm will forward a single packet every so often, with the
period depending on packet size and on the configured rate. All of the
fragments of an IP packets are normally transmitted back-to-back, as a
group. In such a situation, therefore, only one of these fragments
will be forwarded and the rest will be dropped. IP does not provide
any way for the intended recipient to ask for only the remaining frag-
ments. In such a case there are two likely possibilities for what will
happen next: either all of the fragments will eventually be retransmit-
ted (as TCP will do), in which case the same problem will recur, or the
sender will not realize that its packet has been dropped and data will
simply be lost (as some UDP-based protocols will do). Either way, it
is possible that no forward progress will ever occur.
ingress_policing_rate: integer, at least 0
Maximum rate for data received on this interface, in kbps. Data
received faster than this rate is dropped. Set to 0 (the
default) to disable policing.
ingress_policing_burst: integer, at least 0
Maximum burst size for data received on this interface, in kb.
The default burst size if set to 0 is 1000 kb. This value has
no effect if ingress_policing_rate is 0.
Specifying a larger burst size lets the algorithm be more for-
giving, which is important for protocols like TCP that react se-
verely to dropped packets. The burst size should be at least
the size of the interface's MTU. Specifying a value that is
numerically at least as large as 10% of ingress_policing_rate
helps TCP come closer to achieving the full rate.
If you're not using policing, please tell us about your configuration.
_______________________________________________
discuss mailing list
discuss at openvswitch.org
http://openvswitch.org/mailman/listinfo/discuss
More information about the discuss
mailing list