[ovs-discuss] Inactivity Probe configuration not taking effect in OVS 2.14.0
Saurabh Deokate
saurabh.deokate at nutanix.com
Mon Jun 28 14:44:02 UTC 2021
Hi Ben,
Any update on this ? Do you need any other details ?
~Saurabh.
On 15/06/21, 12:54 PM, "Saurabh Deokate" <saurabh.deokate at nutanix.com> wrote:
Hi Ben,
Here is the output for ovs-vsctl list controller
[root at 172-31-64-26-aws-eu-central-1c ~]# ovs-vsctl list controller
_uuid : eb56176a-ad32-4eb0-9cd8-7ab3bd448a68
connection_mode : out-of-band
controller_burst_limit: []
controller_queue_size: []
controller_rate_limit: []
enable_async_messages: []
external_ids : {}
inactivity_probe : 0
is_connected : true
local_gateway : []
local_ip : []
local_netmask : []
max_backoff : []
other_config : {}
role : other
status : {last_error="Connection refused", sec_since_connect="42606", sec_since_disconnect="42614", state=ACTIVE}
target : "tcp:127.0.0.1:6653"
type : []
Let me know if you need any other details.
~Saurabh.
On 11/06/21, 4:03 AM, "Ben Pfaff" <blp at ovn.org> wrote:
On Mon, Jun 07, 2021 at 02:51:58PM +0000, Saurabh Deokate wrote:
> Hi Team,
>
> We are seeing an issue in OVS 2.14.0 after moving from 2.8.0. We first set the controller on the bridge and then set inactivity probe for our controller to 0 to disable new connection attempts by ovs. After this we start our controller to serve request. But in the new version of OVS somehow we still see inactivity probe kicking in after every 5s and trying to reconnect. This issue is triggered when we are in the middle of handling a packet in our controller (i.e. OFController) which is blocked for almost 40s.
>
>
> Kernel version: CentOS Linux release 7.9.2009
> Output of ovs-vsctl list controller command shows inactivity_probe: 0
>
> Below is the snippet from ovs-vswitchd.log
>
> 021-05-11T22:32:55.378Z|00608|rconn|INFO|br0.uvms<->tcp:127.0.0.1:6653: connected
> 2021-05-11T22:33:05.382Z|00609|connmgr|INFO|br0.uvms<->tcp:127.0.0.1:6653: 44 flow_mods 10 s ago (44 adds)
> 2021-05-11T22:33:05.386Z|00610|rconn|ERR|br0.uvms<->tcp:127.0.0.1:6653: no response to inactivity probe after 5 seconds, disconnecting
> 2021-05-11T22:33:06.406Z|00611|rconn|INFO|br0.uvms<->tcp:127.0.0.1:6653: connecting...
> 2021-05-11T22:33:06.438Z|00612|rconn|INFO|br0.uvms<->tcp:127.0.0.1:6653: connected
> 2021-05-11T22:33:16.438Z|00613|rconn|ERR|br0.uvms<->tcp:127.0.0.1:6653: no response to inactivity probe after 5 seconds, disconnecting
> 2021-05-11T22:33:17.921Z|00614|rconn|INFO|br0.uvms<->tcp:127.0.0.1:6653: connecting...
> 2021-05-11T22:33:18.108Z|00615|rconn|INFO|br0.uvms<->tcp:127.0.0.1:6653: connected
> 2021-05-11T22:33:28.110Z|00616|rconn|ERR|br0.uvms<->tcp:127.0.0.1:6653: no response to inactivity probe after 5 seconds, disconnecting
> 2021-05-11T22:33:29.433Z|00617|rconn|INFO|br0.uvms<->tcp:127.0.0.1:6653: connecting...
> 2021-05-11T22:33:29.933Z|00618|rconn|INFO|br0.uvms<->tcp:127.0.0.1:6653: connected
>
>
> Can you please help us find out what could be wrong with this configuration and what is the expected behaviour from ovs switch when the receiver on the controller is blocked for long.
Hmm, I can't reproduce this with current OVS. I do see a problem with
the fail-open implementation; I'll see a patch for that.
Can you show the output of "ovs-vsctl list controller"?
More information about the discuss
mailing list