> So something in this openstack driver is broken, because it does not respond > to server probes.
I must specify, it’s not broken ALL the time. It seems to start breaking when traffic increases on the openstack setup. This increase though is not accompanied by a CPU overload though: it’s happening right now and the load is barely over 1, on a 16 core xeon.The client trying to connect also appears to be ovs-vswitchd. Jean-Philippe Méthot Openstack system administrator Administrateur système Openstack PlanetHoster inc. > Le 27 sept. 2018 à 15:47, Paul Greenberg <[email protected]> a écrit : > > This specific error is triggered by the following. When a client connects to > ovsdb json rpc server, it has to follow certain protocol. In this case, a the > server sends probes, and the client must acknowledge them by sending the > exact message it received from the server back to the server. If a client, > does not do that in time, the server drops the client. > > So something in this openstack driver is broken, because it does not respond > to server probes. > > Best Regards, > Paul Greenberg > > From: 20230277700n behalf of > Sent: Thursday, September 27, 2018 3:40 PM > To: [email protected] > Cc: [email protected] > Subject: Re: [ovs-discuss] "ovs|01253|reconnect|ERR|tcp:127.0.0.1:50814: no > response to inactivity probe after 5.01 seconds, disconnecting" messages and > lost packets > > > ovs-vswitchd is multi-threaded. ovsdb-server is single threaded. > (You did not answer my question about the file from which the logs were > printed in your email) > > Who is at 127.0.0.1:45928 <http://127.0.0.1:45928/> and 127.0.0.1:45930 > <http://127.0.0.1:45930/>? > > On Thu, 27 Sep 2018 at 11:14, Jean-Philippe Méthot > <[email protected] <mailto:[email protected]>> wrote: > Thank you for your reply. > > This is Openstack with ml2 plugin. There’s no other 3rd party application > used with our network, so no OVN or anything of the sort. Essentially, to > give a quick idea of the topology, we have our vms on our compute nodes going > through GRE tunnels toward network nodes where they are routed in network > namespace toward a flat external network. > >> Generally, the above indicates that a daemon fronting a Open vSwitch >> database hasn't been able to connect to its client. Usually happens when CPU >> consumption is very high. > > Our network nodes CPU are literally sleeping. Is openvswitch single-thread or > multi-thread though? If ovs overloaded a single thread, it’s possible I may > have missed it. > > Jean-Philippe Méthot > Openstack system administrator > Administrateur système Openstack > PlanetHoster inc. > > > > >> Le 27 sept. 2018 à 14:04, Guru Shetty <[email protected] <mailto:[email protected]>> a >> écrit : >> >> >> >> On Wed, 26 Sep 2018 at 12:59, Jean-Philippe Méthot via discuss >> <[email protected] <mailto:[email protected]>> wrote: >> Hi, >> >> I’ve been using openvswitch for my networking backend on openstack for >> several years now. Lately, as our network has grown, we’ve started noticing >> some intermittent packet drop accompanied with the following error message >> in openvswitch: >> >> 2018-09-26T04:15:20.676Z|00005|reconnect|ERR|tcp:127.0.0.1:45928 >> <http://127.0.0.1:45928/>: no response to inactivity probe after 5 seconds, >> disconnecting >> 2018-09-26T04:15:20.677Z|00006|reconnect|ERR|tcp:127.0.0.1:45930 >> <http://127.0.0.1:45930/>: no response to inactivity probe after 5 seconds, >> disconnecting >> >> Open vSwitch is a project with multiple daemons. Since you are using >> OpenStack, it is not clear from your message, what type of networking plugin >> you are using. Do you use OVN? >> Also, you did not mention from which file you have gotten the above errors. >> >> Generally, the above indicates that a daemon fronting a Open vSwitch >> database hasn't been able to connect to its client. Usually happens when CPU >> consumption is very high. >> >> >> 2018-09-26T04:15:30.409Z|00007|reconnect|ERR|tcp:127.0.0.1:45874 >> <http://127.0.0.1:45874/>: no response to inactivity probe after 5 seconds, >> disconnecting >> 2018-09-26T04:15:33.661Z|00008|reconnect|ERR|tcp:127.0.0.1:45934 >> <http://127.0.0.1:45934/>: no response to inactivity probe after 5 seconds, >> disconnecting >> 2018-09-26T04:15:33.847Z|00009|reconnect|ERR|tcp:127.0.0.1:45894 >> <http://127.0.0.1:45894/>: no response to inactivity probe after 5 seconds, >> disconnecting >> 2018-09-26T04:16:03.247Z|00010|reconnect|ERR|tcp:127.0.0.1:45958 >> <http://127.0.0.1:45958/>: no response to inactivity probe after 5 seconds, >> disconnecting >> 2018-09-26T04:16:21.534Z|00011|reconnect|ERR|tcp:127.0.0.1:45956 >> <http://127.0.0.1:45956/>: no response to inactivity probe after 5 seconds, >> disconnecting >> 2018-09-26T04:16:21.786Z|00012|reconnect|ERR|tcp:127.0.0.1:45974 >> <http://127.0.0.1:45974/>: no response to inactivity probe after 5 seconds, >> disconnecting >> 2018-09-26T04:16:47.085Z|00013|reconnect|ERR|tcp:127.0.0.1:45988 >> <http://127.0.0.1:45988/>: no response to inactivity probe after 5 seconds, >> disconnecting >> 2018-09-26T04:16:49.618Z|00014|reconnect|ERR|tcp:127.0.0.1:45982 >> <http://127.0.0.1:45982/>: no response to inactivity probe after 5 seconds, >> disconnecting >> 2018-09-26T04:16:53.321Z|00015|reconnect|ERR|tcp:127.0.0.1:45964 >> <http://127.0.0.1:45964/>: no response to inactivity probe after 5 seconds, >> disconnecting >> 2018-09-26T04:17:15.543Z|00016|reconnect|ERR|tcp:127.0.0.1:45986 >> <http://127.0.0.1:45986/>: no response to inactivity probe after 5 seconds, >> disconnecting >> 2018-09-26T04:17:24.767Z|00017|reconnect|ERR|tcp:127.0.0.1:45990 >> <http://127.0.0.1:45990/>: no response to inactivity probe after 5 seconds, >> disconnecting >> 2018-09-26T04:17:31.735Z|00018|reconnect|ERR|tcp:127.0.0.1:45998 >> <http://127.0.0.1:45998/>: no response to inactivity probe after 5 seconds, >> disconnecting >> 2018-09-26T04:20:12.593Z|00019|reconnect|ERR|tcp:127.0.0.1:46014 >> <http://127.0.0.1:46014/>: no response to inactivity probe after 5 seconds, >> disconnecting >> 2018-09-26T04:23:51.996Z|00020|reconnect|ERR|tcp:127.0.0.1:46028 >> <http://127.0.0.1:46028/>: no response to inactivity probe after 5 seconds, >> disconnecting >> 2018-09-26T04:25:12.187Z|00021|reconnect|ERR|tcp:127.0.0.1:46022 >> <http://127.0.0.1:46022/>: no response to inactivity probe after 5 seconds, >> disconnecting >> 2018-09-26T04:25:28.871Z|00022|reconnect|ERR|tcp:127.0.0.1:46056 >> <http://127.0.0.1:46056/>: no response to inactivity probe after 5 seconds, >> disconnecting >> 2018-09-26T04:27:11.663Z|00023|reconnect|ERR|tcp:127.0.0.1:46046 >> <http://127.0.0.1:46046/>: no response to inactivity probe after 5 seconds, >> disconnecting >> 2018-09-26T04:29:56.161Z|00024|jsonrpc|WARN|tcp:127.0.0.1:46018 >> <http://127.0.0.1:46018/>: receive error: Connection reset by peer >> 2018-09-26T04:29:56.161Z|00025|reconnect|WARN|tcp:127.0.0.1:46018 >> <http://127.0.0.1:46018/>: connection dropped (Connection reset by peer) >> >> This definitely kills the connection for a few seconds before it reconnects. >> So, I’ve been wondering, what is this probe and what is really happening >> here? What’s the cause and is there a way to fix this? >> >> Openvswitch version is 2.9.0-3 on CentOS 7 with Openstack Pike running on it >> (but the issues show up on Queens too). >> >> >> Jean-Philippe Méthot >> Openstack system administrator >> Administrateur système Openstack >> PlanetHoster inc. >> >> >> >> >> _______________________________________________ >> discuss mailing list >> [email protected] <mailto:[email protected]> >> https://mail.openvswitch.org/mailman/listinfo/ovs-discuss >> <https://mail.openvswitch.org/mailman/listinfo/ovs-discuss>
_______________________________________________ discuss mailing list [email protected] https://mail.openvswitch.org/mailman/listinfo/ovs-discuss
