Hello JamO, I haven’t setup any configuration limiting either CPU or memory for the containers I’m spawning, although I just found a great link [0], allowing me to see pretty much everything regarding container’s metrics.
Yes I will open a bugzilla with detailed instruction of how I’m running my test. I will also commit what I have. I think I have a robot test somewhere leveraging some stuff I have to spawn the containers. I will do that certainly next week, though. Thanks, Alexis [0]: https://www.datadoghq.com/blog/how-to-collect-docker-metrics/ <https://www.datadoghq.com/blog/how-to-collect-docker-metrics/> > On Mar 3, 2016, at 12:41 AM, Jamo Luhrsen <[email protected]> wrote: > > > > On 02/19/2016 02:10 PM, Alexis de Talhouët wrote: >> So far my results are: >> >> OVS 2.4.0: ODL configure with 2G of mem —> max is ~50 switches connected >> OVS 2.3.1: ODL configure with 256MG of mem —> I currently have 150 switches >> connected, can’t scale more due to infra limits. > > Alexis, I think this is probably worth putting a bugzilla up. > > How much horsepower do you need per docker ovs instance? We need to get this > automated in CSIT. Marcus from ovsdb wants to do similar tests with ovsdb. > > JamO > > >> I will pursue me testing next week. >> >> Thanks, >> Alexis >> >>> On Feb 19, 2016, at 5:06 PM, Abhijit Kumbhare <[email protected] >>> <mailto:[email protected]>> wrote: >>> >>> Interesting. I wonder - why that would be? >>> >>> On Fri, Feb 19, 2016 at 1:19 PM, Alexis de Talhouët >>> <[email protected] <mailto:[email protected]>> wrote: >>> >>> OVS 2.3.x scales fine >>> OVS 2.4.x doesn’t scale well. >>> >>> Here is also the docker file for ovs 2.4.1 >>> >>> >>> >>>> On Feb 19, 2016, at 11:20 AM, Alexis de Talhouët >>>> <[email protected] <mailto:[email protected]>> wrote: >>>> >>>>> can I use your containers? do you have any scripts/tools to bring >>>>> things up/down? >>>> >>>> Sure, attached a tar file containing all scripts / config / dockerfile >>>> I’m using to setup docker containers >>>> emulating OvS. >>>> FYI: it’s ovs 2.3.0 and not 2.4.0 anymore >>>> >>>> Also, forget about this whole mail thread, something in my private >>>> container must be breaking OVS behaviour, I >>>> don’t know what yet. >>>> >>>> With the docker file attached here, I can scale 90+ without any >>>> trouble... >>>> >>>> Thanks, >>>> Alexis >>>> >>>> <ovs_scalability_setup.tar.gz> >>>> >>>>> On Feb 18, 2016, at 6:07 PM, Jamo Luhrsen <[email protected] >>>>> <mailto:[email protected]>> wrote: >>>>> >>>>> inline... >>>>> >>>>> On 02/18/2016 02:58 PM, Alexis de Talhouët wrote: >>>>>> I’m running OVS 2.4, against stable/lithium, openflowplugin-li >>>>> >>>>> >>>>> so this is one difference between CSIT and your setup, in addition to >>>>> the whole >>>>> containers vs mininet. >>>>> >>>>>> I never scaled up to 1k, this was in the CSIT job. >>>>>> In a real scenario, I scaled to ~400. But it was even before >>>>>> clustering came into play in ofp lithium. >>>>>> >>>>>> I think the log I sent have log trace for openflowplugin and >>>>>> openflowjava, it not the case I could resubmit the >>>>>> logs. >>>>>> I removed some of them in openflowjava because it was way to chatty >>>>>> (logging all messages content between ovs >>>>>> <---> odl) >>>>>> >>>>>> Unfortunately those IOException happen after the whole thing blow up. >>>>>> I was able to narrow done some logs in >>>>>> openflowjava >>>>>> to see the first disconnected event. As mentioned in a previous mail >>>>>> (in this mail thread) it’s the device that is >>>>>> issuing the disconnect: >>>>>> >>>>>>> 2016-02-18 16:56:30,018 | DEBUG | entLoopGroup-6-3 | OFFrameDecoder >>>>>>> | 201 - >>>>>>> org.opendaylight.openflowjava.openflow-protocol-impl - >>>>>>> 0.6.4.SNAPSHOT | skipping bytebuf - too few bytes for >>>>>>> header: 0 < 8 >>>>>>> 2016-02-18 16:56:30,018 | DEBUG | entLoopGroup-6-3 | >>>>>>> OFVersionDetector | 201 - >>>>>>> org.opendaylight.openflowjava.openflow-protocol-impl - >>>>>>> 0.6.4.SNAPSHOT | not enough data >>>>>>> 2016-02-18 16:56:30,018 | DEBUG | entLoopGroup-6-3 | >>>>>>> DelegatingInboundHandler | 201 - >>>>>>> org.opendaylight.openflowjava.openflow-protocol-impl - >>>>>>> 0.6.4.SNAPSHOT | Channel inactive >>>>>>> 2016-02-18 16:56:30,018 | DEBUG | entLoopGroup-6-3 | >>>>>>> ConnectionAdapterImpl | 201 - >>>>>>> org.opendaylight.openflowjava.openflow-protocol-impl - >>>>>>> 0.6.4.SNAPSHOT | ConsumeIntern msg on [id: 0x1efab5fb, >>>>>>> /172.18.0.49:36983 <http://172.18.0.49:36983/> :> >>>>>>> /192.168.1.159:6633 <http://192.168.1.159:6633/>] >>>>>>> 2016-02-18 16:56:30,018 | DEBUG | entLoopGroup-6-3 | >>>>>>> ConnectionAdapterImpl | 201 - >>>>>>> org.opendaylight.openflowjava.openflow-protocol-impl - >>>>>>> 0.6.4.SNAPSHOT | ConsumeIntern msg - DisconnectEvent >>>>>>> 2016-02-18 16:56:30,018 | DEBUG | entLoopGroup-6-3 | >>>>>>> ConnectionContextImpl | 205 - >>>>>>> org.opendaylight.openflowplugin.impl - 0.1.4.SNAPSHOT | >>>>>>> disconnecting: node=/172.18.0.49:36983|auxId=0|connection >>>>>>> state = RIP >>>>>> >>>>>> Those logs come from another run, so are not in the logs I sent >>>>>> earlier. Although the behaviour is always the same. >>>>>> >>>>>> Regarding the memory, I don’t want to add more than 2G memory, >>>>>> because, and I tested it, the more memory I add, >>>>>> the more >>>>>> I can scale. But as you pointed out, >>>>>> this issue is not OOM error. Thus I rather like failing at 2G (less >>>>>> docker containers to spawn each run ~50). >>>>> >>>>> so, maybe reduce your memory then to simplify the reproducing steps. >>>>> Since you know that increasing >>>>> memory allows you to scale further, but still hit the problem; let's >>>>> make it easier to hit. how far >>>>> can you go with the max mem set to 500M? if you are only loading >>>>> ofp-li. >>>>> >>>>>> I definitely need some help here, because I can’t sort myself out in >>>>>> the openflowplugin + openflowjava codebase… >>>>>> But I believe I already have Michal’s attention :) >>>>> >>>>> can I use your containers? do you have any scripts/tools to bring >>>>> things up/down? >>>>> I might be able to try and reproduce myself. I like breaking things :) >>>>> >>>>> JamO >>>>> >>>>> >>>>>> >>>>>> Thanks, >>>>>> Alexis >>>>>> >>>>>> >>>>>>> On Feb 18, 2016, at 5:44 PM, Jamo Luhrsen <[email protected] >>>>>>> <mailto:[email protected]> <mailto:[email protected]>> wrote: >>>>>>> >>>>>>> Alexis, don't worry about filing a bug just to give us a common >>>>>>> place to work/comment, even >>>>>>> if we close it later because of something outside of ODL. Email is >>>>>>> fine too. >>>>>>> >>>>>>> what ovs version do you have in your containers? this test sounds >>>>>>> great. >>>>>>> >>>>>>> Luis is right, that if you were scaling well past 1k in the past, >>>>>>> but now it falls over at >>>>>>> 50 it sounds like a bug. >>>>>>> >>>>>>> Oh, you can try increasing the jvm max_mem from default of 2G just >>>>>>> as a data point. The >>>>>>> fact that you don't get OOMs makes me think memory might not be the >>>>>>> final bottleneck. >>>>>>> >>>>>>> you could enable debug/trace logs in the right modules (need ofp >>>>>>> devs to tell us that) >>>>>>> for a little more info. >>>>>>> >>>>>>> I've seen those IOExceptions before and always assumed it was from >>>>>>> an OF switch doing a >>>>>>> hard RST on it's connection. >>>>>>> >>>>>>> >>>>>>> Thanks, >>>>>>> JamO >>>>>>> >>>>>>> >>>>>>> >>>>>>> On 02/18/2016 11:48 AM, Luis Gomez wrote: >>>>>>>> If the same test worked 6-8 months ago this seems like a bug, but >>>>>>>> please feel free to open it whenever you >>>>>>>> are sure. >>>>>>>> >>>>>>>>> On Feb 18, 2016, at 11:45 AM, Alexis de Talhouët >>>>>>>>> <[email protected] <mailto:[email protected]> >>>>>>>>> <mailto:[email protected]> <mailto:[email protected]>> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>> Hello Luis, >>>>>>>>> >>>>>>>>> For sure I’m willing to open a bug but before I want to make sure >>>>>>>>> there is a bug and that I’m not doing >>>>>>>>> something wrong. >>>>>>>>> In ODL’s infra, there is a test to find the maximum number of >>>>>>>>> switches that can be connected to ODL, and >>>>>>>>> this test >>>>>>>>> reach ~ 500 [0] >>>>>>>>> I was able to scale up to 1090 switches [1] using the CSIT job in >>>>>>>>> the sandbox. >>>>>>>>> I believe the CSIT test is different in a way that switches are >>>>>>>>> emulated in one mininet VM, whereas I’m >>>>>>>>> connecting OVS >>>>>>>>> instances from separate containers. >>>>>>>>> >>>>>>>>> 6-8 months ago, I was able to perform the same test, and scale >>>>>>>>> with OVS docker container up to ~400 before >>>>>>>>> ODL start >>>>>>>>> crashing (with some optimization done behind the scene, i.e. >>>>>>>>> ulimit, mem, cpu, GC…) >>>>>>>>> Now I’m not able to scale more than 100 with the same >>>>>>>>> configuration. >>>>>>>>> >>>>>>>>> FYI: I just quickly look at the CSIT test [0] karaf.log, it seems >>>>>>>>> the test is actually failing but it is not >>>>>>>>> correctly >>>>>>>>> advertised… switch connection are dropped. >>>>>>>>> Look for those: >>>>>>>>> 016-02-18 07:07:51,741 | WARN | entLoopGroup-6-6 | OFFrameDecoder >>>>>>>>> | 181 - >>>>>>>>> org.opendaylight.openflowjava.openflow-protocol-impl - >>>>>>>>> 0.6.4.SNAPSHOT | Unexpected exception from downstream. >>>>>>>>> java.io.IOException: Connection reset by peer >>>>>>>>> at sun.nio.ch.FileDispatcherImpl.read0(Native Method)[:1.7.0_85] >>>>>>>>> at >>>>>>>>> sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:39)[:1.7.0_85] >>>>>>>>> at >>>>>>>>> sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:223)[:1.7.0_85] >>>>>>>>> at sun.nio.ch.IOUtil.read(IOUtil.java:192)[:1.7.0_85] >>>>>>>>> at >>>>>>>>> sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:384)[:1.7.0_85] >>>>>>>>> at >>>>>>>>> >>>>>>>>> io.netty.buffer.PooledUnsafeDirectByteBuf.setBytes(PooledUnsafeDirectByteBuf.java:311)[111:io.netty.buffer:4.0.26.Final] >>>>>>>>> at >>>>>>>>> io.netty.buffer.AbstractByteBuf.writeBytes(AbstractByteBuf.java:881)[111:io.netty.buffer:4.0.26.Final] >>>>>>>>> at >>>>>>>>> >>>>>>>>> io.netty.channel.socket.nio.NioSocketChannel.doReadBytes(NioSocketChannel.java:241)[109:io.netty.transport:4.0.26.Final] >>>>>>>>> at >>>>>>>>> >>>>>>>>> io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:119)[109:io.netty.transport:4.0.26.Final] >>>>>>>>> at >>>>>>>>> >>>>>>>>> io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511)[109:io.netty.transport:4.0.26.Final] >>>>>>>>> at >>>>>>>>> >>>>>>>>> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468)[109:io.netty.transport:4.0.26.Final] >>>>>>>>> at >>>>>>>>> >>>>>>>>> io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382)[109:io.netty.transport:4.0.26.Final] >>>>>>>>> at >>>>>>>>> io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:349)[109:io.netty.transport:4.0.26.Final] >>>>>>>>> at >>>>>>>>> >>>>>>>>> io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:111)[110:io.netty.common:4.0.26.Final] >>>>>>>>> at >>>>>>>>> >>>>>>>>> io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:137)[110:io.netty.common:4.0.26.Final] >>>>>>>>> at java.lang.Thread.run(Thread.java:745)[:1.7.0_85] >>>>>>>>> >>>>>>>>> >>>>>>>>> [0]: >>>>>>>>> https://jenkins.opendaylight.org/releng/view/openflowplugin/job/openflowplugin-csit-1node-periodic-scalability-daily-only-stable-lithium/ >>>>>>>>> [1]: https://git.opendaylight.org/gerrit/#/c/33213/ >>>>>>>>> >>>>>>>>>> On Feb 18, 2016, at 2:28 PM, Luis Gomez <[email protected] >>>>>>>>>> <mailto:[email protected]> <mailto:[email protected]>> wrote: >>>>>>>>>> >>>>>>>>>> Alexis, thanks very much for sharing this test. Would you mind to >>>>>>>>>> open a bug with all this info so we can >>>>>>>>>> track this? >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> On Feb 18, 2016, at 7:29 AM, Alexis de Talhouët >>>>>>>>>>> <[email protected] >>>>>>>>>>> <mailto:[email protected]> >>>>>>>>>>> <mailto:[email protected]>> wrote: >>>>>>>>>>> >>>>>>>>>>> Hi Michal, >>>>>>>>>>> >>>>>>>>>>> ODL memory is capped at 2go, the more memory I add, those more >>>>>>>>>>> OVS I can connect. Regarding CPU, it’s >>>>>>>>>>> around 10-20% >>>>>>>>>>> when connecting new OVS, with some peak to 80%. >>>>>>>>>>> >>>>>>>>>>> After some investigation, here is what I observed: >>>>>>>>>>> Let say I have 50 switches connected, stat manager disabled. I >>>>>>>>>>> have one opened socket per switch, plus an >>>>>>>>>>> additional >>>>>>>>>>> one for the controller. >>>>>>>>>>> Then I connect a new switch (2016-02-18 09:35:08,059), 51 >>>>>>>>>>> switches… something is happening causing all >>>>>>>>>>> connection to >>>>>>>>>>> be dropped (by device?) and then ODL >>>>>>>>>>> try to recreate them and goes in a crazy loop where it is never >>>>>>>>>>> able to re-establish communication, but keeps >>>>>>>>>>> creating new sockets. >>>>>>>>>>> I’m suspecting something being garbage collected due to lack of >>>>>>>>>>> memory, although no OOM errors. >>>>>>>>>>> >>>>>>>>>>> Attached the YourKit Java Profiler analysis for the described >>>>>>>>>>> scenario and the logs [1]. >>>>>>>>>>> >>>>>>>>>>> Thanks, >>>>>>>>>>> Alexis >>>>>>>>>>> >>>>>>>>>>> [1]: >>>>>>>>>>> https://www.dropbox.com/sh/dgqeqv4j76zwbh3/AACim0za1fUozc7DlYJ4fsMJa?dl=0 >>>>>>>>>>> >>>>>>>>>>>> On Feb 9, 2016, at 8:59 AM, Michal Rehak -X (mirehak - PANTHEON >>>>>>>>>>>> TECHNOLOGIES at Cisco) <[email protected] >>>>>>>>>>>> <mailto:[email protected]> >>>>>>>>>>>> <mailto:[email protected]>> wrote: >>>>>>>>>>>> >>>>>>>>>>>> Hi Alexis, >>>>>>>>>>>> I am not sure how OVS uses threads - in changelog there is some >>>>>>>>>>>> concurrency related improvement in 2.1.3 >>>>>>>>>>>> and 2.3. >>>>>>>>>>>> Also I guess docker can be forced regarding assigned resources. >>>>>>>>>>>> >>>>>>>>>>>> For you the most important is the amount of cores used by >>>>>>>>>>>> controller. >>>>>>>>>>>> >>>>>>>>>>>> How does your cpu and memory consumption look like when you >>>>>>>>>>>> connect all the OVSs? >>>>>>>>>>>> >>>>>>>>>>>> Regards, >>>>>>>>>>>> Michal >>>>>>>>>>>> >>>>>>>>>>>> ________________________________________ >>>>>>>>>>>> From: Alexis de Talhouët <[email protected] >>>>>>>>>>>> <mailto:[email protected]> >>>>>>>>>>>> <mailto:[email protected]>> >>>>>>>>>>>> Sent: Tuesday, February 9, 2016 14:44 >>>>>>>>>>>> To: Michal Rehak -X (mirehak - PANTHEON TECHNOLOGIES at Cisco) >>>>>>>>>>>> Cc: [email protected] >>>>>>>>>>>> <mailto:[email protected]> >>>>>>>>>>>> <mailto:[email protected]> >>>>>>>>>>>> Subject: Re: [openflowplugin-dev] Scalability issues >>>>>>>>>>>> >>>>>>>>>>>> Hello Michal, >>>>>>>>>>>> >>>>>>>>>>>> Yes, all the OvS instances I’m running has a unique DPID. >>>>>>>>>>>> >>>>>>>>>>>> Regarding the thread limit for netty, I’m running test in a >>>>>>>>>>>> server that has 28 CPU(s). >>>>>>>>>>>> >>>>>>>>>>>> Does each OvS instances is assigned its own thread? >>>>>>>>>>>> >>>>>>>>>>>> Thanks, >>>>>>>>>>>> Alexis >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>>> On Feb 9, 2016, at 3:42 AM, Michal Rehak -X (mirehak - >>>>>>>>>>>>> PANTHEON TECHNOLOGIES at Cisco) >>>>>>>>>>>>> <[email protected] <mailto:[email protected]> >>>>>>>>>>>>> <mailto:[email protected]>> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>> Hi Alexis, >>>>>>>>>>>>> in Li-design there is the stats manager not in form of >>>>>>>>>>>>> standalone app but as part of core of ofPlugin. >>>>>>>>>>>>> You can >>>>>>>>>>>>> disable it via rpc. >>>>>>>>>>>>> >>>>>>>>>>>>> Just a question regarding your ovs setup. Do you have all >>>>>>>>>>>>> DPIDs unique? >>>>>>>>>>>>> >>>>>>>>>>>>> Also there is limit for netty in form of amount of used >>>>>>>>>>>>> threads. By default it uses 2 x >>>>>>>>>>>>> cpu_cores_amount. You >>>>>>>>>>>>> should have as many cores as possible in order to get max >>>>>>>>>>>>> performance. >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> Regards, >>>>>>>>>>>>> Michal >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> ________________________________________ >>>>>>>>>>>>> From: [email protected] >>>>>>>>>>>>> <mailto:[email protected]> >>>>>>>>>>>>> <mailto:[email protected]> >>>>>>>>>>>>> <[email protected] >>>>>>>>>>>>> <mailto:[email protected]> >>>>>>>>>>>>> <mailto:[email protected]>> >>>>>>>>>>>>> on >>>>>>>>>>>>> behalf of Alexis de Talhouët <[email protected] >>>>>>>>>>>>> <mailto:[email protected]> >>>>>>>>>>>>> <mailto:[email protected]>> >>>>>>>>>>>>> Sent: Tuesday, February 9, 2016 00:45 >>>>>>>>>>>>> To: [email protected] >>>>>>>>>>>>> <mailto:[email protected]> >>>>>>>>>>>>> <mailto:[email protected]> >>>>>>>>>>>>> Subject: [openflowplugin-dev] Scalability issues >>>>>>>>>>>>> >>>>>>>>>>>>> Hello openflowplugin-dev, >>>>>>>>>>>>> >>>>>>>>>>>>> I’m currently running some scalability test against >>>>>>>>>>>>> openflowplugin-li plugin, stable/lithium. >>>>>>>>>>>>> Playing with CSIT job, I was able to connect up to 1090 >>>>>>>>>>>>> switches: https://git.opendaylight.org/gerrit/#/c/33213/ >>>>>>>>>>>>> >>>>>>>>>>>>> I’m now running the test against 40 OvS switches, each one of >>>>>>>>>>>>> them is in a docker container. >>>>>>>>>>>>> >>>>>>>>>>>>> Connecting around 30 of them works fine, but then, adding a >>>>>>>>>>>>> new one break completely ODL, it goes crazy and >>>>>>>>>>>>> unresponsible. >>>>>>>>>>>>> Attach a snippet of the karaf.log with log set to DEBUG for >>>>>>>>>>>>> org.opendaylight.openflowplugin, thus it’s a >>>>>>>>>>>>> really >>>>>>>>>>>>> big log (~2.5MB). >>>>>>>>>>>>> >>>>>>>>>>>>> Here it what I observed based on the log: >>>>>>>>>>>>> I have 30 switches connected, all works fine. Then I add a new >>>>>>>>>>>>> one: >>>>>>>>>>>>> - SalRoleServiceImpl starts doing its thing (2016-02-08 >>>>>>>>>>>>> 23:13:38,534) >>>>>>>>>>>>> - RpcManagerImpl Registering Openflow RPCs (2016-02-08 >>>>>>>>>>>>> 23:13:38,546) >>>>>>>>>>>>> - ConnectionAdapterImpl Hello received (2016-02-08 >>>>>>>>>>>>> 23:13:40,520) >>>>>>>>>>>>> - Creation of the transaction chain, … >>>>>>>>>>>>> >>>>>>>>>>>>> Then all starts failing apart with this log: >>>>>>>>>>>>>> 2016-02-08 23:13:50,021 | DEBUG | ntLoopGroup-11-9 | >>>>>>>>>>>>>> ConnectionContextImpl | 190 - >>>>>>>>>>>>>> org.opendaylight.openflowplugin.impl - 0.1.4.SNAPSHOT | >>>>>>>>>>>>>> disconnecting: >>>>>>>>>>>>>> node=/172.31.100.9:46736|auxId=0|connection state = RIP >>>>>>>>>>>>> End then ConnectionContextImpl disconnects one by one the >>>>>>>>>>>>> switches, RpcManagerImpl is unregistered >>>>>>>>>>>>> Then it goes crazy for a while. >>>>>>>>>>>>> But all I’ve done is adding a new switch.. >>>>>>>>>>>>> >>>>>>>>>>>>> Finally, at 2016-02-08 23:14:26,666, exceptions are thrown: >>>>>>>>>>>>>> 2016-02-08 23:14:26,666 | ERROR | lt-dispatcher-85 | >>>>>>>>>>>>>> LocalThreePhaseCommitCohort | 172 - >>>>>>>>>>>>>> org.opendaylight.controller.sal-distributed-datastore - >>>>>>>>>>>>>> 1.2.4.SNAPSHOT | Failed to prepare transaction >>>>>>>>>>>>>> member-1-chn-5-txn-180 on backend >>>>>>>>>>>>>> akka.pattern.AskTimeoutException: Ask timed out on >>>>>>>>>>>>>> [ActorSelection[Anchor(akka://opendaylight-cluster-data/), >>>>>>>>>>>>>> >>>>>>>>>>>>>> Path(/user/shardmanager-operational/member-1-shard-inventory-operational#-1518836725)]] >>>>>>>>>>>>>> after [30000 ms] >>>>>>>>>>>>> And it goes for a while. >>>>>>>>>>>>> >>>>>>>>>>>>> Do you have any input on the same? >>>>>>>>>>>>> >>>>>>>>>>>>> Could you give some advice to be able to scale? (I know >>>>>>>>>>>>> disabling StatisticManager can help for instance) >>>>>>>>>>>>> >>>>>>>>>>>>> Am I doing something wrong? >>>>>>>>>>>>> >>>>>>>>>>>>> I can provide any asked information regarding the issue I’m >>>>>>>>>>>>> facing. >>>>>>>>>>>>> >>>>>>>>>>>>> Thanks, >>>>>>>>>>>>> Alexis >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> openflowplugin-dev mailing list >>>>>>>>>>> [email protected] >>>>>>>>>>> <mailto:[email protected]> >>>>>>>>>>> <mailto:[email protected]> >>>>>>>>>>> >>>>>>>>>>> https://lists.opendaylight.org/mailman/listinfo/openflowplugin-dev >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> openflowplugin-dev mailing list >>>>>>>> [email protected] >>>>>>>> <mailto:[email protected]> >>>>>>>> <mailto:[email protected]> >>>>>>>> https://lists.opendaylight.org/mailman/listinfo/openflowplugin-dev >>>> >>> >>> >>> _______________________________________________ >>> openflowplugin-dev mailing list >>> [email protected] >>> <mailto:[email protected]> >>> https://lists.opendaylight.org/mailman/listinfo/openflowplugin-dev >>> >>> >>
_______________________________________________ openflowplugin-dev mailing list [email protected] https://lists.opendaylight.org/mailman/listinfo/openflowplugin-dev
