Mohamed, Sorry to hear the demo isnt working well for you in Boron. I havent tried it in Boron lately, but have used it successfully recently in master (Carbon). Do you have to use Boron, or can you switch to Carbon?
I'll try running it from Boron now. I'll let you know how it goes for me in Boron. Either way, it looks like the problem is in the OpenflowPlugin, and Im not sure we'll be able to do much from SFC. Regards, Brady -----Original Message----- From: Mohamed ElSerngawy <[email protected]<mailto:mohamed%20elserngawy%20%[email protected]%3e>> To: Jamo Luhrsen <[email protected]<mailto:jamo%20luhrsen%20%[email protected]%3e>>, [email protected] <[email protected]<mailto:%[email protected]%22%20%[email protected]%3e>> Cc: Openflowplugin-dev <[email protected]<mailto:openflowplugin-dev%20%[email protected]%3e>> Subject: Re: [openflowplugin-dev] br-sfc become slave and disconnect from ODL Date: Wed, 8 Mar 2017 15:51:29 -0500 Hi, following up on this thread In order to reproduce it, I attached with the email a vagrant file that will spawn 5 VMs: odl, classifier1/2 and sff1/2. After spawn all the VMs: 1- In the odl VM start odl and install the following features: vagrant@odl:~/$ distribution-karaf-0.5.2-Boron-SR2/bin/karaf opendaylight-user@root>feature:install odl-sfc-ui odl-sfclisp odl-sfc-scf-openflow odl-sfc-sb-rest odl-sfc-ovs odl-sfc-netconf odl-sfcofl2 2- in the other VMs classifier1/2 and sff1/2 set the OVS instance manager and create br-sfc: vagrant@classifier1:~$ sudo ovs-vsctl set-manager tcp:10.10.1.5:6640<http://10.10.1.5:6640> vagrant@classifier1:~$ sudo ovs-vsctl add-br br-sfc vagrant@classifier1:~$ sudo ovs-vsctl set-controller br-sfc tcp:10.10.1.5:6653<http://10.10.1.5:6653> usually it show up after the third VM, check the ODL logs it should show the same error that i sent in the first email [1] if we checked the ovs status the manager is connected but the br-sfc is not vagrant@classifier1:~$ sudo ovs-vsctl show 0b5ce3d8-12f8-4298-98fa-0009450b9b2a Manager "tcp:10.10.1.5:6640<http://10.10.1.5:6640>" is_connected: true Bridge br-sfc Port veth-br Interface veth-br Port br-sfc Interface br-sfc type: internal ovs_version: "2.5.90" Once I stop/start odl it works fine and all br-sfc are connected !!!. I tried to debug and look in deep but didn't really help. Any idea why this happening ? Thanks [1] 2017-03-06 23:27:43,369 | WARN | pool-38-thread-1 | DeviceContextImpl | 312 - org.opendaylight.openflowplugin.impl - 0.3.2.Boron-SR2 | New role BECOMESLAVE was not propagated to device openflow:112250352583491 during 10 sec 2017-03-06 23:27:43,369 | ERROR | pool-38-thread-1 | SalRoleServiceImpl | 312 - org.opendaylight.openflowplugin.impl - 0.3.2.Boron-SR2 | SetRoleService set Role BECOMESLAVE for Node: KeyedInstanceIdentifier{targetType=interface org.opendaylight.yang.gen.v1.urn.opendaylight.inventory.rev130819.nodes.Node, path=[org.opendaylight.yang.gen.v1.urn.opendaylight.inventory.rev130819.Nodes, org.opendaylight.yang.gen.v1.urn.opendaylight.inventory.rev130819.nodes.Node[key=NodeKey [_id=Uri [_value=openflow:112250352583491]]]]} fail . Reason java.util.concurrent.CancellationException: Task was cancelled. 2017-03-06 23:27:43,369 | WARN | pool-38-thread-1 | DeviceManagerImpl | 312 - org.opendaylight.openflowplugin.impl - 0.3.2.Boron-SR2 | Was not able to set role SLAVE to device on node openflow:112250352583491 2017-03-06 23:28:25,569 | INFO | pool-38-thread-1 | StatisticsManagerImpl | 312 - org.opendaylight.openflowplugin.impl - 0.3.2.Boron-SR2 | Statistics collection for node openflow:2913975747405 still in progress even after 50 secs 2017-03-06 23:28:25,577 | WARN | pool-38-thread-1 | StatisticsManagerImpl | 312 - org.opendaylight.openflowplugin.impl - 0.3.2.Boron-SR2 | Statistics gathering for single node openflow:2913975747405 was not successful: Task was cancelled. On Mon, Mar 6, 2017 at 7:01 PM, Jamo Luhrsen <[email protected]<mailto:[email protected]>> wrote: ok, well if it's the same issue then it's something thats been around for a while. The job I was debugging is a recent carbon. JamO On 03/06/2017 04:00 PM, Mohamed ElSerngawy wrote: > I'm using boron sr2 > > On Monday, March 6, 2017, Jamo Luhrsen > <[email protected]<mailto:[email protected]> > <mailto:[email protected]<mailto:[email protected]>>> wrote: > > What distro are you using? > > netvirt CSIT had a bad failure I was trying to debug today that looks > just like this. > email thread here [0] > > I hadn't seen it before so thought it might not be reproducible very > easily... > > JamO > > [0] > https://lists.opendaylight.org/pipermail/netvirt-dev/2017-March/003566.html > > <https://lists.opendaylight.org/pipermail/netvirt-dev/2017-March/003566.html> > > > On 03/06/2017 03:35 PM, Mohamed ElSerngawy wrote: > > Hi Team, > > > > I'm having issue with openflow in the sfc103 demo, the bridge br-sfc is > not connected to ODL raising error at [1]. Can > anyone > > advice, why is the bridge become slave and not connected to ODL ? > > > > > > [1] 2017-03-06 23:27:43,369 | WARN | pool-38-thread-1 | > DeviceContextImpl | 312 - > > org.opendaylight.openflowplugin.impl - 0.3.2.Boron-SR2 | New role > BECOMESLAVE was not propagated to device > > openflow:112250352583491 during 10 sec > > > > 2017-03-06 23:27:43,369 | ERROR | pool-38-thread-1 | SalRoleServiceImpl > | 312 - > > org.opendaylight.openflowplugin.impl - 0.3.2.Boron-SR2 | SetRoleService > set Role BECOMESLAVE for Node: > > KeyedInstanceIdentifier{targetType=interface > org.opendaylight.yang.gen.v1.urn.opendaylight.inventory.rev130819.nodes.Node, > > > path=[org.opendaylight.yang.gen.v1.urn.opendaylight.inventory.rev130819.Nodes, > > > org.opendaylight.yang.gen.v1.urn.opendaylight.inventory.rev130819.nodes.Node[key=NodeKey > [_id=Uri > > [_value=openflow:112250352583491]]]]} fail . Reason > java.util.concurrent.CancellationException: Task was cancelled. > > > > 2017-03-06 23:27:43,369 | WARN | pool-38-thread-1 | DeviceManagerImpl > | 312 - > > org.opendaylight.openflowplugin.impl - 0.3.2.Boron-SR2 | Was not able > to set role SLAVE to device on node > > openflow:112250352583491 > > > > 2017-03-06 23:28:25,569 | INFO | pool-38-thread-1 | > StatisticsManagerImpl | 312 - > > org.opendaylight.openflowplugin.impl - 0.3.2.Boron-SR2 | Statistics > collection for node openflow:2913975747405 still in > > progress even after 50 secs > > > > 2017-03-06 23:28:25,577 | WARN | pool-38-thread-1 | > StatisticsManagerImpl | 312 - > > org.opendaylight.openflowplugin.impl - 0.3.2.Boron-SR2 | Statistics > gathering for single node openflow:2913975747405 > was not > > successful: Task was cancelled. > > > > Thanks > > > > > > > > _______________________________________________ > > openflowplugin-dev mailing list > > > [email protected]<mailto:[email protected]> > <javascript:;> > > https://lists.opendaylight.org/mailman/listinfo/openflowplugin-dev > <https://lists.opendaylight.org/mailman/listinfo/openflowplugin-dev> > > > _______________________________________________ openflowplugin-dev mailing list [email protected]<mailto:[email protected]> https://lists.opendaylight.org/mailman/listinfo/openflowplugin-dev
_______________________________________________ openflowplugin-dev mailing list [email protected] https://lists.opendaylight.org/mailman/listinfo/openflowplugin-dev
