In this case we are purely relying on link state provided by DPDK.
Have you tried to check if same problem exists with DPDK testpmd app?


> On 18 Oct 2019, at 10:26, Chuan Han via Lists.Fd.Io 
> <chuanhan=google....@lists.fd.io> wrote:
> 
> I cleaned up startup config a bit. I can still see eth0 down. 
> 
> See attachment for config file and log. There are some errors in log, but I 
> am not sure they are worrisome or not. 
> 
> On Thu, Oct 17, 2019 at 5:22 PM Florin Coras <fcoras.li...@gmail.com 
> <mailto:fcoras.li...@gmail.com>> wrote:
> This looks like a DPDK issue, but I’ll let Damjan be the judge of that. 
> 
> To see if this is a config issues, could you simplify your startup config by
> - removing “workers 0” from the two nics and adding “num-rx-queues 2” to the 
> nics or to the default stanza, if you’re running with 2 workers
> - comment out the cryptodev config 
> 
> If the two nics don’t come up, check if there’s any obvious dpdk error in 
> “show log”.
> 
> Florin 
> 
>> On Oct 17, 2019, at 4:56 PM, Chuan Han via Lists.Fd.Io <http://lists.fd.io/> 
>> <chuanhan=google....@lists.fd.io <mailto:chuanhan=google....@lists.fd.io>> 
>> wrote:
>> 
>> I tried disabling autoneg on R740 side. It is not allowed too. If vpp cannot 
>> allow two nics to be successfully added to the same vpp instance, it seems 
>> to be a bug. Is it something which can be easily spotted in the code base? 
>> 
>> It is also not possible to enforce symmetricity on internet. The other party 
>> can do anything as long as basic ping works. 
>> 
>> On Thu, Oct 17, 2019 at 3:55 PM Chuan Han <chuan...@google.com 
>> <mailto:chuan...@google.com>> wrote:
>> If I only put one phy nic, i.e., eth0, to vpp, 'sh hardware' shows it is up. 
>> If I put both eth0 and eth1 in vpp, eth0 is always down. It seems something 
>> is wrong with the nic or vpp does not support this type of hardware? 
>> 
>> We tried enabling autoneg on R230. It is not allowed. To avoid asymmetric 
>> settings, disabling autoneg on R740 will help? 
>> 
>> On Thu, Oct 17, 2019 at 3:46 PM Balaji Venkatraman (balajiv) 
>> <bala...@cisco.com <mailto:bala...@cisco.com>> wrote:
>> It plays a role if it is asymmetric at both ends. You could enable it at 
>> both ends and check. 
>> 
>>> On Oct 17, 2019, at 3:15 PM, Chuan Han <chuan...@google.com 
>>> <mailto:chuan...@google.com>> wrote:
>>> 
>>> 
>>> I rebooted the r230 machine and found the phy nic corresponding to eth has 
>>> autoneg off. 
>>> 
>>> root@esdn-relay:~/gnxi/perf_testing/r230# ethtool enp6s0f1
>>> Settings for enp6s0f1:
>>>         Supported ports: [ FIBRE ]
>>>         Supported link modes:   10000baseT/Full 
>>>         Supported pause frame use: Symmetric
>>>         Supports auto-negotiation: No
>>>         Supported FEC modes: Not reported
>>>         Advertised link modes:  10000baseT/Full 
>>>         Advertised pause frame use: Symmetric
>>>         Advertised auto-negotiation: No
>>>         Advertised FEC modes: Not reported
>>>         Speed: 10000Mb/s
>>>         Duplex: Full
>>>         Port: Direct Attach Copper
>>>         PHYAD: 0
>>>         Transceiver: internal
>>>         Auto-negotiation: off
>>>         Supports Wake-on: d
>>>         Wake-on: d
>>>         Current message level: 0x00000007 (7)
>>>                                drv probe link
>>>         Link detected: yes
>>> root@esdn-relay:~/gnxi/perf_testing/r230# 
>>> 
>>> On r740, autoneg is on. It is copper. 
>>> 
>>> root@esdn-lab:~/gnxi/perf_testing/r740/vpp# ethtool eno3
>>> Settings for eno3:
>>>         Supported ports: [ TP ]
>>>         Supported link modes:   100baseT/Full 
>>>                                 1000baseT/Full 
>>>                                 10000baseT/Full 
>>>         Supported pause frame use: Symmetric
>>>         Supports auto-negotiation: Yes
>>>         Supported FEC modes: Not reported
>>>         Advertised link modes:  100baseT/Full 
>>>                                 1000baseT/Full 
>>>                                 10000baseT/Full 
>>>         Advertised pause frame use: Symmetric
>>>         Advertised auto-negotiation: Yes
>>>         Advertised FEC modes: Not reported
>>>         Speed: 10000Mb/s
>>>         Duplex: Full
>>>         Port: Twisted Pair
>>>         PHYAD: 0
>>>         Transceiver: internal
>>>         Auto-negotiation: on
>>>         MDI-X: Unknown
>>>         Supports Wake-on: umbg
>>>         Wake-on: g
>>>         Current message level: 0x00000007 (7)
>>>                                drv probe link
>>>         Link detected: yes
>>> root@esdn-lab:~/gnxi/perf_testing/r740/vpp# 
>>> 
>>> not clear if this plays a role or not. 
>>> 
>>> On Thu, Oct 17, 2019 at 2:41 PM Chuan Han via Lists.Fd.Io 
>>> <http://lists.fd.io/> <chuanhan=google....@lists.fd.io 
>>> <mailto:google....@lists.fd.io>> wrote:
>>> Restarting ixia controller does not help. We ended up with both ixia ports 
>>> having '!'. 
>>> 
>>> We are not sure how ixia port plays a role here. eth0 interfaces are the 
>>> interfaces connecting two servers, not to ixia. 
>>> 
>>> On Thu, Oct 17, 2019 at 11:26 AM Balaji Venkatraman (balajiv) 
>>> <bala...@cisco.com <mailto:bala...@cisco.com>> wrote:
>>> Hi Chuan,
>>> 
>>>  
>>> 
>>> Could you please try to reset the ixia controller connected to port 4?
>>> 
>>> I have seen issues with ‘!’ on ixia. Given the carrier on eth0 is down, I 
>>> suspect the ixia port.
>>> 
>>>  
>>> 
>>> --
>>> 
>>> Regards,
>>> 
>>> Balaji. 
>>> 
>>>  
>>> 
>>>  
>>> 
>>> From: Chuan Han <chuan...@google.com <mailto:chuan...@google.com>>
>>> Date: Thursday, October 17, 2019 at 11:09 AM
>>> To: "Balaji Venkatraman (balajiv)" <bala...@cisco.com 
>>> <mailto:bala...@cisco.com>>
>>> Cc: "vpp-dev@lists.fd.io <mailto:vpp-dev@lists.fd.io>" <vpp-dev@lists.fd.io 
>>> <mailto:vpp-dev@lists.fd.io>>, Arivudainambi Appachi gounder 
>>> <aappa...@google.com <mailto:aappa...@google.com>>, Jerry Cen 
>>> <zhiw...@google.com <mailto:zhiw...@google.com>>
>>> Subject: Re: [vpp-dev] Basic l2 bridging does not work
>>> 
>>>  
>>> 
>>> Yes. It is unidirectional stream from port 1 to port 4.  
>>> 
>>>  
>>> 
>>> Another engineer, Nambi, configured ixia. What he showed me yesterday is 
>>> that xia port connected to port 1 is green and good. ixia port connected to 
>>> port 4 is green but has a red exclamation mark, which means ping does not 
>>> work. 
>>> 
>>>  
>>> 
>>> We also found eth0 on R230 is down shown by "show hardware eth0" command. 
>>> However "show int" shows it is up.
>>> 
>>>  
>>> 
>>>  
>>> 
>>> vpp# sh hardware-interfaces eth0
>>>               Name                Idx   Link  Hardware
>>> eth0                               2    down  eth0
>>>   Link speed: unknown
>>>   Ethernet address b4:96:91:23:1e:d6
>>>   Intel 82599
>>>     carrier down 
>>>     flags: admin-up promisc pmd rx-ip4-cksum
>>>     rx: queues 1 (max 128), desc 512 (min 32 max 4096 align 8)
>>>     tx: queues 3 (max 64), desc 512 (min 32 max 4096 align 8)
>>>     pci: device 8086:154d subsystem 8086:7b11 address 0000:06:00.01 numa 0
>>>     max rx packet len: 15872
>>>     promiscuous: unicast on all-multicast on
>>>     vlan offload: strip off filter off qinq off
>>>     rx offload avail:  vlan-strip ipv4-cksum udp-cksum tcp-cksum tcp-lro 
>>>                        macsec-strip vlan-filter vlan-extend jumbo-frame 
>>> scatter 
>>>                        security keep-crc 
>>>     rx offload active: ipv4-cksum 
>>>     tx offload avail:  vlan-insert ipv4-cksum udp-cksum tcp-cksum 
>>> sctp-cksum 
>>>                        tcp-tso macsec-insert multi-segs security 
>>>     tx offload active: none
>>>     rss avail:         ipv4-tcp ipv4-udp ipv4 ipv6-tcp-ex ipv6-udp-ex 
>>> ipv6-tcp 
>>>                        ipv6-udp ipv6-ex ipv6 
>>>     rss active:        none
>>>     tx burst function: (nil)
>>>     rx burst function: ixgbe_recv_pkts_vec
>>> 
>>>     rx frames ok                                       33278
>>>     rx bytes ok                                      3960082
>>>     extended stats:
>>>       rx good packets                                  33278
>>>       rx good bytes                                  3960082
>>>       rx q0packets                                     33278
>>>       rx q0bytes                                     3960082
>>>       rx size 65 to 127 packets                        33278
>>>       rx multicast packets                             33278
>>>       rx total packets                                 33278
>>>       rx total bytes                                 3960082
>>> vpp# sh int
>>>               Name               Idx    State  MTU (L3/IP4/IP6/MPLS)     
>>> Counter          Count     
>>> eth0                              2      up          9000/0/0/0     rx 
>>> packets                 33279
>>>                                                                     rx 
>>> bytes                 3960201
>>>                                                                     drops   
>>>                        5
>>>                                                                     punt    
>>>                        1
>>>                                                                     
>>> tx-error                   33274
>>> eth1                              1      up          9000/0/0/0     rx 
>>> packets                 33274
>>>                                                                     rx 
>>> bytes                 3959606
>>>                                                                     tx 
>>> packets                 33273
>>>                                                                     tx 
>>> bytes                 3959487
>>>                                                                     drops   
>>>                    33274
>>>                                                                     
>>> tx-error                       3
>>> local0                            0     down          0/0/0/0       
>>> vpp# 
>>> 
>>>  
>>> 
>>> On Thu, Oct 17, 2019 at 10:54 AM Balaji Venkatraman (balajiv) 
>>> <bala...@cisco.com <mailto:bala...@cisco.com>> wrote:
>>> 
>>> Hi Chuan,
>>> 
>>>  
>>> 
>>> I assume u have unidirectional stream ? ixia->1->2->3->4->ixia?
>>> 
>>>  
>>> 
>>> vpp# sh int
>>>               Name               Idx    State  MTU (L3/IP4/IP6/MPLS)     
>>> Counter          Count     
>>> eth0                              2      up          9000/0/0/0     rx 
>>> packets                 30925
>>>                                                                     rx 
>>> bytes                 3680075
>>>                                                                     drops   
>>>                        5
>>>                                                                     punt    
>>>                        1
>>>                                                                     
>>> tx-error                   30920 
>>> eth1                              1      up          9000/0/0/0     rx 
>>> packets                 30920 <<< packets are received on port 3
>>>                                                                     rx 
>>> bytes                 3679480
>>>                                                                     tx 
>>> packets                 30919
>>>                                                                     tx 
>>> bytes                 3679361
>>>                                                                     drops   
>>>                    30920 <<< all dropped at port 3
>>>                                                                     
>>> tx-error                       3
>>> local0                            0     down          0/0/0/0       
>>> 
>>>  
>>> 
>>> On sh error logs on R 230 we see
>>> 
>>>          1             ethernet-input             l3 mac mismatch <<<<
>>>          3               eth1-output              interface is down
>>>      30922               eth0-output              interface is down
>>> 
>>>  
>>> 
>>>  
>>> 
>>> Do u see the arp getting resolved on ixia? The mac on ixia at port with 
>>> 172.16.1.2/24 <http://172.16.1.2/24> should be seen on its other port. Are 
>>> the ixia ports up at both ends?
>>> 
>>>  
>>> 
>>>  
>>> 
>>> --
>>> 
>>> Regards,
>>> 
>>> Balaji. 
>>> 
>>>  
>>> 
>>>  
>>> 
>>> From: <vpp-dev@lists.fd.io <mailto:vpp-dev@lists.fd.io>> on behalf of 
>>> "Chuan Han via Lists.Fd.Io <http://lists.fd.io/>" 
>>> <chuanhan=google....@lists.fd.io <mailto:google....@lists.fd.io>>
>>> Reply-To: "chuan...@google.com <mailto:chuan...@google.com>" 
>>> <chuan...@google.com <mailto:chuan...@google.com>>
>>> Date: Thursday, October 17, 2019 at 9:59 AM
>>> To: "Balaji Venkatraman (balajiv)" <bala...@cisco.com 
>>> <mailto:bala...@cisco.com>>
>>> Cc: "vpp-dev@lists.fd.io <mailto:vpp-dev@lists.fd.io>" <vpp-dev@lists.fd.io 
>>> <mailto:vpp-dev@lists.fd.io>>
>>> Subject: Re: [vpp-dev] Basic l2 bridging does not work
>>> 
>>>  
>>> 
>>> It seems R740 vpp works fine. All packets coming from port 1 go to port 2. 
>>> 
>>>  
>>> 
>>> vpp# sh int
>>>               Name               Idx    State  MTU (L3/IP4/IP6/MPLS)     
>>> Counter          Count     
>>> eth0                              2      up          9000/0/0/0     tx 
>>> packets                 30895
>>>                                                                     tx 
>>> bytes                 3676505
>>> eth1                              1      up          9000/0/0/0     rx 
>>> packets                 30895
>>>                                                                     rx 
>>> bytes                 3676505
>>> local0                            0     down          0/0/0/0       
>>> vpp# sh int
>>>               Name               Idx    State  MTU (L3/IP4/IP6/MPLS)     
>>> Counter          Count     
>>> eth0                              2      up          9000/0/0/0     tx 
>>> packets                 30897
>>>                                                                     tx 
>>> bytes                 3676743
>>> eth1                              1      up          9000/0/0/0     rx 
>>> packets                 30897
>>>                                                                     rx 
>>> bytes                 3676743
>>> local0                            0     down          0/0/0/0       
>>> vpp# sh error
>>>    Count                    Node                  Reason
>>>      30899                l2-output               L2 output packets
>>>      30899                l2-learn                L2 learn packets
>>>          1                l2-learn                L2 learn misses
>>>      30899                l2-input                L2 input packets
>>>      30899                l2-flood                L2 flood packets
>>> vpp# 
>>> 
>>>  
>>> 
>>> The drop happened on R230 vpp. Port 3 dropped all pkts complaining about 
>>> down interface. However, show command shows interfaces are up. 
>>> 
>>>  
>>> 
>>> vpp# sh int
>>>               Name               Idx    State  MTU (L3/IP4/IP6/MPLS)     
>>> Counter          Count     
>>> eth0                              2      up          9000/0/0/0     rx 
>>> packets                 30925
>>>                                                                     rx 
>>> bytes                 3680075
>>>                                                                     drops   
>>>                        5
>>>                                                                     punt    
>>>                        1
>>>                                                                     
>>> tx-error                   30920
>>> eth1                              1      up          9000/0/0/0     rx 
>>> packets                 30920
>>>                                                                     rx 
>>> bytes                 3679480
>>>                                                                     tx 
>>> packets                 30919
>>>                                                                     tx 
>>> bytes                 3679361
>>>                                                                     drops   
>>>                    30920
>>>                                                                     
>>> tx-error                       3
>>> local0                            0     down          0/0/0/0       
>>> vpp# sh error
>>>    Count                    Node                  Reason
>>>          2                llc-input               unknown llc ssap/dsap
>>>      61846                l2-output               L2 output packets
>>>      61846                l2-learn                L2 learn packets
>>>          2                l2-learn                L2 learn misses
>>>      61846                l2-input                L2 input packets
>>>      61846                l2-flood                L2 flood packets
>>>          1             ethernet-input             l3 mac mismatch
>>>          3               eth1-output              interface is down
>>>      30922               eth0-output              interface is down
>>> vpp# 
>>> 
>>>  
>>> 
>>> Not sure how to check mac issues. Can you explain a bit more? Here is what 
>>> I can see on R230 vpp. 
>>> 
>>>  
>>> 
>>> vpp# show bridge-domain 1 detail
>>>   BD-ID   Index   BSN  Age(min)  Learning  U-Forwrd   UU-Flood   Flooding  
>>> ARP-Term  arp-ufwd   BVI-Intf 
>>>     1       1      0     off        on        on       flood        on      
>>>  off       off        N/A    
>>> 
>>>            Interface           If-idx ISN  SHG  BVI  TxFlood        
>>> VLAN-Tag-Rewrite       
>>>              eth0                2     1    0    -      *                 
>>> none             
>>>              eth1                1     1    0    -      *                 
>>> none             
>>> vpp# sh l2fib verbose           
>>>     Mac-Address     BD-Idx If-Idx BSN-ISN Age(min) static filter bvi        
>>>  Interface-Name        
>>>  28:99:3a:f4:3a:a6    1      2      0/1      -       -      -     -         
>>>       eth0             
>>>  28:99:3a:f4:3a:9c    1      1      0/1      -       -      -     -         
>>>       eth1             
>>> L2FIB total/learned entries: 2/2  Last scan time: 0.0000e0sec  Learn limit: 
>>> 4194304 
>>> vpp# 
>>> 
>>>  
>>> 
>>> On Wed, Oct 16, 2019 at 6:01 PM Balaji Venkatraman (balajiv) 
>>> <bala...@cisco.com <mailto:bala...@cisco.com>> wrote:
>>> 
>>> +-------------------------------------------------------------------------+
>>> 
>>> |                                                                         |
>>> 
>>> |                                                                         |
>>> 
>>> |                                   IXIA                                  |
>>> 
>>> |                                                                         |
>>> 
>>> |                                                                         |
>>> 
>>> +-----------------------------------------------------------^-------------+
>>> 
>>>             |172.16.1.1/24 <http://172.16.1.1/24>                           
>>>        |172.16.1.2/24 <http://172.16.1.2/24>
>>>             |                                               |
>>> 
>>>             |                                               |
>>> 
>>>             |eth0                                           | eth0
>>> 
>>> +-----------v-------------+                    +------------+-----------+
>>> 
>>> |           1             |                    |            4           |
>>> 
>>> |                         |                    |                        |
>>> 
>>> |                         |                    |                        |
>>> 
>>> |                         |                    |                        |
>>> 
>>> |                         |eth1          eth1  |                        |
>>> 
>>> |        VPP1           2 +--------------------> 3        VPP 2         |
>>> 
>>> |                         |                    |                        |
>>> 
>>> |                         |                    |                        |
>>> 
>>> |                         |                    |                        |
>>> 
>>> |                         |                    |                        |
>>> 
>>> +-------------------------+                    +------------------------+
>>> 
>>>          R 740                                           R 230
>>> 
>>>  
>>> 
>>>  
>>> 
>>> Might help if you could check if the packet counts at ingress (port 1) & 
>>> egress (port 2) match. Similarly 3 & 4. And the mac entries seen on both 
>>> vpp(s). ARP req/rep tracing might also help.
>>> 
>>>  
>>> 
>>>  
>>> 
>>> /-
>>> 
>>> Balaji
>>> 
>>>  
>>> 
>>>  
>>> 
>>>  
>>> 
>>>  
>>> 
>>> From: <vpp-dev@lists.fd.io <mailto:vpp-dev@lists.fd.io>> on behalf of 
>>> "Damjan Marion via Lists.Fd.Io <http://lists.fd.io/>" 
>>> <dmarion=me....@lists.fd.io <mailto:me....@lists.fd.io>>
>>> Reply-To: "dmar...@me.com <mailto:dmar...@me.com>" <dmar...@me.com 
>>> <mailto:dmar...@me.com>>
>>> Date: Wednesday, October 16, 2019 at 5:12 PM
>>> To: "chuan...@google.com <mailto:chuan...@google.com>" <chuan...@google.com 
>>> <mailto:chuan...@google.com>>
>>> Cc: "vpp-dev@lists.fd.io <mailto:vpp-dev@lists.fd.io>" <vpp-dev@lists.fd.io 
>>> <mailto:vpp-dev@lists.fd.io>>
>>> Subject: Re: [vpp-dev] Basic l2 bridging does not work
>>> 
>>>  
>>> 
>>>  
>>> 
>>> On 16 Oct 2019, at 16:14, Chuan Han via Lists.Fd.Io <http://lists.fd.io/> 
>>> <chuanhan=google....@lists.fd.io <mailto:chuanhan=google....@lists.fd.io>> 
>>> wrote:
>>> 
>>>  
>>> 
>>> Hi, vpp experts, 
>>> 
>>>  
>>> 
>>> We are trying to make basic l2 bridge works within vpp. 
>>> 
>>>  
>>> 
>>> We have two servers: r230 and r740, each of which has two phy nics. Two 
>>> servers are connected via cable. On each server, we bring these two nics 
>>> into the same vpp instance and put them into the same l2 bridge. We tried 
>>> sending traffic using ixia. However, ixia shows ping does not work. 
>>> 
>>>  
>>> 
>>> I attached the topology, vpp conf files, startup conf file, and logs.
>>> 
>>>  
>>> 
>>> Please advise where we could make it wrong.
>>> 
>>>  
>>> 
>>> Thanks.
>>> 
>>> Chuan
>>> 
>>> <r230 vpp.conf><r740 vpp.conf><r230 vpp startup.cfg><r740 vpp 
>>> startup.cfg><r740.log><r230.log><vpp testbed - 
>>> bridge.pdf>-=-=-=-=-=-=-=-=-=-=-=-
>>> Links: You receive all messages sent to this group.
>>> 
>>> View/Reply Online (#14189): https://lists.fd.io/g/vpp-dev/message/14189 
>>> <https://lists.fd.io/g/vpp-dev/message/14189>
>>> Mute This Topic: https://lists.fd.io/mt/34655826/675642 
>>> <https://lists.fd.io/mt/34655826/675642>
>>> Group Owner: vpp-dev+ow...@lists.fd.io <mailto:vpp-dev+ow...@lists.fd.io>
>>> Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub 
>>> <https://lists.fd.io/g/vpp-dev/unsub>  [dmar...@me.com 
>>> <mailto:dmar...@me.com>]
>>> -=-=-=-=-=-=-=-=-=-=-=-
>>> 
>>>  
>>> 
>>> On the 1st look everything look ok including the packet trace.
>>> 
>>> Can you try to clear counters* and enable packet trace on both instances.
>>> 
>>> Then send known number of packets and find put where drop happens by 
>>> looking into same outputs you already shared.....
>>> 
>>>  
>>> 
>>> * "clear int", "clear run", "clear trace"
>>> 
>>>  
>>> 
>>> -- 
>>> Damjan
>>> 
>>>  
>>> 
>>> -=-=-=-=-=-=-=-=-=-=-=-
>>> Links: You receive all messages sent to this group.
>>> 
>>> View/Reply Online (#14211): https://lists.fd.io/g/vpp-dev/message/14211 
>>> <https://lists.fd.io/g/vpp-dev/message/14211>
>>> Mute This Topic: https://lists.fd.io/mt/34655826/1991531 
>>> <https://lists.fd.io/mt/34655826/1991531>
>>> Group Owner: vpp-dev+ow...@lists.fd.io <mailto:vpp-dev%2bow...@lists.fd.io>
>>> Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub 
>>> <https://lists.fd.io/g/vpp-dev/unsub>  [chuan...@google.com 
>>> <mailto:chuan...@google.com>]
>>> -=-=-=-=-=-=-=-=-=-=-=-
>> -=-=-=-=-=-=-=-=-=-=-=-
>> Links: You receive all messages sent to this group.
>> 
>> View/Reply Online (#14219): https://lists.fd.io/g/vpp-dev/message/14219 
>> <https://lists.fd.io/g/vpp-dev/message/14219>
>> Mute This Topic: https://lists.fd.io/mt/34655826/675152 
>> <https://lists.fd.io/mt/34655826/675152>
>> Group Owner: vpp-dev+ow...@lists.fd.io <mailto:vpp-dev+ow...@lists.fd.io>
>> Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub 
>> <https://lists.fd.io/g/vpp-dev/unsub>  [fcoras.li...@gmail.com 
>> <mailto:fcoras.li...@gmail.com>]
>> -=-=-=-=-=-=-=-=-=-=-=-
> 
> <log><vpp startup.conf>-=-=-=-=-=-=-=-=-=-=-=-
> Links: You receive all messages sent to this group.
> 
> View/Reply Online (#14240): https://lists.fd.io/g/vpp-dev/message/14240 
> <https://lists.fd.io/g/vpp-dev/message/14240>
> Mute This Topic: https://lists.fd.io/mt/34655826/675642 
> <https://lists.fd.io/mt/34655826/675642>
> Group Owner: vpp-dev+ow...@lists.fd.io <mailto:vpp-dev+ow...@lists.fd.io>
> Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub 
> <https://lists.fd.io/g/vpp-dev/unsub>  [dmar...@me.com 
> <mailto:dmar...@me.com>]
> -=-=-=-=-=-=-=-=-=-=-=-

-- 
Damjan

-=-=-=-=-=-=-=-=-=-=-=-
Links: You receive all messages sent to this group.

View/Reply Online (#14241): https://lists.fd.io/g/vpp-dev/message/14241
Mute This Topic: https://lists.fd.io/mt/34655826/21656
Group Owner: vpp-dev+ow...@lists.fd.io
Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub  [arch...@mail-archive.com]
-=-=-=-=-=-=-=-=-=-=-=-

Reply via email to