Hi Steven,

I am testing both memif and vhost-virtio, unfortunately memif is not
working as well. I posted question to the list, let me know if
something is wrong. Below is the link

https://lists.fd.io/g/vpp-dev/topic/q_on_memif_between_vpp/20371922?p=,,,20,0,0,0::recentpostdate%2Fsticky,,,20,2,0,20371922

Thanks.

On Wed, May 30, 2018 at 4:41 PM, Ravi Kerur <rke...@gmail.com> wrote:
> Hi Steve,
>
> Thank you for your inputs, I added feature-mask to see if it helps in
> setting up queues correctly, it didn't so I will remove it. I have
> tried following combination
>
> (1) VPP->vhost-user (on host) and DPDK/testpmd->virtio-user (in a
> container)  -- VPP crashes
> (2) DPDK/testpmd->vhost-user (on host) and DPDK/testpmd->virtio-user
> (in a container) -- works fine
>
> To use DPDK vhost-user inside VPP, I defined configuration in
> startup.conf as mentioned by you and it looks as follows
>
> unix {
>   nodaemon
>   log /var/log/vpp/vpp.log
>   full-coredump
>   cli-listen /run/vpp/cli.sock
>   gid vpp
> }
>
> api-segment {
>   gid vpp
> }
>
> cpu {
>         main-core 1
>         corelist-workers 6-9
> }
>
> dpdk {
>         dev 0000:04:10.4
>         dev 0000:04:10.6
>         uio-driver vfio-pci
>         vdev net_vhost0,iface=/var/run/vpp/sock1.sock
>         vdev net_vhost1,iface=/var/run/vpp/sock2.sock
>         huge-dir /dev/hugepages_1GB
>         socket-mem 2048,2048
> }
>
> From VPP logs
> dpdk: EAL init args: -c 3c2 -n 4 --vdev
> net_vhost0,iface=/var/run/vpp/sock1.sock --vdev
> net_vhost1,iface=/var/run/vpp/sock2.sock --huge-dir /dev/hugepages_1GB
> -w 0000:04:10.4 -w 0000:04:10.6 --master-lcore 1 --socket-mem
> 2048,2048
>
> However, VPP doesn't create interface at all
>
> vpp# show interface
>               Name               Idx       State          Counter
>     Count
> VirtualFunctionEthernet4/10/4     1        down
> VirtualFunctionEthernet4/10/6     2        down
> local0                            0        down
>
> since it is a static mapping I am assuming it should be created, correct?
>
> Thanks.
>
> On Wed, May 30, 2018 at 3:43 PM, Steven Luong (sluong) <slu...@cisco.com> 
> wrote:
>> Ravi,
>>
>> First and foremost, get rid of the feature-mask option. I don't know what 
>> 0x40400000 does for you. If that does not help, try testing it with dpdk 
>> based vhost-user instead of VPP native vhost-user to make sure that they can 
>> work well with each other first. To use dpdk vhost-user, add a vdev command 
>> in the startup.conf for each vhost-user device that you have.
>>
>> dpdk { vdev net_vhost0,iface=/var/run/vpp/sock1.sock }
>>
>> dpdk based vhost-user interface is named VhostEthernet0, VhostEthernet1, 
>> etc. Make sure you use the right interface name to set the state to up.
>>
>> If dpdk based vhost-user does not work with testpmd either, it looks like 
>> some problem with the way that you invoke testpmd.
>>
>> If dpdk based vhost-user works well with the same testpmd device driver and 
>> not vpp native vhost-user, I can set up something similar to yours to look 
>> into it.
>>
>> The device driver, testpmd, is supposed to pass the shared memory region to 
>> VPP for TX/RX queues. It looks like VPP vhost-user might have run into a 
>> bump there with using the shared memory (txvq->avail).
>>
>> Steven
>>
>> PS. vhost-user is not an optimum interface for containers. You may want to 
>> look into using memif if you don't already know about it.
>>
>>
>> On 5/30/18, 2:06 PM, "Ravi Kerur" <rke...@gmail.com> wrote:
>>
>>     I am not sure what is wrong with the setup or a bug in vpp, vpp
>>     crashes with vhost<-->virtio communication.
>>
>>     (1) Vhost-interfaces are created and attached to bridge-domain as follows
>>
>>     create vhost socket /var/run/vpp/sock1.sock server feature-mask 
>> 0x40400000
>>     create vhost socket /var/run/vpp/sock2.sock server feature-mask 
>> 0x40400000
>>     set interface state VirtualEthernet0/0/0 up
>>     set interface state VirtualEthernet0/0/1 up
>>
>>     set interface l2 bridge VirtualEthernet0/0/0 1
>>     set interface l2 bridge VirtualEthernet0/0/1 1
>>
>>
>>     (2) DPDK/testpmd is started in a container to talk to vpp/vhost-user
>>     interface as follows
>>
>>     docker run -it --privileged -v
>>     /var/run/vpp/sock1.sock:/var/run/usvhost1 -v
>>     /var/run/vpp/sock2.sock:/var/run/usvhost2 -v
>>     /dev/hugepages:/dev/hugepages dpdk-app-testpmd ./bin/testpmd -c 0x3 -n
>>     4 --log-level=9 -m 64 --no-pci --single-file-segments
>>     --vdev=virtio_user0,path=/var/run/usvhost1,mac=54:00:00:01:01:01
>>     --vdev=virtio_user1,path=/var/run/usvhost2,mac=54:00:00:01:01:02 --
>>     -i
>>
>>     (3) show vhost-user VirtualEthernet0/0/1
>>     Virtio vhost-user interfaces
>>     Global:
>>       coalesce frames 32 time 1e-3
>>       number of rx virtqueues in interrupt mode: 0
>>     Interface: VirtualEthernet0/0/1 (ifindex 4)
>>     virtio_net_hdr_sz 10
>>      features mask (0x40400000):
>>      features (0x0):
>>       protocol features (0x0)
>>
>>      socket filename /var/run/vpp/sock2.sock type server errno "Success"
>>
>>      rx placement:
>>      tx placement: spin-lock
>>        thread 0 on vring 0
>>        thread 1 on vring 0
>>        thread 2 on vring 0
>>        thread 3 on vring 0
>>        thread 4 on vring 0
>>
>>      Memory regions (total 1)
>>      region fd    guest_phys_addr    memory_size        userspace_addr
>>     mmap_offset        mmap_addr
>>      ====== ===== ================== ================== ==================
>>     ================== ==================
>>       0     55    0x00007ff7c0000000 0x0000000040000000 0x00007ff7c0000000
>>     0x0000000000000000 0x00007ffbc0000000
>>
>>     vpp# show vhost-user VirtualEthernet0/0/0
>>     Virtio vhost-user interfaces
>>     Global:
>>       coalesce frames 32 time 1e-3
>>       number of rx virtqueues in interrupt mode: 0
>>     Interface: VirtualEthernet0/0/0 (ifindex 3)
>>     virtio_net_hdr_sz 10
>>      features mask (0x40400000):
>>      features (0x0):
>>       protocol features (0x0)
>>
>>      socket filename /var/run/vpp/sock1.sock type server errno "Success"
>>
>>      rx placement:
>>      tx placement: spin-lock
>>        thread 0 on vring 0
>>        thread 1 on vring 0
>>        thread 2 on vring 0
>>        thread 3 on vring 0
>>        thread 4 on vring 0
>>
>>      Memory regions (total 1)
>>      region fd    guest_phys_addr    memory_size        userspace_addr
>>     mmap_offset        mmap_addr
>>      ====== ===== ================== ================== ==================
>>     ================== ==================
>>       0     51    0x00007ff7c0000000 0x0000000040000000 0x00007ff7c0000000
>>     0x0000000000000000 0x00007ffc00000000
>>
>>     (4) vpp stack trace
>>     Program received signal SIGSEGV, Segmentation fault.
>>     [Switching to Thread 0x7ffd0e090700 (LWP 46570)]
>>     0x00007ffff7414642 in vhost_user_if_input
>>     (mode=VNET_HW_INTERFACE_RX_MODE_POLLING,
>>         node=0x7fffb76bab00, qid=<optimized out>, vui=0x7fffb6739700,
>>         vum=0x7ffff78f4480 <vhost_user_main>, vm=0x7fffb672a9c0)
>>         at 
>> /var/venom/vpp/build-data/../src/vnet/devices/virtio/vhost-user.c:1596
>>     1596      if (PREDICT_FALSE (txvq->avail->flags & 0xFFFE))
>>     (gdb) bt
>>     #0  0x00007ffff7414642 in vhost_user_if_input
>>     (mode=VNET_HW_INTERFACE_RX_MODE_POLLING,
>>         node=0x7fffb76bab00, qid=<optimized out>, vui=0x7fffb6739700,
>>         vum=0x7ffff78f4480 <vhost_user_main>, vm=0x7fffb672a9c0)
>>         at 
>> /var/venom/vpp/build-data/../src/vnet/devices/virtio/vhost-user.c:1596
>>     #1  vhost_user_input (f=<optimized out>, node=<optimized out>,
>>     vm=<optimized out>)
>>         at 
>> /var/venom/vpp/build-data/../src/vnet/devices/virtio/vhost-user.c:1947
>>     #2  vhost_user_input_avx2 (vm=<optimized out>, node=<optimized out>,
>>     frame=<optimized out>)
>>         at 
>> /var/venom/vpp/build-data/../src/vnet/devices/virtio/vhost-user.c:1972
>>     #3  0x00007ffff7954d35 in dispatch_node
>>     (last_time_stamp=12391212490024174, frame=0x0,
>>         dispatch_state=VLIB_NODE_STATE_POLLING, type=VLIB_NODE_TYPE_INPUT,
>>     node=0x7fffb76bab00,
>>         vm=0x7fffb672a9c0) at 
>> /var/venom/vpp/build-data/../src/vlib/main.c:988
>>     #4  vlib_main_or_worker_loop (is_main=0, vm=0x7fffb672a9c0)
>>         at /var/venom/vpp/build-data/../src/vlib/main.c:1507
>>     #5  vlib_worker_loop (vm=0x7fffb672a9c0) at
>>     /var/venom/vpp/build-data/../src/vlib/main.c:1641
>>     #6  0x00007ffff6ad25d8 in clib_calljmp ()
>>         at /var/venom/vpp/build-data/../src/vppinfra/longjmp.S:110
>>     #7  0x00007ffd0e08fdb0 in ?? ()
>>     #8  0x00007fffb4436edd in eal_thread_loop (arg=<optimized out>)
>>         at 
>> /var/venom/vpp/build-root/build-vpp-native/dpdk/dpdk-stable-18.02.1/lib/librte_eal/linuxapp/eal/eal_thread.c:153
>>     #9  0x0000000000000000 in ?? ()
>>     (gdb) frame 0
>>     #0  0x00007ffff7414642 in vhost_user_if_input
>>     (mode=VNET_HW_INTERFACE_RX_MODE_POLLING,
>>         node=0x7fffb76bab00, qid=<optimized out>, vui=0x7fffb6739700,
>>         vum=0x7ffff78f4480 <vhost_user_main>, vm=0x7fffb672a9c0)
>>         at 
>> /var/venom/vpp/build-data/../src/vnet/devices/virtio/vhost-user.c:1596
>>     1596      if (PREDICT_FALSE (txvq->avail->flags & 0xFFFE))
>>     (gdb) p txvq
>>     $1 = (vhost_user_vring_t *) 0x7fffb6739ac0
>>     (gdb) p *txvq
>>     $2 = {cacheline0 = 0x7fffb6739ac0 "?", qsz_mask = 255, last_avail_idx
>>     = 0, last_used_idx = 0,
>>       n_since_last_int = 0, desc = 0x7ffbfff97000, avail = 0x7ffbfff98000,
>>     used = 0x7ffbfff99000,
>>       int_deadline = 0, started = 1 '\001', enabled = 0 '\000', log_used = 0 
>> '\000',
>>       cacheline1 = 0x7fffb6739b00 "????\n", errfd = -1, callfd_idx = 10,
>>     kickfd_idx = 14,
>>       log_guest_addr = 0, mode = 1}
>>     (gdb) p *(txvq->avail)
>>     Cannot access memory at address 0x7ffbfff98000
>>     (gdb)
>>
>>     On Tue, May 29, 2018 at 10:47 AM, Ravi Kerur <rke...@gmail.com> wrote:
>>     > Steve,
>>     >
>>     > Thanks for inputs on debugs and gdb. I am using gdb on my development
>>     > system to debug the issue. I would like to have reliable core
>>     > generation on the system on which I don't have access to install gdb.
>>     > I installed corekeeper and it still doesn't generate core. I am
>>     > running vpp inside a VM (VirtualBox/vagrant), not sure if I need to
>>     > set something inside vagrant config file.
>>     >
>>     >  dpkg -l corekeeper
>>     > Desired=Unknown/Install/Remove/Purge/Hold
>>     > | 
>> Status=Not/Inst/Conf-files/Unpacked/halF-conf/Half-inst/trig-aWait/Trig-pend
>>     > |/ Err?=(none)/Reinst-required (Status,Err: uppercase=bad)
>>     > ||/ Name                 Version         Architecture    Description
>>     > 
>> +++-====================-===============-===============-==============================================
>>     > ii  corekeeper           1.6             amd64           enable core
>>     > files and report crashes to the system
>>     >
>>     > Thanks.
>>     >
>>     > On Tue, May 29, 2018 at 9:38 AM, Steven Luong (sluong) 
>> <slu...@cisco.com> wrote:
>>     >> Ravi,
>>     >>
>>     >> I install corekeeper and the core file is kept in /var/crash. But why 
>> not use gdb to attach to the VPP process?
>>     >> To turn on VPP vhost-user debug, type "debug vhost-user on" at the 
>> VPP prompt.
>>     >>
>>     >> Steven
>>     >>
>>     >> On 5/29/18, 9:10 AM, "vpp-dev@lists.fd.io on behalf of Ravi Kerur" 
>> <vpp-dev@lists.fd.io on behalf of rke...@gmail.com> wrote:
>>     >>
>>     >>     Hi Marco,
>>     >>
>>     >>
>>     >>     On Tue, May 29, 2018 at 6:30 AM, Marco Varlese <mvarl...@suse.de> 
>> wrote:
>>     >>     > Ravi,
>>     >>     >
>>     >>     > On Sun, 2018-05-27 at 12:20 -0700, Ravi Kerur wrote:
>>     >>     >> Hello,
>>     >>     >>
>>     >>     >> I have a VM(16.04.4 Ubuntu x86_64) with 2 cores and 4G RAM. I 
>> have
>>     >>     >> installed VPP successfully on it. Later I have created 
>> vhost-user
>>     >>     >> interfaces via
>>     >>     >>
>>     >>     >> create vhost socket /var/run/vpp/sock1.sock server
>>     >>     >> create vhost socket /var/run/vpp/sock2.sock server
>>     >>     >> set interface state VirtualEthernet0/0/0 up
>>     >>     >> set interface state VirtualEthernet0/0/1 up
>>     >>     >>
>>     >>     >> set interface l2 bridge VirtualEthernet0/0/0 1
>>     >>     >> set interface l2 bridge VirtualEthernet0/0/1 1
>>     >>     >>
>>     >>     >> I then run 'DPDK/testpmd' inside a container which will use
>>     >>     >> virtio-user interfaces using the following command
>>     >>     >>
>>     >>     >> docker run -it --privileged -v
>>     >>     >> /var/run/vpp/sock1.sock:/var/run/usvhost1 -v
>>     >>     >> /var/run/vpp/sock2.sock:/var/run/usvhost2 -v
>>     >>     >> /dev/hugepages:/dev/hugepages dpdk-app-testpmd ./bin/testpmd 
>> -c 0x3 -n
>>     >>     >> 4 --log-level=9 -m 64 --no-pci --single-file-segments
>>     >>     >> 
>> --vdev=virtio_user0,path=/var/run/usvhost1,mac=54:01:00:01:01:01
>>     >>     >> 
>> --vdev=virtio_user1,path=/var/run/usvhost2,mac=54:01:00:01:01:02 --
>>     >>     >> -i
>>     >>     >>
>>     >>     >> VPP Vnet crashes with following message
>>     >>     >>
>>     >>     >> May 27 11:44:00 localhost vnet[6818]: received signal SIGSEGV, 
>> PC
>>     >>     >> 0x7fcca4620187, faulting address 0x7fcb317ac000
>>     >>     >>
>>     >>     >> Questions:
>>     >>     >> I have 'ulimit -c unlimited' and /etc/vpp/startup.conf has
>>     >>     >> unix {
>>     >>     >>   nodaemon
>>     >>     >>   log /var/log/vpp/vpp.log
>>     >>     >>   full-coredump
>>     >>     >>   cli-listen /run/vpp/cli.sock
>>     >>     >>   gid vpp
>>     >>     >> }
>>     >>     >>
>>     >>     >> But I couldn't locate corefile?
>>     >>     > The location of the coredump file depends on your system 
>> configuration.
>>     >>     >
>>     >>     > Please, check "cat /proc/sys/kernel/core_pattern"
>>     >>     >
>>     >>     > If you have systemd-coredump in the output of the above 
>> command, then likely the
>>     >>     > location of the coredump files is "/var/lib/systemd/coredump/"
>>     >>     >
>>     >>     > You can also change the location of where your system places 
>> the coredump files:
>>     >>     > echo '/PATH_TO_YOU_LOCATION/core_%e.%p' | sudo tee 
>> /proc/sys/kernel/core_pattern
>>     >>     >
>>     >>     > See if that helps...
>>     >>     >
>>     >>
>>     >>     Initially '/proc/sys/kernel/core_pattern' was set to 'core'. I 
>> changed
>>     >>     it to 'systemd-coredump'. Still no core generated. VPP crashes
>>     >>
>>     >>     May 29 08:54:34 localhost vnet[4107]: received signal SIGSEGV, PC
>>     >>     0x7f0167751187, faulting address 0x7efff43ac000
>>     >>     May 29 08:54:34 localhost systemd[1]: vpp.service: Main process
>>     >>     exited, code=killed, status=6/ABRT
>>     >>     May 29 08:54:34 localhost systemd[1]: vpp.service: Unit entered 
>> failed state.
>>     >>     May 29 08:54:34 localhost systemd[1]: vpp.service: Failed with 
>> result 'signal'.
>>     >>
>>     >>
>>     >>     cat /proc/sys/kernel/core_pattern
>>     >>     systemd-coredump
>>     >>
>>     >>
>>     >>     ulimit -a
>>     >>     core file size          (blocks, -c) unlimited
>>     >>     data seg size           (kbytes, -d) unlimited
>>     >>     scheduling priority             (-e) 0
>>     >>     file size               (blocks, -f) unlimited
>>     >>     pending signals                 (-i) 15657
>>     >>     max locked memory       (kbytes, -l) 64
>>     >>     max memory size         (kbytes, -m) unlimited
>>     >>     open files                      (-n) 1024
>>     >>     pipe size            (512 bytes, -p) 8
>>     >>     POSIX message queues     (bytes, -q) 819200
>>     >>     real-time priority              (-r) 0
>>     >>     stack size              (kbytes, -s) 8192
>>     >>     cpu time               (seconds, -t) unlimited
>>     >>     max user processes              (-u) 15657
>>     >>     virtual memory          (kbytes, -v) unlimited
>>     >>     file locks                      (-x) unlimited
>>     >>
>>     >>     cd /var/lib/systemd/coredump/
>>     >>     root@localhost:/var/lib/systemd/coredump# ls
>>     >>     root@localhost:/var/lib/systemd/coredump#
>>     >>
>>     >>     >>
>>     >>     >> (2) How to enable debugs? I have used 'make build' but no 
>> additional
>>     >>     >> logs other than those shown below
>>     >>     >>
>>     >>     >>
>>     >>     >> VPP logs from /var/log/syslog is shown below
>>     >>     >> cat /var/log/syslog
>>     >>     >> May 27 11:40:28 localhost vpp[6818]: 
>> vlib_plugin_early_init:361:
>>     >>     >> plugin path /usr/lib/vpp_plugins:/usr/lib64/vpp_plugins
>>     >>     >> May 27 11:40:28 localhost vpp[6818]: load_one_plugin:189: 
>> Loaded
>>     >>     >> plugin: abf_plugin.so (ACL based Forwarding)
>>     >>     >> May 27 11:40:28 localhost vpp[6818]: load_one_plugin:189: 
>> Loaded
>>     >>     >> plugin: acl_plugin.so (Access Control Lists)
>>     >>     >> May 27 11:40:28 localhost vpp[6818]: load_one_plugin:189: 
>> Loaded
>>     >>     >> plugin: avf_plugin.so (Intel Adaptive Virtual Function (AVF) 
>> Device
>>     >>     >> Plugin)
>>     >>     >> May 27 11:40:28 localhost vpp[6818]: load_one_plugin:191: 
>> Loaded
>>     >>     >> plugin: cdp_plugin.so
>>     >>     >> May 27 11:40:28 localhost vpp[6818]: load_one_plugin:189: 
>> Loaded
>>     >>     >> plugin: dpdk_plugin.so (Data Plane Development Kit (DPDK))
>>     >>     >> May 27 11:40:28 localhost vpp[6818]: load_one_plugin:189: 
>> Loaded
>>     >>     >> plugin: flowprobe_plugin.so (Flow per Packet)
>>     >>     >> May 27 11:40:28 localhost vpp[6818]: load_one_plugin:189: 
>> Loaded
>>     >>     >> plugin: gbp_plugin.so (Group Based Policy)
>>     >>     >> May 27 11:40:28 localhost vpp[6818]: load_one_plugin:189: 
>> Loaded
>>     >>     >> plugin: gtpu_plugin.so (GTPv1-U)
>>     >>     >> May 27 11:40:28 localhost vpp[6818]: load_one_plugin:189: 
>> Loaded
>>     >>     >> plugin: igmp_plugin.so (IGMP messaging)
>>     >>     >> May 27 11:40:28 localhost vpp[6818]: load_one_plugin:189: 
>> Loaded
>>     >>     >> plugin: ila_plugin.so (Identifier-locator addressing for IPv6)
>>     >>     >> May 27 11:40:28 localhost vpp[6818]: load_one_plugin:189: 
>> Loaded
>>     >>     >> plugin: ioam_plugin.so (Inbound OAM)
>>     >>     >> May 27 11:40:28 localhost vpp[6818]: load_one_plugin:117: 
>> Plugin
>>     >>     >> disabled (default): ixge_plugin.so
>>     >>     >> May 27 11:40:28 localhost vpp[6818]: load_one_plugin:189: 
>> Loaded
>>     >>     >> plugin: l2e_plugin.so (L2 Emulation)
>>     >>     >> May 27 11:40:28 localhost vpp[6818]: load_one_plugin:189: 
>> Loaded
>>     >>     >> plugin: lacp_plugin.so (Link Aggregation Control Protocol)
>>     >>     >> May 27 11:40:28 localhost vpp[6818]: load_one_plugin:189: 
>> Loaded
>>     >>     >> plugin: lb_plugin.so (Load Balancer)
>>     >>     >> May 27 11:40:28 localhost vpp[6818]: load_one_plugin:189: 
>> Loaded
>>     >>     >> plugin: memif_plugin.so (Packet Memory Interface (experimetal))
>>     >>     >> May 27 11:40:28 localhost vpp[6818]: load_one_plugin:189: 
>> Loaded
>>     >>     >> plugin: nat_plugin.so (Network Address Translation)
>>     >>     >> May 27 11:40:28 localhost vpp[6818]: load_one_plugin:189: 
>> Loaded
>>     >>     >> plugin: pppoe_plugin.so (PPPoE)
>>     >>     >> May 27 11:40:28 localhost vpp[6818]: load_one_plugin:189: 
>> Loaded
>>     >>     >> plugin: srv6ad_plugin.so (Dynamic SRv6 proxy)
>>     >>     >> May 27 11:40:28 localhost vpp[6818]: load_one_plugin:189: 
>> Loaded
>>     >>     >> plugin: srv6am_plugin.so (Masquerading SRv6 proxy)
>>     >>     >> May 27 11:40:28 localhost vpp[6818]: load_one_plugin:189: 
>> Loaded
>>     >>     >> plugin: srv6as_plugin.so (Static SRv6 proxy)
>>     >>     >> May 27 11:40:29 localhost vpp[6818]: load_one_plugin:189: 
>> Loaded
>>     >>     >> plugin: stn_plugin.so (VPP Steals the NIC for Container 
>> integration)
>>     >>     >> May 27 11:40:29 localhost vpp[6818]: load_one_plugin:189: 
>> Loaded
>>     >>     >> plugin: tlsmbedtls_plugin.so (mbedtls based TLS Engine)
>>     >>     >> May 27 11:40:29 localhost vpp[6818]: load_one_plugin:189: 
>> Loaded
>>     >>     >> plugin: tlsopenssl_plugin.so (openssl based TLS Engine)
>>     >>     >> May 27 11:40:29 localhost vpp[6818]: /usr/bin/vpp[6818]:
>>     >>     >> load_one_vat_plugin:67: Loaded plugin: dpdk_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: 
>> load_one_vat_plugin:67:
>>     >>     >> Loaded plugin: dpdk_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost vpp[6818]: /usr/bin/vpp[6818]:
>>     >>     >> load_one_vat_plugin:67: Loaded plugin: lb_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost vpp[6818]: /usr/bin/vpp[6818]:
>>     >>     >> load_one_vat_plugin:67: Loaded plugin: flowprobe_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost vpp[6818]: /usr/bin/vpp[6818]:
>>     >>     >> load_one_vat_plugin:67: Loaded plugin: stn_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost vpp[6818]: /usr/bin/vpp[6818]:
>>     >>     >> load_one_vat_plugin:67: Loaded plugin: nat_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost vpp[6818]: /usr/bin/vpp[6818]:
>>     >>     >> load_one_vat_plugin:67: Loaded plugin: udp_ping_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost vpp[6818]: /usr/bin/vpp[6818]:
>>     >>     >> load_one_vat_plugin:67: Loaded plugin: pppoe_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost vpp[6818]: /usr/bin/vpp[6818]:
>>     >>     >> load_one_vat_plugin:67: Loaded plugin: lacp_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: 
>> load_one_vat_plugin:67:
>>     >>     >> Loaded plugin: lb_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost vpp[6818]: /usr/bin/vpp[6818]:
>>     >>     >> load_one_vat_plugin:67: Loaded plugin: acl_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost vpp[6818]: /usr/bin/vpp[6818]:
>>     >>     >> load_one_vat_plugin:67: Loaded plugin: 
>> ioam_export_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost vpp[6818]: /usr/bin/vpp[6818]:
>>     >>     >> load_one_vat_plugin:67: Loaded plugin: 
>> ioam_trace_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost vpp[6818]: /usr/bin/vpp[6818]:
>>     >>     >> load_one_vat_plugin:67: Loaded plugin:
>>     >>     >> vxlan_gpe_ioam_export_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost vpp[6818]: /usr/bin/vpp[6818]:
>>     >>     >> load_one_vat_plugin:67: Loaded plugin: gtpu_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost vpp[6818]: /usr/bin/vpp[6818]:
>>     >>     >> load_one_vat_plugin:67: Loaded plugin: cdp_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost vpp[6818]: /usr/bin/vpp[6818]:
>>     >>     >> load_one_vat_plugin:67: Loaded plugin: 
>> ioam_vxlan_gpe_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost vpp[6818]: /usr/bin/vpp[6818]:
>>     >>     >> load_one_vat_plugin:67: Loaded plugin: memif_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost vpp[6818]: /usr/bin/vpp[6818]:
>>     >>     >> load_one_vat_plugin:67: Loaded plugin: ioam_pot_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: 
>> load_one_vat_plugin:67:
>>     >>     >> Loaded plugin: flowprobe_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: 
>> load_one_vat_plugin:67:
>>     >>     >> Loaded plugin: stn_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: 
>> load_one_vat_plugin:67:
>>     >>     >> Loaded plugin: nat_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: 
>> load_one_vat_plugin:67:
>>     >>     >> Loaded plugin: udp_ping_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: 
>> load_one_vat_plugin:67:
>>     >>     >> Loaded plugin: pppoe_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: 
>> load_one_vat_plugin:67:
>>     >>     >> Loaded plugin: lacp_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: 
>> load_one_vat_plugin:67:
>>     >>     >> Loaded plugin: acl_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: 
>> load_one_vat_plugin:67:
>>     >>     >> Loaded plugin: ioam_export_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: 
>> load_one_vat_plugin:67:
>>     >>     >> Loaded plugin: ioam_trace_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: 
>> load_one_vat_plugin:67:
>>     >>     >> Loaded plugin: vxlan_gpe_ioam_export_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: 
>> load_one_vat_plugin:67:
>>     >>     >> Loaded plugin: gtpu_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: 
>> load_one_vat_plugin:67:
>>     >>     >> Loaded plugin: cdp_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: 
>> load_one_vat_plugin:67:
>>     >>     >> Loaded plugin: ioam_vxlan_gpe_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: 
>> load_one_vat_plugin:67:
>>     >>     >> Loaded plugin: memif_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: 
>> load_one_vat_plugin:67:
>>     >>     >> Loaded plugin: ioam_pot_test_plugin.so
>>     >>     >> May 27 11:40:29 localhost vpp[6818]: /usr/bin/vpp[6818]: dpdk: 
>> EAL
>>     >>     >> init args: -c 1 -n 4 --no-pci --huge-dir /dev/hugepages 
>> --master-lcore
>>     >>     >> 0 --socket-mem 256,0
>>     >>     >> May 27 11:40:29 localhost /usr/bin/vpp[6818]: dpdk: EAL init 
>> args: -c
>>     >>     >> 1 -n 4 --no-pci --huge-dir /dev/hugepages --master-lcore 0
>>     >>     >> --socket-mem 256,0
>>     >>     >> May 27 11:40:29 localhost vnet[6818]: dpdk_ipsec_process:1019: 
>> not
>>     >>     >> enough DPDK crypto resources, default to OpenSSL
>>     >>     >> May 27 11:43:19 localhost vnet[6818]: show vhost-user: unknown 
>> input `detail
>>     >>     >> May 27 11:44:00 localhost vnet[6818]: received signal SIGSEGV, 
>> PC
>>     >>     >> 0x7fcca4620187, faulting address 0x7fcb317ac000
>>     >>     >> May 27 11:44:00 localhost systemd[1]: vpp.service: Main process
>>     >>     >> exited, code=killed, status=6/ABRT
>>     >>     >> May 27 11:44:00 localhost systemd[1]: vpp.service: Unit 
>> entered failed state.
>>     >>     >> May 27 11:44:00 localhost systemd[1]: vpp.service: Failed with 
>> result
>>     >>     >> 'signal'.
>>     >>     >> May 27 11:44:00 localhost systemd[1]: vpp.service: Service 
>> hold-off
>>     >>     >> time over, scheduling restart
>>     >>     >>
>>     >>
>>     >>     Thanks,
>>     >>     Ravi
>>     >>
>>     >>     >>
>>     >>     >>
>>     >>     >> Thanks.
>>     >>     > Cheers,
>>     >>     > Marco
>>     >>     >>
>>     >>     >>
>>     >>     >>
>>     >>     > --
>>     >>     > Marco V
>>     >>     >
>>     >>     > SUSE LINUX GmbH | GF: Felix Imendörffer, Jane Smithard, Graham 
>> Norton
>>     >>     > HRB 21284 (AG Nürnberg) Maxfeldstr. 5, D-90409, Nürnberg
>>     >>
>>     >>
>>     >>
>>     >>
>>     >>
>>     >
>>     >
>>     >
>>
>>
>
> 
>

-=-=-=-=-=-=-=-=-=-=-=-
Links:

You receive all messages sent to this group.

View/Reply Online (#9467): https://lists.fd.io/g/vpp-dev/message/9467
View All Messages In Topic (9): https://lists.fd.io/g/vpp-dev/topic/20346431
Mute This Topic: https://lists.fd.io/mt/20346431/21656
New Topic: https://lists.fd.io/g/vpp-dev/post

Change Your Subscription: https://lists.fd.io/g/vpp-dev/editsub/21656
Group Home: https://lists.fd.io/g/vpp-dev
Contact Group Owner: vpp-dev+ow...@lists.fd.io
Terms of Service: https://lists.fd.io/static/tos
Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub
-=-=-=-=-=-=-=-=-=-=-=-

Reply via email to