Hi , I am insterested in connecting two instance of vpp by memif (one vpp is running in host and another vpp is running in a lxc container). I have achieved functionality goal by memif but I have problem in performance test. I have done the following steps respectively: 1. first of all I installed lxc on my system 2. then, I made and installed vpp in lxc (I'll call it lxcvpp) 3. I installed vpp on my system (I'll call it hostvpp) 4. I created 2 memif on hostvpp create memif socket /tmp/unix_socket.file1 master create memif socket /tmp/unix_socket.file2 slave 5. I created 2 memif on lxcvpp create memif socket /share/unix_socket.file1 slave create memif socket /share/unix_socket.file2 master 6. I have two physical interface which are binded to hostvpp (I call two interfaces eth1 and eth2 respectively). so, I bridged my input interface (eth1) and memif0 (bridge-domain 1) and also bridged eth2 and memif1 (bridge-domain 2).
7. moreover, I bridged memif0 and memif1 in lxcvpp. 8. I used trex as traffic generator. trafic is transmitted from trex to hostvpp by eth1 and it is recieved from eth2 interface of hostvpp. packets flow of this scenario is shown below. trex---->eth0(hostvpp)----> memif0(hostvpp)---->memif0(lxcvpp)---->memif1(lxcvpp)---->memif1(hostvpp)---->eth2(hostvpp)---->trex After running trex, I got 4 MPPS with 64B packet size. Is it the maximum throughput of memif in this scenario? I expected much more throughput than 4 MPPS. Is there any solution and tuning to obtain more throughput? (I allocated one core to hostvpp and another core to lxcvpp) Cheers, khers
_______________________________________________ vpp-dev mailing list vpp-dev@lists.fd.io https://lists.fd.io/mailman/listinfo/vpp-dev