Hello,

I have tried increasing the workers from 2 to 5 and rx/tx queues from 1024
(default) to 2048 ,also decreasing till 256 but of no use.
As you told, vector is very high (> 100). Please let us know if there is
any other way or reason for the same.


Thread 1 vpp_wk_0 (lcore 2)
Time 41.3, 10 sec internal node vector rate 79.42 loops/sec 3628.25
  vector rates in 1.9388e5, out 6.5319e2, drop 1.9322e5, punt 0.0000e0
             Name                 State         Calls          Vectors
   Suspends         Clocks       Vectors/Call
TenGigabitEthernet4/0/0-output   active              19735           26978
              0          2.12e3            1.37
TenGigabitEthernet4/0/0-tx       active              19735           26969
              0          1.43e3            1.37
dpdk-input                       polling          10241833         8004674
              0          2.81e3             .78
drop                             active              92422         7977696
              0          1.31e3           86.32
error-drop                       active              92422         7977696
              0          6.82e1           86.32
ethernet-input                   active              93109         8004674
              0          1.84e2           85.97
gtpu4-input                      active              93106         8004671
              0          3.29e2           85.97
ip4-drop                         active                  2               2
              0          1.33e4            1.00
ip4-input                        active              93106         8004671
              0          6.82e2           85.97
ip4-input-no-checksum            active              93106         8004673
              0          2.37e2           85.97
ip4-local                        active              93106         8004671
              0          2.66e2           85.97
ip4-lookup                       active             112841         8031651
              0          3.55e2           71.18
ip4-policer-classify             active              93106         8004671
              0          1.35e3           85.97
ip4-rewrite                      active              19735           26978
              0          2.43e3            1.37
ip4-udp-lookup                   active              93106         8004671
              0          3.17e2           85.97
ip6-input                        active                  1               1
              0          1.99e4            1.00
ip6-not-enabled                  active                  1               1
              0          2.59e4            1.00
unix-epoll-input                 polling              9998               0
              0          3.30e3            0.00
---------------
Thread 2 vpp_wk_1 (lcore 3)
Time 41.3, 10 sec internal node vector rate 0.00 loops/sec 988640.31
  vector rates in 0.0000e0, out 0.0000e0, drop 0.0000e0, punt 0.0000e0
             Name                 State         Calls          Vectors
   Suspends         Clocks       Vectors/Call
dpdk-input                       polling          37858263               0
              0          1.14e3            0.00
unix-epoll-input                 polling             36940               0
              0          3.54e3            0.00
---------------
Thread 3 vpp_wk_2 (lcore 4)
Time 41.3, 10 sec internal node vector rate 0.00 loops/sec 983700.23
  vector rates in 4.8441e-2, out 0.0000e0, drop 4.8441e-2, punt 0.0000e0
             Name                 State         Calls          Vectors
   Suspends         Clocks       Vectors/Call
dpdk-input                       polling          37820862               2
              0         2.16e10            0.00
drop                             active                  2               2
              0          9.34e3            1.00
error-drop                       active                  2               2
              0          7.78e3            1.00
ethernet-input                   active                  2               2
              0          2.44e4            1.00
ip6-input                        active                  2               2
              0          1.06e4            1.00
ip6-not-enabled                  active                  2               2
              0          8.75e3            1.00
unix-epoll-input                 polling             36906               0
              0          3.54e3            0.00
---------------
Thread 4 vpp_wk_3 (lcore 5)
Time 41.3, 10 sec internal node vector rate 0.00 loops/sec 985307.67
  vector rates in 0.0000e0, out 0.0000e0, drop 0.0000e0, punt 0.0000e0
             Name                 State         Calls          Vectors
   Suspends         Clocks       Vectors/Call
dpdk-input                       polling          37834354               0
              0          1.14e3            0.00
unix-epoll-input                 polling             36918               0
              0          3.47e3            0.00

Thanks,
Akash

On Tue, Sep 28, 2021 at 4:09 PM Akash S R via lists.fd.io <akashsr.akashsr=
gmail....@lists.fd.io> wrote:

> Thanks mate, that makes sense. Will check it out and get back .
>
>
> Thanks,
> Akash
>
> On Tue, Sep 28, 2021, 16:00 Benoit Ganne (bganne) <bga...@cisco.com>
> wrote:
>
>> Rx-miss means the NIC must drop packets on RX because the rx queue was
>> full, usually because VPP cannot keep up with the incoming packet rate.
>> You can check it with the output of 'show run'. If you see a big average
>> vector size (100 or more) it means VPP is busy.
>> To improve that you must increase the number of VPP workers (and rx
>> queues).
>> Rx-misses can also be caused by traffic spikes. In that case you can
>> increase the rx queues size to absorb the bursts.
>>
>> Best
>> ben
>>
>> > -----Original Message-----
>> > From: vpp-dev@lists.fd.io <vpp-dev@lists.fd.io> On Behalf Of Akash S R
>> > Sent: mardi 28 septembre 2021 11:43
>> > To: vpp-dev <vpp-dev@lists.fd.io>
>> > Subject: [vpp-dev] rx-miss while sending packets to Interface
>> (IMPORTANT)
>> >
>> > Hello mates,
>> >
>> > Its been a long time, me , raising a query out here to you guys :) (Nah,
>> > Please ignore this junk.)
>> >
>> > I have a question on rx-miss. Whenever we fire packets at some high
>> rate,
>> > say 1GBPS or more, I get rx-miss on the interface with some packet
>> count.
>> > I referred to a link below from the VPP forum:
>> https://lists.fd.io/g/vpp-
>> > dev/topic/78179815#17985
>> >
>> > TenGigabitEthernet4/0/1           3      up          1500/0/0/0     rx
>> > packets               8622948
>> >                                                                     rx
>> > bytes              1103737344
>> >                                                                     ip4
>> > 8622948
>> >                                                                     rx-
>> > miss                283721416
>> >
>> > But the buffer is available and allocated is less. I reduced the queues
>> to
>> > 256 and threads to 4.
>> > But still the issue is not resolved.
>> > DBGvpp# show dpdk buffer
>> > name="vpp pool 0"  available =   13984 allocated =    2816 total =
>>  16800
>> >
>> >  May I know the reason for this rx-miss and any kinda resolution for
>> this
>> > issue?
>> >
>> > Thanks,
>> > Akash
>>
>>
> 
>
>
-=-=-=-=-=-=-=-=-=-=-=-
Links: You receive all messages sent to this group.
View/Reply Online (#20219): https://lists.fd.io/g/vpp-dev/message/20219
Mute This Topic: https://lists.fd.io/mt/85920772/21656
Group Owner: vpp-dev+ow...@lists.fd.io
Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub [arch...@mail-archive.com]
-=-=-=-=-=-=-=-=-=-=-=-

Reply via email to