Hi Bob and Tim, Thank you for responding to my question. sometime today will try increasing the window. Since I am testing this in a virtual environment (OpenStack), it is hard to define the wire limitation. I get UDP throughput somewhere between 100-120 Mbps and TCP throughput between 50-100 Mbps (using 1402 (GENEVE) & 1410 (VXLAN) mss since I am using tunneling protocol).
In the case of UDP that Bob saying. so the limitation you said is in the intermediate network, right? so the client will force to send all packet (by increasing window size) but will get dropped in the intermediate net? Best Regards, Zufar Dhiyaulhaq On Mon, Aug 3, 2020 at 2:05 AM Bob McMahon <bob.mcma...@broadcom.com> wrote: > If the bottleneck is the transmitter's wire that means things will back up > behind that. The network stack on the client will queue packets. Since > it's in a state of oversubscription there is no way for the client to ever > drain the bottleneck so-to-speak. A bottleneck is when the service time is > less than the arrival time. So one host to look more closely at the client > host, where the bottleneck is, to understand. There are two things > happening, iperf is issuing writes() and the network stack is sending > packets. While related, they're different. > > The iperf client issues a write() to cause the sending of a packet. If > the operating system has system buffers it will accept the write() > otherwise it has two choices, block or suspend the write until a system > buffer comes available or error on the write. What I suspect you're seeing > is an os blocking on the write(). Increasing the window size will allow > the os to accept the write and pass the packet to the network stack, which > will in turn drop the packet. Then you'll see packet loss. > > Did you try with a bigger window? > > Bob > > On Fri, Jul 31, 2020 at 4:29 PM Zufar Dhiyaulhaq < > zufardhiyaul...@gmail.com> wrote: > >> Hi Bob, >> >> Thanks for replying. In my understanding, when increasing the bitrate >> above the bandwidth/throughput, it will increase the packet loss right? but >> in my case, I increase to 9 Gbps and still not seeing any packet loss. Did >> increasing window size will increasing packet loss? and why that can happen? >> >> I am trying to simulate packet loss. >> >> Thanks >> >> Best Regards, >> Zufar Dhiyaulhaq >> >> >> On Sat, Aug 1, 2020 at 5:28 AM Bob McMahon <bob.mcma...@broadcom.com> >> wrote: >> >>> Try to increase the window size with -w on the client. This will allow >>> the operating system to accept the write and drop packets within the >>> stack. If the window is too small the operating system will block the >>> write until os buffers are available. >>> >>> Bob >>> >>> On Fri, Jul 31, 2020 at 8:56 AM Zufar Dhiyaulhaq < >>> zufardhiyaul...@gmail.com> wrote: >>> >>>> Hi Folks, >>>> >>>> I have a problem with iperf3, I try to simulate packet loss with Iperf3 >>>> with increasing the bitrate above the bandwidth. But packet loss output not >>>> increasing. >>>> >>>> Ubuntu 18.04 >>>> Iperf 3.7.3 >>>> >>>> I don't know why this is happening? Is there any bug with Iperf? this >>>> sounds stupid for me. >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> *ubuntu@vm1:~$ iperf3 -c 192.168.0.92 --udp -t 20 --bitrate 9000m -R >>>> -Viperf 3.7Linux vm1 4.15.0-112-generic #113-Ubuntu SMP Thu Jul 9 23:41:39 >>>> UTC 2020 x86_64Control connection MSS 1390Setting UDP block size to >>>> 1390Time: Fri, 31 Jul 2020 15:53:39 GMTConnecting to host 192.168.0.92, >>>> port 5201Reverse mode, remote host 192.168.0.92 is sending Cookie: >>>> geu5ktrvwtalkelbszen5ym4rzxfp5xgzwdy Target Bitrate: 9000000000[ 5] >>>> local 192.168.0.226 port 47999 connected to 192.168.0.92 port 5201Starting >>>> Test: protocol: UDP, 1 streams, 1390 byte blocks, omitting 0 seconds, 20 >>>> second test, tos 0[ ID] Interval Transfer Bitrate >>>> Jitter Lost/Total Datagrams[ 5] 0.00-1.00 sec 13.8 MBytes 116 >>>> Mbits/sec 0.081 ms 269/10704 (2.5%) [ 5] 1.00-2.00 sec 13.7 MBytes >>>> 115 Mbits/sec 0.085 ms 0/10346 (0%) [ 5] 2.00-3.00 sec 13.6 >>>> MBytes 114 Mbits/sec 0.035 ms 126/10365 (1.2%) [ 5] 3.00-4.00 sec >>>> 12.8 MBytes 107 Mbits/sec 0.033 ms 279/9946 (2.8%) [ 5] 4.00-5.00 >>>> sec 13.5 MBytes 113 Mbits/sec 0.051 ms 262/10427 (2.5%) [ 5] >>>> 5.00-6.00 sec 13.2 MBytes 111 Mbits/sec 0.058 ms 0/9965 (0%) [ 5] >>>> 6.00-7.00 sec 13.3 MBytes 111 Mbits/sec 0.044 ms 32/10047 (0.32%) >>>> [ 5] 7.00-8.00 sec 13.0 MBytes 109 Mbits/sec 0.053 ms 43/9874 >>>> (0.44%) [ 5] 8.00-9.00 sec 13.0 MBytes 109 Mbits/sec 0.042 ms >>>> 34/9847 (0.35%) [ 5] 9.00-10.00 sec 13.6 MBytes 114 Mbits/sec >>>> 0.055 ms 78/10305 (0.76%) [ 5] 10.00-11.00 sec 13.5 MBytes 113 >>>> Mbits/sec 0.070 ms 0/10171 (0%) [ 5] 11.00-12.00 sec 13.1 MBytes >>>> 110 Mbits/sec 0.047 ms 0/9851 (0%) [ 5] 12.00-13.00 sec 13.3 MBytes >>>> 112 Mbits/sec 0.034 ms 0/10055 (0%) [ 5] 13.00-14.00 sec 13.4 >>>> MBytes 112 Mbits/sec 0.040 ms 36/10136 (0.36%) [ 5] 14.00-15.00 sec >>>> 13.9 MBytes 117 Mbits/sec 0.055 ms 437/10921 (4%) [ 5] 15.00-16.00 >>>> sec 13.2 MBytes 111 Mbits/sec 0.043 ms 25/9964 (0.25%) [ 5] >>>> 16.00-17.00 sec 13.2 MBytes 110 Mbits/sec 0.043 ms 21/9942 (0.21%) >>>> [ 5] 17.00-18.00 sec 12.9 MBytes 108 Mbits/sec 0.046 ms 0/9702 >>>> (0%) [ 5] 18.00-19.00 sec 13.4 MBytes 112 Mbits/sec 0.050 ms >>>> 208/10294 (2%) [ 5] 19.00-20.00 sec 13.5 MBytes 113 Mbits/sec >>>> 0.048 ms 0/10152 (0%) - - - - - - - - - - - - - - - - - - - - - - - - >>>> -Test Complete. Summary Results:[ ID] Interval Transfer >>>> Bitrate Jitter Lost/Total Datagrams[ 5] 0.00-20.04 sec 269 >>>> MBytes 113 Mbits/sec 0.000 ms 0/203058 (0%) sender[ 5] 0.00-20.00 >>>> sec 267 MBytes 112 Mbits/sec 0.048 ms 1850/203014 (0.91%) receiver* >>>> >>>> Thank you >>>> >>>> Best Regards, >>>> Zufar Dhiyaulhaq >>>> _______________________________________________ >>>> Iperf-users mailing list >>>> Iperf-users@lists.sourceforge.net >>>> https://lists.sourceforge.net/lists/listinfo/iperf-users >>>> >>>
_______________________________________________ Iperf-users mailing list Iperf-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/iperf-users