Just an FYI, comparing a VO with 1 mpdu per ampdu vs 32 and small packets (minimize airtime.) If VO can break 10K pps it's most likely aggregating.
[root@localhost iperf2-code]# wl -i ap0 ampdu_mpdu 1 root@localhost iperf2-code]# iperf -c 192.168.1.4 -u -S 0xC0 -b 40m -e -i 1 -t 10 -l 20 ------------------------------------------------------------ Client connecting to 192.168.1.4, UDP port 5001 with pid 16463 Sending 20 byte datagrams, IPG target: 4.00 us (kalman adjust) UDP buffer size: 208 KByte (default) ------------------------------------------------------------ [ 3] local 192.168.1.1 port 41010 connected with 192.168.1.4 port 5001 [ ID] Interval Transfer Bandwidth Write/Err PPS [ 3] 0.0000-1.0000 sec 161 KBytes 1.32 Mbits/sec 8243/0 8167 pps [ 3] 1.0000-2.0000 sec 158 KBytes 1.29 Mbits/sec 8085/0 8035 pps [ 3] 2.0000-3.0000 sec 155 KBytes 1.27 Mbits/sec 7928/0 8040 pps [ 3] 3.0000-4.0000 sec 158 KBytes 1.29 Mbits/sec 8075/0 8035 pps [ 3] 4.0000-5.0000 sec 158 KBytes 1.30 Mbits/sec 8094/0 8019 pps [ 3] 5.0000-6.0000 sec 155 KBytes 1.27 Mbits/sec 7954/0 8032 pps [ 3] 6.0000-7.0000 sec 158 KBytes 1.30 Mbits/sec 8097/0 8040 pps [ 3] 7.0000-8.0000 sec 155 KBytes 1.27 Mbits/sec 7930/0 8034 pps [ 3] 8.0000-9.0000 sec 158 KBytes 1.30 Mbits/sec 8095/0 8034 pps [ 3] 0.0000-10.0153 sec 1.54 MBytes 1.29 Mbits/sec 80596/0 8047 pps [ 3] Sent 80596 datagrams [ 3] Server Report: [ 3] 0.0000-10.0293 sec 1.54 MBytes 1.29 Mbits/sec 1.179 ms 0/80596 (0%) 26.031/ 0.709/36.593/ 0.197 ms 8036 pps 6.17 [root@localhost iperf2-code]# wl -i ap0 ampdu_mpdu 32 [root@localhost iperf2-code]# iperf -c 192.168.1.4 -u -S 0xC0 -b 40m -e -i 1 -t 10 -l 20 ------------------------------------------------------------ Client connecting to 192.168.1.4, UDP port 5001 with pid 16467 Sending 20 byte datagrams, IPG target: 4.00 us (kalman adjust) UDP buffer size: 208 KByte (default) ------------------------------------------------------------ [ 3] local 192.168.1.1 port 58139 connected with 192.168.1.4 port 5001 [ ID] Interval Transfer Bandwidth Write/Err PPS [ 3] 0.0000-1.0000 sec 797 KBytes 6.53 Mbits/sec 40826/0 40801 pps [ 3] 1.0000-2.0000 sec 796 KBytes 6.52 Mbits/sec 40737/0 40727 pps [ 3] 2.0000-3.0000 sec 794 KBytes 6.50 Mbits/sec 40652/0 40685 pps [ 3] 3.0000-4.0000 sec 797 KBytes 6.53 Mbits/sec 40815/0 40708 pps [ 3] 4.0000-5.0000 sec 793 KBytes 6.50 Mbits/sec 40613/0 40656 pps [ 3] 5.0000-6.0000 sec 794 KBytes 6.51 Mbits/sec 40663/0 40692 pps [ 3] 6.0000-7.0000 sec 796 KBytes 6.52 Mbits/sec 40779/0 40722 pps [ 3] 7.0000-8.0000 sec 793 KBytes 6.50 Mbits/sec 40624/0 40704 pps [ 3] 8.0000-9.0000 sec 797 KBytes 6.53 Mbits/sec 40813/0 40713 pps [ 3] 0.0000-10.0018 sec 7.77 MBytes 6.51 Mbits/sec 407178/0 40710 pps [ 3] Sent 407178 datagrams [ 3] Server Report: [ 3] 0.0000-10.0022 sec 7.77 MBytes 6.51 Mbits/sec 0.276 ms 0/407178 (0%) 5.159/ 1.293/ 8.469/ 0.091 ms 40708 pps 157.82 Bob On Tue, Nov 20, 2018 at 11:26 AM Bob McMahon <bob.mcma...@broadcom.com> wrote: > VO AC may not use ampdu aggregation so double check that. The reason is > that VO are small packet payloads (~200 bytes) and have low relative > frequency of transmits. A wireless driver waiting around to fill an AMPDU > from a VO queue will cause excessive latency and impact the voice session. > Many drivers won't aggregate them. Since VO isn't likely using ampdus > and since it has the most preferred wme access settings, a 40 Mb/s iperf > flow with VO is going to consume most all the TXOPs. The other access > classes won't have many TXOPS at all. (The whole point of WiFi aggregation > technologies is to amortize the media access cost which is expensive per > collision avoidance. A rough theoretical max estimate for TXOPs is 10,000 > per second.) Maybe measure the packets per second of just VO to get an > empirical number. It's probably better to reduce this to something more > typical of lets say 10 voice calls (or 20 call legs.) > > You might also want to consider measuring "network power" which, while a > bit of a misnomer, is defined as average throughput over delay. Also, > maybe consider changing AIFS too. This can lead to interesting things. > Not sure you're exact goals but just some other things to consider. > > Bob > > On Tue, Nov 20, 2018 at 9:08 AM Kasper Biegun <kas...@alert.pl> wrote: > >> Hi Bruce, >> >> I am making a thesis project and I have to measure influence of a-mpdu >> and TXOPlimit on the 802.11ac network efficiency and I have to obtain >> plot of BK, BE, VI and VO throughputs, when they are running at the same >> time. I have to measure some scenarios with different ampdu sizes and >> when TXOPlimit is enabled and disabled. >> >> Thank you very much for your help, thanks to you I was able to run the >> tests, but I am still having some troubles. >> >> Today I have tried running tests again, but with lower iperf bandwidth >> (and window parameter) and it worked better, because I was able to run >> all four tests at the same time. But I had troubles with the highest >> priority traffic VO - it "takes" whole bandwidth. I mean that when I set >> e.g bandwidth (-b) to 40Mbits, VO traffic takes 40Mbits (the rest >> 20Mbits is divided to the rest of traffics) and it is undesirable during >> my measurements, because it is like manual limitation than network >> limitation. >> >> >> Example command: >> >> #iperf3 -c 10.10.0.1 -p 5025 -b 40M -u -S 0xc0 -i 1 -w 8192B --bind >> 10.10.0.2 --cport 5023 >> >> >> Kasper >> >> >> W dniu 2018-11-20 o 00:24, Bruce A. Mah pisze: >> > If memory serves me right, Kasper Biegun wrote: >> > >> >> to connect server and client I am using WiFi 5GHz (802.11ac), band: >> >> 20MHz with MCS (7 or 8) so I can get 65 to 78 Mbit/s. When I am testing >> >> now, I get around 60Mbit/s for the first test and after that second >> test >> >> starts and reaches around 60Mbit/s. But according to my assumptions it >> >> should work at the same time and divide possible bandwidth (I set >> values >> >> of TXOPlimit according to 802.11ac standard). >> > Hi Kasper-- >> > >> > I'm not sure why the tests are serialized. The only thing I can think >> > of is that the first test has completely blocked the wireless link so >> > that the second test can't even start. >> > >> > Wait. You're testing over wireless? Yet you're doing 1000Mbps = 1Gbps >> > tests. I don't think there's any wireless link that can handle two >> > 1Gbps tests, or even one. I think you really are saturating the link. >> > Because the tests are using UDP they'll just blast out packets at >> > whatever rate you specify regardless of whether the path is actually >> > capable of supporting it. You need to turn down the bitrate on both >> > tests (especially the first). What scenario are you trying to simulate >> > anyway? >> > >> > Bruce. >> > >> >> W dniu 2018-11-19 o 21:50, Bruce A. Mah pisze: >> >> >> >>> If memory serves me right, Kasper Biegun wrote: >> >>> >> >>>> currently I'm working on project connected with testing UDP >> throughput >> >>>> for different QoS values. To get the needed results I have to use >> Best >> >>>> Effort (0x70), Background (0x28), Voice (0xc0) and Video (0xb8) >> traffic >> >>>> at the same time. I have troubles with running them, because one >> stream >> >>>> was waiting for another to finish or if they started - only one was >> >>>> working (the rest had 0 bandwidth). Then I updated iperf to newest >> >>>> version available on github and installed it. Now, for testing >> purposes, >> >>>> I am running only Voice and Video at the same time, and I am still >> >>>> getting the same issue - one transfer is waiting for the second one >> to >> >>>> finish. >> >>>> >> >>>> I am using commands: >> >>>> >> >>>> - on server: #iperf3 -s -p 5015 and #iperf -s -p 5020 >> >>>> >> >>>> - on transmitter: #iperf3 -c 10.10.0.1 -p 5015 -S 192 --bind >> >>>> 10.10.0.2 --cport 5013 -u -b 1000m and >> >>>> >> >>>> #iperf3 -c 10.10.0.1 -p 5020 -S 184 --bind 10.10.0.2 >> --cport >> >>>> 5018 -u -b 1000m >> >>>> >> >>>> Server is working on Ubuntu 16.04 and transmitter is on Raspberry Pi >> 3. >> >>>> >> >>>> >> >>>> Could you please help me with that? >> >>> Hi Kasper-- >> >>> >> >>> Is it possible that one test is starving the other? What is the >> >>> bandwidth on the path between your server and client? >> >>> >> >>> (For what it's worth you're setting up the tests correctly by using >> two >> >>> independent sets of client and server.) >> >>> >> >>> Bruce. >> >>> >> > >> >> >> >> _______________________________________________ >> Iperf-users mailing list >> Iperf-users@lists.sourceforge.net >> https://lists.sourceforge.net/lists/listinfo/iperf-users >> >
_______________________________________________ Iperf-users mailing list Iperf-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/iperf-users