Re: [RFC] Enabling IPFIREWALL_FORWARD in run-time

2012-10-23 Thread n j
 On 10/19/12 4:25 AM, Andrey V. Elsukov wrote:

 Hi All,

 Many years ago i have already proposed this feature, but at that time
 several people were against, because as they said, it could affect
 performance. Now, when we have high speed network adapters, SMP kernel
 and network stack, several locks acquired in the path of each packet,
 and i have an ability to test this in the lab.

 So, i prepared the patch, that removes IPFIREWALL_FORWARD option from
 the kernel and makes this functionality always build-in, but it is
 turned off by default and can be enabled via the sysctl(8) variable
 net.pfil.forward=1.

 http://people.freebsd.org/~ae/pfil_forward.diff

 Also we have done some tests with the ixia traffic generator connected
 via 10G network adapter. Tests have show that there is no visible
 difference, and there is no visible performance degradation.

 Any objections?

Just another me-too mail - this is great news!

I can't really comment on the quality of the patch or the performance
results as I'm neither an expert in low-level coding nor do I have a
test lab to give this patch a go, but if there are no concrete
objections, I really hope this goes forward.

Thanks for the good work.

Regards,
-- 
Nino
___
freebsd-net@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-net
To unsubscribe, send any mail to freebsd-net-unsubscr...@freebsd.org


Re: VIMAGE crashes on 9.x with hotplug net80211 devices

2012-10-23 Thread Marko Zec
On Monday 22 October 2012 23:43:11 Adrian Chadd wrote:
 Hi,

 I don't mind tackling the net80211 clone detach path.

 I do mind how the default for hotplug is argh, it doesn't work. :-)

 So I'd like to come up with something to fix the basic device detach,
 rather than having to actually add CURVNET_*() calls around each
 if_free() in each device detach method.

As already mentioned earlier, I don't terribly object if you'd place 
CURVNET_SET(ifp-if_vnet) inside if_free() and a limited number of similar 
functions, but I don't quite believe this is will enough to solve the 
device_detach() issue without having to touch any of the drivers...

Marko
___
freebsd-net@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-net
To unsubscribe, send any mail to freebsd-net-unsubscr...@freebsd.org


kern/92880: [libc] [patch] almost rewritten inet_network(3) function

2012-10-23 Thread Andrey Simonenko
The following reply was made to PR kern/92880; it has been noted by GNATS.

From: Andrey Simonenko si...@comsys.ntu-kpi.kiev.ua
To: bug-follo...@freebsd.org
Cc:  
Subject: kern/92880: [libc] [patch] almost rewritten inet_network(3) function
Date: Tue, 23 Oct 2012 11:36:04 +0300

 I optimized inet_network() again.
 
 Difference between implementation of inet_network(3) from 9.1-PRERELEASE
 and my implementation.
 
 STRING INET_NETWORKINET_NETWORK_NEW
 0x12 0x0012  0x0012
 127.10x7f01  0x7f01
 127.1.2.30x7f010203  0x7f010203
 0x123456 INADDR_NONE INADDR_NONE
 0x12.0x340x1234  0x1234
 0x12.0x345   INADDR_NONE INADDR_NONE
 1.2.3.4.5INADDR_NONE INADDR_NONE
 1..3.4   INADDR_NONE INADDR_NONE
 .INADDR_NONE INADDR_NONE
 1.   INADDR_NONE INADDR_NONE
 .1   INADDR_NONE INADDR_NONE
 0x   0x  INADDR_NONE ---
 00x  0x
 01.02.07.077 0x0102073f  0x0102073f
 0x1.23.045.0 0x01172500  0x01172500
  INADDR_NONE INADDR_NONE
  INADDR_NONE INADDR_NONE
  f   INADDR_NONE INADDR_NONE
 bar  INADDR_NONE INADDR_NONE
 1.2bar   INADDR_NONE INADDR_NONE
 1.   INADDR_NONE INADDR_NONE
 =CA=C3=D5=CB=C5=CE   INADDR_NONE INADDR_NONE
 255.255.255.255  INADDR_NONE INADDR_NONE
 xINADDR_NONE INADDR_NONE
 0X12 0x0012  0x0012
 078  INADDR_NONE INADDR_NONE
 1 bar0x0001  INADDR_NONE ---
 127.0xabcd   INADDR_NONE INADDR_NONE
 128  0x0080  0x0080
 0.1.20x0102  0x0102
 0xff.010.23.0xa0 0xff0817a0  0xff0817a0
 x10  0x0010  INADDR_NONE ---
 X20  0x0020  INADDR_NONE ---
 x10.x20  0x1020  INADDR_NONE ---
 4294967297   0x0001  INADDR_NONE ---
 0x1000f  0x000f  INADDR_NONE ---
 0403 0x0003  INADDR_NONE ---
 
 #include sys/types.h
 #include sys/socket.h
 
 #include netinet/in.h
 #include arpa/inet.h
 
 #include ctype.h
 #include stdbool.h
 #include stdint.h
 #include stdio.h
 #include stdlib.h
 #include string.h
 
 static in_addr_t
 inet_network_new(const char *s)
 {
u_int d, base, dots;
in_addr_t addr, byte;
u_char c;
bool flag;
 
addr =3D 0;
dots =3D 0;
for (;; ++s) {
byte =3D 0;
flag =3D false;
if (*s =3D=3D '0') {
++s;
if (*s =3D=3D 'x' || *s =3D=3D 'X') {
++s;
base =3D 16;
} else {
base =3D 8;
flag =3D true;
}
} else
base =3D 10;
for (; (c =3D *s) !=3D '\0'; ++s) {
d =3D digittoint(c);
if (c !=3D '0'  (d =3D=3D 0 || d =3D base))
break;
byte =3D byte * base + d;
if (byte  UINT8_MAX)
return (INADDR_NONE);
flag =3D true;
}
if (!flag)
return (INADDR_NONE);
addr =3D (addr  8) | byte;
if (c !=3D '.')
break;
if (++dots =3D=3D 4)
return (INADDR_NONE);
}
return (c =3D=3D '\0' ? addr : INADDR_NONE);
 }
 
 int
 main(void)
 {
const char *const addr_str_tbl[] =3D {
0x12, 127.1, 127.1.2.3, 0x123456, 0x12.0x34,
0x12.0x345, 1.2.3.4.5, 1..3.4, ., 1., .1, 0x,
0, 01.02.07.077, 0x1.23.045.0, ,  ,  f, bar,
1.2bar, 1., =CA=C3=D5=CB=C5=CE, 255.255.255.255, x, 
0X12=
 , 078,
1 bar, 127.0xabcd, 128, 0.1.2, 0xff.010.23.0xa0,
x10, X20, x10.x20, 4294967297, 0x1000f,
0403, NULL };
const char *const *addr_str;
size_t len;
in_addr_t addr1, addr2;
 
printf(STRING\t\t\tINET_NETWORK\tINET_NETWORK_NEW\n);
for (addr_str =3D addr_str_tbl; *addr_str !=3D NULL; ++addr_str) {
printf(\%s\, *addr_str);
len =3D strlen(*addr_str) + 2;
if (len  8)
printf(\t\t\t);
else if (len  16)
printf(\t\t);

Re: VIMAGE crashes on 9.x with hotplug net80211 devices

2012-10-23 Thread Adrian Chadd
On 23 October 2012 00:16, Marko Zec z...@fer.hr wrote:

 As already mentioned earlier, I don't terribly object if you'd place
 CURVNET_SET(ifp-if_vnet) inside if_free() and a limited number of similar
 functions, but I don't quite believe this is will enough to solve the
 device_detach() issue without having to touch any of the drivers...

That's why I'm asking for more/better ideas.

So far my ideas are:

* for hotplug insert - do the same as what you're doing during the
kldload and boot device enumeration pass - call CURVNET_SET(vnet0)
* for device unload (hotplug or otherwise) - if vnet isn't set,
implicitly set it to vnet0
* for the net80211 vaps, they get destroyed in a few places (ioctl
path, device detach path, I'm sure I saw one more) so I have to use
CURVNET_SET(ifp-if_vnet) on those.

Now, that _should_ fix it for ath(4) and net80211, and it should fix
it for all the other non-USB wireless devices out there.

Now as for USB - Hans, what do you think? Should we do something
similar? How does VIMAGE work right now with USB wireless and USB
ethernet devices?

Marko - thanks for persisting with this. I'd like to try and make this
work for 10.0.



Adrian
___
freebsd-net@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-net
To unsubscribe, send any mail to freebsd-net-unsubscr...@freebsd.org


Re: kern/92880: [libc] [patch] almost rewritten inet_network(3) function

2012-10-23 Thread Adrian Chadd
... don't suppose you want to throw this into a test case somewhere in the tree?

The new ATF import would be ideal for this. :)



Adrian

On 23 October 2012 01:40, Andrey Simonenko si...@comsys.ntu-kpi.kiev.ua wrote:
 The following reply was made to PR kern/92880; it has been noted by GNATS.

 From: Andrey Simonenko si...@comsys.ntu-kpi.kiev.ua
 To: bug-follo...@freebsd.org
 Cc:
 Subject: kern/92880: [libc] [patch] almost rewritten inet_network(3) function
 Date: Tue, 23 Oct 2012 11:36:04 +0300

  I optimized inet_network() again.

  Difference between implementation of inet_network(3) from 9.1-PRERELEASE
  and my implementation.

  STRING INET_NETWORKINET_NETWORK_NEW
  0x12 0x0012  0x0012
  127.10x7f01  0x7f01
  127.1.2.30x7f010203  0x7f010203
  0x123456 INADDR_NONE INADDR_NONE
  0x12.0x340x1234  0x1234
  0x12.0x345   INADDR_NONE INADDR_NONE
  1.2.3.4.5INADDR_NONE INADDR_NONE
  1..3.4   INADDR_NONE INADDR_NONE
  .INADDR_NONE INADDR_NONE
  1.   INADDR_NONE INADDR_NONE
  .1   INADDR_NONE INADDR_NONE
  0x   0x  INADDR_NONE ---
  00x  0x
  01.02.07.077 0x0102073f  0x0102073f
  0x1.23.045.0 0x01172500  0x01172500
   INADDR_NONE INADDR_NONE
   INADDR_NONE INADDR_NONE
   f   INADDR_NONE INADDR_NONE
  bar  INADDR_NONE INADDR_NONE
  1.2bar   INADDR_NONE INADDR_NONE
  1.   INADDR_NONE INADDR_NONE
  =CA=C3=D5=CB=C5=CE   INADDR_NONE INADDR_NONE
  255.255.255.255  INADDR_NONE INADDR_NONE
  xINADDR_NONE INADDR_NONE
  0X12 0x0012  0x0012
  078  INADDR_NONE INADDR_NONE
  1 bar0x0001  INADDR_NONE ---
  127.0xabcd   INADDR_NONE INADDR_NONE
  128  0x0080  0x0080
  0.1.20x0102  0x0102
  0xff.010.23.0xa0 0xff0817a0  0xff0817a0
  x10  0x0010  INADDR_NONE ---
  X20  0x0020  INADDR_NONE ---
  x10.x20  0x1020  INADDR_NONE ---
  4294967297   0x0001  INADDR_NONE ---
  0x1000f  0x000f  INADDR_NONE ---
  0403 0x0003  INADDR_NONE ---

  #include sys/types.h
  #include sys/socket.h

  #include netinet/in.h
  #include arpa/inet.h

  #include ctype.h
  #include stdbool.h
  #include stdint.h
  #include stdio.h
  #include stdlib.h
  #include string.h

  static in_addr_t
  inet_network_new(const char *s)
  {
 u_int d, base, dots;
 in_addr_t addr, byte;
 u_char c;
 bool flag;

 addr =3D 0;
 dots =3D 0;
 for (;; ++s) {
 byte =3D 0;
 flag =3D false;
 if (*s =3D=3D '0') {
 ++s;
 if (*s =3D=3D 'x' || *s =3D=3D 'X') {
 ++s;
 base =3D 16;
 } else {
 base =3D 8;
 flag =3D true;
 }
 } else
 base =3D 10;
 for (; (c =3D *s) !=3D '\0'; ++s) {
 d =3D digittoint(c);
 if (c !=3D '0'  (d =3D=3D 0 || d =3D base))
 break;
 byte =3D byte * base + d;
 if (byte  UINT8_MAX)
 return (INADDR_NONE);
 flag =3D true;
 }
 if (!flag)
 return (INADDR_NONE);
 addr =3D (addr  8) | byte;
 if (c !=3D '.')
 break;
 if (++dots =3D=3D 4)
 return (INADDR_NONE);
 }
 return (c =3D=3D '\0' ? addr : INADDR_NONE);
  }

  int
  main(void)
  {
 const char *const addr_str_tbl[] =3D {
 0x12, 127.1, 127.1.2.3, 0x123456, 0x12.0x34,
 0x12.0x345, 1.2.3.4.5, 1..3.4, ., 1., .1, 0x,
 0, 01.02.07.077, 0x1.23.045.0, ,  ,  f, bar,
 1.2bar, 1., =CA=C3=D5=CB=C5=CE, 255.255.255.255, x, 
 0X12=
  , 078,
 1 bar, 127.0xabcd, 128, 0.1.2, 0xff.010.23.0xa0,
 x10, X20, x10.x20, 4294967297, 0x1000f,
 0403, NULL };
 const char *const *addr_str;
 size_t len;
 in_addr_t addr1, addr2;

 

fragmentation problem in FreeBSD 7

2012-10-23 Thread Sebastian Kuzminsky
Hi folks, this is my first post to freebsd-net, and my first bug-fix
submission...  I hope this is the right mailing list for this issue, and
the right format for sending in patches

I'm working on a derivative of FreeBSD 7.

I've run into a problem with IP header checksums when fragmenting to an
e1000 (em) interface, and I've narrowed it down to a very simple test.  The
test setup is like this:

[computer A]---(network 1)---[computer B]---(network 2)---[computer C]

That gorgeous drawing shows computer A connected to computer B via network
1, and computer B connected to computer C via network 2.  Computer B is set
up to forward packets between networks 1 and 2.  A can see B but not C.  C
can see B but not A.  B forwards between A and C.  Pretty simple.

One of B's NICs is a Broadcom, handled by the bce driver; this one works
fine in all my testing.

B's other NIC is an Intel PRO/1000 handled by the em driver.  This is the
one giving me trouble.

The test disables PMTUD on all three hosts.  It then sets the MTU of the
bce and em interfaces to the unrealistically low value of 72 bytes, and
tries to pass TCP packets back and forth using nc on computers A and C
(with computer B acting as a gateway).  This is to force the B gateway to
fragment the TCP frames it forwards.

Receiving on the em and sending on the bce works just fine (as noted
above).  Small TCP frames that fit in the MTU, big TCP frames that get
fragmented, no problems.

Receiving on the bce and sending on the em interface works fine for small
TCP frames that don't need fragmentation, but when B has to fragment the IP
packets before sending them out the em, the IP header checksums in the IP
packets that appear on the em's wires are wrong.  I came to this conclusion
by packet capture and by watching the 'bad header checksums' counter of
'netstat -s -p ip', both running on the computer receiving the fragments.


Ok, those are all my observations, next comes thoughts about the cause  a
proposed fix.


The root of the problem is two-fold:

1.  ip_output.c:ip_fragment() does not clear the CSUM_IP flag in the mbuf
when it does software IP checksum computation, so the mbuf still looks like
it needs IP checksumming.

2. The em driver does not advertise IP checksum offloading, but still
checks the CSUM_IP flag in the mbuf and modifies the packet when that flag
is set (this is in em_transmit_checksum_setup(), called by em_xmit()).
 Unfortunately the em driver gets the checksum wrong in this case, i guess
that's why it doesn't advertise this capability in its if_hwassist!

So the fragments that ip_fastfwd.c:ip_fastforward() gets from
ip_output.c:ip_fragment() have ip-ip_sum set correctly, but the
mbuf-m_pkthdr.csum_flags incorrectly has CSUM_IP still set, and this
causes the em driver to emit incorrect packets.

There are some other callers of ip_fragment(), notably ip_output().
 ip_output() clears CSUM_IP in the mbuf csum_flags itself if it's not in
if_hwassist, so avoids this problem.

So, the fix is simple: clear the mbuf's CSUM_IP when computing ip-ip_sum
in ip_fragment().  The first attached patch (against
gitorious/svn_stable_7) does this.

In looking at this issue, I noticed that ip_output()'s use of sw_csum is
inconsistent.  ip_output() splits the mbuf's csum_flags into two parts: the
stuff that hardware will assist with (these flags get left in the mbuf) and
the stuff that software needs to do (these get moved to sw_csum).  But
later ip_output() calls functions that don't get sw_csum, or that don't
know to look in it and look in the mbuf instead.  My second patch fixes
these kinds of issues and (IMO) simplifies the code by leaving all the
packet's checksumming needs in the mbuf, getting rid of sw_csum entirely.


-- 
Sebastian Kuzminsky
Linerate Systems


0001-Update-the-mbuf-csum_flags-of-IP-fragments-when-comp.patch
Description: Binary data


0002-Simplify-the-tracking-of-mbuf-checksumming-needs.patch
Description: Binary data
___
freebsd-net@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-net
To unsubscribe, send any mail to freebsd-net-unsubscr...@freebsd.org

Re: ixgb TSO performance degrades by ~30% between 7.4 and 8.2/9.0/9.1

2012-10-23 Thread Andre Oppermann

On 24.10.2012 01:13, Garrett Cooper wrote:

On Tue, Oct 23, 2012 at 3:57 PM, Garrett Cooper yaneg...@gmail.com wrote:

Hi,

Doing some poking around at the ixgb driver with a card I have at
$work using netperf and two machines hooked up over crossover, I
discovered that while ixgb's throughput performance was fantastic on
7.3/7.4, thoughput performance of the card is degraded on 8.2/9.0/9.1
by ~30% (9400Mbps on 7.4 - 6294Mbps on 9.0 for example). LRO
performance on the other hand is fantastic and doesn't degrade with
the card across FreeBSD versions. Performance remains constant with
ixgb across 8.2/9.0/9.1. I didn't observe the CPU usage.

More details:

The machines are hooked up in the following configuration:

  --- 
| Machine 1 | cxgb | - 10Gbit fibre - | ix1 | Machine 2 |
  ----

Machine Configuration:

The card in Machine 2 is an 82599EB card according to pciconf -lv.

/boot/loader.conf tunables (most of these are set according to 9.x
defaults in order to establish a sane baseline):

kern.ipc.nmbjumbo9=262144
kern.ipc.nmbjumbo16=262144
kern.ipc.nmbclusters=262144
kern.ipc.nmbjumbop=262144
kern.ipc.maxsockbuf=2097152

/etc/sysctl.conf tunables:

net.inet.tcp.recvspace=65536
net.inet.tcp.recvspace_inc=16384
net.inet.tcp.recvspace_max=2097152
net.inet.tcp.sendspace=32768
net.inet.tcp.sendbuf_max=2097152
net.inet.tcp.sendbuf_inc=8192

Kernel Config:

Machine 1 is running a custom version of FreeBSD. The version has been
constant over the course of my testing. Can give vague details on the
config, but can't give some specific details.
Machine 2 is running 7.4/8.2/9.0/9.1 with a GENERIC kernel.

Networking configuration:

- Machine 1 has an IPv4 address of 10.10.10.1; IPv6 is not configured.
The interface mtu is 1500.
- Machine 2 has an IPv4 address of 10.10.10.2; IPv6 is not configured.
The interface mtu is 1500.

Netperf configuration:

- netserver is run on both machines; I don't add any additional
arguments to the netserver invocation so it just goes off and forks.
- netperf is run like: netperf -cCjt TCP_STREAM -H IP-ADDRESS

I was wondering if this was a known issue and/or others had seen
similar problems with this card. I haven't gone into profiling the
kernel yet with DTrace, but if no one gets back to me before sometime
later on this week/next week that will be my next course of action for
tracking down the source of the performance problem.


A couple more notes:

1. We're not using Intel SFP modules -- they're Finisar based
(shouldn't matter, but I know that some Intel NICs are incredibly
picky when it comes to SFP modules).
2. The performance on 8.2 is actually worse than on 9.x: ~5700Mbps.


There have been a very large number of changes to our network
stack between 7.4 and today.  That makes it very difficult to
pinpoint a specific change.

First of all please test 10-current as well but make sure that
you turn off WITNESS and INVARIANTS in your kernel config.

Second there may be an issue with TSO and ixgb where TSO can
cause packets that are a bit larger than 64K by the ethernet
header size.  ixgb DMA is set up for 64K all inclusive.  While
TSO did generate a IP valid packet it is arguably not very well
digestable for the driver considering that 64K is a very magic
number.  TSO should reduce its largest packet size by max_linkhdr.
A fix for that is in the works and will be available shortly.

--
Andre

___
freebsd-net@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-net
To unsubscribe, send any mail to freebsd-net-unsubscr...@freebsd.org


Re: ixgb TSO performance degrades by ~30% between 7.4 and 8.2/9.0/9.1

2012-10-23 Thread Garrett Cooper
On Tue, Oct 23, 2012 at 4:32 PM, Andre Oppermann opperm...@networx.ch wrote:

...

 Doing some poking around at the ixgb driver with a card I have at
 $work using netperf and two machines hooked up over crossover, I
 discovered that while ixgb's throughput performance was fantastic on
 7.3/7.4, thoughput performance of the card is degraded on 8.2/9.0/9.1
 by ~30% (9400Mbps on 7.4 - 6294Mbps on 9.0 for example). LRO
 performance on the other hand is fantastic and doesn't degrade with
 the card across FreeBSD versions. Performance remains constant with
 ixgb across 8.2/9.0/9.1. I didn't observe the CPU usage.

 More details:

 The machines are hooked up in the following configuration:

   ---
 
 | Machine 1 | cxgb | - 10Gbit fibre - | ix1 | Machine 2 |
   ---
 -

 Machine Configuration:

 The card in Machine 2 is an 82599EB card according to pciconf -lv.

 /boot/loader.conf tunables (most of these are set according to 9.x
 defaults in order to establish a sane baseline):

 kern.ipc.nmbjumbo9=262144
 kern.ipc.nmbjumbo16=262144
 kern.ipc.nmbclusters=262144
 kern.ipc.nmbjumbop=262144
 kern.ipc.maxsockbuf=2097152

 /etc/sysctl.conf tunables:

 net.inet.tcp.recvspace=65536
 net.inet.tcp.recvspace_inc=16384
 net.inet.tcp.recvspace_max=2097152
 net.inet.tcp.sendspace=32768
 net.inet.tcp.sendbuf_max=2097152
 net.inet.tcp.sendbuf_inc=8192

 Kernel Config:

 Machine 1 is running a custom version of FreeBSD. The version has been
 constant over the course of my testing. Can give vague details on the
 config, but can't give some specific details.
 Machine 2 is running 7.4/8.2/9.0/9.1 with a GENERIC kernel.

 Networking configuration:

 - Machine 1 has an IPv4 address of 10.10.10.1; IPv6 is not configured.
 The interface mtu is 1500.
 - Machine 2 has an IPv4 address of 10.10.10.2; IPv6 is not configured.
 The interface mtu is 1500.

 Netperf configuration:

 - netserver is run on both machines; I don't add any additional
 arguments to the netserver invocation so it just goes off and forks.
 - netperf is run like: netperf -cCjt TCP_STREAM -H IP-ADDRESS

 I was wondering if this was a known issue and/or others had seen
 similar problems with this card. I haven't gone into profiling the
 kernel yet with DTrace, but if no one gets back to me before sometime
 later on this week/next week that will be my next course of action for
 tracking down the source of the performance problem.


 A couple more notes:

 1. We're not using Intel SFP modules -- they're Finisar based
 (shouldn't matter, but I know that some Intel NICs are incredibly
 picky when it comes to SFP modules).
 2. The performance on 8.2 is actually worse than on 9.x: ~5700Mbps.

 There have been a very large number of changes to our network
 stack between 7.4 and today.  That makes it very difficult to
 pinpoint a specific change.

 First of all please test 10-current as well but make sure that
 you turn off WITNESS and INVARIANTS in your kernel config.

 Second there may be an issue with TSO and ixgb where TSO can
 cause packets that are a bit larger than 64K by the ethernet
 header size.  ixgb DMA is set up for 64K all inclusive.  While
 TSO did generate a IP valid packet it is arguably not very well
 digestable for the driver considering that 64K is a very magic
 number.  TSO should reduce its largest packet size by max_linkhdr.
 A fix for that is in the works and will be available shortly.

Ok. I'll give it a shot sometime this upcoming week to see if
things improve (of course without INVARIANTS and WITNESS ;)..); I saw
there were some changes in the works on CURRENT -- just haven't gotten
around to testing that yet (melifaro and a few other folks have been
working on performance and removing unnecessary locking too it seems).
Thanks for the feedback!
-Garrett
___
freebsd-net@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-net
To unsubscribe, send any mail to freebsd-net-unsubscr...@freebsd.org


Re: ixgb TSO performance degrades by ~30% between 7.4 and 8.2/9.0/9.1

2012-10-23 Thread Jack Vogel
It you mean the ixgbe driver please call it that, the IS an ixgb driver
which is for
VERY old PCI cards, from the context I assume you mean the newer hardware :)

Jack


On Tue, Oct 23, 2012 at 3:57 PM, Garrett Cooper yaneg...@gmail.com wrote:

 Hi,

 Doing some poking around at the ixgb driver with a card I have at
 $work using netperf and two machines hooked up over crossover, I
 discovered that while ixgb's throughput performance was fantastic on
 7.3/7.4, thoughput performance of the card is degraded on 8.2/9.0/9.1
 by ~30% (9400Mbps on 7.4 - 6294Mbps on 9.0 for example). LRO
 performance on the other hand is fantastic and doesn't degrade with
 the card across FreeBSD versions. Performance remains constant with
 ixgb across 8.2/9.0/9.1. I didn't observe the CPU usage.

 More details:

 The machines are hooked up in the following configuration:

  ---
 
 | Machine 1 | cxgb | - 10Gbit fibre - | ix1 | Machine 2 |
  ---
  -

 Machine Configuration:

 The card in Machine 2 is an 82599EB card according to pciconf -lv.

 /boot/loader.conf tunables (most of these are set according to 9.x
 defaults in order to establish a sane baseline):

 kern.ipc.nmbjumbo9=262144
 kern.ipc.nmbjumbo16=262144
 kern.ipc.nmbclusters=262144
 kern.ipc.nmbjumbop=262144
 kern.ipc.maxsockbuf=2097152

 /etc/sysctl.conf tunables:

 net.inet.tcp.recvspace=65536
 net.inet.tcp.recvspace_inc=16384
 net.inet.tcp.recvspace_max=2097152
 net.inet.tcp.sendspace=32768
 net.inet.tcp.sendbuf_max=2097152
 net.inet.tcp.sendbuf_inc=8192

 Kernel Config:

 Machine 1 is running a custom version of FreeBSD. The version has been
 constant over the course of my testing. Can give vague details on the
 config, but can't give some specific details.
 Machine 2 is running 7.4/8.2/9.0/9.1 with a GENERIC kernel.

 Networking configuration:

 - Machine 1 has an IPv4 address of 10.10.10.1; IPv6 is not configured.
 The interface mtu is 1500.
 - Machine 2 has an IPv4 address of 10.10.10.2; IPv6 is not configured.
 The interface mtu is 1500.

 Netperf configuration:

 - netserver is run on both machines; I don't add any additional
 arguments to the netserver invocation so it just goes off and forks.
 - netperf is run like: netperf -cCjt TCP_STREAM -H IP-ADDRESS

 I was wondering if this was a known issue and/or others had seen
 similar problems with this card. I haven't gone into profiling the
 kernel yet with DTrace, but if no one gets back to me before sometime
 later on this week/next week that will be my next course of action for
 tracking down the source of the performance problem.

 Thanks!
 -Garrett

___
freebsd-net@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-net
To unsubscribe, send any mail to freebsd-net-unsubscr...@freebsd.org


Re: ixgbe TSO performance degrades by ~30% between 7.4 and 8.2/9.0/9.1

2012-10-23 Thread Garrett Cooper
On Tue, Oct 23, 2012 at 4:58 PM, Jack Vogel jfvo...@gmail.com wrote:
 It you mean the ixgbe driver please call it that, the IS an ixgb driver
 which is for
 VERY old PCI cards, from the context I assume you mean the newer hardware :)

Yeah... I meant ixgbe. Subject line fixed :).
Thanks!
-Garrett

 On Tue, Oct 23, 2012 at 3:57 PM, Garrett Cooper yaneg...@gmail.com wrote:

 Hi,

 Doing some poking around at the ixgb driver with a card I have at
 $work using netperf and two machines hooked up over crossover, I
 discovered that while ixgb's throughput performance was fantastic on
 7.3/7.4, thoughput performance of the card is degraded on 8.2/9.0/9.1
 by ~30% (9400Mbps on 7.4 - 6294Mbps on 9.0 for example). LRO
 performance on the other hand is fantastic and doesn't degrade with
 the card across FreeBSD versions. Performance remains constant with
 ixgb across 8.2/9.0/9.1. I didn't observe the CPU usage.

 More details:

 The machines are hooked up in the following configuration:

  ---
 
 | Machine 1 | cxgb | - 10Gbit fibre - | ix1 | Machine 2 |
  ---
 -

 Machine Configuration:

 The card in Machine 2 is an 82599EB card according to pciconf -lv.

 /boot/loader.conf tunables (most of these are set according to 9.x
 defaults in order to establish a sane baseline):

 kern.ipc.nmbjumbo9=262144
 kern.ipc.nmbjumbo16=262144
 kern.ipc.nmbclusters=262144
 kern.ipc.nmbjumbop=262144
 kern.ipc.maxsockbuf=2097152

 /etc/sysctl.conf tunables:

 net.inet.tcp.recvspace=65536
 net.inet.tcp.recvspace_inc=16384
 net.inet.tcp.recvspace_max=2097152
 net.inet.tcp.sendspace=32768
 net.inet.tcp.sendbuf_max=2097152
 net.inet.tcp.sendbuf_inc=8192

 Kernel Config:

 Machine 1 is running a custom version of FreeBSD. The version has been
 constant over the course of my testing. Can give vague details on the
 config, but can't give some specific details.
 Machine 2 is running 7.4/8.2/9.0/9.1 with a GENERIC kernel.

 Networking configuration:

 - Machine 1 has an IPv4 address of 10.10.10.1; IPv6 is not configured.
 The interface mtu is 1500.
 - Machine 2 has an IPv4 address of 10.10.10.2; IPv6 is not configured.
 The interface mtu is 1500.

 Netperf configuration:

 - netserver is run on both machines; I don't add any additional
 arguments to the netserver invocation so it just goes off and forks.
 - netperf is run like: netperf -cCjt TCP_STREAM -H IP-ADDRESS

 I was wondering if this was a known issue and/or others had seen
 similar problems with this card. I haven't gone into profiling the
 kernel yet with DTrace, but if no one gets back to me before sometime
 later on this week/next week that will be my next course of action for
 tracking down the source of the performance problem.
___
freebsd-net@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-net
To unsubscribe, send any mail to freebsd-net-unsubscr...@freebsd.org