18.10.2014 3:11, Tantilov, Emil S пишет: >> -----Original Message----- >> From: Bokhan Artem [mailto:a...@eml.ru] >> Sent: Friday, October 17, 2014 4:53 AM >> To: Fujinaka, Todd; E1000-devel@lists.sourceforge.net >> Subject: Re: [E1000-devel] performance of two 82599 nics at >> a host >> >> Thank you for the answer, Todd. >> >> 15.10.2014 23:07, Fujinaka, Todd пишет: >>> Can you send us the output of "lspci -vvv" (run as root) >> and ethtool -s? >> Attached. >> >>> I'm guessing that you're trying to use memory on remote >> NUMA nodes. You can try running: >>> https://software.intel.com/en-us/articles/intel- >> performance-counter-monitor-a-better-way-to-measure-cpu- >> utilization >> You are right. When using both adapters remote memory is >> used. Remote memory is not used when any of two adapters run alone. >> Results are in attachment too. Why this happen? > This list strips attachments, so I can't see your results, I've added the report here https://sourceforge.net/p/e1000/bugs/435/
> but in general you can see that the driver allocates as many queues as CPUs. > Because of this > you will have queues that are on both NUMA nodes. Configuration is optimized quite with RSS=8 and AtrSampleRate=0. Irqs binded to local cpus. > The idea behind the default behavior of the driver is to use the CPU which is > local > to the user space application, not necessarily the CPU that is local to the > NUMA node of > the device. Performance degradation can be far greater when feeding > application from > a remote queue even if the queue is local to the interface. You can find > multiple threads > regarding NUMA performance if you look at the archives of this mailing list. > > For optimal performance we generally recommend to map the affinity of the > queues to CPUs > using the set_irq_affinity script provided with the driver. You will have to > disable irqbalance. Already done. > Aside from your results I don't know anything about the actual test, or the > type of traffic or > your setup, so it's hard to tell where your bottleneck is or if it’s related > to NUMA at all. I'm trying to maximize pps throughout when host acts as ip forwarder. After some optimizations I have some success with 12.6 Mpps when using single CPU. I would expect doubling throughout with second CPU and adapter but got only degradation of throughout. > > In general when reporting issues I recommend to open a bug on sourceforge as > it makes it easier to > upload files and for us to track it. Thank you. > > Thanks, > Emil > ------------------------------------------------------------------------------ Comprehensive Server Monitoring with Site24x7. Monitor 10 servers for $9/Month. Get alerted through email, SMS, voice calls or mobile push notifications. Take corrective actions from your mobile device. http://p.sf.net/sfu/Zoho _______________________________________________ E1000-devel mailing list E1000-devel@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/e1000-devel To learn more about Intel® Ethernet, visit http://communities.intel.com/community/wired