Hi Bart,

I think that these three links will give you some idea about performance you can get with 3 or more physical NICs.

Maximum Throughput
https://www.kernel.org/doc/Documentation/networking/bonding.txt

Two examples using Gigabit
http://www.scl.ameslab.gov/Projects/MP_Lite/
http://www.ibm.com/developerworks/data/library/techarticle/dm-1202streamslinuxperf/

If the NIC´s are integrated to the motherboard, you can get about 90 percent of performance, if you are using PCI ports the performance is poor up to 50 or 60 percent, also, you know that all connectivity must support Gigabit.

Regards
Ra



El 2014-04-22 14:01, Lars Ellenberg escribió:
On Fri, Apr 18, 2014 at 08:59:10PM +0200, Bart Coninckx wrote:
Hi all,

In the past I read somewhere that by bonding more than two NICs there is a severe penalty on speed as TCP re-ordering needs to happen.

I'm currently building a two-node DRBD cluster that uses Infiniband
for DRBD. The cluster offers SCST targets. I would like to offer the
best speed possible to the iSCSI clients (which are on gigabit NICs). Has anyone tried 3 or 4 card bonding? What performance do you get out of this?

You have to distinguish between "round-robin" balancing
(which is the only mode I know that can give a *single* TCP connection
more than single physical link bandwidth) and any of the "hashing"
balancing modes, in which the single TCP connection will always use just
one physical link, but you get additional *aggregated* bandwidth
over all TCP links, as long as you have "enough" connections
that the balancing algorithms can work their "statistical magic".

With balance-rr, you likely have to measure, the outcome
may depend on a lot of things including link and irq latencies.
Chose your deployment-local "optimum" from
single-TCP session throughput vs # bonding channels.

For other bonding modes, you won't be able to increase single TCP
session beyond single physical link saturation, but your aggregate
throughput will increase with number of bonding channels and number of
communication partners (mapped to different "balance buckets").

_______________________________________________
drbd-user mailing list
[email protected]
http://lists.linbit.com/mailman/listinfo/drbd-user

Reply via email to