Peter Kjellstrom wrote:
On Tuesday 24 February 2009, Jie Cai wrote:
I have implemented a uDAPL program to measure the bandwidth on IB with
multirail connections.
The HCA used in the cluster is Mellanox ConnectX HCA. Each HCA has two
ports.
The program utilize the two port on each node of cluster to build
multirail IB connections.
The peak bandwidth I can get is ~ 1.3 GB/s (not bi-directional), which
is almost the same as single rail connections.
Assuming you have a 2.5 GT/s pci-express x8 that speed is a result of the bus
not being able to keep up with the HCA. Since the bus is holding even a
single DDR IB port back you see no improvement with two ports.
I do connect HCA in a 16x pci-e slot on each node.
However, I am trying to drive 2 ports simultaneously.
The workstation i am using is Sun Ultra 24,
and the HCA is Mellanox ConnectX MHGH28-XTC.
The data for the HCA and Ultra 24 is
MHGH28-XTC
IB ports: Dual Copper 4X 20Gb/s
Host Bus: PCIe 2.0 2.5GT/s
Ultra 24 workstation:
1333 MHz frontside bus with DDR2 memory support upto (10.67 GB per
second bandwidth)
PCI Express Slots
* Two full-length x16 Gen-2 slots (where the HCA has been connected to)
* One full-length x8 slot
* One full-length x1 slot
So, it may not be the problem of bottleneck in bus.
To fully drive a DDR IB port you need either 16x pci-express 2.5 GT/s or a 8x
5 GT/s. For one QDR or two DDR you'll need even more...
The pci-e slot in Ultra 24 is PCI Express Gen2 x16. The data transfer
rate is
5 Gpbs.
Will this be sufficient to drive the 2 ddr ports on MHGH28-XTC ?
Or is there any other possible reasons?
/Peter
Does anyone have similar experience?
_______________________________________________
general mailing list
[email protected]
http://lists.openfabrics.org/cgi-bin/mailman/listinfo/general
To unsubscribe, please visit http://openib.org/mailman/listinfo/openib-general