On 15-Apr-09, at 11:34 AM, Lamar Owen wrote:

On Tuesday 14 April 2009 18:22:03 Jason Lixfeld wrote:
For the life of us, we can't seem to get any more than 60Mbps
sustained across the ATM testing with iperf, so we're just trying to
figure out if the GSR just can't push any more than what it's doing or
if there's something else afoot.
[snip]
We've done our due diligence to ensure the bits of the network between
the test machine and the ATM can support 100Mbps, so we're fairly

Hmm, 60mb/s using a 100mb/s connected box sounds about right. To really strain an OC12 you need a gigabit connected tester that can really do a
gigabit of traffic.  Or multiple test PC's.

In this case, I can iperf 97Mbps between two machines connected together at 100Mb.

I have a 12012 here in production, and have some of the kit necessary to test point to point ATM connections (including a Catalyst 8540MSR with OC12, ARM, and gigabit cards), and have a 4xOC12/ATM/MM, but it will be a few days before
I could have the time to set up a test to see if the 12012 is limited.

We've been wrestling with this for weeks now, but haven't had the means to be able to compare our results to anyone else to see whether or not we're an anomaly, so what's another day or four :)

The LC
engines on the ATM card and the 3GE card will be the limiting factor, and those cards are rated for line rate on four simultaneous OC12's or line rate on two GigE (can't do full line rate on all three with a 2.5Gb/s fabric
connection).

The load is really low, so I'd be very surprised if it was an LC limitation, but what do I know:

bdr1.nyc-hudson-12008#show int a2/0 load

    Interface                   bits/sec     pack/sec
--------------------           ------------  ----------
AT2/0                 Tx           48464000      14099
                     Rx          104808000      18012
bdr1.nyc-hudson-12008#show int a2/1 load

    Interface                   bits/sec     pack/sec
--------------------           ------------  ----------
AT2/1                 Tx           57581000      13032
                     Rx          116319000      14466
bdr1.nyc-hudson-12008#show int g5/0 load

    Interface                   bits/sec     pack/sec
--------------------           ------------  ----------
Gi5/0                 Tx           56851000       8981
                     Rx           35082000       7833
bdr1.nyc-hudson-12008#show int g5/1 load

    Interface                   bits/sec     pack/sec
--------------------           ------------  ----------
Gi5/1                 Tx          166072000      23424
                     Rx           70951000      19116
bdr1.nyc-hudson-12008#

So:

Total Throughput: 656128000
Total PPS:        118963
Average Size (B): 689.4

The GRP CPU is not involved in the data plane on a GSR; the LC
engine CPU's/ASICs do dCEF and talk directly over the fabric. Unless you have serious fabric issues preventing full bandwidth, in which case you have bigger
problems.

Again, the bandwidth going over the entire box is like 650Mbps spread more or less evenly across the two LCs.

So I'd first check to see if your iperf test box can really generate sufficient
traffic.

Here's one of the tests we've done, and we were able to get ~97Mbps here:

Macbook Pro -> Linksys 100Mb -> 1811 -> 7609 -> 10GE -> 7609 -> 3550 - > PC 100Mb NIC.

What sort of ATM switch or router is on the other end of those multimode short reach OC12's? What sort of router is terminating them? How are your PVC's
set up?

A2/0 and A2/1 on the GSR connect to two ports on a Fore ASX200BX.

The ASX200BX connects into the provider's SONET network. On the Z side, we're taking the OC12 into a Fore ASX1000.

ATM2/0.100 (vpi/vci 0/100) on side A ultimately terminates on an OSM-2OC12-ATM-MM in 7609-A on side Z. ATM2/1.110 (vpi/vci 0/110) on side A ultimately terminates on an OSM-2OC12-ATM-MM in 7609-B on side Z. 7609-A and 7609-B on the Z side are connected by an OC12 ATM on own own dark fiber.

Fores at the A and Z side are both clean as a whistle. No error seconds anywhere.

_______________________________________________
cisco-nsp mailing list  [email protected]
https://puck.nether.net/mailman/listinfo/cisco-nsp
archive at http://puck.nether.net/pipermail/cisco-nsp/


_______________________________________________
cisco-nsp mailing list  [email protected]
https://puck.nether.net/mailman/listinfo/cisco-nsp
archive at http://puck.nether.net/pipermail/cisco-nsp/

Reply via email to