joshua sahala wrote:
drew,

it may or may not be related, but...check the output of 'sh counter
int <int> [delta]' and look at the qos[1-21][In|Out]lost counters.

i was experiencing various drops due to the default interface (qos)
buffer allocation:  basically, all of my traffic was hitting the 76xx
swouter in the q0 buffer and overrunning it (there were no drops in
any of the other qos queues because no traffic was ever hitting them).
 i ended up having to rewrite the buffer mapping to allocate
everything to q0 and the random discards stopped (at least the ones
caused by this issue).

I want to revive an old thread if I can. I'm facing a similar issue now. Gi1/1 on my 6724s in my core 7600s (3BXL) connect to one of my border routers, a 7206 G1. Both interfaces on both 6724s show large volumes of input drops and flushes. Gi1/2 on the same 6724s connect to a 3845 which is my other border and it shows significantly lower drops and flushes (4 digits instead of 7 or 8). All 4 links are SX. 'sh counters' didn't yield anything terribly interesting either.

7613-1.clr#sh counters interface gi1/1 delta | e = 0
Time since last clear
---------------------
never

64 bit counters:
 0.                      rxHCTotalPkts = 123760873738
 1.                      txHCTotalPkts = 45947101814
 2.                    rxHCUnicastPkts = 123747989684
 3.                    txHCUnicastPkts = 45941233718
 4.                  rxHCMulticastPkts = 12883997
 5.                  txHCMulticastPkts = 5868073
 6.                  rxHCBroadcastPkts = 57
 7.                  txHCBroadcastPkts = 23
 8.                         rxHCOctets = 101377579108374
 9.                         txHCOctets = 16976124978053
10.                 rxTxHCPkts64Octets = 8893600878
11.            rxTxHCPkts65to127Octets = 57698604883
12.           rxTxHCPkts128to255Octets = 20633513794
13.           rxTxHCPkts256to511Octets = 7123204457
14.          rxTxHCpkts512to1023Octets = 6652027912
15.         rxTxHCpkts1024to1518Octets = 26440990980

32 bit counters:
 2.                    rxOversizedPkts = 2492150694
13.                         linkChange = 2
All Port Counters
 1.                          InPackets = 123760839646
 2.                           InOctets = 101377556782449
 3.                        InUcastPkts = 123747955595
 4.                        InMcastPkts = 12883994
 5.                        InBcastPkts = 57
 6.                         OutPackets = 45947087810
 7.                          OutOctets = 16976121260975
 8.                       OutUcastPkts = 45941219715
 9.                       OutMcastPkts = 5868072
10.                       OutBcastPkts = 23
22.                             Giants = 2492143293
35.                 rxTxHCPkts64Octets = 8893600875
36.            rxTxHCPkts65to127Octets = 57698582793
37.           rxTxHCPkts128to255Octets = 20633505929
38.           rxTxHCPkts256to511Octets = 7123201908
39.          rxTxHCpkts512to1023Octets = 6652026348
40.         rxTxHCpkts1024to1518Octets = 26440984821
44.                      OversizedPkts = 2492143293


The giants are explained by the MTU I have on those links. I run 9000 on all infrastructure links. Other than that I don't see anything else wrong. All the QoS Lost lines were 0. All infrastructure interfaces are also MPLS enabled. The 7206 carries the bulk of the Internet traffic as does 7600 #1 so it's not a big surprise to see its links affected much more so than the 3845 links.

I'm graphing interface errors/discards with Cacti. I have to question the numbers it's giving me though. They have never seemed to be accurate to me on any of my interfaces.

Are my queues not deep enough to carry the traffic flow? Peak Mbps on through the 7206 is about 120Mbps and if Cacti is right then we're also only talking about 17,000 pps on the upstream-facing interface of the 7206, most of which would come from 7600 #1. Thoughts?

Thanks
 Justin
_______________________________________________
cisco-nsp mailing list  cisco-nsp@puck.nether.net
https://puck.nether.net/mailman/listinfo/cisco-nsp
archive at http://puck.nether.net/pipermail/cisco-nsp/

Reply via email to