On Wed, May 28, 2014 at 3:08 PM, Matthew W. Ross
<[email protected]> wrote:
>>  Latency is the real killer.
>
> How do you reduce latency?

  Make the cable shorter.

  Ha ha, only serious.

  Network latency is how long it takes for a bit to go from point A to
point B.  The longer the path, the longer it takes.

  There are other factors.  Signals propagate at different rates
depending on the medium (e.g., CAT5 is very roughly half the speed of
light).  Processing (routing, switching, conversion, etc.) are
typically slower than the wire, and can add delay, especially if the
equipment is heavily loaded.

  My comment was directed more at WAN/Internet links.  People see the
ads for 42 bajigabits per second from cheap ISPs, and don't appreciate
that there is more to speed than how wide the road is.

> I decried to bypass the 2510G switch they
> were aggregated through directly to our 5308xl. Guess what, the problem
> all-but disappeared. Obviously, the 2510 can't handle the traffic.

  I wouldn't be so sure.  On paper, the 2510 has a higher backplane
bandwidth, and a higher forwarding rate.  Comparing specs between
vendors can be dicey, because of different test methodologies, but
vendors are usually (*usually*) consistent within themselves.  I'm
more inclined to believe a configuration issue, physical problem (bad
cable/port/etc.), or a local overload you bypassed.

  For example: If you've got six servers, each with a gigabit link,
plugged into your 2510, and then have a single gigabit link connecting
your 2510 to your 5308xl, the traffic from those servers could
overwhelm the uplink.  The buffer on the 2510 will fill and then start
dropping frames.  The servers go into retransmit and make the problem
worse.  Everything goes into exponential backoff until aggregate
transmit rate drops below the uplink speed.  Things start working
again, so transmit rate increases.  Lather rinse repeat.

  Another scenario would you've got something defined on the switch
that forces the forwarding decision to go off the switch ASIC and onto
the general-purpose CPU, which is orders of magnitude slower.

  Not saying these were your problem.  They're just scenarios for the
sake of discussion.

  Did you check port statistics on switches and servers?  Check the
logs on the switches?  Are all the fault finders enabled on the
switches?

  HP tech support should be able to help, too.

> On a similar note, We used to be running HP 2910s as our iSCSI traffic
> connection switches between our 3 VMware hosts and our EqualLogic, but we
> were again having latency issues.

  Define "latency issues".  Please note that "latency" does not mean
"network trouble".  :)

> As recommended by our vendor, we upgraded
> to Dell PowerConnect 6224s. On paper, these two switches are quite similar,
> but the Dell switches handle the latency where the HPs were not.

  Uh.... "handle the latency"?  :)  Latency isn't something switches
handle.  It's something they cause, but usually not significantly.

  The specs for those switches do look pretty similar, so I too would
expect them to handle similar loads.

  When you get into performance tuning of networks, configuration can
matter a lot.  Stuff that was wrong before but didn't cause a problem
because it was small becomes a problem because now it's big.  And
different vendors each have different design assumptions, defaults,
and quirks.  As a result, you can get situations where vendor X works
with vendor Y "out of the box", but fails with vendor Z.  So you can
replace all your X with Z, or you can adjust configuration.  And then
what do you do when you discover Z doesn't work with W?

-- Ben


Reply via email to