Curtis Villamizar and Mr Katz,

        First, let me state as I did earlier, I do feel more resources
        dealing with handling of Updates pkts containing refresh age
        LSAs WOULD yield a higher benefit. And since we are not able
        to speed up a REMOTE router's execution of of various tasks, we
        need to concentrate on our on local tasks to allow US to LOCALLY
        converge ASAP.

        But, we are in this forwarding table vs OSPF routing table
        costs, benefit analysis..

        Lets see if we agree on these things.

        1) Most LSAs recieved are just age refresh LSAs, thus they
           have no effect on SPF calcs.

        2) Most SPF calcs do not result in a lower cost route or
           in some implementations a additional equal cost route.

        3) Even if a OSPF routing table will yield a better cost
           route than earlier in time, due to administrative
           distances that newer cost, may not still be used.

           Then the forwarding table mods are rarely even done!
           So, why would you want to spend time to decrease the
           latency of a operation on such a infrequent event?????

        If subsec hellos or equivalent operation was used, I would
        first verify that nbr lookups are efficent..

        The freq/repeativeness of a operation, if speeded up, should
        yield the best results..

        So, if we aggregate the number of repeativeness operations, we
        need to do 1 lookup per recived LSA within our routing table.
        Most of these will result in age refreshes. So, first, the
        LSA lookup is stressed the most. 

        Then if we had a extremely fast SPF type calcs based on different
        LSA types / triggers, we COULD re-run repetive calcs withoout forcing
        a delay. A delay is standand in our industry. However, if we agree
        with #2 and #3, then the only reason for the delay is due to slower
        SPF calcs. I agree some minimal delay is necessary to verify that
        all new LSAs within a single Update, given that a full ajacency has
        already been established, should be implemented.

        Lastly, bringing back up the forwarding table. The table COULD be
        considered somewhat like a cache. Specific atomic operations are
normally
        implimented because of single change entries. A operation could
        simply Invalidate an entry. This is a single small amount of data that
        needs to be transfered and even slow PCI buses are capable of handling
        a infrequent data with minimal latency. Most of this is done in
        hardware. Sorry, it is a don't care whether their are multiple readers
        for a entry if that entry needs to be invalidated or updated. We don't
        wait for the readers to finish. We need to pre-empt them NOW.

        Thus, given appropriate resouces, I would creare a set of equal LM
Bench
        micro benchmarks, against the routing table and identify what items are
        given higher priorities. I would not be surprised if these items had
        not already been given quite a bit of fine tuning..

        The only major issues are the best case versus worse case numbers,
        and whether those best case numbers can be improved while not
        increasing the worse case numbers.

        Lastly, the assumption is whether BGP is in the environment? If it is,
        IMO TCP would be the limiting factor. TCP will not allow a large burst
        of data on a periodiclly idle connection. Most implmentations will fall
        into a slow start methodology. This is the same problem with LDP even
        when a path has been established with a specified bandwidth. TCP most
        likely will initially get in the way and only after a sustained number
        of segments transmitted will it allow the bandwidth to increase 1
        segment per round-trip-time (rtt). This is known as the congestion
        avoidance (CA) phase.


        Mitchell Erblich
        
        -------------------
        

        



Curtis Villamizar wrote:
> 
> In message <[EMAIL PROTECTED]>
> Erblichs writes:
> >
> >
> >       The biggest reason for this is because the forwarding
> >       tables need faster memory. If we assume that forwarding
> >       table memory is 4x faster than routing table memory,
> 
> Nice try but ...
> 
> The forwarding memory is on another card in a reasonably big router
> and therefore the information has to be trasferred from one processor
> to another and then installed in the forwarding memory.
> 
> Also the forwarding memory is primarily used for forwarding and in a
> big busy router not too many cycles are left over for the writes that
> update the forwarding memory due to the reads that are occurring
> concurrently.  In some architectures this is minimized.  For example,
> some may use memory caching techniques others dual memory banks, etc.
> 
> But perhaps the biggest factor is that the few hundred or few thousand
> IGP routes are the basis for BGP next hops so after they are computed
> the few 100,000 BGP routes and forwarding entries are mapped onto the
> SPF results.  Since there are usually more than 100 times as many BGP
> routes as IGP routes doubling the speed of the SPF has no effect.
> 
> In any case, Dave is right.  The SPF typically takes far less time
> than the things that happen as a result of the SPF.
> 
> Curtis
> 
> ps - OTOH the speed of the MPLS/TE CSPF is very significant.

_______________________________________________
OSPF mailing list
[email protected]
https://www1.ietf.org/mailman/listinfo/ospf

Reply via email to