> On Oct 5, 2017, at 6:19 PM, allison via cctalk <[email protected]> wrote:
>
> Moore's law only worked for hardware, software lagged typically two
> years behind.
There's a more cynical view, sometimes called "the virtual disease", which is
that software performance is constant because all the gains from Moore's law
have been consumed by increased software inefficiency and complexity.
A related aspect is an observation from my boss at DEC: interrupts (context
switches) always take 10 microseconds no matter how fast the CPU is.
Both of these are slightly unfair but much too close to the truth for comfort.
paul