Addendum to my previous rant/rave/ramble...
I do think that the rise in popularity of the Graph500 over the LINPACK
benchmarks is an important admission that not only have computer
architectures changed, but the characteristics of "interesting problems"
being run on them.
I remember it quite well...
http://tech.slashdot.org/story/13/01/03/232259/supercomputer-repossessed-by-state-may-be-sold-in-pieces
With conflicted (but not mixed) feelings. I (all but) bid on
providing the visualization gateways for the system as my first major
project after I left LANL in 08... I"m glad i saw the writing on the
dysfunctional wall of state procurement as well as the ill conceived
nature of the whole project. I never cared much for our former
governor and didn't think he understood much outside of politics, this
project being one of the more obvious follies (to me). But don't
mistake this for me agreeing with Martinez's knee-jerk attempts to
dismantle everything he did. I think the Supercomputer initiative
was well intentioned, and maybe at one level of understanding, very
promising.
But frankly the main things "Big Iron" are good for are creating an
expensive sandbox to spend money in. You have to have a lot of money
to spend to even begin to use them effectively... mostly programmers
and scientists-cum-programmers to make proper use of them (and machine
rooms with lots of power and environmental control and...) and staff
to keep them running and up to date and ... The state (whether as a
government, a collection of academic institutions, a budding
commercial venue for high tech, thousands of entrepreneurs, etc.)
simply didn't (and doesn't) have the kind of oomph it needs. Sandia
and LANL and NCGR all use UberComputing about as effectively as
anyone, but have huge staff and budgets to make that happen.
I had worked in "big iron" shops most of my career, never really
believing in them. While I *do* think some important things were done
because there was big iron at places like LANL, I think the bulk of
the "important things" happened first on the smaller machines (Vax
with VMS or BSD) in the 80's and then the plethora of Scientific
Workstations (e.g. Sun, SGI, Apollo, HP, NeXT, etc.) in the 90's and
ultimately the PCs running Linux and the mini-clusters that grew from
them.
Even though I worked with and on the big iron of different Generations
(CDC/Cray/TMC/*NIX-cluster-of-the-month) and even built a few utility
Linux clusters, I never believed that the roughly single or double
order of magnitude increases led to many qualitative advances in
computing or science. There certainly have been *some* important
advances made, the most obvious (in my uneducated opinion) might have
been in bioinformatics. Generally, the value seems to have been in
embarassingly parallel problems where there was funding to pay for the
"big iron" and a clear value to shortening the time to an answer by a
couple of orders of magnitude (like getting an answer in a day that
otherwise might take a week or even a few months).
I think some Science was accellerated quite well by that kind of
leverage. But in other fields it simply became an excuse for bloated
budgets and distracting scientists from their science and making
(letting?) them become computer scientists. There is plenty of
precedent for this as early as the 40's and 50's which I respect..
modern computing might not exist were it not for those early "ACS"
(MANIAC, ILLIAC, etc)
It may seem contradictory, but I *do* believe all that flailing that I
observed (and too often participated in) with big iron and hordes of
small iron (clusters) and DYI/NIH development (from OSs to hardware to
text editors for crimeny sake!) was an important early seed for much
of our current consumer, entertainment and hobbyist-driven
computing. While *games* may have really fueled the graphics cards,
it was SGI that really got it all moving in the right direction in the
first place... getting over the early hurdles.
I'm not an expert on the Space Program but while Tang, Space Blankets
and zero-G ballpoints might be the more obvious but mundane (trite?)
spinoffs, there are also more impactful spinoff technologies like
Velcro and photovoltaics and heat pipes and bears, oh my! to refer
to. Similarly the huge (ginormous?) budgets that Defense and Energy
put into uber-computing over decades, have had valuable
side-effects... but I never believed that a *State* could achieve the
same thing. Maybe the Japanese or Chinese "state", but not NM...
Nevertheless I *am* sympathetic with those who really, really (really)
wanted it to work. But I am not sympathetic with the Martinez gang
who have been using every opportunity to bash the previous
administration. I think this particular failure is real, but I think
the fanfare around *demolitioning* it is totally politics-driven hype
of the worst kind.
Yes, the gear is vintage if not antique and there is unlikely any
*commercial* market for it. I'm not sure of all the implications of
"selling" it in pieces to the Universities (State run) but it seems
likely the funding to "buy" it comes out of the same pocket that it
goes back into when sold. This might be a useful a useful
bookkeeping fiction, but I suspect it is another
Richardson-bashing/Martinez-grandstanding opportunity.
I don't really agree with Owen on the presumption that such resources
can't be used effectively without a fat pipe all the way into our
houses (or offices)... Remote X, VNC, etc. make it pretty easy to
do 99% of what you need to do without ever bringing the bulk of the
data back over the net. I too romanticize having a direct drop on a
Tbit/sec at my dinner table but I don't think the lack of it explains
my lack of utilization of big iron (whether in Rio Rancho, Los Alamos
or Mountain View).
The availability of the Amazon Cloud and/or relatively affordable
price of a densely packed GPU/CPU mini-cluster challenges us all to
put our projects where our mouth is and actually implement some
effective parallel algorithms that can do the heavy lifting. The
tools are there to make this 100 times easier than it ever was when I
was learning/developing the tricks of the trade... and it is still
hard. My only words of wisdom on the topic might be that instead of
limiting ourselves to well known "embarrassingly parallel" algorithms
or swimming upstream trying to force fit intrinsically serial
algorithms into parallel environments, we should look to discovering
(recognizing, inventing?) uniquely different approaches. This is
what the nonlinear and complexity science movement of the 80's did in
it's own way to reconfigure formerly intractable (intellectually as
well as computational) problems into tractable, and sometimes even
*elegant* problems with similarly elegant solutions.
- Steve
============================================================
FRIAM Applied Complexity Group listserv
Meets Fridays 9a-11:30 at cafe at St. John's College
to unsubscribe http://redfish.com/mailman/listinfo/friam_redfish.com
============================================================
FRIAM Applied Complexity Group listserv
Meets Fridays 9a-11:30 at cafe at St. John's College
to unsubscribe http://redfish.com/mailman/listinfo/friam_redfish.com