On 04/15/2013 02:04 PM, Ashley Pittman wrote: > On 14 Apr 2013, at 10:01, Jonathan Dursi wrote: > >> I cannot agree with this piece highly enough. >> >> Widespread cloud availability, GPU, etc, has enabled all sorts of weird, >> wacky, and _useful_ large-scale technical computing use cases, and arguing >> about whether new use case X is "really" HPC has long since lost whatever >> novelty it had. I'm pleased to see Jeff Layton using the broader term >> "Research Computing"; in my corner of the world I've been pushing for the >> term Advanced R&D Computing (ARC) as a catch all for any sort of >> technical/numerical computing that requires you to do something "special" >> (e.g., do something different than run naive serial code on a desktop). >> Someone else can probably come up with a better name, but I actually think >> that holding on to terms with existing pretty strong connotations is hurting >> more than helping at this point. > I've taken to saying I work in "Computing" as a distinct field from "IT". > The difference being that Computing is about using computers for > calculations/analytics rather than as a tool for doing something else. >
I do the same thing. The term "IT" makes my skin crawl. I've always considered myself a scientist more than an IT person. Prentice _______________________________________________ Beowulf mailing list, [email protected] sponsored by Penguin Computing To change your subscription (digest mode or unsubscribe) visit http://www.beowulf.org/mailman/listinfo/beowulf
