On 10/21/10 06:43, Eugen Leitl wrote:
The cloud is real. The idea that you need a physical machine close by to get
any work done is completely out the window at this point. My only machine at
Google is a Mac laptop (with a big honking monitor and wireless keyboard and
trackpad when I am at my desk). I do all of my development work on a virtual
Linux machine running in a datacenter somewhere -- I am not sure exactly
where, not that it matters. I ssh into the virtual machine to do pretty much
everything: edit code, fire off builds, run tests, etc. The systems I build
are running in various datacenters and I rarely notice or care where they are
physically located. Wide-area network latencies are low enough that this
works fine for interactive use, even when I'm at home on my cable modem.

The fact that the author is using a Mac and doing development work on a virtual Linux machine in an unknown location highlights the underlying theme of the article, the resultant thread, and perhaps even this entire mailing list:

Different setups work better for different workloads.

Clearly, the author feels that the inconvenience incurred by having to use a virtual Linux machine to perform his development is less than the inconvenience of running Linux as his main OS. Otherwise, he would simply use Linux on his machine and sit at home in his pajamas, sipping a hot cup of Earl Grey and working out his HPC problem locally. Nonetheless, there are numerous examples of workloads in the scientific community (used here in reference to the physical sciences) and in HPC development, which unfortunately do not play nicely with such a remote and fluctuating setup.

For instance, in my research, it is far easier to own the machines one runs on (or at least have root access) to develop and test breakthroughs in systems development. Often messing with the kernel, toolchain, or arbitrary libraries in the distribution is required to effect and test the change in which one is interested. It goes without saying that we have quite a bit of difficulty convincing "IT" types (even being computer science persons ourselves) that this is a reasonable thing to do on the average cluster, even in a university setting. Certainly, in "the cloud" alterations at this level are not tolerated. Further, it is extremely rare to find clusters tailored to system development such that they have master nodes that reboot all the slave nodes with new images and new file systems (and however many more system specific parameters, specified by the developer) for every run.

That said, I do recognize that system development is a small sector in HPC and not by any means the most influential customer. However, I do feel it furthers the mantra that we should not all be pigeon-holed into one particularly "efficient" setup just because it works well for the common case (or because Google "invented it").

As totally off-topic points: it is great to see that RGB-Bot has been rebooted (even if only with limited bandwidth), and I absolutely have no idea why Eugen Leitl posted a blog entry from Matt Welsh. I scanned and scanned for some comment from Eugen or the way in which it somehow wrapped into recent conversation, but at this point I'm lost on why it originally got posted (besides being quite the fire-starter).

Best,

ellis
_______________________________________________
Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing
To change your subscription (digest mode or unsubscribe) visit 
http://www.beowulf.org/mailman/listinfo/beowulf

Reply via email to