On 10/01/2015 09:27 AM, Orion Poplawski wrote:
We may be looking a getting a couple new compute nodes. I'm leery
though of
going too high in processor core counts. Does anyone have any general
experiences with performance scaling up to 12 cores per processor with
general
models like CM1/WRF/RAMS on the current crop of Xeon processors?
Lately I have been working on a system with >512Gb of RAM and a lot of
processors.
This wouldn't be at all a cost effective beowulf node, but it is a
godsend when the problems being addressed require huge amounts of memory
and do not partition easily to run on multiple nodes. The one I'm using
is NUMA rather than SMP, and careful placement of the data with respect
to the nodes is sometimes required for optimal performance. That is
likely to be the case on some of the machines you may be looking at.
One must also be careful using multiple processors on this machine
because some of them share cache, and others don't, so that adding too
many processes in the wrong place can reduce throughput because these
start having to go to slower main memory inside tight loops. Some of
this is discussed in this longish thread:
http://www.beowulf.org/pipermail/beowulf/2015-July/033263.html
This machine is also prone to locking up (to the point it doesn't answer
terminal keystrokes from a remote X11 terminal) when writing huge files
back to disk. I have not tracked this one down yet, it seems to be
related to unmapping a memory mapped 10.5 Gb file. A bit difficult to
debug because when it is happening it isn't possible to look at what the
machine is doing.
Regards,
David Mathog
[email protected]
Manager, Sequence Analysis Facility, Biology Division, Caltech
_______________________________________________
Beowulf mailing list, [email protected] sponsored by Penguin Computing
To change your subscription (digest mode or unsubscribe) visit
http://www.beowulf.org/mailman/listinfo/beowulf