Thank you for the replies, and now I am aware of three different tools for 
parallel/distributed computing:

  1. GNU Parallel (https://www.gnu.org/software/parallel/) - 'Works great for 
me. Easy to install. Easy to use. Given well-written programs, can be used to 
take advantage of all the cores on your computer.

  2. Apache Hadoop (https://hadoop.apache.org) - 'Seems to be optimized for 
map/reduced computing, but can take advantage of many computers with many cores.

  3. Apache Spark (http://spark.apache.org) - 'Seems to have learned lessons 
from Hadoop. 'Seems easier to install, and can also take advantage of many 
computers with many cores, but also comes with a few programming interfaces 
such as Python & Java. Also seems a bit easier to install than Hadoop.

If I had one computer with 132 cores, then I'd probably stick with Parallel. If 
I had multiple computers with multiple cores, then I'd probably look more 
closely at Spark. Both Hadoop & Spark are "heavy" frameworks, compared to 
Parallel which is one Perl script.

--
Eric Morgan

Reply via email to