Lennart Augustsson and Thomas Johnsson got some encouraging results fifteen years ago with their nu-G-machine. They compiled Lazy ML for a shared memory multiprocessor, and benchmarked against the sequential LML compiler, the precursor of hbc and at that time the best compiler for a lazy functional language. They observed speed-ups of at least 20% even with only two processors, and with four the speed-ups were in the range 2-3.3x. The idea was to use "sparks" to start parallel computations, which were ignored, and thus cheap, unless a processor was actually available. OK, then benchmarks were tiny (n-queens), and the LML compiler had higher overheads for ordinary computation than GHC, but still, it's an encouraging result for exploiting multi-core machines.

Here's the reference:

@inproceedings{99386,
author = {Lennart Augustsson and Thomas Johnsson},
title = {Parallel graph reduction with the (v , G)-machine},
booktitle = {FPCA '89: Proceedings of the fourth international conference on 
Functional programming languages and computer architecture},
year = {1989},
isbn = {0-89791-328-0},
pages = {202--213},
location = {Imperial College, London, United Kingdom},
doi = {http://doi.acm.org/10.1145/99370.99386},
publisher = {ACM Press},
}

John



I thought the "lazy functional languages are great for implicit parallelism" thing died out some time ago - at least as far as running the programs on conventional hardware is concerned.


Designing an algorithm that breaks apart a "sequential" lazy program into parallel chunks of the appropriate size is **HARD** (with double asterixes). The time and space behavior of a lazy program is complex enough for the _programmer_ to reason about, let alone an automated analysis - which has no knowledge of what the program is actually trying to do.

I think a more likely approach lies in the direction of the so called "parallel strategies". If you haven't already, I would strongly suggest reading: Algorithm + Strategy = Parallelism, 1998, PW Trinder, et al.
You can get this paper from Simon Peyton Jones's homepage.


Also, at the end of Hans Wolfgang-Loidl's thesis he develops a granularity analysis for a Haskell subset - one of the first steps in any kind of implicit parallelism. It's a pretty good effort, but at the end of it all it still relies on a pre-existing table of information about recursive functions. I think that these kind of analyses tend suffer from uncomputability problems more than most.

If you've still got your heart set on implicit parallelism, then there's a (very slow) simulator you might want to poke around with. I wrote it based around Clem Baker-Finch's "Abstract machine for parallel lazy evaluation", which supports fully speculative implicit parallelism.

There's a link to it on my homepage at http://cs.anu.edu.au/people/Ben.Lippmeier


Keean Schupke wrote:

I have to say I disagree... I feel Haskell is highly suited to implicit parallel execution... The key to "implicit" parallelisation is that it is implicit - not explicit, so the programmer should feel like they are programming a sequential language. If we can assume little memory access penalties for threads running on other CPUs (shared cache model), it seems to be a matter of putting locks on the right structures, and allowing any worker-thread to take the next function ready to run from the scheduler.
Keean.


_______________________________________________
Haskell mailing list
[email protected]
http://www.haskell.org/mailman/listinfo/haskell


_______________________________________________
Haskell mailing list
[email protected]
http://www.haskell.org/mailman/listinfo/haskell


_______________________________________________
Haskell mailing list
[email protected]
http://www.haskell.org/mailman/listinfo/haskell

Reply via email to