Eray Ozkural (exa) wrote:

>LAM seems to be more practical, and since it doesn't bring performance
>penalties it's our implementation of choice.
>
That's about what I thought.

>Nevertheless, I would not suggest
>switching to lam unless we have the 6.5.2 stable version available in
>debian which we do not. The 6.5.2 will require packaging changes,
>the debian diffs for 6.3.2 won't work. I will present new versions for
>testing soon hopefully. Anyway, the new lam version builds and installs
>with ./configure, make and make install so it's quite trivial.
>
>The technical justification for mpich would be
> 1. a more complete / up-to-date implementation
> 2. better performance
>
>Generally, I haven't observed (2). However, for custom hardware mpich
>may be better (i.e. myrinet) And note that it has been used and extended
>by many vendors as a reference implementation. (1) Standards compliance
>is surely a more important matter. For instance, there is the new
>
>MPI_AlltoallW(... many args ...)
>
>function intended for matrix computations. If the new mpich has it while lam
>lacking, then for the sake of standard mpich would be better.
>
That makes sense.  I'll leave it with mpich at least until lam 6.5.2 is 
in unstable.

Regarding standards, since it is built to work with both, I imagine 
PETSc either doesn't use mpich-specific interfaces, or else has 
workarounds for interface differences, so "whichever is faster" is the 
correct choice.  But this means I probably should do more extensive 
testing than just a few nonlinear finite difference problems, maybe a 
couple of dense matrix tests too, which could give mpich an advantage if 
PETSc does use MPI_AlltoallW().

Also, since we're largely an x86 community, and Linux has only recently 
started to scale to larger SMP machines, I imagine most Debian PETSc 
users are of the Beowulf variety, with relatively inexpensive 
networking, i.e. not myrinet.  But I could be wrong...  In any case, 
"whichever is faster" should be the fastest implementation for the most 
people, and let the rest just do (assuming lam is chosen):

    apt-get source petsc
    cd petsc-2.1.0
    debian/rules PETSC_MPI=mpich binary

Thank you for bringing these issues to my attention, I'm pretty much 
"just a user" with barely enough knowledge of internals to make stuff 
work. :-)

>Hmm, I think I should look at the neat petscgraphics ;)
>
Now that it's uploaded for more than just powerpc, you should be able to 
just apt-get install petscgraphics1-demo to drag in its dependencies, 
and then run chts or "mpirun -np X /usr/bin/chts".  I'm working out a 
divide by zero issue, then will upload for alpha too.

At this point, the slow steps are sending the triangles to geomview, and 
rendering them there; the triangulation generation itself is fast even 
on one CPU.  And it exposes the geomview transparency bug, so for all 
those reasons I need to get away from the current architecture.  I'm 
thinking of using Evas to render the semi-transparent triangles into a 
pixmap on each processor, then just have node 0 layer those 
semi-transparent pixmaps.  Or is there a good, fast alternative to Evas? :-)

Zeen,
-- 

-Adam P.

GPG fingerprint: D54D 1AEE B11C CE9B A02B  C5DD 526F 01E8 564E E4B6

Welcome to the best software in the world today cafe! 
<http://lyre.mit.edu/%7Epowell/The_Best_Stuff_In_The_World_Today_Cafe.ogg>




--  
To UNSUBSCRIBE, email to [EMAIL PROTECTED]
with a subject of "unsubscribe". Trouble? Contact [EMAIL PROTECTED]

Reply via email to