Hi Victor

I'd suggest 3 seconds of CPU time is far, far to small a problem to do
scaling tests with.  Even with only 2 CPUs, I wouldn't go below 100
times that.


On Mon, 2007-06-11 at 01:10 -0700, victor marian wrote:
> Hi Jeff
> 
> I ran the NAS Parallel Bechmark and it gives for me
> -bash%/export/home/vmarian/fortran/benchmarks/NPB3.2/NPB3.2-MPI/bin$
> mpirun -np 1 cg.A.1
> --------------------------------------------------------------------------
> [0,1,0]: uDAPL on host SERVSOLARIS was unable to find
> any NICs.
> Another transport will be used instead, although this
> may result in
> lower performance.
> --------------------------------------------------------------------------
>  NAS Parallel Benchmarks 3.2 -- CG Benchmark
> 
>  Size:      14000
>  Iterations:    15
>  Number of active processes:     1
>  Number of nonzeroes per row:       11
>  Eigenvalue shift: .200E+02
>  Benchmark completed
>  VERIFICATION SUCCESSFUL
>  Zeta is      0.171302350540E+02
>  Error is     0.512264003323E-13
> 
> 
>  CG Benchmark Completed.
>  Class           =                        A
>  Size            =                    14000
>  Iterations      =                       15
>  Time in seconds =                     3.02
>  Total processes =                        1
>  Compiled procs  =                        1
>  Mop/s total     =                   495.93
>  Mop/s/process   =                   495.93
>  Operation type  =           floating point
>  Verification    =               SUCCESSFUL
>  Version         =                      3.2
>  Compile date    =              11 Jun 2007
> 
> 
> -bash%/export/home/vmarian/fortran/benchmarks/NPB3.2/NPB3.2-MPI/bin$
> mpirun -np 2 cg.A.2
> --------------------------------------------------------------------------
> [0,1,0]: uDAPL on host SERVSOLARIS was unable to find
> any NICs.
> Another transport will be used instead, although this
> may result in
> lower performance.
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> [0,1,1]: uDAPL on host SERVSOLARIS was unable to find
> any NICs.
> Another transport will be used instead, although this
> may result in
> lower performance.
> --------------------------------------------------------------------------
> 
> 
>  NAS Parallel Benchmarks 3.2 -- CG Benchmark
> 
>  Size:      14000
>  Iterations:    15
>  Number of active processes:     2
>  Number of nonzeroes per row:       11
>  Eigenvalue shift: .200E+02
> 
>  Benchmark completed
>  VERIFICATION SUCCESSFUL
>  Zeta is      0.171302350540E+02
>  Error is     0.522633719989E-13
> 
> 
>  CG Benchmark Completed.
>  Class           =                        A
>  Size            =                    14000
>  Iterations      =                       15
>  Time in seconds =                     2.47
>  Total processes =                        2
>  Compiled procs  =                        2
>  Mop/s total     =                   606.32
>  Mop/s/process   =                   303.16
>  Operation type  =           floating point
>  Verification    =               SUCCESSFUL
>  Version         =                      3.2
>  Compile date    =              11 Jun 2007
> 
> 
>     You can remark that the scalling is not so good
> like yours. Maibe I am having comunications problems
> between processors.
>    You can also remark that I am faster on one process
> concared to your processor.
> 
>                                        Victor
> 
> 
> 
> 
> 
> --- Jeff Pummill <jpum...@uark.edu> wrote:
> 
> > Perfect! Thanks Jeff!
> > 
> > The NAS Parallel Benchmark on a dual core AMD
> > machine now returns this...
> > [jpummil@localhost bin]$ mpirun -np 1 cg.A.1
> > NAS Parallel Benchmarks 3.2 -- CG Benchmark
> > CG Benchmark Completed.
> >  Class           =                        A
> >  Size            =                    14000
> >  Iterations      =                       15
> >  Time in seconds =                     4.75
> >  Total processes =                        1
> >  Compiled procs  =                        1
> >  Mop/s total     =                   315.32
> > 
> > ...and...
> > 
> > [jpummil@localhost bin]$ mpirun -np 2 cg.A.2
> > NAS Parallel Benchmarks 3.2 -- CG Benchmark
> >  CG Benchmark Completed.
> >  Class           =                        A
> >  Size            =                    14000
> >  Iterations      =                       15
> >  Time in seconds =                     2.48
> >  Total processes =                        2
> >  Compiled procs  =                        2
> >  Mop/s total     =                   604.46
> > 
> > Not quite linear, but one must account for all of
> > the OS traffic that 
> > one core or the other must deal with.
> > 
> > 
> > Jeff F. Pummill
> > Senior Linux Cluster Administrator
> > University of Arkansas
> > Fayetteville, Arkansas 72701
> > (479) 575 - 4590
> > http://hpc.uark.edu
> > 
> > "A supercomputer is a device for turning
> > compute-bound
> > problems into I/O-bound problems." -Seymour Cray
> > 
> > 
> > Jeff Squyres wrote:
> > > Just remove the -L and -l arguments -- OMPI's
> > "mpif90" (and other  
> > > wrapper compilers) will do all that magic for you.
> > >
> > > Many -L/-l arguments in MPI application Makefiles
> > are throwbacks to  
> > > older versions of MPICH wrapper compilers that
> > didn't always work  
> > > properly.  Those days are long gone; most (all?)
> > MPI wrapper  
> > > compilers do not need you to specify -L/-l these
> > days.
> > >
> > >
> > >
> > > On Jun 10, 2007, at 3:08 PM, Jeff Pummill wrote:
> > >
> > >   
> > >> Maybe the "dumb question" of the week, but here
> > goes...
> > >>
> > >> I am trying to compile a piece of code (NPB)
> > under OpenMPI and I am  
> > >> having a problem with specifying the right
> > library. Possibly  
> > >> something I need to define in a LD_LIBRARY_PATH
> > statement?
> > >>
> > >> Using Gnu mpich, the line looked like this...
> > >>
> > >> FMPI_LIB  = -L/opt/mpich/gnu/lib/ -lmpich
> > >>
> > >> I tried to replace this with...
> > >>
> > >> FMPI_LIB  = -L/usr/lib/openmpi/ -llibmpi
> > >>
> > >> to which the make responded...
> > >> mpif90 -O -o ../bin/cg.A.2 cg.o
> > ../common/print_results.o ../common/ 
> > >> randdp.o ../common/timers.o -L/usr/lib/openmpi/
> > -llibmpi
> > >> /usr/bin/ld: cannot find -llibmpi
> > >> collect2: ld returned 1 exit status
> > >>
> > >> Wrong library file? Setup or path issue?
> > >>
> > >> -- 
> > >> Jeff F. Pummill
> > >> Senior Linux Cluster Administrator
> > >> University of Arkansas
> > >>
> > >> _______________________________________________
> > >> users mailing list
> > >> us...@open-mpi.org
> > >>
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
> > >>     
> > >
> > >
> > >   
> > > _______________________________________________
> > users mailing list
> > us...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> 
> 
>  
> ____________________________________________________________________________________
> Get your own web address.  
> Have a HUGE year through Yahoo! Small Business.
> http://smallbusiness.yahoo.com/domains/?p=BESTDEAL
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to