Dear All, I compiled the latest version of the code on a cluster made up of several Dual Core AMD Opteron nodes by ifort and mpif90. There is no any error or warning in the SRC_*/compile.msg files. The code runs well on the nodes, if we only parallelize the k-points by for example the following .machines file: ?lapw0:node3 node20 1:node3 1:node20 granularity:1 extrafine:1
But, the program stops with the following error:? ifort: command line warning #10159: invalid argument for option '-m' ifort: command line error: option '-n' is ambiguous once we use the fine grained parallelization by for example the following .machines file: ?lapw0:node3:2 node20:2 1:node3:2 1:node20:2 granularity:1 extrafine:1 I have used l_cprof_p_11.1.073_intel64, fftw-2.1.5, mpich2 for compiling the code, and the following settings. ----------------------------------------- Current settings: ?O???Compiler options:? ? ? ??-FR -mp1 -w -prec_div -pc80 -pad -ip -DINTEL_VML -traceback ?L???Linker Flags:? ? ? ? ? ??$(FOPT) -L/home/softs/intel/ifort11/mkl/lib/em64t -pthread ?P???Preprocessor flags?? ? ??'-DParallel' ?R???R_LIB (LAPACK+BLAS):?? ??-lmkl_lapack -lmkl_intel_lp64 -lmkl_intel_thread -lmkl_core -lpthread -lguide ------------------------------------------- ?? ??RP??RP_LIB(SCALAPACK+PBLAS): -lmkl_scalapack_lp64 -lmkl_solver_lp64 -lmkl_blacs_lp64 -L/home/softs/mpich2/lib -lmpich -L/home/softs/fftw-2.1.5/mpi/.libs/ -lfftw_mpi -L/home/softs/fftw-2.1.5/fftw/.libs/ -lfftw $(R_LIBS) ?? ??FP??FPOPT(par.comp.options): -FR -mp1 -w -prec_div -pc80 -pad -ip -DINTEL_VML -traceback -I/home/softs/mpch2/include ?? ??MP??MPIRUN commando? ? ? ??: /home/softs/mpich2/bin/mpif90 -machinefile _HOSTS_ -n _NP_ _EXEC_ --------------------------------------------- The parallel_options is:? --------------------------------------------- setenv USE_REMOTE 1 setenv MPI_REMOTE 1 setenv WIEN_GRANULARITY 1 setenv WIEN_MPIRUN "/home/softs/mpich2/bin/mpif90 -machinefile _HOSTS_ -n _NP_ _EXEC_" ------------------------------- I changed the mpif90 to mpirun only in the parallel options (just as a test) but I did not recompile the code by mpirun. The result is as follows: LAPW0 END ?LAPW0 END Fatal error in PMPI_Comm_size: Invalid communicator, error stack: PMPI_Comm_size(111): MPI_Comm_size(comm=0x5b, size=0x8aa96c) failed PMPI_Comm_size(69).: Invalid communicator real0m0.050s user0m0.010s sys0m0.038s test.scf1_1: No such file or directory. FERMI - Error cp: cannot stat `.in.tmp': No such file or directory rm: cannot remove `.in.tmp': No such file or directory rm: cannot remove `.in.tmp1': No such file or directory Similar error was occurred when I used mpiexec in the parallel options without recompiling the code.? I found that the "Invalid communicator" originates from incompatible mpi.h of mpirun or mpiexec with that of mpif90. So I changed back it to mpif90. Since I guessed that the problem originates from the version of mpi, I tried different versions of mpi, i.e., mpich2-1.0.6, mpich2-1.3.1, mpich2-1.4, openmpi-1.4.2. Any comment on why the code is compiled with no error or warning, but it stops with error is highly appreciated.? Is there any restrictions for compiling the code by mpif90 on AMD systems, as discussed above? ? Sincerely yours, S. Jalali /_/_/_/_/_/_/_/_/_/_/_/_/_/_/_/_/_/_/_/_/_/_/_/_/_/ Saeid Jalali Asadabadi, Department of Physics, Faculty of Science, University of Isfahan (UI), Hezar Gerib Avenue, 81744 Isfahan, Iran. Phones: Dep. of Phys. ? :+98-0311-793 2435 Office ? ? ? ? ? ? ? :+98-0311-793 4776 Fax No. ? ? ? ? ? ?:+98-0311-793 4800 E-mail ? ? ? ? ? ? ?:sjalali at phys.ui.ac.ir ? ? ? ? ? ? ? ? ? ? ?? ?? :sjalali at sci.ui.ac.ir ? ? ? ? ? ? ? ? ? ? ? ? ? :sjalali at mailaps.org ? ? ? ? ? ? ? ? ? ? ? ? ? :saeid.jalali.asadabadi at gmail.com ? ? ? ? ? ? ? ? ? ? ? ? ? :s_jalali_a at yahoo.com Homepage ? ? ? ?:http://sci.ui.ac.ir/~sjalali www ? ? ? ? ? ? ? ?? :http://www.ui.ac.ir /_/_/_/_/_/_/_/_/_/_/_/_/_/_/_/_/_/_/_/_/_/_/_/_/_/ -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://zeus.theochem.tuwien.ac.at/pipermail/wien/attachments/20120531/06000eb4/attachment.htm>