One more thing i forgot to mention in my previous e-mail. In the output file I 
get the following message:


2 total processes killed (some possibly by mpirun during cleanup)


Thanks



________________________________
From: users <users-boun...@open-mpi.org> on behalf of Elio Physics 
<elio-phys...@live.com>
Sent: Saturday, April 23, 2016 3:07 AM
To: Open MPI Users
Subject: Re: [OMPI users] MPIRUN SEGMENTATION FAULT


I have used valgrind and this is what i got:


valgrind mpirun ~/Elie/SPRKKR/bin/kkrscf6.3MPI Fe_SCF.inp > 
scf-51551.jlborges.fisica.ufmg.br.out
==8135== Memcheck, a memory error detector
==8135== Copyright (C) 2002-2012, and GNU GPL'd, by Julian Seward et al.
==8135== Using Valgrind-3.8.1 and LibVEX; rerun with -h for copyright info
==8135== Command: mpirun /home/emoujaes/Elie/SPRKKR/bin/kkrscf6.3MPI Fe_SCF.inp
==8135==
--------------------------------------------------------------------------
mpirun noticed that process rank 0 with PID 8147 on node 
jlborges.fisica.ufmg.br exited on signal 11 (Segmentation fault).
--------------------------------------------------------------------------
==8135==
==8135== HEAP SUMMARY:
==8135==     in use at exit: 485,683 bytes in 1,899 blocks
==8135==   total heap usage: 7,723 allocs, 5,824 frees, 12,185,660 bytes 
allocated
==8135==
==8135== LEAK SUMMARY:
==8135==    definitely lost: 34,944 bytes in 34 blocks
==8135==    indirectly lost: 26,613 bytes in 58 blocks
==8135==      possibly lost: 0 bytes in 0 blocks
==8135==    still reachable: 424,126 bytes in 1,807 blocks
==8135==         suppressed: 0 bytes in 0 blocks
==8135== Rerun with --leak-check=full to see details of leaked memory
==8135==
==8135== For counts of detected and suppressed errors, rerun with: -v
==8135== ERROR SUMMARY: 0 errors from 0 contexts (suppressed: 6 from 6)


What does that supposed to mean?


Regards

________________________________
From: users <users-boun...@open-mpi.org> on behalf of Ralph Castain 
<r...@open-mpi.org>
Sent: Saturday, April 23, 2016 1:36:50 AM
To: Open MPI Users
Subject: Re: [OMPI users] MPIRUN SEGMENTATION FAULT

All I can say is that your program segfault'd during execution - you might want 
to look at the core file using a debugger like gdb to see why it failed.


On Apr 22, 2016, at 8:32 PM, Elio Physics 
<elio-phys...@live.com<mailto:elio-phys...@live.com>> wrote:

Dear all,

I have successfully compiled a code where the executable have been produced. 
However when I started using the executable with mpirun, the code stopped with 
the following error:

"mpirun noticed that process rank 0 with PID 570 on node compute-1-9.local 
exited on signal 11 (Segmentation fault)."

What is that error due to? and How can i solve it?

I will post the make.inc compilation file:

BUILD_TYPE ?=
#BUILD_TYPE := debug

VERSION = 6.3

ifeq ($(BUILD_TYPE), debug)
 VERSION := $(VERSION)$(BUILD_TYPE)
endif

BIN =~/Elie/SPRKKR/bin
#BIN=~/bin
#BIN=/tmp/$(USER)

LIB =  -L/opt/intel/mkl/lib/intel64/libmkl_blas95_ilp64 
-L/opt/intel/mkl/lib/intel64/libmkl_lapack95_ilp64 -L/opt/intel/mkl/lib/intel64 
-lmkl_scalapack_ilp64 -lmkl_intel_ilp64 -lmkl_core -lmkl_sequential 
-lmkl_blacs_i
ntelmpi_ilp64 -lpthread -lm -ldl

#Include mpif.h
INCLUDE = -I/opt/intel/mkl/include/intel64/ilp64  -I/opt/intel/mkl/lib/include

#FFLAGS
FFLAGS = -O2

FC   = mpif90 -c $(FFLAGS) $(INCLUDE)
LINK = mpif90   $(FFLAGS) $(INCLUDE)

MPI=MPI

Thanks in advance

Elio
University of Rondonia, brazil

_______________________________________________
users mailing list
us...@open-mpi.org<mailto:us...@open-mpi.org>
Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
Link to this post: 
http://www.open-mpi.org/community/lists/users/2016/04/29000.php

Reply via email to