Re: [SIESTA-L] Compiling and running Siesta with openMPI-2.0.2, gfortran-6.3.0 (Debian 9)

2021-06-30 Por tôpico Nick Papior
Hmm... That's quite annoying... :(

Thanks for reporting back!

Den ons. 30. jun. 2021 kl. 15.39 skrev Karen Fidanyan <
karen.fidan...@mpsd.mpg.de>:

> Dear Nick,
>
> I've tried Divide-and-Conquer, Expert and QR, they all fail with the same
> backtrace.
> I couldn't compile MRRR, I think my ScaLAPACK misses some routines.
>
> But, following your idea about a bug in ScaLAPACK, I recompiled Siesta
> with the MKL libraries from the Debian-9 repo. They are from 2019, not that
> old.
> It also failed for Divide-and-Conquer, but MRRR, Expert and QR work fine.
> I think it's enough for my purposes, I'll use MRRR. But I don't know what
> is wrong with D I attach the output of the D with MKL, maybe you find
> it useful.
>
> Thank you for help!
>
> Best,
> Karen
> On 6/30/21 11:25 AM, Nick Papior wrote:
>
> I have now tried to rerun it with 4.1.
> And I get no error, even in debug mode.
>
> My bet is that the scalapack library is an old and buggy one. But I could
> be wrong.
>
> Could you rerun with the different possibilities for Diag.Algorithm?
> I.e. try them all and see which ones works, and which doesn't, then report
> back.
>
> Den ons. 30. jun. 2021 kl. 11.16 skrev Karen Fidanyan <
> karen.fidan...@mpsd.mpg.de>:
>
>> Dear Nick,
>>
>> thanks for helping!
>>
>> I redid it with -Og flag. The input, *.psf and the output are attached. I
>> also attach debug.* files obtained with -DDEBUG.
>> I run as `mpirun -np 2 ~/soft/siesta-4.1/Obj-dbg-Og/siesta control.fdf
>> 2>&1 | tee siesta.out`.
>>
>> Sincerely,
>> Karen Fidanyan
>> On 6/28/21 10:22 PM, Nick Papior wrote:
>>
>> I can't rerun without psf files.
>>
>> Could you try and compile with -Og -g -fbacktrace (without fcheck=all).
>>
>> Then try again. :)
>>
>> Den man. 28. jun. 2021 kl. 22.01 skrev Karen Fidanyan <
>> karen.fidan...@mpsd.mpg.de>:
>>
>>> Dear Siesta users,
>>>
>>> I'm having a hard time trying to run SIESTA on my Debian-9 laptop.
>>> I have:
>>>
>>> GNU Fortran (Debian 6.3.0-18+deb9u1) 6.3.0 20170516
>>> OpenMPI-2.0.2-2
>>> libblas 3.7.0-2, liblapack 3.7.0-2
>>> libscalapack-openmpi1 1.8.0-13
>>>
>>> My arch.make is the following:
>>> **
>>> .SUFFIXES:
>>> .SUFFIXES: .f .F .o .a .f90 .F90 .c
>>>
>>> SIESTA_ARCH = gfortran_openMPI
>>>
>>> FPP = $(FC) -E -P -x c
>>> FC = mpifort
>>> FC_SERIAL = gfortran
>>> FFLAGS = -O0 -g -fbacktrace -fcheck=all #-Wall
>>> FFLAGS_DEBUG = -g -O0
>>>
>>> PP = gcc -E -P -C
>>> CC = gcc
>>> CFLAGS = -O0 -g -Wall
>>>
>>> AR = ar
>>> RANLIB = ranlib
>>> SYS = nag
>>>
>>> LDFLAGS = -static-libgcc -ldl
>>>
>>> BLASLAPACK_LIBS = -llapack  -lblas \
>>>  -lscalapack-openmpi -lblacs-openmpi
>>> -lblacsF77init-openmpi \
>>>  -lblacsCinit-openmpi \
>>>  -lpthread -lm
>>>
>>> MPI_INTERFACE = libmpi_f90.a
>>> MPI_INCLUDE   = .
>>>
>>> FPPFLAGS_MPI = -DMPI -DMPI_TIMING -D_DIAG_WORK
>>> FPPFLAGS = $(DEFS_PREFIX) -DFC_HAVE_FLUSH -DFC_HAVE_ABORT $(FPPFLAGS_MPI)
>>>
>>> INCFLAGS = $(MPI_INCLUDE)
>>>
>>> LIBS = $(BLASLAPACK_LIBS) $(MPI_LIBS)
>>>
>>> atom.o: atom.F
>>>  $(FC) -c $(FFLAGS_DEBUG) $(INCFLAGS) $(FPPFLAGS)
>>> $(FPPFLAGS_fixed_F) $<
>>>
>>>
>>> .c.o:
>>>  $(CC) -c $(CFLAGS) $(INCFLAGS) $(CPPFLAGS) $<
>>> .F.o:
>>>  $(FC) -c $(FFLAGS) $(INCFLAGS) $(FPPFLAGS) $(FPPFLAGS_fixed_F)  $<
>>> .F90.o:
>>>  $(FC) -c $(FFLAGS) $(INCFLAGS) $(FPPFLAGS) $(FPPFLAGS_free_F90) $<
>>> .f.o:
>>>  $(FC) -c $(FFLAGS) $(INCFLAGS) $(FCFLAGS_fixed_f)  $<
>>> .f90.o:
>>>  $(FC) -c $(FFLAGS) $(INCFLAGS) $(FCFLAGS_free_f90)  $<
>>> **
>>>
>>> The code compiles without errors.
>>> If I run with Diag.ParallelOverK  True, I can run on multiple cores, no
>>> errors.
>>> With Diag.ParallelOverK  False, I can run `mpirun -np 1` without errors,
>>> but if I try to use >=2 cores, it fails with:
>>> **
>>> Program received signal SIGSEGV: Segmentation fault - invalid memory
>>> reference.
>>>
>>> Backtrace for this error:
>>> #0  0x2ba6eb754d1d in ???
>>> #1  0x2ba6eb753f7d in ???
>>> #2  0x2ba6ec95405f in ???
>>> #3  0x2ba70ec1cd8c in ???
>>> #4  0x2ba6eab438a4 in ???
>>> #5  0x2ba6eab44336 in ???
>>> #6  0x563b3f1cfead in __m_diag_MOD_diag_c
>>>  at /home/fidanyan/soft/siesta-4.1/Src/diag.F90:709
>>> #7  0x563b3f1d2ef9 in cdiag_
>>>  at /home/fidanyan/soft/siesta-4.1/Src/diag.F90:2253
>>> #8  0x563b3ebc7c8d in diagk_
>>>  at /home/fidanyan/soft/siesta-4.1/Src/diagk.F:195
>>> #9  0x563b3eb9d714 in __m_diagon_MOD_diagon
>>>  at /home/fidanyan/soft/siesta-4.1/Src/diagon.F:265
>>> #10  0x563b3ed897cb in __m_compute_dm_MOD_compute_dm
>>>  at /home/fidanyan/soft/siesta-4.1/Src/compute_dm.F:172
>>> #11  0x563b3edbfaa5 in __m_siesta_forces_MOD_siesta_forces
>>>  at /home/fidanyan/soft/siesta-4.1/Src/siesta_forces.F:315
>>> #12  0x563b3f9a4005 in siesta
>>>  

Re: [SIESTA-L] Compiling and running Siesta with openMPI-2.0.2, gfortran-6.3.0 (Debian 9)

2021-06-30 Por tôpico Nick Papior
I have now tried to rerun it with 4.1.
And I get no error, even in debug mode.

My bet is that the scalapack library is an old and buggy one. But I could
be wrong.

Could you rerun with the different possibilities for Diag.Algorithm?
I.e. try them all and see which ones works, and which doesn't, then report
back.

Den ons. 30. jun. 2021 kl. 11.16 skrev Karen Fidanyan <
karen.fidan...@mpsd.mpg.de>:

> Dear Nick,
>
> thanks for helping!
>
> I redid it with -Og flag. The input, *.psf and the output are attached. I
> also attach debug.* files obtained with -DDEBUG.
> I run as `mpirun -np 2 ~/soft/siesta-4.1/Obj-dbg-Og/siesta control.fdf
> 2>&1 | tee siesta.out`.
>
> Sincerely,
> Karen Fidanyan
> On 6/28/21 10:22 PM, Nick Papior wrote:
>
> I can't rerun without psf files.
>
> Could you try and compile with -Og -g -fbacktrace (without fcheck=all).
>
> Then try again. :)
>
> Den man. 28. jun. 2021 kl. 22.01 skrev Karen Fidanyan <
> karen.fidan...@mpsd.mpg.de>:
>
>> Dear Siesta users,
>>
>> I'm having a hard time trying to run SIESTA on my Debian-9 laptop.
>> I have:
>>
>> GNU Fortran (Debian 6.3.0-18+deb9u1) 6.3.0 20170516
>> OpenMPI-2.0.2-2
>> libblas 3.7.0-2, liblapack 3.7.0-2
>> libscalapack-openmpi1 1.8.0-13
>>
>> My arch.make is the following:
>> **
>> .SUFFIXES:
>> .SUFFIXES: .f .F .o .a .f90 .F90 .c
>>
>> SIESTA_ARCH = gfortran_openMPI
>>
>> FPP = $(FC) -E -P -x c
>> FC = mpifort
>> FC_SERIAL = gfortran
>> FFLAGS = -O0 -g -fbacktrace -fcheck=all #-Wall
>> FFLAGS_DEBUG = -g -O0
>>
>> PP = gcc -E -P -C
>> CC = gcc
>> CFLAGS = -O0 -g -Wall
>>
>> AR = ar
>> RANLIB = ranlib
>> SYS = nag
>>
>> LDFLAGS = -static-libgcc -ldl
>>
>> BLASLAPACK_LIBS = -llapack  -lblas \
>>  -lscalapack-openmpi -lblacs-openmpi
>> -lblacsF77init-openmpi \
>>  -lblacsCinit-openmpi \
>>  -lpthread -lm
>>
>> MPI_INTERFACE = libmpi_f90.a
>> MPI_INCLUDE   = .
>>
>> FPPFLAGS_MPI = -DMPI -DMPI_TIMING -D_DIAG_WORK
>> FPPFLAGS = $(DEFS_PREFIX) -DFC_HAVE_FLUSH -DFC_HAVE_ABORT $(FPPFLAGS_MPI)
>>
>> INCFLAGS = $(MPI_INCLUDE)
>>
>> LIBS = $(BLASLAPACK_LIBS) $(MPI_LIBS)
>>
>> atom.o: atom.F
>>  $(FC) -c $(FFLAGS_DEBUG) $(INCFLAGS) $(FPPFLAGS)
>> $(FPPFLAGS_fixed_F) $<
>>
>>
>> .c.o:
>>  $(CC) -c $(CFLAGS) $(INCFLAGS) $(CPPFLAGS) $<
>> .F.o:
>>  $(FC) -c $(FFLAGS) $(INCFLAGS) $(FPPFLAGS) $(FPPFLAGS_fixed_F)  $<
>> .F90.o:
>>  $(FC) -c $(FFLAGS) $(INCFLAGS) $(FPPFLAGS) $(FPPFLAGS_free_F90) $<
>> .f.o:
>>  $(FC) -c $(FFLAGS) $(INCFLAGS) $(FCFLAGS_fixed_f)  $<
>> .f90.o:
>>  $(FC) -c $(FFLAGS) $(INCFLAGS) $(FCFLAGS_free_f90)  $<
>> **
>>
>> The code compiles without errors.
>> If I run with Diag.ParallelOverK  True, I can run on multiple cores, no
>> errors.
>> With Diag.ParallelOverK  False, I can run `mpirun -np 1` without errors,
>> but if I try to use >=2 cores, it fails with:
>> **
>> Program received signal SIGSEGV: Segmentation fault - invalid memory
>> reference.
>>
>> Backtrace for this error:
>> #0  0x2ba6eb754d1d in ???
>> #1  0x2ba6eb753f7d in ???
>> #2  0x2ba6ec95405f in ???
>> #3  0x2ba70ec1cd8c in ???
>> #4  0x2ba6eab438a4 in ???
>> #5  0x2ba6eab44336 in ???
>> #6  0x563b3f1cfead in __m_diag_MOD_diag_c
>>  at /home/fidanyan/soft/siesta-4.1/Src/diag.F90:709
>> #7  0x563b3f1d2ef9 in cdiag_
>>  at /home/fidanyan/soft/siesta-4.1/Src/diag.F90:2253
>> #8  0x563b3ebc7c8d in diagk_
>>  at /home/fidanyan/soft/siesta-4.1/Src/diagk.F:195
>> #9  0x563b3eb9d714 in __m_diagon_MOD_diagon
>>  at /home/fidanyan/soft/siesta-4.1/Src/diagon.F:265
>> #10  0x563b3ed897cb in __m_compute_dm_MOD_compute_dm
>>  at /home/fidanyan/soft/siesta-4.1/Src/compute_dm.F:172
>> #11  0x563b3edbfaa5 in __m_siesta_forces_MOD_siesta_forces
>>  at /home/fidanyan/soft/siesta-4.1/Src/siesta_forces.F:315
>> #12  0x563b3f9a4005 in siesta
>>  at /home/fidanyan/soft/siesta-4.1/Src/siesta.F:73
>> #13  0x563b3f9a408a in main
>>  at /home/fidanyan/soft/siesta-4.1/Src/siesta.F:10
>> --
>> mpirun noticed that process rank 0 with PID 0 on node fenugreek exited
>> on signal 11 (Segmentation fault).
>> **
>>
>> I ran it by
>> `mpirun -np 2 ~/soft/siesta-4.1/Obj-debug-O0/siesta control.fdf | tee
>> siesta.out`
>>
>> The header of the broken calculation:
>> 
>>
>>
>> Siesta Version  : v4.1.5-1-g384057250
>> Architecture: gfortran_openMPI
>> Compiler version: GNU Fortran (Debian 6.3.0-18+deb9u1) 6.3.0 20170516
>> Compiler flags  : mpifort -O0 -g -fbacktrace -fcheck=all
>> PP flags:  -DFC_HAVE_FLUSH -DFC_HAVE_ABORT -DMPI -DMPI_TIMING
>> 

Re: [SIESTA-L] Compiling and running Siesta with openMPI-2.0.2, gfortran-6.3.0 (Debian 9)

2021-06-29 Por tôpico Nick Papior
I can't rerun without psf files.

Could you try and compile with -Og -g -fbacktrace (without fcheck=all).

Then try again. :)

Den man. 28. jun. 2021 kl. 22.01 skrev Karen Fidanyan <
karen.fidan...@mpsd.mpg.de>:

> Dear Siesta users,
>
> I'm having a hard time trying to run SIESTA on my Debian-9 laptop.
> I have:
>
> GNU Fortran (Debian 6.3.0-18+deb9u1) 6.3.0 20170516
> OpenMPI-2.0.2-2
> libblas 3.7.0-2, liblapack 3.7.0-2
> libscalapack-openmpi1 1.8.0-13
>
> My arch.make is the following:
> **
> .SUFFIXES:
> .SUFFIXES: .f .F .o .a .f90 .F90 .c
>
> SIESTA_ARCH = gfortran_openMPI
>
> FPP = $(FC) -E -P -x c
> FC = mpifort
> FC_SERIAL = gfortran
> FFLAGS = -O0 -g -fbacktrace -fcheck=all #-Wall
> FFLAGS_DEBUG = -g -O0
>
> PP = gcc -E -P -C
> CC = gcc
> CFLAGS = -O0 -g -Wall
>
> AR = ar
> RANLIB = ranlib
> SYS = nag
>
> LDFLAGS = -static-libgcc -ldl
>
> BLASLAPACK_LIBS = -llapack  -lblas \
>  -lscalapack-openmpi -lblacs-openmpi
> -lblacsF77init-openmpi \
>  -lblacsCinit-openmpi \
>  -lpthread -lm
>
> MPI_INTERFACE = libmpi_f90.a
> MPI_INCLUDE   = .
>
> FPPFLAGS_MPI = -DMPI -DMPI_TIMING -D_DIAG_WORK
> FPPFLAGS = $(DEFS_PREFIX) -DFC_HAVE_FLUSH -DFC_HAVE_ABORT $(FPPFLAGS_MPI)
>
> INCFLAGS = $(MPI_INCLUDE)
>
> LIBS = $(BLASLAPACK_LIBS) $(MPI_LIBS)
>
> atom.o: atom.F
>  $(FC) -c $(FFLAGS_DEBUG) $(INCFLAGS) $(FPPFLAGS)
> $(FPPFLAGS_fixed_F) $<
>
>
> .c.o:
>  $(CC) -c $(CFLAGS) $(INCFLAGS) $(CPPFLAGS) $<
> .F.o:
>  $(FC) -c $(FFLAGS) $(INCFLAGS) $(FPPFLAGS) $(FPPFLAGS_fixed_F)  $<
> .F90.o:
>  $(FC) -c $(FFLAGS) $(INCFLAGS) $(FPPFLAGS) $(FPPFLAGS_free_F90) $<
> .f.o:
>  $(FC) -c $(FFLAGS) $(INCFLAGS) $(FCFLAGS_fixed_f)  $<
> .f90.o:
>  $(FC) -c $(FFLAGS) $(INCFLAGS) $(FCFLAGS_free_f90)  $<
> **
>
> The code compiles without errors.
> If I run with Diag.ParallelOverK  True, I can run on multiple cores, no
> errors.
> With Diag.ParallelOverK  False, I can run `mpirun -np 1` without errors,
> but if I try to use >=2 cores, it fails with:
> **
> Program received signal SIGSEGV: Segmentation fault - invalid memory
> reference.
>
> Backtrace for this error:
> #0  0x2ba6eb754d1d in ???
> #1  0x2ba6eb753f7d in ???
> #2  0x2ba6ec95405f in ???
> #3  0x2ba70ec1cd8c in ???
> #4  0x2ba6eab438a4 in ???
> #5  0x2ba6eab44336 in ???
> #6  0x563b3f1cfead in __m_diag_MOD_diag_c
>  at /home/fidanyan/soft/siesta-4.1/Src/diag.F90:709
> #7  0x563b3f1d2ef9 in cdiag_
>  at /home/fidanyan/soft/siesta-4.1/Src/diag.F90:2253
> #8  0x563b3ebc7c8d in diagk_
>  at /home/fidanyan/soft/siesta-4.1/Src/diagk.F:195
> #9  0x563b3eb9d714 in __m_diagon_MOD_diagon
>  at /home/fidanyan/soft/siesta-4.1/Src/diagon.F:265
> #10  0x563b3ed897cb in __m_compute_dm_MOD_compute_dm
>  at /home/fidanyan/soft/siesta-4.1/Src/compute_dm.F:172
> #11  0x563b3edbfaa5 in __m_siesta_forces_MOD_siesta_forces
>  at /home/fidanyan/soft/siesta-4.1/Src/siesta_forces.F:315
> #12  0x563b3f9a4005 in siesta
>  at /home/fidanyan/soft/siesta-4.1/Src/siesta.F:73
> #13  0x563b3f9a408a in main
>  at /home/fidanyan/soft/siesta-4.1/Src/siesta.F:10
> --
> mpirun noticed that process rank 0 with PID 0 on node fenugreek exited
> on signal 11 (Segmentation fault).
> **
>
> I ran it by
> `mpirun -np 2 ~/soft/siesta-4.1/Obj-debug-O0/siesta control.fdf | tee
> siesta.out`
>
> The header of the broken calculation:
> 
>
>
> Siesta Version  : v4.1.5-1-g384057250
> Architecture: gfortran_openMPI
> Compiler version: GNU Fortran (Debian 6.3.0-18+deb9u1) 6.3.0 20170516
> Compiler flags  : mpifort -O0 -g -fbacktrace -fcheck=all
> PP flags:  -DFC_HAVE_FLUSH -DFC_HAVE_ABORT -DMPI -DMPI_TIMING
> -D_DIAG_WORK
> Libraries   :  -llapack -lblas -lscalapack-openmpi -lblacs-openmpi
> -lblacsF77init-openmpi -lblacsCinit-openmpi -lpthread -lm
> PARALLEL version
>
> * Running on 2 nodes in parallel
> 
>
>
>
> I also attach the fdf file and the full output with an error.
> Do you have an idea what is wrong?
>
> Sincerely,
> Karen Fidanyan
> PhD student
> Max Planck Institute for the Structure and Dynamics of Matter
> Hamburg, Germany
>
>
> --
> SIESTA is supported by the Spanish Research Agency (AEI) and by the
> European H2020 MaX Centre of Excellence (http://www.max-centre.eu/)
>


-- 
Kind regards Nick

-- 
SIESTA is supported by the Spanish Research Agency (AEI) and by the European 
H2020 MaX Centre of Excellence (http://www.max-centre.eu/)


[SIESTA-L] Compiling and running Siesta with openMPI-2.0.2, gfortran-6.3.0 (Debian 9)

2021-06-28 Por tôpico Karen Fidanyan

Dear Siesta users,

I'm having a hard time trying to run SIESTA on my Debian-9 laptop.
I have:

GNU Fortran (Debian 6.3.0-18+deb9u1) 6.3.0 20170516
OpenMPI-2.0.2-2
libblas 3.7.0-2, liblapack 3.7.0-2
libscalapack-openmpi1 1.8.0-13

My arch.make is the following:
**
.SUFFIXES:
.SUFFIXES: .f .F .o .a .f90 .F90 .c

SIESTA_ARCH = gfortran_openMPI

FPP = $(FC) -E -P -x c
FC = mpifort
FC_SERIAL = gfortran
FFLAGS = -O0 -g -fbacktrace -fcheck=all #-Wall
FFLAGS_DEBUG = -g -O0

PP = gcc -E -P -C
CC = gcc
CFLAGS = -O0 -g -Wall

AR = ar
RANLIB = ranlib
SYS = nag

LDFLAGS = -static-libgcc -ldl

BLASLAPACK_LIBS = -llapack  -lblas \
    -lscalapack-openmpi -lblacs-openmpi 
-lblacsF77init-openmpi \

    -lblacsCinit-openmpi \
    -lpthread -lm

MPI_INTERFACE = libmpi_f90.a
MPI_INCLUDE   = .

FPPFLAGS_MPI = -DMPI -DMPI_TIMING -D_DIAG_WORK
FPPFLAGS = $(DEFS_PREFIX) -DFC_HAVE_FLUSH -DFC_HAVE_ABORT $(FPPFLAGS_MPI)

INCFLAGS = $(MPI_INCLUDE)

LIBS = $(BLASLAPACK_LIBS) $(MPI_LIBS)

atom.o: atom.F
    $(FC) -c $(FFLAGS_DEBUG) $(INCFLAGS) $(FPPFLAGS) 
$(FPPFLAGS_fixed_F) $<



.c.o:
    $(CC) -c $(CFLAGS) $(INCFLAGS) $(CPPFLAGS) $<
.F.o:
    $(FC) -c $(FFLAGS) $(INCFLAGS) $(FPPFLAGS) $(FPPFLAGS_fixed_F)  $<
.F90.o:
    $(FC) -c $(FFLAGS) $(INCFLAGS) $(FPPFLAGS) $(FPPFLAGS_free_F90) $<
.f.o:
    $(FC) -c $(FFLAGS) $(INCFLAGS) $(FCFLAGS_fixed_f)  $<
.f90.o:
    $(FC) -c $(FFLAGS) $(INCFLAGS) $(FCFLAGS_free_f90)  $<
**

The code compiles without errors.
If I run with Diag.ParallelOverK  True, I can run on multiple cores, no 
errors.
With Diag.ParallelOverK  False, I can run `mpirun -np 1` without errors, 
but if I try to use >=2 cores, it fails with:

**
Program received signal SIGSEGV: Segmentation fault - invalid memory 
reference.


Backtrace for this error:
#0  0x2ba6eb754d1d in ???
#1  0x2ba6eb753f7d in ???
#2  0x2ba6ec95405f in ???
#3  0x2ba70ec1cd8c in ???
#4  0x2ba6eab438a4 in ???
#5  0x2ba6eab44336 in ???
#6  0x563b3f1cfead in __m_diag_MOD_diag_c
    at /home/fidanyan/soft/siesta-4.1/Src/diag.F90:709
#7  0x563b3f1d2ef9 in cdiag_
    at /home/fidanyan/soft/siesta-4.1/Src/diag.F90:2253
#8  0x563b3ebc7c8d in diagk_
    at /home/fidanyan/soft/siesta-4.1/Src/diagk.F:195
#9  0x563b3eb9d714 in __m_diagon_MOD_diagon
    at /home/fidanyan/soft/siesta-4.1/Src/diagon.F:265
#10  0x563b3ed897cb in __m_compute_dm_MOD_compute_dm
    at /home/fidanyan/soft/siesta-4.1/Src/compute_dm.F:172
#11  0x563b3edbfaa5 in __m_siesta_forces_MOD_siesta_forces
    at /home/fidanyan/soft/siesta-4.1/Src/siesta_forces.F:315
#12  0x563b3f9a4005 in siesta
    at /home/fidanyan/soft/siesta-4.1/Src/siesta.F:73
#13  0x563b3f9a408a in main
    at /home/fidanyan/soft/siesta-4.1/Src/siesta.F:10
--
mpirun noticed that process rank 0 with PID 0 on node fenugreek exited 
on signal 11 (Segmentation fault).

**

I ran it by
`mpirun -np 2 ~/soft/siesta-4.1/Obj-debug-O0/siesta control.fdf | tee 
siesta.out`


The header of the broken calculation:
 


Siesta Version  : v4.1.5-1-g384057250
Architecture    : gfortran_openMPI
Compiler version: GNU Fortran (Debian 6.3.0-18+deb9u1) 6.3.0 20170516
Compiler flags  : mpifort -O0 -g -fbacktrace -fcheck=all
PP flags    :  -DFC_HAVE_FLUSH -DFC_HAVE_ABORT -DMPI -DMPI_TIMING 
-D_DIAG_WORK
Libraries   :  -llapack -lblas -lscalapack-openmpi -lblacs-openmpi 
-lblacsF77init-openmpi -lblacsCinit-openmpi -lpthread -lm

PARALLEL version

* Running on 2 nodes in parallel
 



I also attach the fdf file and the full output with an error.
Do you have an idea what is wrong?

Sincerely,
Karen Fidanyan
PhD student
Max Planck Institute for the Structure and Dynamics of Matter
Hamburg, Germany

Siesta Version  : v4.1.5-1-g384057250
Architecture: gfortran_openMPI
Compiler version: GNU Fortran (Debian 6.3.0-18+deb9u1) 6.3.0 20170516
Compiler flags  : mpifort -O0 -g -fbacktrace -fcheck=all
PP flags:  -DFC_HAVE_FLUSH -DFC_HAVE_ABORT -DMPI -DMPI_TIMING 
-D_DIAG_WORK
Libraries   :  -llapack -lblas -lscalapack-openmpi -lblacs-openmpi 
-lblacsF77init-openmpi -lblacsCinit-openmpi -lpthread -lm
PARALLEL version

* Running on 2 nodes in parallel
>> Start of run:  28-JUN-2021  18:42:45

   ***   
   *  WELCOME TO SIESTA  *   
   ***   

reinit: Reading from control.fdf

reinit: ---