Dear all,
after a very useful email exchange with Jed Brown quite a while ago, I
was able to find the lowest eigenvalue of a large matrix which is
constructed as a tensor product. Admittedly the solution is a bit
hacked, but is based on a Matrix shell and Armadillo and therefore
reasonably fast. The problem seems to work well for smaller systems, but
once the vectors reach a certain size, I get "out of memory" errors. I
have tested the initialization of a vector of that size and
multiplication by the matrix. This works fine and takes roughly 20GB of
memory. There are 256 GB available, so I see no reason why the esp
solvers should complain. Does anyone have an idea what goes wrong here?
The error message is not very helpful and claims that a memory is
requested that is way beyond any reasonable number: /Memory requested
18446744056529684480./
Thanks and all the best, Mathis Friesdorf
*Output of the Program:*
/mathis@n180:~/localisation$ ./local_plus 27/
/System Size: 27
--------------------------------------------------------------------------
[[30558,1],0]: A high-performance Open MPI point-to-point messaging module
was unable to find any relevant network interfaces:
Module: OpenFabrics (openib)
Host: n180
Another transport will be used instead, although this may result in
lower performance.
--------------------------------------------------------------------------
[0]PETSC ERROR: --------------------- Error Message
------------------------------------
[0]PETSC ERROR: Out of memory. This could be due to allocating
[0]PETSC ERROR: too large an object or bleeding by not properly
[0]PETSC ERROR: destroying unneeded objects.
[0]PETSC ERROR: Memory allocated 3221286704 Memory used by process
3229827072
[0]PETSC ERROR: Try running with -malloc_dump or -malloc_log for info.
[0]PETSC ERROR: Memory requested 18446744056529684480!
[0]PETSC ERROR:
------------------------------------------------------------------------
[0]PETSC ERROR: Petsc Release Version 3.4.4, Mar, 13, 2014
[0]PETSC ERROR: See docs/changes/index.html for recent updates.
[0]PETSC ERROR: See docs/faq.html for hints about trouble shooting.
[0]PETSC ERROR: See docs/index.html for manual pages.
[0]PETSC ERROR:
------------------------------------------------------------------------
[0]PETSC ERROR: ./local_plus on a arch-linux2-cxx-debug named n180 by
mathis Wed Jun 25 12:23:01 2014
[0]PETSC ERROR: Libraries linked from /home/mathis/bin_nodebug/lib
[0]PETSC ERROR: Configure run at Wed Jun 25 00:03:34 2014
[0]PETSC ERROR: Configure options PETSC_DIR=/home/mathis/petsc-3.4.4
--with-debugging=1 COPTFLAGS="-O3 -march=p4 -mtune=p4" --with-fortran=0
-with-mpi=1 --with-mpi-dir=/usr/lib/openmpi --with-clanguage=cxx
--prefix=/home/mathis/bin_nodebug
[0]PETSC ERROR:
------------------------------------------------------------------------
[0]PETSC ERROR: PetscMallocAlign() line 46 in
/home/mathis/petsc-3.4.4/src/sys/memory/mal.c
[0]PETSC ERROR: PetscTrMallocDefault() line 189 in
/home/mathis/petsc-3.4.4/src/sys/memory/mtr.c
[0]PETSC ERROR: VecDuplicateVecs_Contiguous() line 62 in
src/vec/contiguous.c
[0]PETSC ERROR: VecDuplicateVecs() line 589 in
/home/mathis/petsc-3.4.4/src/vec/vec/interface/vector.c
[0]PETSC ERROR: EPSAllocateSolution() line 51 in src/eps/interface/mem.c
[0]PETSC ERROR: EPSSetUp_KrylovSchur() line 141 in
src/eps/impls/krylov/krylovschur/krylovschur.c
[0]PETSC ERROR: EPSSetUp() line 147 in src/eps/interface/setup.c
[0]PETSC ERROR: EPSSolve() line 90 in src/eps/interface/solve.c
[0]PETSC ERROR: main() line 48 in local_plus.cpp
--------------------------------------------------------------------------
MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD
with errorcode 55.
NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
You may or may not see output from other processes, depending on
exactly when Open MPI kills them.
--------------------------------------------------------------------------
/
*Output of make:
*/mathis@n180:~/localisation$ make local_plus
mpicxx -o local_plus.o -c -Wall -Wwrite-strings -Wno-strict-aliasing
-Wno-unknown-pragmas -g -fPIC
-I/home/mathis/armadillo-4.300.8/include -lblas -llapack
-L/home/mathis/armadillo-4.300.8 -O3 -larmadillo -fomit-frame-pointer
-I/home/mathis/bin_nodebug/include -I/home/mathis/bin_nodebug/include
-I/usr/lib/openmpi/include -I/usr/lib/openmpi/include/openmpi
-D__INSDIR__= -I/home/mathis/bin_nodebug
-I/home/mathis/bin_nodebug//include -I/home/mathis/bin_nodebug/include
local_plus.cpp
local_plus.cpp:22:0: warning: "__FUNCT__" redefined [enabled by default]
In file included from /home/mathis/bin_nodebug/include/petscvec.h:10:0,
from local_plus.cpp:10:
/home/mathis/bin_nodebug/include/petscviewer.h:386:0: note: this is the
location of the previous definition
g++ -o local_plus local_plus.o -Wl,-rpath,/home/mathis/bin_nodebug//lib
-L/home/mathis/bin_nodebug//lib -lslepc
-Wl,-rpath,/home/mathis/bin_nodebug/lib -L/home/mathis/bin_nodebug/lib
-lpetsc -llapack -lblas -lpthread -Wl,-rpath,/usr/lib/openmpi/lib
-L/usr/lib/openmpi/lib -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.7
-L/usr/lib/gcc/x86_64-linux-gnu/4.7 -Wl,-rpath,/usr/lib/x86_64-linux-gnu
-L/usr/lib/x86_64-linux-gnu -Wl,-rpath,/lib/x86_64-linux-gnu
-L/lib/x86_64-linux-gnu -lmpi_cxx -lstdc++ -ldl -lmpi -lopen-rte
-lopen-pal -lnsl -lutil -lgcc_s -lpthread -ldl
/bin/rm -f local_plus.o
/
*Code:
*///Author: Mathis Friesdorf
//[email protected] <mailto:[email protected]>
static char help[] = "1D chain\n";
#include <iostream>
#include <typeinfo>
#include <armadillo>
#include <petscsys.h>
#include <petscvec.h>
#include <petscmat.h>
#include <slepcsys.h>
#include <slepceps.h>
#include <math.h>
#include <assert.h>
PetscErrorCode BathMult(Mat H, Vec x,Vec y);
PetscInt L=30,d=2,dsys;
PetscErrorCode ierr;
arma::mat hint = "1.0 0 0 0.0; 0 -1.0 2.0 0; 0 2.0 -1.0 0; 0 0 0 1.0;";
#define __FUNCT__ "main"
int main(int argc, char **argv)
{
Mat H;
EPS eps;
Vec xr,xi;
PetscScalar kr,ki;
PetscInt j, nconv;
L = strtol(argv[1],NULL,10);
dsys = pow(d,L);
printf("%s","System Size: ");
printf("%i",L);
printf("%s","\n");
SlepcInitialize(&argc,&argv,(char*)0,help);
MatCreateShell(PETSC_COMM_WORLD,dsys,dsys,dsys,dsys,NULL,&H);
MatShellSetOperation(H,MATOP_MULT,(void(*)())BathMult);
ierr = MatGetVecs(H,NULL,&xr); CHKERRQ(ierr);
ierr = MatGetVecs(H,NULL,&xi); CHKERRQ(ierr);
ierr = EPSCreate(PETSC_COMM_WORLD, &eps); CHKERRQ(ierr);
ierr = EPSSetOperators(eps, H, NULL); CHKERRQ(ierr);
ierr = EPSSetProblemType(eps, EPS_HEP); CHKERRQ(ierr);
ierr = EPSSetWhichEigenpairs(eps,EPS_SMALLEST_REAL); CHKERRQ(ierr);
ierr = EPSSetFromOptions( eps ); CHKERRQ(ierr);
ierr = EPSSolve(eps); CHKERRQ(ierr);
ierr = EPSGetConverged(eps, &nconv); CHKERRQ(ierr);
for (j=0; j<1; j++) {
EPSGetEigenpair(eps, j, &kr, &ki, xr, xi);
printf("%s","Lowest Eigenvalue: ");
PetscPrintf(PETSC_COMM_WORLD,"%9F",kr);
PetscPrintf(PETSC_COMM_WORLD,"\n");
}
EPSDestroy(&eps);
ierr = SlepcFinalize();
return 0;
}
#undef __FUNCT__
#define __FUNCT__ "BathMult"
PetscErrorCode BathMult(Mat H, Vec x, Vec y)
{
PetscInt l;
uint slice;
PetscScalar *arrayin,*arrayout;
VecGetArray(x,&arrayin);
VecGetArray(y,&arrayout);
arma::cube A = arma::cube(arrayin,1,1,pow(d,L),
/*copy_aux_mem*/false,/*strict*/true);
arma::mat result = arma::mat(arrayout,pow(d,L),1,
/*copy_aux_mem*/false,/*strict*/true);
for (l=0;l<L-1;l++){
A.reshape(pow(d,L-2-l),pow(d,2),pow(d,l));
result.reshape(pow(d,L-l),pow(d,l));
for (slice=0;slice<A.n_slices;slice++){
result.col(slice) += vectorise(A.slice(slice)*hint);
}
}
arrayin = A.memptr();
ierr = VecRestoreArray(x,&arrayin); CHKERRQ(ierr);
arrayout = result.memptr();
ierr = VecRestoreArray(y,&arrayout); CHKERRQ(ierr);
PetscFunctionReturn(0);
}
#undef __FUNCT__/*
*