Thanks for your answer;

so counting all these sizes (H,S, hpanel, spanel, spanelus...) is good way to 
estimate memory per one thread.


Ad: "This guess indicates that you should be OK, but do your nodes really have 
10Gb/core? That would be unusually large."  Good point, there is some 
restriction, I think 2gb/core. Again, I have to check it with cluster admin.


Best,


Miro


________________________________
From: Wien <wien-boun...@zeus.theochem.tuwien.ac.at> on behalf of Laurence 
Marks <laurence.ma...@gmail.com>
Sent: 04 July 2023 18:17:15
To: A Mailing list for WIEN2k users
Subject: Re: [Wien] xxmr2d:out of memory / estimate memory consumption

If you look at the output you provided, each local copy of H and S is 3Gb. 
Adding another 3Gb for luck suggests that you would need about 360 Gb, assuming 
that you are only running on k-point with 36 cores.

This guess indicates that you should be OK, but do your nodes really have 
10Gb/core? That would be unusually large.

Also, 36 cores is really small, a point that Peter made some time ago. 120-256.
--
Professor Laurence Marks (Laurie)
Department of Materials Science and Engineering, Northwestern University
www.numis.northwestern.edu<http://www.numis.northwestern.edu>
"Research is to see what everybody else has seen, and to think what nobody else 
has thought" Albert Szent-Györgyi

On Tue, Jul 4, 2023, 10:02 Ilias, Miroslav 
<m.il...@gsi.de<mailto:m.il...@gsi.de>> wrote:

Greetings,


I have the big system of 
https://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/msg22609.html .


After fixing the proper openmpi compilation of wien2k I proceed further into 
the lapw1_mpi module. But here I got the error "xxmr2d:out of memory" for 
SBATCH parameters  N=1,  n=36 (that is 36 MPI threads), and --mem=450GB.


Now I am looking for a method for estimating the total memory consumption, 
--mem.


In the  TsOOQg.output1_1 file I see some info about memory spending in 
MB...Would it be please possible to estimate the SBATCH memory need from this 
info ?




MPI-parallel calculation using    36 processors
Scalapack processors array (row,col):   6   6
Matrix size        84052
Optimum Blocksize for setup 124 Excess %  0.428D-01
Optimum Blocksize for diag  30 Excess %  0.143D-01
Base Blocksize   64 Diagonalization   32
         allocate H      2997.6 MB          dimensions 14016 14016
         allocate S      2997.6 MB          dimensions 14016 14016
    allocate spanel        13.7 MB          dimensions 14016    64
    allocate hpanel        13.7 MB          dimensions 14016    64
  allocate spanelus        13.7 MB          dimensions 14016    64
      allocate slen         6.8 MB          dimensions 14016    64
        allocate x2         6.8 MB          dimensions 14016    64
  allocate legendre        89.0 MB          dimensions 14016    13    64
allocate al,bl (row)         4.7 MB          dimensions 14016    11
allocate al,bl (col)         0.0 MB          dimensions    64    11
        allocate YL         3.4 MB          dimensions    15 14016     1
number of local orbitals, nlo (hamilt)     1401
      allocate YL          20.5 MB          dimensions    15 84052     1
      allocate phsc         1.3 MB          dimensions 84052

Time for al,bl    (hamilt, cpu/wall) :        24.49       24.82
Time for legendre (hamilt, cpu/wall) :         5.19        5.21


And lapw1.error  :

**  Error in Parallel LAPW1
**  LAPW1 STOPPED at Tue 04 Jul 2023 10:36:17 AM CEST
**  check ERROR FILES!
Error in LAPW1

Best,


Miro







_______________________________________________
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at<mailto:Wien@zeus.theochem.tuwien.ac.at>
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html
_______________________________________________
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html

Reply via email to