This problem is easily solvable (it again means: You MUST READ the UG (parallelization), otherwise you will run VERY badly).

For such a small problem (14 atoms, matrixsize 2600) it is NOT necessary (in fact probably even quite bad) to use mpi-parallelization.


Instead use k-parallelization (and maybe   export OMP_NUM_THREAD=2).

simply put eg. 24 lines like:

1:localhost

into the .machines file, and you will run 24 parallel lapw1 (each using 2 cores when OMP_NUM_THREAD=2 is set).

--------------
With respect to your other questions:
I don't know what:  lapw1para_mpi -p -band is ??

lapw1 should be always invoked using:

x lapw1 -p      or    x lapw1 -p -band

The difference is just that you are using either case.klist or case.klist_band. Checkout how many k-points are in these 2 files (250 was just an "input", it seems to have made a 13x13x1 mesh and then still applied symmetry, so you may have just ~ 30 k-points in case.klist ...)

-----------------
Another question: do you have 48 "physical cores", or only 24 ???
Do you have 2 or 4 Xeons (with 12 cores each) in your computer ??

If you have only 24 "real" cores:
The "virtual cores" which Intel gives you "for free" due to their "hyperthreading", are usually not very effective. You can at most expect an improvement of 10-20% when using 48 instead of 24 cores, but sometimes, this can also degrade performance by 30% because the memorybus gets overloaded. So test it ....



On 11/29/18 1:10 PM, Coriolan TIUSAN wrote:
Thanks for the suggestion of dividing the band calculation.

Actually, I would like to make a 'zoom' around the Gamma point (for X-G-X direction) with a resolution of about 0.001 Bohr-1 (to get enough accuracy for small Rasba splittings, k_0< 0.01 Bohr-1). I guess I could simply make the 'zoom' calculation?

The .machines, file, having in view that I have only one node (computer) with 48 available CPUs is:

-------------------------------------

1:localhost:48
granularity:1
extrafine:1
lapw0:localhost:48
dstart:localhost:48
nlvdw:localhost:48

--------------------------------------

For a supercell here attached, I was trying to make a bandstructure calculations along the X-G-X direction with at least 200 points....which corresponds to a step of only 0.005 Bohr-1, not enough for Rashba in same order of magnitude.

For my calculations I get: MATRIX SIZE  2606LOs: 138  RKM= 6.99 and the RAM of 64Gk is 100% filles plus about 100G of swap...

Beyond all aspects, what I would like to understand is also why in scf calculation I have no memory 'overload'  FOR 250K POINTS (13 13 1)... while when running  'lapw1para_mpi -p -band ' the memory issue seem more dramatic?

If necessary, my struct file is:

------------------

VFeMgO-vid                               s-o calc. M||  1.00 0.00  0.00
P 14
RELA
   5.725872  5.725872 61.131153 90.000000 90.000000 90.000000
ATOM  -1: X=0.50000000 Y=0.50000000 Z=0.01215444
           MULT= 1          ISPLIT= 8
V 1        NPT=  781  R0=.000050000 RMT=   2.18000   Z: 23.00000
LOCAL ROT MATRIX:    1.0000000 0.0000000 0.0000000
                      0.0000000 1.0000000 0.0000000
                      0.0000000 0.0000000 1.0000000
ATOM  -2: X=0.00000000 Y=0.00000000 Z=0.05174176
           MULT= 1          ISPLIT= 8
V 2        NPT=  781  R0=.000050000 RMT=   2.18000   Z: 23.00000
LOCAL ROT MATRIX:    1.0000000 0.0000000 0.0000000
                      0.0000000 1.0000000 0.0000000
                      0.0000000 0.0000000 1.0000000
ATOM  -3: X=0.50000000 Y=0.50000000 Z=0.09885823
           MULT= 1          ISPLIT= 8
V 3        NPT=  781  R0=.000050000 RMT=   2.18000   Z: 23.00000
LOCAL ROT MATRIX:    1.0000000 0.0000000 0.0000000
                      0.0000000 1.0000000 0.0000000
                      0.0000000 0.0000000 1.0000000
ATOM  -4: X=0.00000000 Y=0.00000000 Z=0.13971867
           MULT= 1          ISPLIT= 8
Fe1        NPT=  781  R0=.000050000 RMT=   1.95000   Z: 26.00000
LOCAL ROT MATRIX:    1.0000000 0.0000000 0.0000000
                      0.0000000 1.0000000 0.0000000
                      0.0000000 0.0000000 1.0000000
ATOM  -5: X=0.50000000 Y=0.50000000 Z=0.18164479
           MULT= 1          ISPLIT= 8
Fe2        NPT=  781  R0=.000050000 RMT=   1.95000   Z: 26.00000
LOCAL ROT MATRIX:    1.0000000 0.0000000 0.0000000
                      0.0000000 1.0000000 0.0000000
                      0.0000000 0.0000000 1.0000000
ATOM  -6: X=0.00000000 Y=0.00000000 Z=0.22284885
           MULT= 1          ISPLIT= 8
Fe3        NPT=  781  R0=.000050000 RMT=   1.95000   Z: 26.00000
LOCAL ROT MATRIX:    1.0000000 0.0000000 0.0000000
                      0.0000000 1.0000000 0.0000000
                      0.0000000 0.0000000 1.0000000
ATOM  -7: X=0.50000000 Y=0.50000000 Z=0.26533335
           MULT= 1          ISPLIT= 8
Fe4        NPT=  781  R0=.000050000 RMT=   1.95000   Z: 26.00000
LOCAL ROT MATRIX:    1.0000000 0.0000000 0.0000000
                      0.0000000 1.0000000 0.0000000
                      0.0000000 0.0000000 1.0000000
ATOM  -8: X=0.00000000 Y=0.00000000 Z=0.30245527
           MULT= 1          ISPLIT= 8
Fe5        NPT=  781  R0=.000050000 RMT=   1.95000   Z: 26.00000
LOCAL ROT MATRIX:    1.0000000 0.0000000 0.0000000
                      0.0000000 1.0000000 0.0000000
                      0.0000000 0.0000000 1.0000000
ATOM  -9: X=0.00000000 Y=0.00000000 Z=0.36627712
           MULT= 1          ISPLIT= 8
O 1        NPT=  781  R0=.000100000 RMT=   1.68000   Z: 8.00000
LOCAL ROT MATRIX:    1.0000000 0.0000000 0.0000000
                      0.0000000 1.0000000 0.0000000
                      0.0000000 0.0000000 1.0000000
ATOM -10: X=0.50000000 Y=0.50000000 Z=0.36416415
           MULT= 1          ISPLIT= 8
Mg1        NPT=  781  R0=.000100000 RMT=   1.87000   Z: 12.00000
LOCAL ROT MATRIX:    1.0000000 0.0000000 0.0000000
                      0.0000000 1.0000000 0.0000000
                      0.0000000 0.0000000 1.0000000
ATOM -11: X=0.50000000 Y=0.50000000 Z=0.43034285
           MULT= 1          ISPLIT= 8
O 2        NPT=  781  R0=.000100000 RMT=   1.68000   Z: 8.00000
LOCAL ROT MATRIX:    1.0000000 0.0000000 0.0000000
                      0.0000000 1.0000000 0.0000000
                      0.0000000 0.0000000 1.0000000
ATOM -12: X=0.00000000 Y=0.00000000 Z=0.43127365
           MULT= 1          ISPLIT= 8
Mg2        NPT=  781  R0=.000100000 RMT=   1.87000   Z: 12.00000
LOCAL ROT MATRIX:    1.0000000 0.0000000 0.0000000
                      0.0000000 1.0000000 0.0000000
                      0.0000000 0.0000000 1.0000000
ATOM -13: X=0.00000000 Y=0.00000000 Z=0.49684798
           MULT= 1          ISPLIT= 8
O 3        NPT=  781  R0=.000100000 RMT=   1.68000   Z: 8.00000
LOCAL ROT MATRIX:    1.0000000 0.0000000 0.0000000
                      0.0000000 1.0000000 0.0000000
                      0.0000000 0.0000000 1.0000000
ATOM -14: X=0.50000000 Y=0.50000000 Z=0.49541730
           MULT= 1          ISPLIT= 8
Mg3        NPT=  781  R0=.000100000 RMT=   1.87000   Z: 12.00000
LOCAL ROT MATRIX:    1.0000000 0.0000000 0.0000000
                      0.0000000 1.0000000 0.0000000
                      0.0000000 0.0000000 1.0000000
    4      NUMBER OF SYMMETRY OPERATIONS
-1 0 0 0.00000000
  0 1 0 0.00000000
  0 0 1 0.00000000
        1   A   1 so. oper.  type  orig. index
  1 0 0 0.00000000
  0 1 0 0.00000000
  0 0 1 0.00000000
        2   A   2
-1 0 0 0.00000000
  0-1 0 0.00000000
  0 0 1 0.00000000
        3   B   3
  1 0 0 0.00000000
  0-1 0 0.00000000
  0 0 1 0.00000000
        4   B   4
---------------------------


La 29/11/2018 13:05, Peter Blaha a scris:
You never listed your .machines file, nor do we know how many k-points are in the scf and the bandstructure cases and what the matrix size(:RKM)/ real/ complex details are.

The memory leakage of intels mpi seems to be very version dependent, but there's nothing we can do against from the wien2k side.

Besides installing a different mpi version, one could more easily run the bandstructure in pieces. Simply divide your klist_band file into several pieces and calculate one after the other.

The resulting case.outputso_1,2,3.. files can simply be concatenated (cat file1 file2 file3 > file) together.



On 11/28/18 1:41 PM, Coriolan TIUSAN wrote:
Dear wien2k users,

I am running wien 18.2 on Ubuntu 18.04 , installed on a HP station: 64GB,  Intel® Xeon(R) Gold 5118 CPU @ 2.30GHz × 48.

The fortran compiler/math library are ifc and intel mkl library. For parallel execution I have MPI+SCALAPACK, FFTW.

For parallel execution  (-p options +.machines), I have dimensioned NMATMAX/NUME according to user guide. Therefore, standard calculations in SCF loops turn well, without any memory paging issues, about 90% of physical RAM being used.

However, in supercells, once getting case.vector files,  when calculating bands (lapw1c -bands  -p) with fine k structure (e.g. above 150-200k on line X-G-X), necessary because I am looking to small Rashba shifts at metel-insulator interfaces...all available physical memory plus a huge amount of swap (>100G) are filled/used...

Any suggestion/ideea for overcoming this issue...without adding additional RAM?

Why in lapw1 -p for selfonsistance memory looks enough while with switch -band overload memory?

With thanks in advance,

C. Tiusan



_______________________________________________
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at: http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html


--

                                      P.Blaha
--------------------------------------------------------------------------
Peter BLAHA, Inst.f. Materials Chemistry, TU Vienna, A-1060 Vienna
Phone: +43-1-58801-165300             FAX: +43-1-58801-165982
Email: bl...@theochem.tuwien.ac.at    WIEN2k: http://www.wien2k.at
WWW:   http://www.imc.tuwien.ac.at/TC_Blaha
--------------------------------------------------------------------------
_______________________________________________
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html

Reply via email to