Dear Prof. Peter Blaha and WIEN2k users,

Thank you. Now my .machines file is:

lapw0:e0591:4

1:e0591:4

1:e0591:4

granularity:1

extrafine:1

I have installed WIEN2k in my user in the cluster. When I use this script “srun
--pty /bin/bash” then it goes to one node of the cluster, the “ls -als
$WIENROOT/lapw0”, “x lapw0” and “lapw0 lapw0.def” commands are executed
but, “x lapw0 –p” is not executed. The following error appears:

There are not enough slots available in the system to satisfy the 4

slots that were requested by the application:

  /home/users/mollabashi/v19.2/lapw0_mpi

Either request fewer slots for your application, or make more slots

available for use.

A "slot" is the Open MPI term for an allocatable unit where we can

launch a process.  The number of slots available are defined by the

environment in which Open MPI processes are run:

  1. Hostfile, via "slots=N" clauses (N defaults to number of

     processor cores if not provided)

  2. The --host command line parameter, via a ":N" suffix on the

     hostname (N defaults to 1 if not provided)

  3. Resource manager (e.g., SLURM, PBS/Torque, LSF, etc.)

  4. If none of a hostfile, the --host command line parameter, or an

     RM is present, Open MPI defaults to the number of processor cores

In all the above cases, if you want Open MPI to default to the number

of hardware threads instead of the number of processor cores, use the

--use-hwthread-cpus option.

Alternatively, you can use the --oversubscribe option to ignore the

number of available slots when deciding the number of processes to

launch.

--------------------------------------------------------------------------

[1]    Exit 1                        mpirun -np 4 -machinefile .machine0
/home/users/mollabashi/v19.2/lapw0_mpi lapw0.def >> .time00

0.067u 0.091s 0:02.97 5.0%      0+0k 52272+144io 54pf+0w

mollabashi@eagle:~/test1/cein$ cat .machines

Sincerely yours,

Leila Mollabashi

On Sun, Apr 11, 2021 at 9:40 PM Peter Blaha <pbl...@theochem.tuwien.ac.at>
wrote:

> Your script is still wrong.
> The .machines file should show:
>
> lapw0:e0150:4
>
> not
> lapw0:e0150
> :4
>
> Therefore it tries to execute lapw0 instead of lapw0_mpi.
> -----------
> Anyway, the first thing is to make the sequential wien2k running. You
> claimed the WIENROOT is known in the batch job.
> Please do:
> ls -als $WIENROOT/lapw0
>
> Does it have execute permission ?
>
> If yes, execute lapw0 explicitly:
>
> x lapw0
>
> and a second time:
>
> lapw0 lapw0.def
>
>
> Am 11.04.2021 um 13:17 schrieb leila mollabashi:
> > Dear Prof. Peter Blaha,
> >
> > Thank you for your guides. You are right. I edited the script and added
> > “source ~/.bashrc, echo 'lapw0:'`hostname`' :'$nproc >> .machines” to it.
> >
> > The crated .machines file is as follows:
> >
> > lapw0:e0150
> >
> > :4
> >
> > 1:e0150:4
> >
> > 1:e0150:4
> >
> > granularity:1
> >
> > extrafine:1
> >
> > The slurm.out file is:
> >
> > e0150
> >
> > # .machines
> >
> > bash: lapw0: command not found
> >
> > real 0m0.001s
> >
> > user 0m0.001s
> >
> > sys 0m0.000s
> >
> > grep: *scf1*: No such file or directory
> >
> > grep: lapw2*.error: No such file or directory
> >
> >>  stop error
> >
> > When I used the following commands:
> >
> > echo $WIENROOT
> > which lapw0
> > which lapw0_mpi
> >
> > The following paths were printed:
> >
> > /home/users/mollabashi/v19.2
> >
> > /home/users/mollabashi/v19.2/lapw0
> >
> > /home/users/mollabashi/v19.2/lapw0_mpi
> >
> > But the error is still exists:
> >
> > bash: lapw0: command not found
> >
> > When I used your script in (faq page), one time the .machines file was
> > generated.
> >
> > But it stopped due to an error.
> >
> > test.scf1_1: No such file or directory.
> >
> > grep: *scf1*: No such file or directory
> >
> > FERMI - Error
> >
> > When I loaded openmpi and ifort as well as icc in the script this error
> > appeared:
> >
> >>SLURM_NTASKS_PER_NODE:  Undefined variable.
> >
> > Every time after that the
> >
> >>SLURM_NTASKS_PER_NODE:  Undefined variable
> >
> >   error happened when I used your scripts without changing it. I have
> > tried several times even in a new directory with no positive effect.
> >
> >>SLURM_NTASKS_PER_NODE:  Undefined variable.
> >
> > Sincerely yours,
> >
> > Leila Mollabashi
> >
> >
> > _______________________________________________
> > Wien mailing list
> > Wien@zeus.theochem.tuwien.ac.at
> > http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
> > SEARCH the MAILING-LIST at:
> http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html
> >
>
> --
> --------------------------------------------------------------------------
> Peter BLAHA, Inst.f. Materials Chemistry, TU Vienna, A-1060 Vienna
> Phone: +43-1-58801-165300             FAX: +43-1-58801-165982
> Email: bl...@theochem.tuwien.ac.at    WIEN2k: http://www.wien2k.at
> WWW:   http://www.imc.tuwien.ac.at
> -------------------------------------------------------------------------
> _______________________________________________
> Wien mailing list
> Wien@zeus.theochem.tuwien.ac.at
> http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
> SEARCH the MAILING-LIST at:
> http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html
>
_______________________________________________
Wien mailing list
Wien@zeus.theochem.tuwien.ac.at
http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
SEARCH the MAILING-LIST at:  
http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html

Reply via email to