[QE-users] Projected band structure

2023-11-30 Thread Mina Hemati via users
Dear all I would like to calculate projected band structure. could you please 
guide me regarding this?
Best regardsMinaPHD of solid state physicsIran University of science and 
technology. ___
The Quantum ESPRESSO community stands by the Ukrainian
people and expresses its concerns about the devastating
effects that the Russian military offensive has on their
country and on the free and peaceful scientific, cultural,
and economic cooperation amongst peoples
___
Quantum ESPRESSO is supported by MaX (www.max-centre.eu)
users mailing list users@lists.quantum-espresso.org
https://lists.quantum-espresso.org/mailman/listinfo/users

Re: [QE-users] [QE-GPU] High GPU oversubscription detected

2023-11-30 Thread Paolo Giannozzi

On 11/30/23 10:24, Yin-Ying Ting wrote:

Could you please provide guidance on resolving the oversubscription 
issue? 


if your job runs on 4 GPUs while pretending to run on one, there is no 
real issue: just ignore the message. If you instruct your job to run on 
4 GPUs but it doesn't, you should report it to your system administrator.


Paolo






Kind regards,

Yin-Ying Ting


On 29.11.23 15:53, Paolo Giannozzi wrote:

On 11/27/23 11:32, Yin-Ying Ting wrote:

Based on the *environment.f90* file, this message is triggered when 
/nproc > ndev * nnode * 2/. If I understand correctly, I have nproc 
(Number of parallel processe)=4, ndev(Number of GPU Devices per Node) 
=4 and nnode (Number of Nodes)=1. This condition seems to be false (4 
> 8). Despite this, the message still appears. All 4 GPUs were active 
during the run.


funny. Even funnier, the number of GPUs actually used does not seem to 
be written anywhere on output.


Add a line printing nproc, ndev, nnode just before the warning is 
issued, recompile and re-run. One (at least) of those numbers is not 
what you expect. Computers are not among the most reliable machines, 
but they should be able to find out who is larger between 4 and 8


Paolo

--

Forschungszentrum Jülich GmbH
Institute of Energy and Climate Research
Theory and Computation of Energy Materials (IEK-13)
E-mail: y.t...@fz-juelich.de





Forschungszentrum Jülich GmbH
52425 Jülich
Sitz der Gesellschaft: Jülich
Eingetragen im Handelsregister des Amtsgerichts Düren Nr. HR B 3498
Vorsitzender des Aufsichtsrats: MinDir Stefan Müller
Geschäftsführung: Prof. Dr. Astrid Lambrecht (Vorsitzende),
Karsten Beneke (stellv. Vorsitzender), Dr. Ir. Pieter Jansens




--
Paolo Giannozzi, DMIF, Univ. Udine, Italy
*** AVAILABLE POST-DOC POSITION:
*** https://physicslab.uniud.it/persone/paolo-giannozzi/advert
___
The Quantum ESPRESSO community stands by the Ukrainian
people and expresses its concerns about the devastating
effects that the Russian military offensive has on their
country and on the free and peaceful scientific, cultural,
and economic cooperation amongst peoples
___
Quantum ESPRESSO is supported by MaX (www.max-centre.eu)
users mailing list users@lists.quantum-espresso.org
https://lists.quantum-espresso.org/mailman/listinfo/users

Re: [QE-users] [QE-GPU] High GPU oversubscription detected

2023-11-30 Thread Yin-Ying Ting

Dear Paolo,

Thank you for your prompt response. Your suggestion was very helpful.

I have reviewed the numbers and discovered that regardless of the value set in 
--gres=gpu:X, the number of ndev remains consistently at 1. Our HPC 
documentation indicates that --gres=gpu:X is the correct method to set GPUs, 
with each node having 4 GPUs. Here is the output when I set --gres=gpu:4:

GPU acceleration is ACTIVE.
GPU-aware MPI enabled

nproc (MPI process): 4
ndev (GPU per Node): 1
nnode (Nodes):   1
Message from routine print_cuda_info:
High GPU oversubscription detected. Are you sure this is what you want?

I monitored GPU core usage every 10 seconds, it appears that all 4 GPU cores 
are activated when setting --gres=gpu:4:

utilization.gpu [%], utilization.memory [%]
96 %, 37 %
95 %, 76 %
95 %, 50 %
95 %, 76 %
time = 70 s

For reference, here is my sbatch submission script:

---

#!/bin/bash -x
#SBATCH --gres=gpu:4 --partition=dc-gpu
#SBATCH --nodes=1
#SBATCH --ntasks-per-node=4
#SBATCH --time=00:00:20

export OMP_NUM_THREADS=1

module load NVHPC/23.7-CUDA-12
module load CUDA/12
module load OpenMPI/4.1.5
module load mpi-settings/CUDA
module load imkl/2023.2.0

monitor_gpu_usage() {
   while true; do
   nvidia-smi --query-gpu=utilization.gpu,utilization.memory --format=csv 
>> gpu_usage_$SLURM_JOB_ID.csv
   sleep 10
   done
}
monitor_gpu_usage &
srun -n 4 pw.x -nk 4 -nd 1 -nb 1 -nt 1 < inp_pwscf > out_pwscf

-


Could you please provide guidance on resolving the oversubscription issue? 
Thank you very much in advance.

Kind regards,

Yin-Ying Ting


On 29.11.23 15:53, Paolo Giannozzi wrote:
On 11/27/23 11:32, Yin-Ying Ting wrote:

Based on the *environment.f90* file, this message is triggered when /nproc > ndev 
* nnode * 2/. If I understand correctly, I have nproc (Number of parallel 
processe)=4, ndev(Number of GPU Devices per Node) =4 and nnode (Number of Nodes)=1. 
This condition seems to be false (4 > 8). Despite this, the message still appears. 
All 4 GPUs were active during the run.

funny. Even funnier, the number of GPUs actually used does not seem to be 
written anywhere on output.

Add a line printing nproc, ndev, nnode just before the warning is issued, 
recompile and re-run. One (at least) of those numbers is not what you expect. 
Computers are not among the most reliable machines, but they should be able to 
find out who is larger between 4 and 8

Paolo
--

Forschungszentrum Jülich GmbH
Institute of Energy and Climate Research
Theory and Computation of Energy Materials (IEK-13)
E-mail: y.t...@fz-juelich.de




Forschungszentrum Jülich GmbH
52425 Jülich
Sitz der Gesellschaft: Jülich
Eingetragen im Handelsregister des Amtsgerichts Düren Nr. HR B 3498
Vorsitzender des Aufsichtsrats: MinDir Stefan Müller
Geschäftsführung: Prof. Dr. Astrid Lambrecht (Vorsitzende),
Karsten Beneke (stellv. Vorsitzender), Dr. Ir. Pieter Jansens


___
The Quantum ESPRESSO community stands by the Ukrainian
people and expresses its concerns about the devastating
effects that the Russian military offensive has on their
country and on the free and peaceful scientific, cultural,
and economic cooperation amongst peoples
___
Quantum ESPRESSO is supported by MaX (www.max-centre.eu)
users mailing list users@lists.quantum-espresso.org
https://lists.quantum-espresso.org/mailman/listinfo/users