[QE-users] Poor GPU scaling for Gamma-point-only calculation on multiple GPUs
Wang Xing
xing.wang at psi.ch
Tue May 13 08:47:24 CEST 2025
Hi all,
I’m running a Gamma-point-only SCF calculation for a nanoparticle system using Quantum ESPRESSO with GPU support. The HPC node has 4 NVIDIA GH200 GPUs, and I observe very poor scaling behavior when increasing the number of GPUs.
Setup
I use 1 MPI task per GPU, with 72 OpenMP threads per task:
*
OMP_NUM_THREADS=$SLURM_CPUS_PER_TASK
*
QE version: 7.4 (GPU build with OpenMP)
*
MPI: OpenMPI (GPU-aware)
Here is the relevant SLURM configuration:
#SBATCH --nodes=1
#SBATCH --ntasks-per-node=4
#SBATCH --cpus-per-task=72
#SBATCH --gpus=4
#SBATCH --gpu-bind=map_gpu:0,1,2,3
#SBATCH --time=00:30:00
module use Spack
module load nvhpc/24.11 openmpi/main-7zgw-GH200-gpu quantum-espresso/7.4-gpu-omp
srun pw.x -pd .true. -npool 1 -in aiida.in > aiida.out
Performance Results (time per SCF iteration)
Configuration
Time (sec)
1 task, 1 GPU
11.9
2 tasks, 2 GPUs
17.0
3 tasks, 3 GPUs
15.3
4 tasks, 4 GPUs
9.7
As you can see, the best performance is with just 1 task + 1 GPU. Increasing the number of GPUs initially worsens the performance, and only with 4 GPUs does it slightly improve.
>From the output:
GPU acceleration is ACTIVE. 1 visible GPUs per MPI rank
GPU-aware MPI enabled
Message from routine print_cuda_info:
High GPU oversubscription detected. Are you sure this is what you want?
I tried --gpu-bind=map_gpu:0,1,2,3 to explicitly bind GPUs to ranks, but the warning still appears, and performance doesn’t change. I’ve also experimented with -nb and -ndiag parameters, but they either don’t help or make things worse.
Question:
Does anyone have experience optimizing Gamma-point-only calculations on multiple GPUs? Is there a known bottleneck or best practice for using multiple GPUs efficiently in such a case?
Any insights would be greatly appreciated.
Best,
Xing
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.quantum-espresso.org/pipermail/users/attachments/20250513/d759aaf7/attachment.html>
More information about the users
mailing list