[QE-users] [ge-GPU] compiling q-e-gpu-qe-gpu-6.7
Pietro Bonfa'
pietro.bonfa at unipr.it
Wed Dec 16 14:37:35 CET 2020
Dear Mohammad,
CUDA may be installed somewhere else, anyway, if you want to skip the
environment check, you may configure QE with this command
./configure FC=pgf90 F90=pgf90 CC=pgcc --with-cuda=yes
--enable-cuda-env-check=no --with-cuda-runtime=SETME --with-cuda-cc=70
--enable-openmp
Remember to set the cuda runtime according to what is provided by your
setup.
Hope this helps,
best,
Pietro
---
Pietro Bonfà
University of Parma
On 12/16/20 9:10 AM, Mohammad Moaddeli wrote:
> Dear all,
> I am trying to compile the 6.7 version of the code using PGI 2020.
> I followed these steps:
>
> *1) NVIDIA driver (NVIDIA-Linux-x86_64-450.80.02.rpm) is installed.*
> *the output of nvidia-smi:*
>
> Wed Dec 16 09:07:11 2020
> +-----------------------------------------------------------------------------+
> | NVIDIA-SMI 450.80.02 Driver Version: 450.80.02 CUDA Version:
> 11.0 |
> |-------------------------------+----------------------+----------------------+
> | GPU Name Persistence-M| Bus-Id Disp.A | Volatile
> Uncorr. ECC |
> | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util
> Compute M. |
> | | |
> MIG M. |
> |===============================+======================+======================|
> | 0 TITAN V Off | 00000000:06:00.0 Off |
> N/A |
> | 27% 37C P0 32W / 250W | 0MiB / 12066MiB | 0%
> Default |
> | | |
> N/A |
> +-------------------------------+----------------------+----------------------+
> | 1 TITAN V Off | 00000000:07:00.0 Off |
> N/A |
> | 25% 37C P0 35W / 250W | 0MiB / 12066MiB | 0%
> Default |
> | | |
> N/A |
> +-------------------------------+----------------------+----------------------+
>
> +-----------------------------------------------------------------------------+
> | Processes:
> |
> | GPU GI CI PID Type Process name GPU
> Memory |
> | ID ID Usage
> |
> |=============================================================================|
> | No running processes found
> |
> +-----------------------------------------------------------------------------+
>
> *The output of pgaccelinfo:*
>
> CUDA Driver Version: 11000
> NVRM version: NVIDIA UNIX x86_64 Kernel Module
> 450.80.02 Wed Sep 23 01:13:39 UTC 2020
>
> Device Number: 0
> Device Name: TITAN V
> Device Revision Number: 7.0
> Global Memory Size: 12652838912
> Number of Multiprocessors: 80
> Concurrent Copy and Execution: Yes
> Total Constant Memory: 65536
> Total Shared Memory per Block: 49152
> Registers per Block: 65536
> Warp Size: 32
> Maximum Threads per Block: 1024
> Maximum Block Dimensions: 1024, 1024, 64
> Maximum Grid Dimensions: 2147483647 x 65535 x 65535
> Maximum Memory Pitch: 2147483647B
> Texture Alignment: 512B
> Clock Rate: 1455 MHz
> Execution Timeout: No
> Integrated Device: No
> Can Map Host Memory: Yes
> Compute Mode: default
> Concurrent Kernels: Yes
> ECC Enabled: No
> Memory Clock Rate: 850 MHz
> Memory Bus Width: 3072 bits
> L2 Cache Size: 4718592 bytes
> Max Threads Per SMP: 2048
> Async Engines: 7
> Unified Addressing: Yes
> Managed Memory: Yes
> Concurrent Managed Memory: Yes
> Preemption Supported: Yes
> Cooperative Launch: Yes
> Multi-Device: Yes
> Default Target: cc70
>
> Device Number: 1
> Device Name: TITAN V
> Device Revision Number: 7.0
> Global Memory Size: 12652838912
> Number of Multiprocessors: 80
> Concurrent Copy and Execution: Yes
> Total Constant Memory: 65536
> Total Shared Memory per Block: 49152
> Registers per Block: 65536
> Warp Size: 32
> Maximum Threads per Block: 1024
> Maximum Block Dimensions: 1024, 1024, 64
> Maximum Grid Dimensions: 2147483647 x 65535 x 65535
> Maximum Memory Pitch: 2147483647B
> Texture Alignment: 512B
> Clock Rate: 1455 MHz
> Execution Timeout: No
> Integrated Device: No
> Can Map Host Memory: Yes
> Compute Mode: default
> Concurrent Kernels: Yes
> ECC Enabled: No
> Memory Clock Rate: 850 MHz
> Memory Bus Width: 3072 bits
> L2 Cache Size: 4718592 bytes
> Max Threads Per SMP: 2048
> Async Engines: 7
> Unified Addressing: Yes
> Managed Memory: Yes
> Concurrent Managed Memory: Yes
> Preemption Supported: Yes
> Cooperative Launch: Yes
> Multi-Device: Yes
> Default Target: cc70
>
> *2) PGI compiler is installed:*
> *yum install nvhpc-20-11-20.11-1.x86_64.rpm nvhpc-2020-20.11-1.x86_64.rpm
> *
> *PATHs that are set in ~/.bashrc file:
> *
> *
> *
> export PATH=/opt/nvidia/hpc_sdk/Linux_x86_64/20.11/cuda/11.1/bin:$PATH
> export PATH=/opt/nvidia/hpc_sdk/Linux_x86_64/20.11/cuda/11.1/include:$PATH
> export
> LD_LIBRARY_PATH=/opt/nvidia/hpc_sdk/Linux_x86_64/20.11/cuda/11.1/lib64:$LD_LIBRARY_PATH
> export
> LD_LIBRARY_PATH=/opt/nvidia/hpc_sdk/Linux_x86_64/20.11/cuda/11.1/extras/CUPTI/lib64:$LD_LIBRARY_PATH
> export
> LD_LIBRARY_PATH=/opt/nvidia/hpc_sdk/Linux_x86_64/20.11/cuda/11.1/lib64/stubs:$LD_LIBRARY_PATH
> NVARCH=`uname -s`_`uname -m`; export NVARCH
> NVCOMPILERS=/opt/nvidia/hpc_sdk; export NVCOMPILERS
> MANPATH=$MANPATH:$NVCOMPILERS/$NVARCH/20.11/compilers/man; export MANPATH
> PATH=$NVCOMPILERS/$NVARCH/20.11/compilers/bin:$PATH; export PATH
> PATH=$NVCOMPILERS/$NVARCH/20.11/compilers/include:$PATH; export PATH
> LD_LIBRARY_PATH=$NVCOMPILERS/$NVARCH/20.11/compilers/lib:$PATH; export
> LD_LIBRARY_PATH
> export PATH=/opt/nvidia/hpc_sdk/Linux_x86_64/20.11/comm_libs/mpi/bin:$PATH
> export
> PATH=/opt/nvidia/hpc_sdk/Linux_x86_64/20.11/comm_libs/mpi/include:$PATH
> export
> LD_LIBRARY_PATH=/opt/nvidia/hpc_sdk/Linux_x86_64/20.11/comm_libs/mpi/lib:$LD_LIBRARY_PATH
> export
> LD_LIBRARY_PATH=/opt/nvidia/hpc_sdk/Linux_x86_64/20.11/math_libs/11.1/lib64:$LD_LIBRARY_PATH
> export
> LD_LIBRARY_PATH=/opt/nvidia/hpc_sdk/Linux_x86_64/20.11/math_libs/11.1/lib64/stubs:$LD_LIBRARY_PATH*
> *
> *
> *
> *3) compiling the code using:*
> *./configure FC=pgf90 F90=pgf90 F77=pgf90 CC=pgcc MPIF90=mpif90
> --with-cuda=/opt/nvidia/hpc_sdk/Linux_x86_64/20.11/cuda
> --with-cuda-runtime=11.1 --with-cuda-cc=70 --enable-openmp
> --with-scalapack=no*
>
> checking build system type... x86_64-pc-linux-gnu
> checking ARCH... x86_64
> checking setting AR... ... ar
> checking setting ARFLAGS... ... ruv
> checking whether the Fortran compiler works... yes
> checking for Fortran compiler default output file name... a.out
> checking for suffix of executables...
> checking whether we are cross compiling... no
> checking for suffix of object files... o
> checking whether we are using the GNU Fortran compiler... no
> checking whether pgf90 accepts -g... yes
> configure: WARNING: F90 value is set to be consistent with value of MPIF90
> checking for mpif90... mpif90
> checking whether we are using the GNU Fortran compiler... no
> checking whether mpif90 accepts -g... yes
> checking version of mpif90... nvfortran 20.11-0
> checking for Fortran flag to compile .f90 files... none
> setting F90... nvfortran
> setting MPIF90... mpif90
> checking whether we are using the GNU C compiler... yes
> checking whether pgcc accepts -g... yes
> checking for pgcc option to accept ISO C89... none needed
> setting CC... pgcc
> setting CFLAGS... -fast -Mpreprocess
> using F90... nvfortran
> setting FFLAGS... -O1
> setting F90FLAGS... $(FFLAGS)
> setting FFLAGS_NOOPT... -O0
> setting CPP... cpp
> setting CPPFLAGS... -P -traditional -Uvector
> setting LD... mpif90
> setting LDFLAGS...
> checking for Fortran flag to compile .f90 files... (cached) none
> checking whether Fortran compiler accepts -Mcuda=cuda11.1... yes
> checking for nvcc...
> /opt/nvidia/hpc_sdk/Linux_x86_64/20.11/compilers/bin/nvcc
> checking whether nvcc works... no
> configure: WARNING: CUDA compiler has problems.
> checking for cuInit in -lcuda... no
> configure: error: in `/codes/qe_6.7_GPU/q-e-gpu-qe-gpu-6.7':
> configure: error: Couldn't find libcuda
> See `config.log' for more details*
> *
> *
> *
> *
> *
> *
> *
> *Any Help will be greatly appreciated.*
> *
> *
> *
> *
> *
> *
> *P.S.*
> *When I run nvcc in terminal, the following error appears:*
> $ which nvcc
> /opt/nvidia/hpc_sdk/Linux_x86_64/20.11/compilers/bin/nvcc
> **
> $ nvcc
> nvcc-Error-CUDA version 10.2 was not installed with this HPC SDK:
> /opt/nvidia/hpc_sdk/Linux_x86_64/20.11/cuda/10.2/bin
>
>
> *Best,*
> *Mohammad Moaddeli*
> *Shiraz University*
> **
> **
>
> _______________________________________________
> Quantum ESPRESSO is supported by MaX (www.max-centre.eu)
> users mailing list users at lists.quantum-espresso.org
> https://lists.quantum-espresso.org/mailman/listinfo/users
>
More information about the users
mailing list