[QE-users] [ge-GPU] compiling q-e-gpu-qe-gpu-6.7

Mohammad Moaddeli mohammad.moaddeli at gmail.com
Tue Dec 22 08:38:23 CET 2020


Dear Pietro,

Thanks for sharing the link. I got the error of missing v0.3.1.tar.gz file
in the archive folder, so I downloaded it manually. Everything went well
and I could run jobs. However, there are a few things that are unclear to
me.

config options:

./configure FC=pgf90 F90=pgf90 CC=pgcc --with-cuda=yes
--enable-cuda-env-check=no --with-cuda-runtime=11.0 --with-cuda-cc=70
--enable-openmp --with-scalapack=no

The make.sys file is changed based on the install instructions on this
page: https://gitlab.com/QEF/q-e-gpu/-/wikis/home

DFLAGS         =  -D__CUDA -D__DFTI -D__MPI
DFLAGS         =  -D__CUDA -D__DFTI -D__MPI__GPU_MPI

Since two cards are installed on the mainboard, I ran the jobs in this form:

mpirun -np 2 pw.x -input file.in | tee file.out

The hardware specifications in the output file:

     GPU used by master process:

        Device Number: 0
        Device name: TITAN V
        Compute capability : 70
        Ratio of single to double precision performance  : 2
        Memory Clock Rate (KHz): 850000
        Memory Bus Width (bits): 3072
        Peak Memory Bandwidth (GB/s): 652.80

I just wonder why the second card (device # 1) is not printed in the output.

Since -D__DFTI is printed in the make.sys file, the MKL and FFTW of Intel
Parallel Studio is utilized. Is this an appropriate config in order to get
the best performance?

Is it possible to compile the code for multiple graphic cards with
different cuda-cc ?

I’m sorry for asking so many questions. I appreciate your time in
responding to whatever you can, when you are able to find the time.

Best,

Mohammad
ShirazU

On Sun, Dec 20, 2020 at 9:08 PM Pietro Bonfa' <pietro.bonfa at unipr.it> wrote:

> Dear Mohammad,
>
> for some reason you are having trouble accessing gitlab. I uploaded a
> package that includes all dependencies and can be compiled without
> network access. You can find it here:
>
>
> https://univpr-my.sharepoint.com/:u:/g/personal/pietro_bonfa_unipr_it/EV-nHENjf1lFkat0RvJypFIBap2o92v9BzG75po06z48WA?e=uiDjDD
>
> Best wishes,
> Pietro
>
> --
> Pietro Bonfà,
> University of Parma
>
> On 12/19/20 7:27 AM, Mohammad Moaddeli wrote:
> > Dear Louis and Pietro,
> >
> > *With the config options as following:*
> > *
> > *
> > *./configure FC=pgf90 F90=pgf90 CC=pgcc
> > --with-cuda=/opt/nvidia/hpc_sdk/Linux_x86_64/20.11/cuda/11.0
> > --with-cuda-runtime=11.0 --with-cuda-cc=70 --enable-openmp
> > --with-scalapack=no*
> > *
> > *
> > *results is:*
> >
> > checking build system type... x86_64-pc-linux-gnu
> > checking ARCH... x86_64
> > checking setting AR... ... ar
> > checking setting ARFLAGS... ... ruv
> > checking whether the Fortran compiler works... yes
> > checking for Fortran compiler default output file name... a.out
> > checking for suffix of executables...
> > checking whether we are cross compiling... no
> > checking for suffix of object files... o
> > checking whether we are using the GNU Fortran compiler... no
> > checking whether pgf90 accepts -g... yes
> > configure: WARNING: F90 value is set to be consistent with value of
> MPIF90
> > checking for mpiifort... no
> > checking for mpif90... mpif90
> > checking whether we are using the GNU Fortran compiler... no
> > checking whether mpif90 accepts -g... yes
> > checking version of mpif90... nvfortran 20.11-0
> > checking for Fortran flag to compile .f90 files... none
> > setting F90... nvfortran
> > setting MPIF90... mpif90
> > checking whether we are using the GNU C compiler... yes
> > checking whether pgcc accepts -g... yes
> > checking for pgcc option to accept ISO C89... none needed
> > setting CC... pgcc
> > setting CFLAGS... -fast -Mpreprocess
> > using F90... nvfortran
> > setting FFLAGS... -O1
> > setting F90FLAGS... $(FFLAGS)
> > setting FFLAGS_NOOPT... -O0
> > setting CPP... cpp
> > setting CPPFLAGS... -P -traditional -Uvector
> > setting LD... mpif90
> > setting LDFLAGS...
> > checking for Fortran flag to compile .f90 files... (cached) none
> > checking whether Fortran compiler accepts -Mcuda=cuda11.0... yes
> > checking for nvcc...
> > /opt/nvidia/hpc_sdk/Linux_x86_64/20.11/compilers/bin/nvcc
> > checking whether nvcc works... yes
> > checking for cuInit in -lcuda... no
> > configure: error: in `/codes/qe_6.7_GPU/q-e-gpu-qe-gpu-6.7':
> > configure: error: Couldn't find libcuda
> > See `config.log' for more details
> >
> > *Although the option --enable-cuda-env-check=no resulted in the
> > configuration finished:*
> > *
> > *
> > *./configure FC=pgf90 F90=pgf90 CC=pgcc --with-cuda=yes
> > --enable-cuda-env-check=no --with-cuda-runtime=11.0 --with-cuda-cc=70
> > --enable-openmp --with-scalapack=no*
> >
> > checking build system type... x86_64-pc-linux-gnu
> > checking ARCH... x86_64
> > checking setting AR... ... ar
> > checking setting ARFLAGS... ... ruv
> > checking whether the Fortran compiler works... yes
> > checking for Fortran compiler default output file name... a.out
> > checking for suffix of executables...
> > checking whether we are cross compiling... no
> > checking for suffix of object files... o
> > checking whether we are using the GNU Fortran compiler... no
> > checking whether pgf90 accepts -g... yes
> > configure: WARNING: F90 value is set to be consistent with value of
> MPIF90
> > checking for mpiifort... no
> > checking for mpif90... mpif90
> > checking whether we are using the GNU Fortran compiler... no
> > checking whether mpif90 accepts -g... yes
> > checking version of mpif90... nvfortran 20.11-0
> > checking for Fortran flag to compile .f90 files... none
> > setting F90... nvfortran
> > setting MPIF90... mpif90
> > checking whether we are using the GNU C compiler... yes
> > checking whether pgcc accepts -g... yes
> > checking for pgcc option to accept ISO C89... none needed
> > setting CC... pgcc
> > setting CFLAGS... -fast -Mpreprocess
> > using F90... nvfortran
> > setting FFLAGS... -O1
> > setting F90FLAGS... $(FFLAGS)
> > setting FFLAGS_NOOPT... -O0
> > setting CPP... cpp
> > setting CPPFLAGS... -P -traditional -Uvector
> > setting LD... mpif90
> > setting LDFLAGS...
> > checking for Fortran flag to compile .f90 files... (cached) none
> > checking whether Fortran compiler accepts -Mcuda=cuda11.0... yes
> > checking for /usr/local/cuda/... no
> > checking for /usr/local/cuda/include... no
> > checking for /usr/local/cuda/lib64... no
> > checking for nvcc...
> > /opt/nvidia/hpc_sdk/Linux_x86_64/20.11/compilers/bin/nvcc
> > checking whether nvcc works... yes
> > checking for cusolverDnZhegvdx_bufferSize in -lcusolver... no
> > configure: WARNING: Using legacy custom solver.
> > checking whether make sets $(MAKE)... yes
> > checking whether Fortran files must be preprocessed... yes
> > checking whether we are using the GNU Fortran 77 compiler... no
> > checking whether nvfortran accepts -g... yes
> > checking for library containing dgemm... no
> > MKL not found
> > in /opt/intel/mkl/lib/intel64: checking for library containing dgemm...
> > -lmkl_intel_lp64
> > setting BLAS_LIBS... -L/opt/intel/mkl/lib/intel64 -lmkl_intel_lp64
> > -lmkl_intel_thread -lmkl_core
> > checking FFT...
> > checking MASS...
> > checking for library containing mpi_init... none required
> > checking ELPA...
> > checking BEEF... -lbeef
> > setting BEEF_LIBS... $(TOPDIR)/LIBBEEF/libbeef.a
> > checking for ranlib... ranlib
> > checking for wget... wget -O
> > setting WGET... wget -O
> > setting DFLAGS... -D__CUDA -D__DFTI -D__MPI
> > setting IFLAGS... -I$(TOPDIR)/include -I$(TOPDIR)/FoX/finclude
> > -I/opt/intel/mkl/include
> > configure: creating ./config.status
> > config.status: creating install/make_lapack.inc
> > config.status: creating include/configure.h
> > config.status: creating make.inc
> > config.status: creating configure.msg
> > config.status: creating install/make_wannier90.inc
> > config.status: creating include/qe_cdefs.h
> > --------------------------------------------------------------------
> > ESPRESSO can take advantage of several optimized numerical libraries
> > (essl, fftw, mkl...).  This configure script attempts to find them,
> > but may fail if they have been installed in non-standard locations.
> > If a required library is not found, the local copy will be compiled.
> >
> > The following libraries have been found:
> >    BLAS_LIBS=-L/opt/intel/mkl/lib/intel64 -lmkl_intel_lp64
> >   -lmkl_intel_thread -lmkl_core
> >    LAPACK_LIBS=
> >    FFT_LIBS=
> >
> > Please check if this is what you expect.
> >
> > If any libraries are missing, you may specify a list of directories
> > to search and retry, as follows:
> >    ./configure LIBDIRS="list of directories, separated by spaces"
> >
> > Parallel environment detected successfully.\
> > Configured for compilation of parallel executables.
> >
> > For more info, read the ESPRESSO User's Guide (Doc/users-guide.tex).
> > --------------------------------------------------------------------
> > configure: success
> >
> > *make pw results in:*
> >
> > ....
> > .....
> > ....
> > cpp -P -traditional -Uvector -D__CUDA -D__DFTI -D__MPI
> > -I/codes/qe_6.7_GPU/q-e-gpu-qe-gpu-6.7//include
> > -I/codes/qe_6.7_GPU/q-e-gpu-qe-gpu-6.7//FoX/finclude
> >   -I/opt/intel/mkl/include thread_util.f90 -o thread_util_tmp.f90 ; \
> >          mpif90 -O1   -D__CUDA -D__DFTI -D__MPI  -Mcuda=cc70,cuda11.0
> > -I/codes/qe_6.7_GPU/q-e-gpu-qe-gpu-6.7//external/devxlib/src
> > -I/codes/qe_6.7_GPU/q-e-gpu-qe-gpu-6.7//external/devxlib/include
> > -I/codes/qe_6.7_GPU/q-e-gpu-qe-gpu-6.7//EIGENSOLVER_GPU/lib_eigsolve
> > -I/codes/qe_6.7_GPU/q-e-gpu-qe-gpu-6.7//external/devxlib/src
> > -I/codes/qe_6.7_GPU/q-e-gpu-qe-gpu-6.7//include
> > -I/codes/qe_6.7_GPU/q-e-gpu-qe-gpu-6.7//FoX/finclude
> >   -I/opt/intel/mkl/include -I. -c thread_util_tmp.f90 -o thread_util.o
> > ar ruv libutil.a clocks_handler.o device_helper.o divide.o data_buffer.o
> > error_handler.o export_gstart_2_solvers.o find_free_unit.o
> > fletcher32_mod.o mem_counter.o mp.o mp_base.o mp_base_gpu.o
> > mp_bands_util.o parallel_include.o set_mpi_comm_4_solvers.o util_param.o
> > thread_util.o
> > ar: creating libutil.a
> > a - clocks_handler.o
> > a - device_helper.o
> > a - divide.o
> > a - data_buffer.o
> > a - error_handler.o
> > a - export_gstart_2_solvers.o
> > a - find_free_unit.o
> > a - fletcher32_mod.o
> > a - mem_counter.o
> > a - mp.o
> > a - mp_base.o
> > a - mp_base_gpu.o
> > a - mp_bands_util.o
> > a - parallel_include.o
> > a - set_mpi_comm_4_solvers.o
> > a - util_param.o
> > a - thread_util.o
> > ranlib libutil.a
> > make[1]: Leaving directory
> `/codes/qe_6.7_GPU/q-e-gpu-qe-gpu-6.7/UtilXlib'
> > cd install ; make -f extlibs_makefile liblapack
> > make[1]: Entering directory
> `/codes/qe_6.7_GPU/q-e-gpu-qe-gpu-6.7/install'
> > make[1]: Nothing to be done for `liblapack'.
> > make[1]: Leaving directory `/codes/qe_6.7_GPU/q-e-gpu-qe-gpu-6.7/install'
> > cd install ; make -f extlibs_makefile libcuda
> > make[1]: Entering directory
> `/codes/qe_6.7_GPU/q-e-gpu-qe-gpu-6.7/install'
> > cd ../external/devxlib; \
> >      if test ! -e configure; then \
> >      wget
> > "
> https://gitlab.com/max-centre/components/devicexlib/-/archive/master/devicexlib-master.tar.gz
> > <
> https://gitlab.com/max-centre/components/devicexlib/-/archive/master/devicexlib-master.tar.gz>"
>
> > -O devxlib.tar.gz || curl
> > "
> https://gitlab.com/max-centre/components/devicexlib/-/archive/master/devicexlib-master.tar.gz
> > <
> https://gitlab.com/max-centre/components/devicexlib/-/archive/master/devicexlib-master.tar.gz>"
>
> > -o devxlib.tar.gz ; \
> >      tar xzf devxlib.tar.gz --strip-components=1 -C . ; \
> >      rm devxlib.tar.gz ; \
> >      fi; \
> >      touch make.inc; \
> >      make clean; \
> >      export F90FLAGS=""; \
> >      ./configure FC=nvfortran CC=pgcc \
> >                  --with-cuda= \
> >                  --with-cuda-cc=70 \
> >                  --with-cuda-runtime=11.0 \
> >                  --disable-parallel \
> >                  --enable-cuda-env-check=no; \
> >      make all
> > --2020-12-19 09:29:12--
> >
> https://gitlab.com/max-centre/components/devicexlib/-/archive/master/devicexlib-master.tar.gz
> > <
> https://gitlab.com/max-centre/components/devicexlib/-/archive/master/devicexlib-master.tar.gz
> >
> > Resolving gitlab.com
> > <http://gitlab.com/>
> > (gitlab.com
> > <http://gitlab.com/>)...
> > 172.65.251.78, 2606:4700:90:0:f22e:fbec:5bed:a9b9
> > Connecting to gitlab.com
> > <http://gitlab.com/>
> > (gitlab.com
> > <http://gitlab.com/>)|172.65.251.78|:443...
> > connected.
> > HTTP request sent, awaiting response... 403 Forbidden
> > 2020-12-19 09:29:23 ERROR 403: Forbidden.
> >
> >    % Total    % Received % Xferd  Average Speed   Time    Time     Time
> >   Current
> >                                   Dload  Upload   Total   Spent    Left
> >   Speed
> > 100    16  100    16    0     0     27      0 --:--:-- --:--:-- --:--:--
> >     27
> >
> > gzip: stdin: not in gzip format
> > tar: Child returned status 1
> > tar: Error is not recoverable: exiting now
> > make[2]: Entering directory
> > `/codes/qe_6.7_GPU/q-e-gpu-qe-gpu-6.7/external/devxlib'
> > make[2]: *** No rule to make target `clean'.  Stop.
> > make[2]: Leaving directory
> > `/codes/qe_6.7_GPU/q-e-gpu-qe-gpu-6.7/external/devxlib'
> > /bin/sh: line 9: ./configure: No such file or directory
> > make[2]: Entering directory
> > `/codes/qe_6.7_GPU/q-e-gpu-qe-gpu-6.7/external/devxlib'
> > make[2]: *** No rule to make target `all'.  Stop.
> > make[2]: Leaving directory
> > `/codes/qe_6.7_GPU/q-e-gpu-qe-gpu-6.7/external/devxlib'
> > make[1]: *** [libcuda_devxlib] Error 2
> > make[1]: Leaving directory `/codes/qe_6.7_GPU/q-e-gpu-qe-gpu-6.7/install'
> > make: *** [libcuda] Error 2
> >
> > *Moreover, "which libcuda.so" results in:*
> > *
> > *
> > /opt/nvidia/hpc_sdk/Linux_x86_64/20.11/cuda/11.0/lib64/stubs/libcuda.so
> > *
> > *
> > *Any help will be greatly appreciated.*
> > *
> > *
> > *Best regards,*
> > *
> > *
> > *Mohammad*
> > *ShirazU*
> >
> > On Wed, Dec 16, 2020 at 5:08 PM Pietro Bonfa' <pietro.bonfa at unipr.it
> > <mailto:pietro.bonfa at unipr.it>> wrote:
> >
> >     Dear Mohammad,
> >
> >     CUDA may be installed somewhere else, anyway, if you want to skip the
> >     environment check, you may configure QE with this command
> >
> >     ./configure FC=pgf90 F90=pgf90 CC=pgcc --with-cuda=yes
> >     --enable-cuda-env-check=no  --with-cuda-runtime=SETME
> --with-cuda-cc=70
> >     --enable-openmp
> >
> >     Remember to set the cuda runtime according to what is provided by
> your
> >     setup.
> >
> >     Hope this helps,
> >     best,
> >     Pietro
> >
> >     ---
> >     Pietro Bonfà
> >     University of Parma
> >
> >
> >     On 12/16/20 9:10 AM, Mohammad Moaddeli wrote:
> >      > Dear all,
> >      > I am trying to compile the 6.7 version of the code using PGI 2020.
> >      > I followed these steps:
> >      >
> >      > *1) NVIDIA driver (NVIDIA-Linux-x86_64-450.80.02.rpm) is
> installed.*
> >      > *the output of nvidia-smi:*
> >      >
> >      > Wed Dec 16 09:07:11 2020
> >      >
> >
>  +-----------------------------------------------------------------------------+
> >      > | NVIDIA-SMI 450.80.02    Driver Version: 450.80.02    CUDA
> Version:
> >      > 11.0     |
> >      >
> >
>  |-------------------------------+----------------------+----------------------+
> >      > | GPU  Name        Persistence-M| Bus-Id        Disp.A | Volatile
> >      > Uncorr. ECC |
> >      > | Fan  Temp  Perf  Pwr:Usage/Cap|         Memory-Usage | GPU-Util
> >      >   Compute M. |
> >      > |                               |                      |
> >      > MIG M. |
> >      >
> >
>  |===============================+======================+======================|
> >      > |   0  TITAN V             Off  | 00000000:06:00.0 Off |
> >      >   N/A |
> >      > | 27%   37C    P0    32W / 250W |      0MiB / 12066MiB |      0%
> >      >   Default |
> >      > |                               |                      |
> >      >   N/A |
> >      >
> >
>  +-------------------------------+----------------------+----------------------+
> >      > |   1  TITAN V             Off  | 00000000:07:00.0 Off |
> >      >   N/A |
> >      > | 25%   37C    P0    35W / 250W |      0MiB / 12066MiB |      0%
> >      >   Default |
> >      > |                               |                      |
> >      >   N/A |
> >      >
> >
>  +-------------------------------+----------------------+----------------------+
> >      >
> >      >
> >
>  +-----------------------------------------------------------------------------+
> >      > | Processes:
> >      >       |
> >      > |  GPU   GI   CI        PID   Type   Process name
> >       GPU
> >      > Memory |
> >      > |        ID   ID
> >        Usage
> >      >       |
> >      >
> >
>  |=============================================================================|
> >      > |  No running processes found
> >      >        |
> >      >
> >
>  +-----------------------------------------------------------------------------+
> >      >
> >      > *The output of pgaccelinfo:*
> >      >
> >      > CUDA Driver Version:           11000
> >      > NVRM version:                  NVIDIA UNIX x86_64 Kernel Module
> >      >   450.80.02  Wed Sep 23 01:13:39 UTC 2020
> >      >
> >      > Device Number:                 0
> >      > Device Name:                   TITAN V
> >      > Device Revision Number:        7.0
> >      > Global Memory Size:            12652838912
> >      > Number of Multiprocessors:     80
> >      > Concurrent Copy and Execution: Yes
> >      > Total Constant Memory:         65536
> >      > Total Shared Memory per Block: 49152
> >      > Registers per Block:           65536
> >      > Warp Size:                     32
> >      > Maximum Threads per Block:     1024
> >      > Maximum Block Dimensions:      1024, 1024, 64
> >      > Maximum Grid Dimensions:       2147483647 x 65535 x 65535
> >      > Maximum Memory Pitch:          2147483647B
> >      > Texture Alignment:             512B
> >      > Clock Rate:                    1455 MHz
> >      > Execution Timeout:             No
> >      > Integrated Device:             No
> >      > Can Map Host Memory:           Yes
> >      > Compute Mode:                  default
> >      > Concurrent Kernels:            Yes
> >      > ECC Enabled:                   No
> >      > Memory Clock Rate:             850 MHz
> >      > Memory Bus Width:              3072 bits
> >      > L2 Cache Size:                 4718592 bytes
> >      > Max Threads Per SMP:           2048
> >      > Async Engines:                 7
> >      > Unified Addressing:            Yes
> >      > Managed Memory:                Yes
> >      > Concurrent Managed Memory:     Yes
> >      > Preemption Supported:          Yes
> >      > Cooperative Launch:            Yes
> >      >    Multi-Device:                Yes
> >      > Default Target:                cc70
> >      >
> >      > Device Number:                 1
> >      > Device Name:                   TITAN V
> >      > Device Revision Number:        7.0
> >      > Global Memory Size:            12652838912
> >      > Number of Multiprocessors:     80
> >      > Concurrent Copy and Execution: Yes
> >      > Total Constant Memory:         65536
> >      > Total Shared Memory per Block: 49152
> >      > Registers per Block:           65536
> >      > Warp Size:                     32
> >      > Maximum Threads per Block:     1024
> >      > Maximum Block Dimensions:      1024, 1024, 64
> >      > Maximum Grid Dimensions:       2147483647 x 65535 x 65535
> >      > Maximum Memory Pitch:          2147483647B
> >      > Texture Alignment:             512B
> >      > Clock Rate:                    1455 MHz
> >      > Execution Timeout:             No
> >      > Integrated Device:             No
> >      > Can Map Host Memory:           Yes
> >      > Compute Mode:                  default
> >      > Concurrent Kernels:            Yes
> >      > ECC Enabled:                   No
> >      > Memory Clock Rate:             850 MHz
> >      > Memory Bus Width:              3072 bits
> >      > L2 Cache Size:                 4718592 bytes
> >      > Max Threads Per SMP:           2048
> >      > Async Engines:                 7
> >      > Unified Addressing:            Yes
> >      > Managed Memory:                Yes
> >      > Concurrent Managed Memory:     Yes
> >      > Preemption Supported:          Yes
> >      > Cooperative Launch:            Yes
> >      >    Multi-Device:                Yes
> >      > Default Target:                cc70
> >      >
> >      > *2) PGI compiler is installed:*
> >      > *yum install nvhpc-20-11-20.11-1.x86_64.rpm
> >     nvhpc-2020-20.11-1.x86_64.rpm
> >      > *
> >      > *PATHs that are set in ~/.bashrc file:
> >      > *
> >      > *
> >      > *
> >      > export
> >     PATH=/opt/nvidia/hpc_sdk/Linux_x86_64/20.11/cuda/11.1/bin:$PATH
> >      > export
> >     PATH=/opt/nvidia/hpc_sdk/Linux_x86_64/20.11/cuda/11.1/include:$PATH
> >      > export
> >      >
> >
>  LD_LIBRARY_PATH=/opt/nvidia/hpc_sdk/Linux_x86_64/20.11/cuda/11.1/lib64:$LD_LIBRARY_PATH
> >      > export
> >      >
> >
>  LD_LIBRARY_PATH=/opt/nvidia/hpc_sdk/Linux_x86_64/20.11/cuda/11.1/extras/CUPTI/lib64:$LD_LIBRARY_PATH
> >      > export
> >      >
> >
>  LD_LIBRARY_PATH=/opt/nvidia/hpc_sdk/Linux_x86_64/20.11/cuda/11.1/lib64/stubs:$LD_LIBRARY_PATH
> >      > NVARCH=`uname -s`_`uname -m`; export NVARCH
> >      > NVCOMPILERS=/opt/nvidia/hpc_sdk; export NVCOMPILERS
> >      > MANPATH=$MANPATH:$NVCOMPILERS/$NVARCH/20.11/compilers/man; export
> >     MANPATH
> >      > PATH=$NVCOMPILERS/$NVARCH/20.11/compilers/bin:$PATH; export PATH
> >      > PATH=$NVCOMPILERS/$NVARCH/20.11/compilers/include:$PATH; export
> PATH
> >      > LD_LIBRARY_PATH=$NVCOMPILERS/$NVARCH/20.11/compilers/lib:$PATH;
> >     export
> >      > LD_LIBRARY_PATH
> >      > export
> >     PATH=/opt/nvidia/hpc_sdk/Linux_x86_64/20.11/comm_libs/mpi/bin:$PATH
> >      > export
> >      >
> >
>  PATH=/opt/nvidia/hpc_sdk/Linux_x86_64/20.11/comm_libs/mpi/include:$PATH
> >      > export
> >      >
> >
>  LD_LIBRARY_PATH=/opt/nvidia/hpc_sdk/Linux_x86_64/20.11/comm_libs/mpi/lib:$LD_LIBRARY_PATH
> >      > export
> >      >
> >
>  LD_LIBRARY_PATH=/opt/nvidia/hpc_sdk/Linux_x86_64/20.11/math_libs/11.1/lib64:$LD_LIBRARY_PATH
> >      > export
> >      >
> >
>  LD_LIBRARY_PATH=/opt/nvidia/hpc_sdk/Linux_x86_64/20.11/math_libs/11.1/lib64/stubs:$LD_LIBRARY_PATH*
> >      > *
> >      > *
> >      > *
> >      > *3) compiling the code using:*
> >      > *./configure FC=pgf90 F90=pgf90 F77=pgf90 CC=pgcc MPIF90=mpif90
> >      > --with-cuda=/opt/nvidia/hpc_sdk/Linux_x86_64/20.11/cuda
> >      > --with-cuda-runtime=11.1 --with-cuda-cc=70 --enable-openmp
> >      > --with-scalapack=no*
> >      >
> >      > checking build system type... x86_64-pc-linux-gnu
> >      > checking ARCH... x86_64
> >      > checking setting AR... ... ar
> >      > checking setting ARFLAGS... ... ruv
> >      > checking whether the Fortran compiler works... yes
> >      > checking for Fortran compiler default output file name... a.out
> >      > checking for suffix of executables...
> >      > checking whether we are cross compiling... no
> >      > checking for suffix of object files... o
> >      > checking whether we are using the GNU Fortran compiler... no
> >      > checking whether pgf90 accepts -g... yes
> >      > configure: WARNING: F90 value is set to be consistent with value
> >     of MPIF90
> >      > checking for mpif90... mpif90
> >      > checking whether we are using the GNU Fortran compiler... no
> >      > checking whether mpif90 accepts -g... yes
> >      > checking version of mpif90... nvfortran 20.11-0
> >      > checking for Fortran flag to compile .f90 files... none
> >      > setting F90... nvfortran
> >      > setting MPIF90... mpif90
> >      > checking whether we are using the GNU C compiler... yes
> >      > checking whether pgcc accepts -g... yes
> >      > checking for pgcc option to accept ISO C89... none needed
> >      > setting CC... pgcc
> >      > setting CFLAGS... -fast -Mpreprocess
> >      > using F90... nvfortran
> >      > setting FFLAGS... -O1
> >      > setting F90FLAGS... $(FFLAGS)
> >      > setting FFLAGS_NOOPT... -O0
> >      > setting CPP... cpp
> >      > setting CPPFLAGS... -P -traditional -Uvector
> >      > setting LD... mpif90
> >      > setting LDFLAGS...
> >      > checking for Fortran flag to compile .f90 files... (cached) none
> >      > checking whether Fortran compiler accepts -Mcuda=cuda11.1... yes
> >      > checking for nvcc...
> >      > /opt/nvidia/hpc_sdk/Linux_x86_64/20.11/compilers/bin/nvcc
> >      > checking whether nvcc works... no
> >      > configure: WARNING: CUDA compiler has problems.
> >      > checking for cuInit in -lcuda... no
> >      > configure: error: in `/codes/qe_6.7_GPU/q-e-gpu-qe-gpu-6.7':
> >      > configure: error: Couldn't find libcuda
> >      > See `config.log' for more details*
> >      > *
> >      > *
> >      > *
> >      > *
> >      > *
> >      > *
> >      > *
> >      > *Any Help will be greatly appreciated.*
> >      > *
> >      > *
> >      > *
> >      > *
> >      > *
> >      > *
> >      > *P.S.*
> >      > *When I run nvcc in terminal, the following error appears:*
> >      > $ which nvcc
> >      > /opt/nvidia/hpc_sdk/Linux_x86_64/20.11/compilers/bin/nvcc
> >      > **
> >      > $ nvcc
> >      > nvcc-Error-CUDA version 10.2 was not installed with this HPC SDK:
> >      > /opt/nvidia/hpc_sdk/Linux_x86_64/20.11/cuda/10.2/bin
> >      >
> >      >
> >      > *Best,*
> >      > *Mohammad Moaddeli*
> >      > *Shiraz University*
> >      > **
> >      > **
> >      >
> >      > _______________________________________________
> >      > Quantum ESPRESSO is supported by MaX (www.max-centre.eu
> >     <http://www.max-centre.eu/>)
> >      > users mailing list users at lists.quantum-espresso.org
> >     <mailto:users at lists.quantum-espresso.org>
> >      > https://lists.quantum-espresso.org/mailman/listinfo/users
> >     <https://lists.quantum-espresso.org/mailman/listinfo/users>
> >      >
> >     _______________________________________________
> >     Quantum ESPRESSO is supported by MaX (www.max-centre.eu
> >     <http://www.max-centre.eu/>)
> >     users mailing list users at lists.quantum-espresso.org
> >     <mailto:users at lists.quantum-espresso.org>
> >     https://lists.quantum-espresso.org/mailman/listinfo/users
> >     <https://lists.quantum-espresso.org/mailman/listinfo/users>
> >
> >
> > _______________________________________________
> > Quantum ESPRESSO is supported by MaX (www.max-centre.eu)
> > users mailing list users at lists.quantum-espresso.org
> > https://lists.quantum-espresso.org/mailman/listinfo/users
> >
> _______________________________________________
> Quantum ESPRESSO is supported by MaX (www.max-centre.eu)
> users mailing list users at lists.quantum-espresso.org
> https://lists.quantum-espresso.org/mailman/listinfo/users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.quantum-espresso.org/pipermail/users/attachments/20201222/cd9bed4c/attachment.html>


More information about the users mailing list