[Pw_forum] On the run of GPU version of QE
Alexander G. Kvashnin
agkvashnin at gmail.com
Tue Mar 25 13:29:02 CET 2014
Dear QE users and developers,
Recently I compiled QE-GPU on my system with GPU NVidia Tesla K20Xm.
I have CUDA ver. 5.5.
Intel MPI 4.1.3.048
Intel Compiler 14.0
I compiled GPU-QE using the instruction on the website. I downloaded
espresso-5.0.2, than downloaded QE-GPU-14.01.0 with patch file. After the
patching I type a configuration command:
./configure
LAPACK_LIBS=/opt/intel/composer_xe_2013_sp1.2.144/mkl/lib/intel64/libmkl_lapack95_lp64.a
--enable-openmp --enable-parallel --enable-cuda --with-gpu-arch=30
--with-cuda-dir=/opt/cuda/5.5 --enable-phigemm --enable-magma
After successful configuration I compile pw-gpu.x and everything was fine.
Then I tried to test the performance of this system compared with common
HPC.
But I have troubles. As I understand commonly 1 process goes to 1GPU, so I
need to run 2 MPI processes on my host with 2 GPUs. If I am wrong, please,
tell me where? After the start I have the error message:
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
task # 0
from cdiaghg : error # 61
diagonalization (ZHEGV*) failed
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
And it does no matter what file I ran. I double checked all the files I
run. In any case I had the same error. Moreover when I ran the same file
using QE without GPU I had successful running. Would you be so kind to tell
me where I did the mistake? And what should I do to achieve successful
running?
The output of ldd command for pw-gpu.x is shown below:
linux-vdso.so.1 => (0x00007ffff1dbb000)
libmkl_scalapack_lp64.so =>
/opt/intel/composer_xe_2013_sp1.2.144/mkl/lib/intel64/libmkl_scalapack_lp64.so
(0x00007f549bc79000)
libmkl_blacs_intelmpi_lp64.so =>
/opt/intel/composer_xe_2013_sp1.2.144/mkl/lib/intel64/libmkl_blacs_intelmpi_lp64.so
(0x00007f549ba3d000)
libmkl_gf_lp64.so =>
/opt/intel/composer_xe_2013_sp1.2.144/mkl/lib/intel64/libmkl_gf_lp64.so
(0x00007f549b2f7000)
libmkl_gnu_thread.so =>
/opt/intel/composer_xe_2013_sp1.2.144/mkl/lib/intel64/libmkl_gnu_thread.so
(0x00007f549a811000)
libmkl_core.so =>
/opt/intel/composer_xe_2013_sp1.2.144/mkl/lib/intel64/libmkl_core.so
(0x00007f5499133000)
libcudart.so.5.5 => /opt/cuda/5.5/lib64/libcudart.so.5.5
(0x00007f5498ee6000)
libcublas.so.5.5 => /opt/cuda/5.5/lib64/libcublas.so.5.5
(0x00007f5495a2e000)
libcufft.so.5.5 => /opt/cuda/5.5/lib64/libcufft.so.5.5
(0x00007f5490f0d000)
libmpigf.so.4 => /opt/intel//impi/
4.1.3.048/intel64/lib/libmpigf.so.4 (0x00007f5490cdd000)
libmpi_dbg_mt.so.4 => /opt/intel//impi/
4.1.3.048/intel64/lib/libmpi_dbg_mt.so.4 (0x00007f549044c000)
libdl.so.2 => /lib64/libdl.so.2 (0x0000003c8d800000)
librt.so.1 => /lib64/librt.so.1 (0x0000003c8e800000)
libpthread.so.0 => /lib64/libpthread.so.0 (0x0000003c8e000000)
libgfortran.so.3 => /usr/lib64/libgfortran.so.3 (0x00007f549013c000)
libm.so.6 => /lib64/libm.so.6 (0x0000003c8dc00000)
libgomp.so.1 => /usr/lib64/libgomp.so.1 (0x0000003c8fa00000)
libgcc_s.so.1 => /lib64/libgcc_s.so.1 (0x0000003c91000000)
libc.so.6 => /lib64/libc.so.6 (0x0000003c8d400000)
libstdc++.so.6 => /usr/lib64/libstdc++.so.6 (0x0000003c94000000)
/lib64/ld-linux-x86-64.so.2 (0x0000003c8d000000)
Thank you in advance!
*--*
*Sincerely yours,*
*Alexander G. Kvashnin *
*=====================================================PhD Student Moscow
Institute of Physics and Technology http://mipt.ru/
<http://s.wisestamp.com/links?url=http%3A%2F%2Fmipt.ru%2F&sn=>*
*141700, Institutsky lane 9, Dolgoprudny, Moscow Region, Russia*
*Junior research scientistTechnological Institute for Superhard and Novel
Carbon Materials http://www.tisnum.ru/
<http://s.wisestamp.com/links?url=http%3A%2F%2Fwww.tisnum.ru%2F&sn=>
<http://s.wisestamp.com/links?url=http%3A%2F%2Fwww.ntcstm.troitsk.ru%2F&sn=>142190,
Central'naya St. 7a, Troitsk, Moscow Region,
Russia=====================================================*
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.quantum-espresso.org/pipermail/users/attachments/20140325/ca2fb0a9/attachment.html>
More information about the users
mailing list