[Pw_forum] run PWscf on national grid service in UK
Pieremanuele Canepa
pc229 at kent.ac.uk
Thu May 14 11:08:30 CEST 2009
Dear All,
I am writing to you since I'm attempting to run PWscf jobs on the NGS
(national grid service http://www.grid-support.ac.uk/) provide here in UK
without any good result. Under the help of one person in charge of the ngs
I was able to compile PWscf (release 4.0.5) on it. This grid use particular
and non free mpi called Platform-mpi (
http://www.scali.com/Products/platform-mpi). Then, I compiled espresso
using Intel FORTRAN and C compiler and these mpi libraries.
Now when I try to run a job on the grid, what I get out is reported as
below:
Program PWSCF v.4.0.5 starts ...
Today is 14May2009 at 9:23: 5
Parallel version (MPI)
Number of processors in use: 8
R & G space division: proc/pool = 8
For Norm-Conserving or Ultrasoft (Vanderbilt) Pseudopotentials or PAW
Current dimensions of program pwscf are:
Max number of different atomic species (ntypx) = 10
Max number of k-points (npk) = 40000
Max angular momentum in pseudopotentials (lmaxx) = 3
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
from read_namelists : error # 17
reading namelist control
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
stopping ...
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
from read_namelists : error # 17
from read_namelists : error # 17
reading namelist control
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
TID HOST_NAME COMMAND_LINE STATUS TERMINATION_TIME
===== ========== ================ =======================
===================
00000 cn54.ngs.r /home/ngs0802/bi Exit (2) 05/14/2009
09:23:05
00001 cn54.ngs.r /home/ngs0802/bi Exit (2) 05/14/2009
09:23:06
00002 cn54.ngs.r /home/ngs0802/bi Exit (2) 05/14/2009
09:23:05
00003 cn54.ngs.r /home/ngs0802/bi Exit (2) 05/14/2009
09:23:05
00004 cn54.ngs.r /home/ngs0802/bi Exit (2) 05/14/2009
09:23:05
00005 cn54.ngs.r /home/ngs0802/bi Exit (2) 05/14/2009
09:23:05
00006 cn54.ngs.r /home/ngs0802/bi Exit (2) 05/14/2009
09:23:05
00007 cn54.ngs.r /home/ngs0802/bi Exit (2) 05/14/2009
09:23:05
Initially, I thought it was related to my input and I tried to run it on my
normal cluster PC, that use OpenMPI, and it seems to work as normally it
should be. What I am suppose to do? Do you have any suggestion? Anyone do
you know how to manage it ?
Best Regards, Piero
--
Pieremanuele Canepa
Room 230
School of Physical Sciences, Ingram Building,
University of Kent, Canterbury, Kent,
CT2 7NH
United Kingdom
-----------------------------------------------------------
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.quantum-espresso.org/pipermail/users/attachments/20090514/fa7dc001/attachment.html>
More information about the users
mailing list