[Pw_forum] Problem with PWCOND

vborisov vborisov at mpi-halle.mpg.de
Wed Mar 26 17:01:02 CET 2014


Dear all,

I noticed a problem while using the PWCOND code for calculating
the transmission for more than one k-point within the two-dimensional 
BZ.
Whereas there is no problem with the conductance calculation for
a single k-point, the following error message appears once I provide
a list with n>=2 points in the input file for conductance calculation.
It is shown below on a standard example for an Al nanowire (same 
behavior
also observed for different systems):


forrtl: severe (151): allocatable array is already allocated
Image              PC                Routine            Line        
Source
pwcond.x           0000000000BEC50D  Unknown               Unknown  
Unknown
pwcond.x           0000000000BEB015  Unknown               Unknown  
Unknown
pwcond.x           0000000000B866D0  Unknown               Unknown  
Unknown
pwcond.x           0000000000B1EDCF  Unknown               Unknown  
Unknown
pwcond.x           0000000000B5D7CB  Unknown               Unknown  
Unknown
pwcond.x           000000000047E8C4  init_gper_.A               72  
init_gper.f90
pwcond.x           000000000045CAB9  do_cond_.A                462  
do_cond.f90
pwcond.x           000000000045388A  MAIN__                     22  
condmain.f90
pwcond.x           0000000000437FDC  Unknown               Unknown  
Unknown
libc.so.6          00002B17BB725CDD  Unknown               Unknown  
Unknown
pwcond.x           0000000000437ED9  Unknown               Unknown  
Unknown
forrtl: severe (151): allocatable array is already allocated
Image              PC                Routine            Line        
Source
pwcond.x           0000000000BEC50D  Unknown               Unknown  
Unknown
pwcond.x           0000000000BEB015  Unknown               Unknown  
Unknown
pwcond.x           0000000000B866D0  Unknown               Unknown  
Unknown
pwcond.x           0000000000B1EDCF  Unknown               Unknown  
Unknown
pwcond.x           0000000000B5D7CB  Unknown               Unknown  
Unknown
pwcond.x           000000000047E8C4  init_gper_.A               72  
init_gper.f90
pwcond.x           000000000045CAB9  do_cond_.A                462  
do_cond.f90
pwcond.x           000000000045388A  MAIN__                     22  
condmain.f90
pwcond.x           0000000000437FDC  Unknown               Unknown  
Unknown
libc.so.6          00002B76A0D2CCDD  Unknown               Unknown  
Unknown
pwcond.x           0000000000437ED9  Unknown               Unknown  
Unknown
--------------------------------------------------------------------------
mpirun has exited due to process rank 1 with PID 17595 on
node node327.cruncher exiting without calling "finalize". This may
have caused other processes in the application to be
terminated by signals sent by mpirun (as reported here).
--------------------------------------------------------------------------

The input for PWSCF:

  &control
     calculation='scf'
     restart_mode='from_scratch',
     pseudo_dir = '/scratch/fhgfs/vladislav/espresso-5.0.2/pseudo/',
     outdir='/scratch/fhgfs/vladislav/espresso-5.0.2/tmp/test/',
     prefix='al'
  /
  &system
     ibrav = 6,
     celldm(1) =5.3,
     celldm(3) =1.414,
     nat= 2,
     ntyp= 1,
     ecutwfc = 15.0,
     occupations='smearing',
     smearing='methfessel-paxton',
     degauss=0.01
  /
  &electrons
     conv_thr = 1.0e-8
     mixing_beta = 0.7
  /
ATOMIC_SPECIES
  Al 26.98 Al.pz-vbc.UPF
ATOMIC_POSITIONS
  Al 0. 0. 0.0
  Al 0.5 0.5 0.707
K_POINTS (automatic)
  4 4 4 1 1 1


The input for PWCOND:

  &inputcond
     outdir='/scratch/fhgfs/vladislav/espresso-5.0.2/tmp/test/',
     prefixl='al'
     prefixs='al'
     tran_file ='trans.al.Ef'
     ikind = 1
     iofspin = 1
     energy0 = 0.00d0
     denergy = -0.01d0
     ewind = 1.d0
     epsproj = 1.d-3
     delgep = 1.d-12
     cutplot = 3.d0
     nz1 = 22
  &
     2
     0.0  0.0  0.5
     0.5  0.5  0.5
     1

************************************************************
************************************************************

The calculation stops after the first k-point and gives
the aforementioned message. Has anyone encountered the same problem
and possibly know how to solve it?

Another issue is related to the use of a distributed network
filesystem. The problem is a heavy workload noticed in our
filesystem that leads to performance problems on our high-performance
cluster. In particular, the meta servers report a very high work
request. This isn't observed using any other code. Exemplarily,
below there is the time report of PWCOND during one of such runs:

PWCOND       :     0h18m CPU        5h19m WALL

init         :    432.25s CPU  18519.07s WALL (       1 calls)
poten        :      0.49s CPU      0.52s WALL (       2 calls)
local        :     49.24s CPU     51.44s WALL (       1 calls)

scatter_forw :    598.01s CPU    599.04s WALL (       2 calls)

compbs       :     12.88s CPU     12.89s WALL (       1 calls)
compbs_2     :     10.84s CPU     10.84s WALL (       1 calls)

Notice the large difference between the CPU and the WALL times
for the init subroutine. This was observed during the parallel
execution with different number of processors both for 5.0.1 and
5.0.2 versions, and on different architectures using OPENMPI 
environment.

I would very much appreciate any help with these matters.

With kind regards,
Vladislav Borisov


Max Planck Institute of Microstructure Physics
Weinberg 2, 06120, Halle (Saale), Germany
Tel No: +49 345 5525448
Fax No: +49 345 5525446
Email: vborisov at mpi-halle.mpg.de




More information about the users mailing list