[Pw_forum] EPW Superconductivity

Gargee Bhattacharyya bhattacharyya.gargee at gmail.com
Thu Jul 28 09:32:40 CEST 2016


Dear all,
             I am trying to calculate *electron-phonon coupling strangth,
anisotropic Elishberg spectral function , anisotrotic superconducting gap*
using EPW as instructed from the site given below

http://epw.org.uk/Main/About

http://epw.org.uk/Documentation/Tutorial

I am trying to run EPW on bulk FCC Lead. I shall be highly obliged if
anyone help me to find my error. I have calculated using the following
steps:

mpirun -np 4 ../../../../bin/pw.x < scf.in > scf.out
mpirun -np 4 ../../../../bin/ph.x < ph.in > ph.out
python pp.py < pp.in
mpirun -np 4 ../../../../bin/pw.x < scf_epw.in > scf_epw.out
mpirun -np 4 ../../../../bin/pw.x -npool 4 < nscf_epw.in > nscf_epw.out
mpirun -np 4 ../../../src/epw.x -npool 4 < epw.in > epw.out

all the calculation were completed successfully except the last epw
calculation

*mpirun -np 4 ../../../src/epw.x -npool 4 < epw.in <http://epw.in> >
epw.out*

It is showing error :

Program EPW v.4.0.0 starts on 28Jul2016 at 10:27:15

     This program is part of the open-source Quantum ESPRESSO suite
     for quantum simulation of materials; please cite
         "P. Giannozzi et al., J. Phys.:Condens. Matter 21 395502 (2009);
          URL http://www.quantum-espresso.org",
     in publications or presentations arising from this work. More details
at
     http://www.quantum-espresso.org/quote

     Parallel version (MPI), running on     4 processors
     R & G space division:  proc/nbgrp/npool/nimage =       4

 %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
     *Error in routine epw_readin (1):*
*     reading input_epw namelist*
 %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%

     stopping ...
--------------------------------------------------------------------------
MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD
with errorcode 1.

NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
You may or may not see output from other processes, depending on
exactly when Open MPI kills them.


*My input file is :*

&inputepw
  prefix      = 'pb'
  amass(1)    = 207.2
  outdir      = './'

  elph        = .true.
  kmaps       = .false.
  epbwrite    = .true.
  epbread     = .false.

  epf_mem     = .true.
  etf_mem     = .true.

  epwwrite    = .true.
  epwread     = .false.

  nbndsub     =  4
  nbndskip    =  0

  wannierize  = .true.
  num_iter    = 300
  dis_win_max = 18
  dis_froz_max= 8
  proj(1)     = 'Pb:sp3'

  iverbosity  = 0

  elinterp    = .true.
  phinterp    = .true.

  tshuffle2   = .true.
  tphases     = .false.

  elecselfen  = .true.
  phonselfen  = .true.
  a2f         = .true.

  parallel_k  = .true.
  parallel_q  = .false.

  fsthick     = 0.5
  eptemp      = 0.045
  degaussw    = 0.045
  degaussq    = 0.05

  dvscf_dir   = './save'
  filukk      = './pb.ukk'

  nk1         = 6
  nk2         = 6
  nk3         = 6

  nq1         = 6
  nq2         = 6
  nq3         = 6

  nqf1         = 4
  nqf2         = 4
  nqf3         = 4
  nkf1         = 30
  nkf2         = 30
  nkf3         = 30
 /
      16 cartesian
    0.0000000   0.0000000   0.0000000   0.0909090
   -0.1666667   0.1666667  -0.1666667   0.0909090
   -0.3333333   0.3333333  -0.3333333   0.0909090
    0.5000000  -0.5000000   0.5000000   0.0909090
    0.0000000   0.3333333   0.0000000   0.0909090
   -0.1666667   0.5000000  -0.1666667   0.0909090
    0.6666667  -0.3333333   0.6666667   0.0909090
    0.5000000  -0.1666667   0.5000000   0.0909090
    0.3333333   0.0000000   0.3333333   0.0909090
    0.0000000   0.6666667   0.0000000   0.0909090
    0.8333333  -0.1666667   0.8333333   0.0909090
    0.6666667   0.0000000   0.6666667   0.0909090
    0.0000000  -1.0000000   0.0000000   0.0909090
    0.6666667  -0.3333333   1.0000000   0.0909090
    0.5000000  -0.1666667   0.8333333   0.0909090
   -0.3333333  -1.0000000   0.0000000   0.0909090


-- 
Yours sincerely

Gargee Bhattacharyya
​PhD Pursuing
Metallurgy Engineering & Materials Science
IIT Indore ​

M.Tech (VLSI Design & Microelectronics Technology)
Department of ETCE
Jadavpur University
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.quantum-espresso.org/pipermail/users/attachments/20160728/89337bdb/attachment.html>


More information about the users mailing list