[QE-users] crash of fs.x

patrizio.graziosi at cnr.it patrizio.graziosi at cnr.it
Wed Feb 3 11:56:24 CET 2021


  Dear all,

  since I have noticed the attachments might have been removed, I  
repeat the message and report the text files at the end of the message  
instead of as attachments. So sorry for the duplicate.

  I am experiencing crashes of fs.x when using versions 6.3, 6.4, 6.5  
and 6.6 either.
  I'm running Quantum Espresso on the supercomputing cluster "galileo"  
at CINECA, Italy, equipped with Intel Xeon E5-2697 v4 (Broadwell)  
nodes, 1022 36-core compute nodes, each one contains 2 18-cores Intel  
Xeon E5-2697 v4 (Broadwell) at 2.30 GHz.


  I do as follows, all the input files and job submission scripts are  
reported at the end of themessage:


  1) run a scf calculation

  2) run a nscf calculation

  3) run fs.x


  at step 3) QE crashes and I get in the .out file

      Error in routine fill_fs_grid (1):
      cannot locate  k point


  and in the CRASH file

      task #        19
      from fill_fs_grid : error #         1
      cannot locate  k point



  Please, note that I asked to colleagues of mine at the University of  
Warwick, where they use QE on a HPC infrastructure (Tinis) with Lenovo  
NeXtScale nx360 M5 servers with 2 x Intel Xeon E5-2630 v3 2.4 GHz  
(Haswell) 8-core processors, to run QE with the very same input files,  
and they
  told me that everything run smoothly with NO crash...


  So, I suppose I'm doing a silly error in the submission script, or,  
as suggested by my colleagues at Warwick, there is a compiling issue  
on the CINECA machine I'm using.


  Can you support me in understanding where or what I have to look at,  
in order to deal with this issue, please?



  BTW in step 2) I also used occupations = 'tetrahedra' or 'fixed',  
with the same crash for fs.x


  I thank you in advance for your help ,


  Patrizio



1) scf calculation

  input file

  &control
     prefix='Mg3Sb2',
     pseudo_dir = './',
     outdir='./'
     wf_collect=.true.
     etot_conv_thr = 1.0d-8,
/
  &system
     ibrav=  4,
     celldm(1) = 8.6767152,
     celldm(3) = 1.5818114,
     nat=  5,
     ntyp= 2,
     ecutwfc = 100.0,
  /
  &electrons
     conv_thr = 1.0d-8,
     mixing_beta = 0.7,
  /
ATOMIC_SPECIES
  Mg  24.305  Mg.upf
  Sb  121.76  Sb.upf
ATOMIC_POSITIONS crystal
  Mg 0      0      0
  Mg 1/3 2/3 0.3679670359
  Mg 2/3 1/3 0.6320329641
  Sb 1/3 2/3 0.7745210172
  Sb 2/3 1/3 0.2254789828
K_POINTS automatic
    13 13 9 0 0 0



  submission script

#!/bin/bash
#SBATCH --time=24:00:00        # Walltime in hh:mm:ss
#SBATCH --nodes=4              # Number of nodes
#SBATCH --ntasks-per-node=6   # Number of MPI ranks per node
#SBATCH --cpus-per-task=6      # Number of OpenMP threads for each MPI  
process/rank
#SBATCH --mem=118000           # Per nodes memory request (MB)
#SBATCH --account=IscrC_MEETH
#SBATCH --job-name=mg3sb2
#SBATCH --partition=gll_usr_prod

module purge
module load profile/phys
module load autoload qe/6.6

export OMP_NUM_THREADS=$SLURM_CPUS_PER_TASK
export MKL_NUM_THREADS=${OMP_NUM_THREADS}

mpirun pw.x -npool 24 -input scf.in > Mg3Sb2.out



2) nscf calculation

  input file

  &control
     prefix='Mg3Sb2',
     pseudo_dir = './',
     outdir='./'
     wf_collect=.true.
     etot_conv_thr = 1.0d-8,
     calculation = 'nscf',
/
  &system
     ibrav=  4,
     celldm(1) = 8.6767152,
     celldm(3) = 1.5818114,
     nat=  5,
     ntyp= 2,
     ecutwfc = 100.0,
  /
  &electrons
     conv_thr = 1.0d-8,
     mixing_beta = 0.7,
  /
ATOMIC_SPECIES
  Mg  24.305  Mg.upf
  Sb  121.76  Sb.upf
ATOMIC_POSITIONS crystal
  Mg 0      0      0
  Mg 1/3 2/3 0.3679670359
  Mg 2/3 1/3 0.6320329641
  Sb 1/3 2/3 0.7745210172
  Sb 2/3 1/3 0.2254789828
K_POINTS automatic
    61 61 41 0 0 0


  submission script, as above but last line

mpirun pw.x -npool 24 -input nscf.in > Mg3Sb2_nscf.out


3) FS calculation

  input file

&fermi
	outdir = './',
	prefix = 'Mg3Sb2',
	DeltaE = 3
/


  submission script, as above but last line

mpirun fs.x -npool 24 -input fs.in > Mg3Sb2_fermi.out

















More information about the users mailing list