[QE-users] Error in routine ggen too many g-vectors
Thomas Brumme
thomas.brumme at uni-leipzig.de
Fri Apr 27 17:21:35 CEST 2018
Dear Sohail Ahmad,
I don't know what is the origin of the "too many g-vectors"
error but I see several errors in your input.
- you want to calculate a 2-dimensional system, so why do
you need k-points in the z direction? You should actually see
no dispersion at all in this direction if you correctly setup your
system
- as mentioned in the dipole example for PP and in all papers
using such a correction, the dipole needs to be in the vacuum
region, i.e., there should be no interaction with the system
and the dipole. Your system starts at the origin and the dipole
is from 0.9 till 1.1 (i.e. 0.1) - it is not only too close but even
within the system
- if your system does not converge within - lets say - 300
electronic steps you have a real problem; 900 makes no sense.
Kind regards
Thomas
On 27.04.2018 16:46, Sohail Ahmad wrote:
> Dear QE users and experts
>
> I am trying to relax the bilayer of PdSe2 after introducing electric
> field,,\
> everytime i am getting the following error . I am using QE 6.0. Its
> perfectly running on other machine..
> %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
> Error in routine ggen 1 (226332):
> too many g-vectors
> %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
>
> Input/output/batchfile are attached
> ----------------------------------------------------------------------------------------------------
> &control
> calculation = 'relax',
> restart_mode = 'from_scratch',
> pseudo_dir= '/home/sohail/scratch/pseudo',
> outdir='./OUT',
> prefix='PdSe2bAAel004',
> tefield = .true.,
> dipfield = .true.,
> etot_conv_thr = 1.0d-5,
> forc_conv_thr = 1.0d-4,
> /
> &system
> ibrav = 4, a = 3.82, b = 3.82, c = 20.0, cosAB = -0.5, cosAC = 0.0,
> cosBC = 0.0,
> nat = 6, ntyp = 2,
> ecutwfc = 80,
> ecutrho = 400,
> nbnd = 40,
> occupations = 'smearing', smearing = 'gaussian', degauss = 0.001,
> edir = 3,
> eamp = 0.004,
> emaxpos = 0.9,
> eopreg = 0.2,
> /
> &electrons
> mixing_beta = 0.3,
> conv_thr = 1.0d-9,
> electron_maxstep = 900,
> /
> &ions
> ion_dynamics = 'bfgs',
> /
> ATOMIC_SPECIES
> Pd 106.42 Pd.pbe-mt_fhi.UPF
> Se 78.96 Se.pbe-mt_fhi.UPF
> ATOMIC_POSITIONS angstrom
> Pd 0.000000000 0.000000000 0.000000000
> Se -1.908913627 1.102119324 1.282624049
> Se 1.908913627 -1.102119324 -1.282624049
> Pd 0.000000000 0.000000000 5.000000000
> Se -1.908913627 1.102119324 6.282624049
> Se 1.908913627 -1.102119324 3.717375951
> K_POINTS AUTOMATIC
> 16 16 4 0 0 0
> --------------------------------------------------------------------------------------------------------------------
> #!/bin/bash
> #SBATCH --nodes=4
> #SBATCH --tasks-per-node=16
> #SBATCH --time=24:00:00
> #SBATCH --job-name=test
> #SBATCH --mem-per-cpu=10000
>
> echo "Nodes I am on:"
> cat $SLURM_JOB_NODELIST
> echo "Current working directory is `pwd`"
> echo "Running on `hostname`"
>
> module load quantumespresso/6.0
>
> echo "Starting run at: `date`"
> srun pw.x < PdSe2bAAel004.rx.in > PdSe2bAAel004.rx.out
>
> echo "Finished run at: `date`"
>
> ---------------------------------------------------------------------------------------------------------------------
>
> Program PWSCF v.6.0 (svn rev. 13079) starts on 27Apr2018 at 7:15:21
>
> This program is part of the open-source Quantum ESPRESSO suite
> for quantum simulation of materials; please cite
> "P. Giannozzi et al., J. Phys.:Condens. Matter 21 395502 (2009);
> URL http://www.quantum-espresso.org",
> in publications or presentations arising from this work. More
> details at
> http://www.quantum-espresso.org/quote
>
> Parallel version (MPI), running on 64 processors
> R & G space division: proc/nbgrp/npool/nimage = 64
> Waiting for input...
> Reading input from standard input
>
> Current dimensions of program PWSCF are:
> Max number of different atomic species (ntypx) = 10
> Max number of k-points (npk) = 40000
> Max angular momentum in pseudopotentials (lmaxx) = 3
> Presently no symmetry can be used with electric field
>
> file Pd.pbe-mt_fhi.UPF: wavefunction(s) 4f renormalized
> file Se.pbe-mt_fhi.UPF: wavefunction(s) 4f renormalized
>
> Subspace diagonalization in iterative solution of the eigenvalue
> problem:
> one sub-group per band group will be used
> scalapack distributed-memory algorithm (size of sub-group: 5* 5
> procs)
>
> Message from routine setup:
> no reason to have ecutrho>4*ecutwfc
>
> Parallelization info
> --------------------
> sticks: dense smooth PW G-vecs: dense smooth PW
> Min 22 18 5 3524 2574 417
> Max 23 19 6 3555 2586 438
> Sum 1417 1159 349 226331 165041 27287
> :
> :
> :
> :
> :
> :
> number of k points= 516 gaussian smearing, width (Ry)= 0.0010
>
> Number of k-points >= 100: set verbosity='high' to print them.
>
> Dense grid: 226331 G-vectors FFT dimensions: ( 45, 45, 243)
>
> Smooth grid: 165041 G-vectors FFT dimensions: ( 45, 45, 216)
>
> Estimated max dynamical RAM per process > 104.83Mb
>
> Estimated total allocated dynamical RAM > 6709.07Mb
>
> %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
> Error in routine ggen 1 (226332):
> too many g-vectors
> %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
>
> stopping ...
>
> ---------------------------------------------------------------------
>
> *Sohail Ahmad*
> Department of Physics
> King Khalid University
> Abha, Saudi Arabia
> --------------------------------------------------------------------
>
>
> _______________________________________________
> users mailing list
> users at lists.quantum-espresso.org
> https://lists.quantum-espresso.org/mailman/listinfo/users
--
Dr. rer. nat. Thomas Brumme
Wilhelm-Ostwald-Institute for Physical and Theoretical Chemistry
Leipzig University
Phillipp-Rosenthal-Strasse 31
04103 Leipzig
Tel: +49 (0)341 97 36456
email: thomas.brumme at uni-leipzig.de
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.quantum-espresso.org/pipermail/users/attachments/20180427/e53782f3/attachment.html>
More information about the users
mailing list