[QE-users] Optimal pw command line for large systems and only Gamma point
Antonio Cammarata
cammaant at fel.cvut.cz
Mon May 13 15:15:02 CEST 2024
I did some tests. For 1000 Si atoms, I use 2010 bands because I need to
get the band gap value; moreover, being a cluster, the surface states of
the truncated bonds might close the gap, especially at the first steps
of the geometry optimization, so it's better I use few empty bands. I
managed to run the calculation by using 10 nodes and a max of 40 cores
per node. My question now is: can you suggest me optimal command line
options and/or input settings to speed up the calculation? And, if
possible, also to reduce the number of nodes? The relevant parameters in
the input file are the following:
input_dft= 'pz'
ecutwfc= 25
occupations= 'smearing'
smearing= 'cold'
degauss= 0.05 ! I know it's quite large, but necessary to stabilize
the SCF at this preliminary stage (no geometry step done yet)
nbnd= 2010
diagonalization= 'ppcg'
mixing_mode= 'plain'
mixing_beta= 0.4
The actual time spent per scf cycle is about 33 minutes. I use QE v. 7.3
compiled with openmpi and scalapack. I have access to the intel
compilers too but I did some tests and the difference is just tens of
seconds. I have only the Gamma point; please, here you have some info
about the grid and the estimated RAM usage:
Dense grid: 24616397 G-vectors FFT dimensions: ( 375, 375, 375)
Dynamical RAM for wfc: 235.91 MB
Dynamical RAM for wfc (w. buffer): 235.91 MB
Dynamical RAM for str. fact: 0.94 MB
Dynamical RAM for local pot: 0.00 MB
Dynamical RAM for nlocal pot: 2112.67 MB
Dynamical RAM for qrad: 0.80 MB
Dynamical RAM for rho,v,vnew: 6.04 MB
Dynamical RAM for rhoin: 2.01 MB
Dynamical RAM for rho*nmix: 15.03 MB
Dynamical RAM for G-vectors: 3.99 MB
Dynamical RAM for h,s,v(r/c): 0.46 MB
Dynamical RAM for <psi|beta>: 552.06 MB
Dynamical RAM for wfcinit/wfcrot: 1305.21 MB
Estimated static dynamical RAM per process > 2.31 GB
Estimated max dynamical RAM per process > 3.60 GB
Estimated total dynamical RAM > 1441.34 GB
Thanks a lot in advance for your kind help.
All the best
Antonio
On 10. 05. 24 12:01, Paolo Giannozzi wrote:
> On 5/10/24 08:58, Antonio Cammarata via users wrote:
>
>> pw.x -nk 1 -nt 1 -nb 1 -nd 768 -inp qe.in > qe.out
>
> too many processors for linear-algebra parallelization. 1000 Si atoms
> = 2000 bands (assuming an insulator with no spin polarization). Use a
> few tens of processors at most
>
>> "some processors have no G-vectors for symmetrization".
>
> which sounds strange to me: with the Gamma point symmetrization is not
> even needed
>
>
>> Dense grid: 30754065 G-vectors FFT dimensions: ( 400, 400, 400)
>
> This is what a 256-atom Si supercell with 30 Ry cutoff yields:
>
> Dense grid: 825897 G-vectors FFT dimensions: ( 162, 162, 162)
>
> I guess you may reduce the size of your supercell
>
> Paolo
>
>> Dynamical RAM for wfc: 153.50 MB
>> Dynamical RAM for wfc (w. buffer): 153.50 MB
>> Dynamical RAM for str. fact: 0.61 MB
>> Dynamical RAM for local pot: 0.00 MB
>> Dynamical RAM for nlocal pot: 1374.66 MB
>> Dynamical RAM for qrad: 0.87 MB
>> Dynamical RAM for rho,v,vnew: 5.50 MB
>> Dynamical RAM for rhoin: 1.83 MB
>> Dynamical RAM for rho*nmix: 9.78 MB
>> Dynamical RAM for G-vectors: 2.60 MB
>> Dynamical RAM for h,s,v(r/c): 0.25 MB
>> Dynamical RAM for <psi|beta>: 552.06 MB
>> Dynamical RAM for wfcinit/wfcrot: 977.20 MB
>> Estimated static dynamical RAM per process > 1.51 GB
>> Estimated max dynamical RAM per process > 2.47 GB
>> Estimated total dynamical RAM > 1900.41 GB
>>
>> I managed to run the simulation with 512 atoms, cg diagonalization
>> and 3 nodes on the same machine with command line
>>
>> pw.x -nk 1 -nt 1 -nd 484 -inp qe.in > qe.out
>>
>> Please, do you have any suggestion on how to set optimal
>> parallelization parameters to avoid the memory issue and run the
>> calculation? I am also planning to run simulations on nanoclusters
>> with more than 1000 atoms.
>>
>> Thanks a lot in advance for your kind help.
>>
>> Antonio
>>
>>
>
--
_______________________________________________
Antonio Cammarata, PhD in Physics
Associate Professor in Applied Physics
Advanced Materials Group
Department of Control Engineering - KN:G-204
Faculty of Electrical Engineering
Czech Technical University in Prague
Karlovo Náměstí, 13
121 35, Prague 2, Czech Republic
Phone: +420 224 35 5711
Fax: +420 224 91 8646
ORCID: orcid.org/0000-0002-5691-0682
WoS ResearcherID: A-4883-2014
More information about the users
mailing list