[Pw_forum] Large difference between the CPU and wall time in the c_bands and sum_bands routines
Harsha Vardhan
harshavardhan9510 at gmail.com
Thu Jun 29 09:57:45 CEST 2017
Dear QE users,
I am running an scf calculation using the pw.x executable for a distorted
graphene sheet, using the ultrasoft pseudopotential file
C_pbe_v1.2.uspp.F.UPF. I am using a large supercell (128 atoms).
I am using the following values for the energy cutoffs: 60.0 ecutwfc, 500.0
ecutrho, with gaussian smearing and Davidson diagonalization. I am using a
9x9x1 Monkhorst-Pack k-point grid.
I am running the calculations on a node in a cluster, which contains 16
processors. I am running it with a 16-process MPIrun.
I have observed that the c_bands and sum_bands routines are taking up a
huge amount of wall time, as compared to the CPU time. I am attaching the
time report for the completed calculation below:
init_run : 1327.31s CPU 1554.01s WALL ( 1 calls)
electrons : 130443.79s CPU 771483.34s WALL ( 1 calls)
Called by init_run:
wfcinit : 1299.34s CPU 1524.78s WALL ( 1 calls)
potinit : 5.56s CPU 5.65s WALL ( 1 calls)
Called by electrons:
c_bands : 97779.04s CPU 441468.41s WALL ( 29 calls)
sum_band : 28359.60s CPU 311790.57s WALL ( 29 calls)
v_of_rho : 3421.30s CPU 4610.39s WALL ( 30 calls)
newd : 58.45s CPU 1180.47s WALL ( 30 calls)
mix_rho : 559.21s CPU 9112.55s WALL ( 29 calls)
Called by c_bands:
init_us_2 : 157.78s CPU 6714.76s WALL ( 2950 calls)
cegterg : 72929.24s CPU 101098.23s WALL ( 1463 calls)
Called by sum_band:
sum_band:bec : 4.65s CPU 39.29s WALL ( 1450 calls)
addusdens : 546.07s CPU 26711.20s WALL ( 29 calls)
Called by *egterg:
h_psi : 49261.91s CPU 58153.26s WALL ( 7375 calls)
s_psi : 5325.93s CPU 5437.34s WALL ( 7375 calls)
g_psi : 65.05s CPU 72.50s WALL ( 5862 calls)
cdiaghg : 3142.76s CPU 3801.60s WALL ( 7325 calls)
Called by h_psi:
add_vuspsi : 5378.26s CPU 5626.14s WALL ( 7375 calls)
General routines
calbec : 10792.25s CPU 15224.73s WALL ( 8825 calls)
fft : 4271.25s CPU 4843.82s WALL ( 920 calls)
ffts : 337.97s CPU 1110.36s WALL ( 118 calls)
fftw : 42573.09s CPU 47318.16s WALL ( 2685986 calls)
interpolate : 598.65s CPU 3372.52s WALL ( 118 calls)
davcio : 0.02s CPU 442.60s WALL ( 50 calls)
Parallel routines
fft_scatter : 37432.00s CPU 42982.58s WALL ( 2687024 calls)
PWSCF : 1d 13h 3m CPU 9d 3h33m WALL
This run was terminated on: 4:55:36 26Jun2017
=------------------------------------------------------------------------------=
JOB DONE.
=------------------------------------------------------------------------------=
As it can be seen from the report, there is a large discrepancy between the
CPU, and wall time.
When I run the same calculation, with a smaller k-point grid (6x6x1), and
smaller ecutwfc = 40.0 and ecutrho= 400.0, the CPU time is only slightly
lesser than the wall time (just 1 hour or so difference between the two).
Kindly inform me what might be causing the large delay between the CPU time
and wall time, and how I might minimize it.
Thanking You,
Yours Sincerely,
M Harshavardhan
Fourth Year Undergraduate
Engineering Physics
IIT Madras
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.quantum-espresso.org/pipermail/users/attachments/20170629/85d833ad/attachment.html>
More information about the users
mailing list