[Pw_forum] Benchmarking: openmpi-1.3 vs openmpi-1.2.8

Mahmoud Payami mpayami at aeoi.org.ir
Sun Feb 15 14:06:47 CET 2009




 Dear ALL,

 I performed a benchmarking test using openmpi-1.2.8 and openmpi-1.3 for a
 2-atom Au cluster.
 The out files show that in all parts "1.3" takes shorter time than "1.2.8".
 However, the wall time in "1.3" is slightly increased. I do not know how to
 resolve it. Is the bottleneck the connection cables?
 Bests,
          Mahmoud Payami

 ----------------------------------------
 #out_1.2.8
 PWSCF : 2m 5.29s CPU time, 4m43.10s wall time

 init_run : 9.50s CPU

 electrons : 108.91s CPU

 forces : 5.90s CPU

 Called by init_run:

 wfcinit : 1.58s CPU

 potinit : 3.84s CPU

 Called by electrons:

 c_bands : 32.38s CPU ( 10 calls, 3.238 s avg)

 sum_band : 29.74s CPU ( 10 calls, 2.974 s avg)

 v_of_rho : 25.64s CPU ( 10 calls, 2.564 s avg)

 newd : 12.15s CPU ( 10 calls, 1.215 s avg)

 mix_rho : 5.26s CPU ( 10 calls, 0.526 s avg)

 Called by c_bands:

 init_us_2 : 0.12s CPU ( 21 calls, 0.006 s avg)

 regterg : 32.10s CPU ( 10 calls, 3.210 s avg)

 Called by *egterg:

 h_psi : 31.51s CPU ( 43 calls, 0.733 s avg)

 s_psi : 0.20s CPU ( 43 calls, 0.005 s avg)

 g_psi : 0.06s CPU ( 32 calls, 0.002 s avg)

 rdiaghg : 1.19s CPU ( 41 calls, 0.029 s avg)

 Called by h_psi:

 add_vuspsi : 0.17s CPU ( 43 calls, 0.004 s avg)

 General routines

 calbec : 2.26s CPU ( 57 calls, 0.040 s avg)

 cft3s : 82.62s CPU ( 595 calls, 0.139 s avg)

 interpolate : 9.35s CPU ( 20 calls, 0.467 s avg)

 davcio : 0.00s CPU ( 9 calls, 0.000 s avg)


 Parallel routines

 fft_scatter : 64.88s CPU ( 595 calls, 0.109 s avg)

 ---------------------------------------------------------------------------------------- #out-1.3 PWSCF : 1m55.59s CPU time, 4m44.04s wall time init_run : 8.69s CPU electrons : 100.47s CPU forces : 5.55s CPU Called by init_run: wfcinit : 1.46s CPU potinit : 3.41s CPU Called by electrons: c_bands : 30.14s CPU ( 10 calls, 3.014 s avg) sum_band : 27.11s CPU ( 10 calls, 2.711 s avg) v_of_rho : 23.62s CPU ( 10 calls, 2.362 s avg) newd : 11.95s CPU ( 10 calls, 1.195 s avg) mix_rho : 4.47s CPU ( 10 calls, 0.447 s avg) Called by c_bands: init_us_2 : 0.10s CPU ( 21 calls, 0.005 s avg) regterg : 29.87s CPU ( 10 calls, 2.987 s avg) Called by *egterg: h_psi : 29.42s CPU ( 43 calls, 0.684 s avg) s_psi : 0.18s CPU ( 43 calls, 0.004 s avg) g_psi : 0.06s CPU ( 32 calls, 0.002 s avg) rdiaghg : 1.05s CPU ( 41 calls, 0.026 s avg) Called by h_psi: add_vuspsi : 0.16s CPU ( 43 calls, 0.004 s avg) General routines calbec : 1.63s CPU ( 57 calls, 0.029 s avg) cft3s : 75.61s CPU ( 595 calls, 0.127 s avg) interpolate : 8.18s CPU ( 20 calls, 0.409 s avg) davcio : 0.00s CPU ( 9 calls, 0.000 s avg) Parallel routines fft_scatter : 57.22s CPU ( 595 calls, 0.096 s avg)>>>> Has anybody succeeded running QE using openmpi-1.3?>>> I get the following error message: [...]>>> from read_namelists: error  #1>>> reading namelist control [...]>>> On the other hand it is ok with openmpi-1.2.8.>>>> try pw.x/cp.x -inp "input file name" . It happens more often than>> not that MPI installations are confused by input redirection>> (pw.x/cp.x < "input file name"). Maybe your newer version is>> not installed in exactly the same way as your previous one>>>> Paolo>> --->> Paolo Giannozzi, Democritos and University of Udine, Italy>>>>>> _______________________________________________>> Pw_forum mailing list>> Pw_forum at pwscf.org>> http://www.democritos.it/mailman/listinfo/pw_forum>>>>>




More information about the users mailing list