[QE-users] Parallel computing of QE7.1 vc-relax crashes when using large number of processors

Xin Jin xin.tlg.jin at outlook.com
Fri Oct 28 14:45:59 CEST 2022


Dear Quantum Espresso Forum,

I encountered a problem related to the parallel computing using QE7.1 
for vc-relax.

I was trying to perform a vc-relax for a 3*3*3 BCC tungsten super cell. 
The code works fine for non-parallel computing, also works fine for 
parallel computing if the number of processors is smaller than 10.

However, if the number of processors is larger than 10, I will get 
following MPI error:
/*** An error occurred in MPI_Comm_free//
//*** reported by process [3585895498,2]//
//*** on communicator MPI_COMM_WORLD//
//*** MPI_ERR_COMM: invalid communicator//
//*** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,//
//***    and potentially your MPI job)/

For parallel computing, I am using /OpenMPI/3.1.4-gcccuda/. (In 
addition, it seems that If I use OpenMPI V4, the simulation speed will 
be much slower than that of V3.)

Another thing is that, if I decrease the size of the supper cell, for 
example to 2*2*2, then there is no problem in the parallel computing 
even if I use more than 30 processors.

Could you help me look at this problem, please?

The input for QE can be found below.

Thank you in advance!

Xin Jin

/&control//
//
//    calculation='vc-relax' //
//    restart_mode='from_scratch', //
//    prefix='W_relax', //
//    pseudo_dir="../../PP_files",//
//    outdir='./'//
//
// ///
////
//
// &system//
//    ibrav= 0, //
//    celldm(1)=5.972,//
//    nat=  54, //
//    ntyp= 1,//
//    ecutwfc = 50,//
//    ecutrho = 500,//
//    occupations='smearing', smearing='mp', degauss=0.06//
// ///
//
// &electrons//
//    diagonalization='david',//
//    conv_thr =  1.0d-8,//
//    mixing_beta = 0.5,//
// ///
////
// &ions//
// ///
//
// &cell//
//    press = 0.0,//
// ///
////
//ATOMIC_SPECIES//
// W  183.84 W.pbe-spn-kjpaw_psl.1.0.0.UPF//
////
//CELL_PARAMETERS {alat}//
//   3.0  0.0  0.0//
//   0.0  3.0  0.0//
//   0.0  0.0  3.0 //
////
//ATOMIC_POSITIONS {alat}//
//W 0.00000 0.00000 0.00000//
//W 0.50000 0.50000 0.50000//
//W 1.00000 0.00000 0.00000//
//W 1.50000 0.50000 0.50000//
//W 2.00000 0.00000 0.00000//
//W 2.50000 0.50000 0.50000//
//W 0.00000 1.00000 0.00000//
//W 0.50000 1.50000 0.50000//
//W 1.00000 1.00000 0.00000//
//W 1.50000 1.50000 0.50000//
//W 2.00000 1.00000 0.00000//
//W 2.50000 1.50000 0.50000//
//W 0.00000 2.00000 0.00000//
//W 0.50000 2.50000 0.50000//
//W 1.00000 2.00000 0.00000//
//W 1.50000 2.50000 0.50000//
//W 2.00000 2.00000 0.00000//
//W 2.50000 2.50000 0.50000//
//W 0.00000 0.00000 1.00000//
//W 0.50000 0.50000 1.50000//
//W 1.00000 0.00000 1.00000//
//W 1.50000 0.50000 1.50000//
//W 2.00000 0.00000 1.00000//
//W 2.50000 0.50000 1.50000//
//W 0.00000 1.00000 1.00000//
//W 0.50000 1.50000 1.50000//
//W 1.00000 1.00000 1.00000//
//W 1.50000 1.50000 1.50000//
//W 2.00000 1.00000 1.00000//
//W 2.50000 1.50000 1.50000//
//W 0.00000 2.00000 1.00000//
//W 0.50000 2.50000 1.50000//
//W 1.00000 2.00000 1.00000//
//W 1.50000 2.50000 1.50000//
//W 2.00000 2.00000 1.00000//
//W 2.50000 2.50000 1.50000//
//W 0.00000 0.00000 2.00000//
//W 0.50000 0.50000 2.50000//
//W 1.00000 0.00000 2.00000//
//W 1.50000 0.50000 2.50000//
//W 2.00000 0.00000 2.00000//
//W 2.50000 0.50000 2.50000//
//W 0.00000 1.00000 2.00000//
//W 0.50000 1.50000 2.50000//
//W 1.00000 1.00000 2.00000//
//W 1.50000 1.50000 2.50000//
//W 2.00000 1.00000 2.00000//
//W 2.50000 1.50000 2.50000//
//W 0.00000 2.00000 2.00000//
//W 0.50000 2.50000 2.50000//
//W 1.00000 2.00000 2.00000//
//W 1.50000 2.50000 2.50000//
//W 2.00000 2.00000 2.00000//
//W 2.50000 2.50000 2.50000//
//
//K_POINTS {automatic}//
//4 4 4 0 0 0//
/
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.quantum-espresso.org/pipermail/users/attachments/20221028/87b8c648/attachment.html>


More information about the users mailing list