<div>Dear all<br></div><div><br></div><div>I tried to use full paths, but it didn't give positive results. It wrote an error message </div><div><br></div><div>application called MPI_Abort(MPI_COMM_WORLD, 0) - process 0</div>
<div><br></div><br><div class="gmail_quote">On 7 March 2011 10:30, Alexander Kvashnin <span dir="ltr"><<a href="mailto:agkvashnin@gmail.com">agkvashnin@gmail.com</a>></span> wrote:<br><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex;">
<div><div class="h5">Thanks, I tried to use "<" instead of "-in" it also didn't work.<br>
OK,I will try to use full paths for input and output, and answer about result.<br>
<br>
----- Исходное сообщение -----<br>
От: Omololu Akin-Ojo <<a href="mailto:prayerz.omo@gmail.com">prayerz.omo@gmail.com</a>><br>
Отправлено: 7 марта 2011 г. 9:56<br>
Кому: PWSCF Forum <<a href="mailto:pw_forum@pwscf.org">pw_forum@pwscf.org</a>><br>
Тема: Re: [Pw_forum] НА: Re: problem in MPI running of QE (16 processors)<br>
<br>
Try to see if specifying the full paths help.<br>
E.g., try something like:<br>
<br>
mpiexec /home/MyDir/bin/pw.x -in /scratch/MyDir/graph.inp ><br>
/scratch/MyDir/graph.out<br>
<br>
(where /home/MyDir/bin is the full path to your pw.x and<br>
/scratch/MyDir/graph.inp is the full path to your output ....)<br>
<br>
( I see you use "-in" instead of "<" to indicate the input. I don't<br>
know too much but _perhaps_ you could also _try_ using "<" instead of<br>
"-in") .<br>
<br>
o.<br>
<br>
On Mon, Mar 7, 2011 at 7:31 AM, Alexander Kvashnin <<a href="mailto:agkvashnin@gmail.com">agkvashnin@gmail.com</a>> wrote:<br>
> Yes, I wrote<br>
><br>
> #PBS -l nodes=16:ppn=4<br>
><br>
> And in userguide of MIPT-60 wrote,that mpiexec must choose number of<br>
> processors automatically, that's why I didn't write anything else<br>
><br>
><br>
> ________________________________<br>
> От: Huiqun Zhou <<a href="mailto:hqzhou@nju.edu.cn">hqzhou@nju.edu.cn</a>><br>
> Отправлено: 7 марта 2011 г. 7:52<br>
> Кому: PWSCF Forum <<a href="mailto:pw_forum@pwscf.org">pw_forum@pwscf.org</a>><br>
> Тема: Re: [Pw_forum] problem in MPI running of QE (16 processors)<br>
><br>
> How did you apply number of node, procs per node in your job<br>
> script?<br>
><br>
> #PBS -l nodes=?:ppn=?<br>
><br>
> zhou huiqun<br>
> @earth sciences, nanjing university, china<br>
><br>
><br>
> ----- Original Message -----<br>
> From: Alexander G. Kvashnin<br>
> To: PWSCF Forum<br>
> Sent: Saturday, March 05, 2011 2:53 AM<br>
> Subject: Re: [Pw_forum] problem in MPI running of QE (16 processors)<br>
> I create PBS task on supercomputer MIPT-60 where I write<br>
><br>
> mpiexec ../../espresso-4.2.1/bin/pw.x -in graph.inp > output.opt<br>
> all other<br>
<br>
[Включен не весь текст исходного сообщения]</div></div></blockquote></div><br><br clear="all"><br>-- <br><div>Sincerely yours</div><div>Alexander G. Kvashnin</div><div>--------------------------------------------------------------------------------------------------------------------------------</div>
<div>Student</div><div>Moscow Institute of Physics and Technology <a href="http://mipt.ru/" target="_blank">http://mipt.ru/</a></div><div>141700, Institutsky lane 9, Dolgoprudny, Moscow Region, Russia<br></div><div>
<br></div><div>Junior research scientist<br></div><div>Technological Institute for Superhard </div><div>and Novel Carbon Materials <a href="http://www.ntcstm.troitsk.ru/" target="_blank">http://www.ntcstm.troitsk.ru/</a></div>
<div>142190, Central'naya St. 7a, Troitsk, Moscow Region, Russia</div><div>================================================================</div><br>