[gmx-users] parallel run breaks !!

Mark Abraham Mark.Abraham at anu.edu.au
Tue Apr 19 02:43:16 CEST 2011


On 4/18/2011 4:39 PM, delara aghaie wrote:
>
>
> Dear Gromacs users
> I run DPPC monolayer on Tip4p2005 water layer system with gromacs 3.3
> I had no problem before but when I run the system with command
> grompp_mpi -c .gro -f .mdp -n .ndx -p .top -o .tpr -np 8
> and then qsub .ll
> the run starts but it has break down several times, each ime after for 
> ex 987654 steps or more steps or less.
>

If you've changed nothing, and it used to work, and now doesn't, then 
something about your computer has changed. We can't really help there - 
and there must be better people to ask.

Mark

> I get in .o.ii file this message:
> ----
> PBS has allocated the following nodes:
> cx1-25-8-4
> cx1-25-8-4
> cx1-25-8-4
> cx1-25-8-4
> cx1-25-8-4
> cx1-25-8-4
> cx1-25-8-4
> cx1-25-8-4
> ---
> mpiexec.intel -genv I_MPI_DEVICE ssm -genv I_MPI_DEBUG 0 -genv 
> I_MPI_PIN yes -genv I_MPI_PIN_MODE lib -genv I_MPI_FALLBACK_DEVICE 
> disable -genv DAPL_MAX_CM_RETRIES 500 -genv DAPL_MAX_CM_RESPONSE_TIME 
> 300 -genv I_MPI_DAPL_CONNECTION_TIMEOUT 300 -machinefile 
> /tmp/pbs.5455562.cx1/tmp.VRhkM11825 -n 8 -wdir /tmp/pbs.5455562.cx1 
> mdrun_mpi -v -s topol.tpr -np 8
> Job output begins below
> -----------------------
> mpdallexit: cannot connect to local mpd 
> (/tmp/pbs.5455562.cx1/mpd2.console_dmohamma_110411.074928); possible 
> causes:
>   1. no mpd is running on this host
>   2. an mpd is running but was started without a "console" (-n option)
> ---
> what can be the solvation of the possible problem?
> Thanks
> D. Aghaie
>




-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://maillist.sys.kth.se/pipermail/gromacs.org_gmx-users/attachments/20110419/48519c74/attachment.html>


More information about the gromacs.org_gmx-users mailing list