[gmx-users] Gromacs 4.5.4 on multi-node cluster

Mark Abraham Mark.Abraham at anu.edu.au
Wed Dec 7 15:53:49 CET 2011


On 8/12/2011 1:26 AM, Nikos Papadimitriou wrote:
> Dear All,
>
> I had been running Gromacs 4.0.7 on a 12-node cluster (Intel i7-920 
> 4-cores) with OS Rocks 5.4.2. Recently, I have upgraded the cluster OS 
> to Rocks 5.4.3 and I have installed Gromacs 4.5.4 from the Bio Roll 
> repository. When running in parallel on the same node, everything 
> works fine. However, when I am trying to run on more than one nodes 
> the run stalls immediately with the following message:
>
> [gromacs at tornado Test]$ /home/gromacs/.Installed/openmpi/bin/mpirun 
> -np 2 -machinefile machines 
> /home/gromacs/.Installed/gromacs/bin/mdrun_mpi -s md_run.tpr -o 
> md_traj.trr -c md_confs.gro -e md.edr -g md.log -v
> NNODES=2, MYRANK=0, HOSTNAME=compute-1-1.local
> NNODES=2, MYRANK=1, HOSTNAME=compute-1-2.local
> NODEID=0 argc=12
> NODEID=1 argc=12
>
> The mdrun_mpi thread seems to start in both nodes but the run does not 
> go on and no file is produced. It seems that the nodes are waiting for 
> some kind of communication between them. The problem occurs even for 
> the simplest case (i.e. NVT simulation of 1000 Argon atoms without 
> Coulombic interactions). Openmpi and networking between the nodes seem 
> to work fine since there are not any problems with other software that 
> run with MPI.

Can you run 2-processor MPI test program with that machine file?

Mark

>
> In an attempt to find a solution, I have manually compiled and 
> installed Gromacs 4.5.5 (with --enable-mpi) after having installed the 
> latest version of openmpi and fftw3 and no error occurred during the 
> installation. However, when trying to run on two different nodes 
> exactly the same problem appears.
>
> Have you any idea what might cause this situation?
> Thank you in advance!
>
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://maillist.sys.kth.se/pipermail/gromacs.org_gmx-users/attachments/20111208/f9c3f4f5/attachment.html>


More information about the gromacs.org_gmx-users mailing list