[gmx-users] gigabit scaling, DPPC benchmark
Jason de Joannis
jdejoan at emory.edu
Mon Jun 6 20:57:43 CEST 2005
> Date: Mon, 06 Jun 2005 19:12:24 +0200
> From: David <spoel at xray.bmc.uu.se>
> Subject: Re: [gmx-users] gigabit scaling, DPPC benchmark
> To: Discussion list for GROMACS users <gmx-users at gromacs.org>
> Message-ID: <1118077944.5498.2.camel at localhost.localdomain>
> Content-Type: text/plain
>
> On Mon, 2005-06-06 at 12:48 -0400, Jason de Joannis wrote:
> > Hi, I would like to improve the scaling on my cluster so it is more
> > like that published on the Gromacs website.
> >
> > Cluster Nodes: Dual Xeon 3.06 GHz (hyperthreaded)
> > Network: Gbit ethernet
> > System Installion:
> > o Linux 2.4.20-64GB-SMP
> > o LAM 6.5.9/MPI 2 C++/ROMIO
> > o fftw-2.1.5
> > o Gromacs 3.2.1
> >
> > For the DPPC benchmark with sort/shuffle:
> >
> > NP 1 2 4
> > ps/day 158 404 517
> > scaling 128% 81%
> >
> > There is a big drop from 2 to 4 as the Gigabit network comes into play.
> > Compare this to the Xeon 2800 benchmark on Gromacs.org where 4 processors
> > get 637 ps/day.
> >
> > Of course the problem becomes more pronounced when I use PME on this
> > benchmark:
> >
> > NP 1 2 4
> > ps/day 115 229 201
> > scaling 100% 44%
> >
> > I suspect my network may be running a 100 rather than 1000 Mbit as
> > suggested by the published 87% scaling for Xeon2800/100Mbit.
> >
> > How can a troubleshoot this problem?
> I'm afraid this is a GROMACS problem. We are working on resolving this,
> but the problem is largely due to the Linux TCP layer which is very
> inefficient as regards the start-up time for a communication.
>
Then why does the Xeon2800/1000Gbit system on gromacs.org scale so well?
I didn't see any special comments about its network.
> It basically means you can not do much about it (except buy SMP machines
> with more than two processors).
> >
> > /Jason
> >
> --
> David.
/Jason
More information about the gromacs.org_gmx-users
mailing list