[gmx-users] hyperthreading

Berk Hess gmx3 at hotmail.com
Mon Aug 9 17:52:55 CEST 2010


Hi,

In Gromacs 4.5 there is no difference, since it does not use real thread parallelization.
Gromacs 4.5 has a built-in threaded MPI library, but openmpi also has an efficient
MPI implementation for shared memory machines. But even with proper thread
parallelization I expect the same 15 to 20% performance improvement.
(I guess the lower 8% number is because of loss when going from 8 to 16 processes
for mdrun for that particular system)

Berk

> Date: Mon, 9 Aug 2010 11:36:35 -0400
> From: chris.neale at utoronto.ca
> To: gmx-users at gromacs.org
> Subject: [gmx-users] hyperthreading
> 
> I haven't tried mdrun -nt based hyperthreading, but I have tested  
> using -np 16 on an 8 core box. I get an 8% to 18% performance increase  
> when using -np 16 and optimizing -npme as compared to -np 8 and  
> optimizing -npme. This is on a cluster of Intel Xeon E5540 aka  
> "Nehalem" with 2 quad cores in a single box and using gromacs 4.0.7. I  
> now regularly overload the number of processes.
> 
> selected examples:
> System A with 250,000 atoms:
> mdrun -np 8  -npme -1    1.15 ns/day
> mdrun -np 8  -npme  2    1.02 ns/day
> mdrun -np 16 -npme  2    0.99 ns/day
> mdrun -np 16 -npme  4    1.36 ns/day <-- 118 % performance vs 1.15 ns/day
> mdrun -np 15 -npme  3    1.32 ns/day
> 
> System B with 35,000 atoms (4 fs timestep):
> mdrun -np 8  -npme -1    22.66 ns/day
> mdrun -np 8  -npme  2    23.06 ns/day
> mdrun -np 16 -npme -1    22.69 ns/day
> mdrun -np 16 -npme  4    24.90 ns/day <-- 108 % performance vs 23.06 ns/day
> mdrun -np 56 -npme 16    14.15 ns/day
> 
> Cutoffs and timesteps differ between these runs, but both use PME and  
> explicit water.
> 
> I'd be interested in hearing about any comparisons between -np based  
> process overloading and -nt based hyperthreading.
> 
> Hope it helps,
> Chris.
> 
> -- original message --
> 
> Hi,
> 
> These are nehalem Xeons I presume?
> Then you get 15 to 20% more performance in Gromacs running 2 vs 1  
> thread or process per physical core.
> 
> Berk
> 
> > Date: Fri, 6 Aug 2010 09:24:11 -0500
> > From: dmobley at gmail.com
> > To: gmx-users at gromacs.org
> > Subject: [gmx-users] hyperthreading
> >
> > Dear All,
> >
> > I'm putting together a new Dell Xeon cluster running ROCKS 5.3 which
> > uses CENTOS (6 I believe). This is currently ~20 dual quad-cores with
> > roughly 16 GB of RAM each.
> >
> > In any case, I wanted to inquire about hyperthreading. Does anyone
> > have experience on similar machines with vs. without hyperthreading?
> > The ROCKS users list suggests that hyperthreading ought always be off
> > for HPC applications, which sounds overly simplistic to me, though I
> > more or less follow the logic of this.
> >
> > So, has anyone done any benchmarking yet in a similar setting, and
> > what thoughts do you have? I obviously can do some benchmarking myself
> > as well but I thought I'd check in with the list first.
> >
> > Thanks so much,
> > David
> >
> >
> > -- 
> > David Mobley
> > dmobley at gmail.com
> > 504-383-3662
> 
> 
> -- 
> gmx-users mailing list    gmx-users at gromacs.org
> http://lists.gromacs.org/mailman/listinfo/gmx-users
> Please search the archive at http://www.gromacs.org/search before posting!
> Please don't post (un)subscribe requests to the list. Use the 
> www interface or send it to gmx-users-request at gromacs.org.
> Can't post? Read http://www.gromacs.org/mailing_lists/users.php
 		 	   		  
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://maillist.sys.kth.se/pipermail/gromacs.org_gmx-users/attachments/20100809/c91285b0/attachment.html>


More information about the gromacs.org_gmx-users mailing list