[gmx-users] Maximising Hardware Performance on Local node: Optimal settings
Kevin Boyd
kevin.boyd at uconn.edu
Sat Dec 7 23:45:40 CET 2019
Hi,
I also wrote up some examples on optimizing for multiple simulations on the
same node, see
https://mailman-1.sys.kth.se/pipermail/gromacs.org_gmx-users/2019-July/126007.html
On Wed, Dec 4, 2019 at 9:36 AM Christian Blau <blau at kth.se> wrote:
> Hi Matt,
>
>
> Here are a few bullet points that might help you, maybe other experts can
> contribute more.
>
>
> If you're running on a single machine, using thread-mpi over mpi is a good
> choice.
>
> "-pin on" might help you.
>
> 60k atoms is not very large, here are some other systems ready to
> benchmark https://www.mpibpc.mpg.de/grubmueller/bench
> that be able to tell you more about your performance on a range of systems.
>
> It is normal that the GPU is not fully utilized; the newest GROMACS
> release should be able to make more use of the GPU,
> so you might want to try out the beta-3 version to get an idea, but please
> don't use for production, but wait till
> January when GROMACS-2020 is released.
>
> If you want to maximise sampling, incorporate running multiple simulations
> simultaneously in your benchmark set (mdrun
> -multidir makes things easy here), most often this is what you actually
> want and can give you a drastic increase in
> output from your hardware (guessing a long shot, you might get 4 * 150
> ns/day)
>
>
> I assume you had already a look at this, but for reference check here:
>
>
> http://manual.gromacs.org/documentation/current/user-guide/mdrun-performance.html
>
> http://manual.gromacs.org/documentation/current/onlinehelp/gmx-mdrun.html
>
> http://manual.gromacs.org/documentation/current/user-guide/mdrun-features.html
>
> https://onlinelibrary.wiley.com/doi/abs/10.1002/jcc.26011
>
>
> Best,
>
> Christian
>
> On 2019-12-04 17:53, Matthew Fisher wrote:
> > Dear all,
> >
> > We're currently running some experiments with a new hardware
> configuration and attempting to maximise performance from it. Our system
> contains 1x V100 and 2x 12 core (24 logical) Xeon Silver 4214 CPUs which,
> after optimisation of CUDA drivers & kernels etc., we've been able to get a
> performance of 210 ns/day for 60k atoms with GROMACS 2019.3 (allowing mdrun
> to select threads, which has surprised us as it only creates 24 OpenMP
> threads for our 48 logical core system). Furthermore we have a surprising
> amount of wasted GPU time. Therefore, we were wondering if anyone had any
> advice on how we could maximise our hardware output? We've enclosed the
> real cycle and time accounting display below.
> >
> > Any help will be massively appreciated
> >
> > Thanks,
> > Matt
> >
> > R E A L C Y C L E A N D T I M E A C C O U N T I N G
> >
> > On 1 MPI rank, each using 24 OpenMP threads
> >
> > Computing: Num Num Call Wall time
> Giga-Cycles
> > Ranks Threads Count (s) total sum
> %
> >
> -----------------------------------------------------------------------------
> > Neighbor search 1 24 12501 32.590 1716.686
> 3.2
> > Launch GPU ops. 1 24 2500002 105.169 5539.764
> 10.2
> > Force 1 24 1250001 140.283 7389.414
> 13.6
> > Wait PME GPU gather 1 24 1250001 79.714 4198.902
> 7.7
> > Reduce GPU PME F 1 24 1250001 25.159 1325.260
> 2.4
> > Wait GPU NB local 1 24 1250001 264.961 13956.769
> 25.7
> > NB X/F buffer ops. 1 24 2487501 177.862 9368.871
> 17.3
> > Write traj. 1 24 252 5.748 302.799
> 0.6
> > Update 1 24 1250001 81.151 4274.601
> 7.9
> > Constraints 1 24 1250001 70.231 3699.389
> 6.8
> > Rest 47.521 2503.167
> 4.6
> >
> -----------------------------------------------------------------------------
> > Total 1030.389 54275.623
> 100.0
> >
> -----------------------------------------------------------------------------
> >
> > Core t (s) Wall t (s) (%)
> > Time: 24729.331 1030.389 2400.0
> > (ns/day) (hour/ns)
> > Performance: 209.630 0.114
> --
> Gromacs Users mailing list
>
> * Please search the archive at
> http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before
> posting!
>
> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>
> * For (un)subscribe requests visit
> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or
> send a mail to gmx-users-request at gromacs.org.
>
More information about the gromacs.org_gmx-users
mailing list