[gmx-users] Gromacs 5.0 compilation slower than 4.6.5. What went wrong ?

Carsten Kutzner ckutzne at gwdg.de
Fri Sep 5 15:58:49 CEST 2014


Hi,

you might want to use g_tune_pme to find out the optimal number
of PME nodes for 4.6 and 5.0.

Carsten



On 05 Sep 2014, at 15:39, David McGiven <davidmcgivenn at gmail.com> wrote:

> What is even more strange is that I tried with 10 pme nodes (mdrun -ntmpi
> 48 -v -c TEST_md.gro -npme 16), got a 15,8% performance loss and ns/day are
> very similar : 33 ns/day
> 
> D.
> 
> 2014-09-05 14:54 GMT+02:00 David McGiven <davidmcgivenn at gmail.com>:
> 
>> Hi Abhi,
>> 
>> Yes I noticed that imbalance but I thought gromacs knew better than the
>> user how to split PP/PME!!
>> 
>> How is it possible that 4.6.5 guesses better than 5.0 ?
>> 
>> Anyway, I tried :
>> mdrun -nt 48 -v -c test.out
>> 
>> Exits with an error "You need to explicitly specify the number of MPI
>> threads (-ntmpi) when using separate PME ranks"
>> 
>> Then:
>> mdrun -ntmpi 48 -v -c TEST_md.gro -npme 12
>> 
>> Then again 35 ns/day with the warning :
>> NOTE: 8.5 % performance was lost because the PME ranks
>>      had less work to do than the PP ranks.
>>      You might want to decrease the number of PME ranks
>>      or decrease the cut-off and the grid spacing.
>> 
>> 
>> I don't know much about Gromacs so I am puzzled.
>> 
>> 
>> 
>> 
>> 2014-09-05 14:32 GMT+02:00 Abhi Acharya <abhi117acharya at gmail.com>:
>> 
>>> Hello,
>>> 
>>> From the log files it is clear that out of 48 cores, the 5.0 run had 8
>>> cores allocated to PME while the 4.6.5 run had 12 cores. This seems to
>>> have
>>> caused a greater load imbalance in case of the 5.0 run.
>>> 
>>> If you notice the last table in both .mdp files, you will notice that the
>>> PME spread/gather wall time values for 5.0 is more than double the wall
>>> time value in case of the 4.6.5.
>>> 
>>> Try running the simulation by explicitly setting the -npme flag as 12.
>>> 
>>> Regards,
>>> Abhishek Acharya
>>> 
>>> 
>>> On Fri, Sep 5, 2014 at 4:43 PM, David McGiven <davidmcgivenn at gmail.com>
>>> wrote:
>>> 
>>>> Thanks Szilard, here it goes! :
>>>> 
>>>> 4.6.5 : http://pastebin.com/nqBn3FKs
>>>> 5.0 : http://pastebin.com/kR4ntHtK
>>>> 
>>>> 2014-09-05 12:47 GMT+02:00 Szilárd Páll <pall.szilard at gmail.com>:
>>>> 
>>>>> mdrun writes a log file, named md.log by default, which contains among
>>>>> other things results of hardware detection and performance
>>>>> measurements. The list does not accept attachments, please upload it
>>>>> somewhere (dropbox, pastebin, etc.) and post a link.
>>>>> 
>>>>> Cheers,
>>>>> --
>>>>> Szilárd
>>>>> 
>>>>> 
>>>>> On Fri, Sep 5, 2014 at 12:37 PM, David McGiven <
>>> davidmcgivenn at gmail.com>
>>>>> wrote:
>>>>>> Command line in both cases is :
>>>>>> 1st :     grompp -f grompp.mdp -c conf.gro -n index.ndx
>>>>>> 2nd :    mdrun -nt 48 -v -c test.out
>>>>>> 
>>>>>> Log file you mean the standard output/error ? Attached to the email
>>> ?
>>>>>> 
>>>>>> Thanks
>>>>>> 
>>>>>> 2014-09-05 12:30 GMT+02:00 Szilárd Páll <pall.szilard at gmail.com>:
>>>>>> 
>>>>>>> Please post the command lines you used to invoke mdrun as well as
>>> the
>>>>>>> log files of the runs you are comparing.
>>>>>>> 
>>>>>>> Cheers,
>>>>>>> --
>>>>>>> Szilárd
>>>>>>> 
>>>>>>> 
>>>>>>> On Fri, Sep 5, 2014 at 12:10 PM, David McGiven <
>>>> davidmcgivenn at gmail.com
>>>>>> 
>>>>>>> wrote:
>>>>>>>> Dear Gromacs users,
>>>>>>>> 
>>>>>>>> I just compiled gromacs 5.0 with the same compiler (gcc 4.7.2),
>>> same
>>>>> OS
>>>>>>>> (RHEL 6) same configuration options and basically everything
>>> than my
>>>>>>>> previous gromacs 4.6.5 compilation and when doing one of our
>>> typical
>>>>>>>> simulations, I get worst performance.
>>>>>>>> 
>>>>>>>> 4.6.5 does 45 ns/day
>>>>>>>> 5.0 does 35 ns/day
>>>>>>>> 
>>>>>>>> Do you have any idea of what could be happening ?
>>>>>>>> 
>>>>>>>> Thanks.
>>>>>>>> 
>>>>>>>> Best Regards,
>>>>>>>> D.
>>>>>>>> --
>>>>>>>> Gromacs Users mailing list
>>>>>>>> 
>>>>>>>> * Please search the archive at
>>>>>>> http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before
>>>>>>> posting!
>>>>>>>> 
>>>>>>>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>>>>>>> 
>>>>>>>> * For (un)subscribe requests visit
>>>>>>>> 
>>> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users
>>>> or
>>>>>>> send a mail to gmx-users-request at gromacs.org.
>>>>>>> --
>>>>>>> Gromacs Users mailing list
>>>>>>> 
>>>>>>> * Please search the archive at
>>>>>>> http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before
>>>>>>> posting!
>>>>>>> 
>>>>>>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>>>>>> 
>>>>>>> * For (un)subscribe requests visit
>>>>>>> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users
>>> or
>>>>>>> send a mail to gmx-users-request at gromacs.org.
>>>>>>> 
>>>>>> --
>>>>>> Gromacs Users mailing list
>>>>>> 
>>>>>> * Please search the archive at
>>>>> http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before
>>>>> posting!
>>>>>> 
>>>>>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>>>>> 
>>>>>> * For (un)subscribe requests visit
>>>>>> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users
>>> or
>>>>> send a mail to gmx-users-request at gromacs.org.
>>>>> --
>>>>> Gromacs Users mailing list
>>>>> 
>>>>> * Please search the archive at
>>>>> http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before
>>>>> posting!
>>>>> 
>>>>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>>>> 
>>>>> * For (un)subscribe requests visit
>>>>> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or
>>>>> send a mail to gmx-users-request at gromacs.org.
>>>>> 
>>>> --
>>>> Gromacs Users mailing list
>>>> 
>>>> * Please search the archive at
>>>> http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before
>>>> posting!
>>>> 
>>>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>>> 
>>>> * For (un)subscribe requests visit
>>>> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or
>>>> send a mail to gmx-users-request at gromacs.org.
>>>> 
>>> --
>>> Gromacs Users mailing list
>>> 
>>> * Please search the archive at
>>> http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before
>>> posting!
>>> 
>>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>> 
>>> * For (un)subscribe requests visit
>>> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or
>>> send a mail to gmx-users-request at gromacs.org.
>>> 
>> 
>> 
> -- 
> Gromacs Users mailing list
> 
> * Please search the archive at http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before posting!
> 
> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
> 
> * For (un)subscribe requests visit
> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or send a mail to gmx-users-request at gromacs.org.


--
Dr. Carsten Kutzner
Max Planck Institute for Biophysical Chemistry
Theoretical and Computational Biophysics
Am Fassberg 11, 37077 Goettingen, Germany
Tel. +49-551-2012313, Fax: +49-551-2012302
http://www.mpibpc.mpg.de/grubmueller/kutzner
http://www.mpibpc.mpg.de/grubmueller/sppexa



More information about the gromacs.org_gmx-users mailing list