[gmx-users] GPU and MPI
Carsten Kutzner
ckutzne at gwdg.de
Mon Sep 1 17:14:33 CEST 2014
On 01 Sep 2014, at 15:58, Da-Wei Li <lidawei at gmail.com> wrote:
> No. With GPU, both domain and PME domain are decomposited by 4X1X1, because
> I use 4 MPI ranks, in line with 4 GPUs. W/o GPU, domain decomposition is
> 20X1X1 and PME is 4X1X1.
So the difference in performance could be due to the different DD and
PME/PP settings. I would try to use exactly the same settings with and
without GPU. With GPUs, you then would need to specify something like
mpirun -n 24 mdrun -dd 20 1 1 -npme 4 -gpu_id 0000011111
So you get 10 DD plus 2 PME ranks per node and map the first 5 DD ranks
to GPU id 0, and the last 5 DD ranks to GPU id 1.
Carsten
>
> dawei
>
>
> On Mon, Sep 1, 2014 at 8:39 AM, Carsten Kutzner <ckutzne at gwdg.de> wrote:
>
>> Hi Dawei,
>>
>> on two nodes, regarding the cases with and without GPUs,
>> do you use the same domain decomposition in both cases?
>>
>> Carsten
>>
>>
>> On 01 Sep 2014, at 14:30, Da-Wei Li <lidawei at gmail.com> wrote:
>>
>>> I have added " -resethway" but still the same result. Use two GPU and 12
>>> cores distributed in 2 nodes will result 33 ns/day, that is, it is about
>> 3
>>> time slower than MD run on one node (2GPU+12core).
>>>
>>> I have no idea what is wrong.
>>>
>>>
>>> dawei
>>>
>>>
>>> On Mon, Sep 1, 2014 at 5:34 AM, Carsten Kutzner <ckutzne at gwdg.de> wrote:
>>>
>>>> Hi,
>>>>
>>>> take a look at mdrun’s hidden but sometimes useful options:
>>>>
>>>> mdrun -h -hiddden
>>>>
>>>> Carsten
>>>>
>>>>
>>>> On 01 Sep 2014, at 11:07, Oliver Schillinger <
>> o.schillinger at fz-juelich.de>
>>>> wrote:
>>>>
>>>>> Hi,
>>>>> I did not know about the -resethway command line switch to mdrun.
>>>>> Why is it nowhere documented?
>>>>> Or am I blind/stupid?
>>>>> Cheers,
>>>>> Oliver
>>>>>
>>>>> On 08/29/2014 05:15 PM, Carsten Kutzner wrote:
>>>>>> Hi Dawei,
>>>>>>
>>>>>> On 29 Aug 2014, at 16:52, Da-Wei Li <lidawei at gmail.com> wrote:
>>>>>>
>>>>>>> Dear Carsten
>>>>>>>
>>>>>>> Thanks for the clarification. Here it is my benchmark for a small
>>>> protein
>>>>>>> system (18k atoms).
>>>>>>>
>>>>>>> (1) 1 node (12 cores/node, no GPU): 50 ns/day
>>>>>>> (2) 2 nodes (12 cores/node, no GPU): 80 ns/day
>>>>>>> (3) 1 node (12 cores/node, 2 K40 GPUs/node): 100 ns/day
>>>>>>> (4) 2 nodes (12 cores/node, 2 K40 GPUs/node): 40 ns/day
>>>>>>>
>>>>>>>
>>>>>>> I send out this question because the benchmark 4 above is very
>>>> suspicious.
>>>>>> Indeed, if you get 80 ns/day without GPUs, then it should not be less
>>>>>> with GPUs. For how many time steps do you run each of the
>>>>>> benchmarks? Do you use the -resethway command line switch to mdrun
>>>>>> to disregard the first half of the run (where initialization and
>>>>>> balancing is done, you don’t want to count that in a benchmark)?
>>>>>>
>>>>>> Carsten
>>>>>>
>>>>>>> But I agree size of my system may play a role.
>>>>>>>
>>>>>>> best,
>>>>>>>
>>>>>>> dawei
>>>>>>>
>>>>>>>
>>>>>>> On Fri, Aug 29, 2014 at 10:36 AM, Carsten Kutzner <ckutzne at gwdg.de>
>>>> wrote:
>>>>>>>
>>>>>>>> Hi Dawei,
>>>>>>>>
>>>>>>>> the mapping of GPUs to PP ranks is printed for the Master node only,
>>>>>>>> but if this node reports two GPUs, then all other PP ranks will also
>>>>>>>> use two GPUs (or an error is reported).
>>>>>>>>
>>>>>>>> The scaling will depend also on your system size, if this is too
>>>> small,
>>>>>>>> then you might be better off by using a single node.
>>>>>>>>
>>>>>>>> Carsten
>>>>>>>>
>>>>>>>>
>>>>>>>> On 29 Aug 2014, at 16:24, Da-Wei Li <lidawei at gmail.com> wrote:
>>>>>>>>
>>>>>>>>> Dear users,
>>>>>>>>>
>>>>>>>>> I recently try to run Gromacs on two nodes, each of them has 12
>> cores
>>>>>>>> and 2
>>>>>>>>> GPUs. The nodes are connected with infiniband and scaling is pretty
>>>> good
>>>>>>>>> when no GPU is evolved.
>>>>>>>>>
>>>>>>>>> My command is like this:
>>>>>>>>>
>>>>>>>>> mpiexec -npernode 2 -np 4 mdrun_mpi -ntomp 6
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> However, it looks like Gromacs only detected 2 GPUs on node 0, then
>>>> skip
>>>>>>>>> node 1. Part of the output looks like:
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> ************************
>>>>>>>>>
>>>>>>>>> Using 4 MPI processes
>>>>>>>>>
>>>>>>>>> Using 6 OpenMP threads per MPI process
>>>>>>>>>
>>>>>>>>> 2 GPUs detected on host n0316.ten:
>>>>>>>>>
>>>>>>>>> #0: NVIDIA Tesla M2070, compute cap.: 2.0, ECC: yes, stat:
>> compatible
>>>>>>>>>
>>>>>>>>> #1: NVIDIA Tesla M2070, compute cap.: 2.0, ECC: yes, stat:
>> compatible
>>>>>>>>>
>>>>>>>>> 2 GPUs user-selected for this run.
>>>>>>>>>
>>>>>>>>> Mapping of GPUs to the 2 PP ranks in this node: #0, #1
>>>>>>>>>
>>>>>>>>> ****************************
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> The performance is about only 40% of the run, where I use only 1
>>>> node (12
>>>>>>>>> cores+2GPUs).
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Does I miss something?
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> thanks.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> dawei
>>>>>>>>> --
>>>>>>>>> Gromacs Users mailing list
>>>>>>>>>
>>>>>>>>> * Please search the archive at
>>>>>>>> http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before
>>>>>>>> posting!
>>>>>>>>>
>>>>>>>>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>>>>>>>>
>>>>>>>>> * For (un)subscribe requests visit
>>>>>>>>> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users
>>>> or
>>>>>>>> send a mail to gmx-users-request at gromacs.org.
>>>>>>>>
>>>>>>>> --
>>>>>>>> Gromacs Users mailing list
>>>>>>>>
>>>>>>>> * Please search the archive at
>>>>>>>> http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before
>>>>>>>> posting!
>>>>>>>>
>>>>>>>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>>>>>>>
>>>>>>>> * For (un)subscribe requests visit
>>>>>>>> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users
>> or
>>>>>>>> send a mail to gmx-users-request at gromacs.org.
>>>>>>>>
>>>>>>> --
>>>>>>> Gromacs Users mailing list
>>>>>>>
>>>>>>> * Please search the archive at
>>>> http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before
>>>> posting!
>>>>>>>
>>>>>>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>>>>>>
>>>>>>> * For (un)subscribe requests visit
>>>>>>> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users
>> or
>>>> send a mail to gmx-users-request at gromacs.org.
>>>>>>
>>>>>>
>>>>>> --
>>>>>> Dr. Carsten Kutzner
>>>>>> Max Planck Institute for Biophysical Chemistry
>>>>>> Theoretical and Computational Biophysics
>>>>>> Am Fassberg 11, 37077 Goettingen, Germany
>>>>>> Tel. +49-551-2012313, Fax: +49-551-2012302
>>>>>> http://www.mpibpc.mpg.de/grubmueller/kutzner
>>>>>> http://www.mpibpc.mpg.de/grubmueller/sppexa
>>>>>>
>>>>>
>>>>> --
>>>>> Oliver Schillinger
>>>>> PhD student
>>>>>
>>>>> ICS-6 - Structural Biochemistry
>>>>> Building 5.8v, Room 3010
>>>>> Phone: +49 2461-61-9532
>>>>> Mobile: +49 172 53 27 914
>>>>>
>>>>> Forschungszentrum Juelich GmbH
>>>>> 52425 Juelich
>>>>> Sitz der Gesellschaft: Juelich
>>>>> Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498
>>>>> Vorsitzender des Aufsichtsrats: MinDir Dr. Karl Eugen Huthmacher
>>>>> Geschaeftsfuehrung: Prof. Dr. Achim Bachem (Vorsitzender),
>>>>> Karsten Beneke (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt,
>>>>> Prof. Dr. Sebastian M. Schmidt
>>>>> --
>>>>> Gromacs Users mailing list
>>>>>
>>>>> * Please search the archive at
>>>> http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before
>>>> posting!
>>>>>
>>>>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>>>>
>>>>> * For (un)subscribe requests visit
>>>>> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or
>>>> send a mail to gmx-users-request at gromacs.org.
>>>>
>>>>
>>>> --
>>>> Dr. Carsten Kutzner
>>>> Max Planck Institute for Biophysical Chemistry
>>>> Theoretical and Computational Biophysics
>>>> Am Fassberg 11, 37077 Goettingen, Germany
>>>> Tel. +49-551-2012313, Fax: +49-551-2012302
>>>> http://www.mpibpc.mpg.de/grubmueller/kutzner
>>>> http://www.mpibpc.mpg.de/grubmueller/sppexa
>>>>
>>>> --
>>>> Gromacs Users mailing list
>>>>
>>>> * Please search the archive at
>>>> http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before
>>>> posting!
>>>>
>>>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>>>
>>>> * For (un)subscribe requests visit
>>>> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or
>>>> send a mail to gmx-users-request at gromacs.org.
>>>>
>>> --
>>> Gromacs Users mailing list
>>>
>>> * Please search the archive at
>> http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before
>> posting!
>>>
>>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>>
>>> * For (un)subscribe requests visit
>>> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or
>> send a mail to gmx-users-request at gromacs.org.
>>
>>
>> --
>> Dr. Carsten Kutzner
>> Max Planck Institute for Biophysical Chemistry
>> Theoretical and Computational Biophysics
>> Am Fassberg 11, 37077 Goettingen, Germany
>> Tel. +49-551-2012313, Fax: +49-551-2012302
>> http://www.mpibpc.mpg.de/grubmueller/kutzner
>> http://www.mpibpc.mpg.de/grubmueller/sppexa
>>
>> --
>> Gromacs Users mailing list
>>
>> * Please search the archive at
>> http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before
>> posting!
>>
>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>
>> * For (un)subscribe requests visit
>> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or
>> send a mail to gmx-users-request at gromacs.org.
>>
> --
> Gromacs Users mailing list
>
> * Please search the archive at http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before posting!
>
> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>
> * For (un)subscribe requests visit
> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or send a mail to gmx-users-request at gromacs.org.
--
Dr. Carsten Kutzner
Max Planck Institute for Biophysical Chemistry
Theoretical and Computational Biophysics
Am Fassberg 11, 37077 Goettingen, Germany
Tel. +49-551-2012313, Fax: +49-551-2012302
http://www.mpibpc.mpg.de/grubmueller/kutzner
http://www.mpibpc.mpg.de/grubmueller/sppexa
More information about the gromacs.org_gmx-users
mailing list