[gmx-users] g_tune_pme for multiple nodes
Carsten Kutzner
ckutzne at gwdg.de
Tue Dec 4 14:09:12 CET 2012
Hi Chandan,
the number of separate PME nodes in Gromacs must be larger than two and
smaller or equal to half the number of MPI processes (=np). Thus, g_tune_pme
checks only up to npme = np/2 PME nodes.
Best,
Carsten
On Dec 4, 2012, at 1:54 PM, Chandan Choudhury <iitdckc at gmail.com> wrote:
> Dear Carsten and Florian,
>
> Thanks for you useful suggestions. It did work. I still have a doubt
> regarding the execution :
>
> export MPIRUN=`which mpirun`
> export MDRUN="/cm/shared/apps/gromacs/4.5.5/single/bin/mdrun_mpi_4.5.5"
> g_tune_pme_4.5.5 -np $NPROCS -s md0-200.tpr -c tune.pdb -x tune.xtc -e
> tune.edr -g tune.log
>
> I am suppling $NPROCS as 24 [2 (nodes)*12(ppn)], so that g_tune_pme tunes
> the no. of pme nodes. As I am executing it on a single node, mdrun never
> checks pme for greater than 12 ppn. So, how do I understand that the pme is
> tuned for 24 ppn spanning across the two nodes.
>
> Chandan
>
>
> --
> Chandan kumar Choudhury
> NCL, Pune
> INDIA
>
>
> On Thu, Nov 29, 2012 at 8:32 PM, Carsten Kutzner <ckutzne at gwdg.de> wrote:
>
>> Hi Chandan,
>>
>> On Nov 29, 2012, at 3:30 PM, Chandan Choudhury <iitdckc at gmail.com> wrote:
>>
>>> Hi Carsten,
>>>
>>> Thanks for your suggestion.
>>>
>>> I did try to pass to total number of cores with the np flag to the
>>> g_tune_pme, but it didnot help. Hopefully I am doing something silliy. I
>>> have pasted the snippet of the PBS script.
>>>
>>> #!/bin/csh
>>> #PBS -l nodes=2:ppn=12:twelve
>>> #PBS -N bilayer_tune
>>> ....
>>> ....
>>> ....
>>>
>>> cd $PBS_O_WORKDIR
>>> export MDRUN="/cm/shared/apps/gromacs/4.5.5/single/bin/mdrun_mpi_4.5.5"
>> from here on you job file should read:
>>
>> export MPIRUN=`which mpirun`
>> g_tune_pme_4.5.5 -np $NPROCS -s md0-200.tpr -c tune.pdb -x tune.xtc -e
>> tune.edr -g tune.log
>>
>>> mpirun -np $NPROCS g_tune_pme_4.5.5 -np 24 -s md0-200.tpr -c tune.pdb -x
>>> tune.xtc -e tune.edr -g tune.log -nice 0
>> this way you will get $NPROCS g_tune_pme instances, each trying to run an
>> mdrun job on 24 cores,
>> which is not what you want. g_tune_pme itself is a serial program, it just
>> spawns the mdrun's.
>>
>> Carsten
>>>
>>>
>>> Then I submit the script using qsub.
>>> When I login to the compute nodes there I donot find and mdrun executable
>>> running.
>>>
>>> I also tried using nodes=1 and np 12. It didnot work through qsub.
>>>
>>> Then I logged in to the compute nodes and executed g_tune_pme_4.5.5 -np
>> 12
>>> -s md0-200.tpr -c tune.pdb -x tune.xtc -e tune.edr -g tune.log -nice 0
>>>
>>> It worked.
>>>
>>> Also, if I just use
>>> $g_tune_pme_4.5.5 -np 12 -s md0-200.tpr -c tune.pdb -x tune.xtc -e
>> tune.edr
>>> -g tune.log -nice 0
>>> g_tune_pme executes on the head node and writes various files.
>>>
>>> Kindly let me know what am I missing when I submit through qsub.
>>>
>>> Thanks
>>>
>>> Chandan
>>> --
>>> Chandan kumar Choudhury
>>> NCL, Pune
>>> INDIA
>>>
>>>
>>> On Mon, Sep 3, 2012 at 3:31 PM, Carsten Kutzner <ckutzne at gwdg.de> wrote:
>>>
>>>> Hi Chandan,
>>>>
>>>> g_tune_pme also finds the optimal number of PME cores if the cores
>>>> are distributed on multiple nodes. Simply pass the total number of
>>>> cores to the -np option. Depending on the MPI and queue environment
>>>> that you use, the distribution of the cores over the nodes may have
>>>> to be specified in a hostfile / machinefile. Check g_tune_pme -h
>>>> on how to set that.
>>>>
>>>> Best,
>>>> Carsten
>>>>
>>>>
>>>> On Aug 28, 2012, at 8:33 PM, Chandan Choudhury <iitdckc at gmail.com>
>> wrote:
>>>>
>>>>> Dear gmx users,
>>>>>
>>>>> I am using 4.5.5 of gromacs.
>>>>>
>>>>> I was trying to use g_tune_pme for a simulation. I intend to execute
>>>>> mdrun at multiple nodes with 12 cores each. Therefore, I would like to
>>>>> optimize the number of pme nodes. I could execute g_tune_pme -np 12
>>>>> md.tpr. But this will only find the optimal PME nodes for single nodes
>>>>> run. How do I find the optimal PME nodes for multiple nodes.
>>>>>
>>>>> Any suggestion would be helpful.
>>>>>
>>>>> Chandan
>>>>>
>>>>> --
>>>>> Chandan kumar Choudhury
>>>>> NCL, Pune
>>>>> INDIA
>>>>> --
>>>>> gmx-users mailing list gmx-users at gromacs.org
>>>>> http://lists.gromacs.org/mailman/listinfo/gmx-users
>>>>> * Please search the archive at
>>>> http://www.gromacs.org/Support/Mailing_Lists/Search before posting!
>>>>> * Please don't post (un)subscribe requests to the list. Use the
>>>>> www interface or send it to gmx-users-request at gromacs.org.
>>>>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>>>
>>>>
>>>> --
>>>> Dr. Carsten Kutzner
>>>> Max Planck Institute for Biophysical Chemistry
>>>> Theoretical and Computational Biophysics
>>>> Am Fassberg 11, 37077 Goettingen, Germany
>>>> Tel. +49-551-2012313, Fax: +49-551-2012302
>>>> http://www3.mpibpc.mpg.de/home/grubmueller/ihp/ckutzne
>>>>
>>>> --
>>>> gmx-users mailing list gmx-users at gromacs.org
>>>> http://lists.gromacs.org/mailman/listinfo/gmx-users
>>>> * Please search the archive at
>>>> http://www.gromacs.org/Support/Mailing_Lists/Search before posting!
>>>> * Please don't post (un)subscribe requests to the list. Use the
>>>> www interface or send it to gmx-users-request at gromacs.org.
>>>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>>>
>>> --
>>> gmx-users mailing list gmx-users at gromacs.org
>>> http://lists.gromacs.org/mailman/listinfo/gmx-users
>>> * Please search the archive at
>> http://www.gromacs.org/Support/Mailing_Lists/Search before posting!
>>> * Please don't post (un)subscribe requests to the list. Use the
>>> www interface or send it to gmx-users-request at gromacs.org.
>>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>
>>
>> --
>> Dr. Carsten Kutzner
>> Max Planck Institute for Biophysical Chemistry
>> Theoretical and Computational Biophysics
>> Am Fassberg 11, 37077 Goettingen, Germany
>> Tel. +49-551-2012313, Fax: +49-551-2012302
>> http://www.mpibpc.mpg.de/grubmueller/kutzner
>>
>> --
>> gmx-users mailing list gmx-users at gromacs.org
>> http://lists.gromacs.org/mailman/listinfo/gmx-users
>> * Please search the archive at
>> http://www.gromacs.org/Support/Mailing_Lists/Search before posting!
>> * Please don't post (un)subscribe requests to the list. Use the
>> www interface or send it to gmx-users-request at gromacs.org.
>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>
> --
> gmx-users mailing list gmx-users at gromacs.org
> http://lists.gromacs.org/mailman/listinfo/gmx-users
> * Please search the archive at http://www.gromacs.org/Support/Mailing_Lists/Search before posting!
> * Please don't post (un)subscribe requests to the list. Use the
> www interface or send it to gmx-users-request at gromacs.org.
> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
--
Dr. Carsten Kutzner
Max Planck Institute for Biophysical Chemistry
Theoretical and Computational Biophysics
Am Fassberg 11, 37077 Goettingen, Germany
Tel. +49-551-2012313, Fax: +49-551-2012302
http://www.mpibpc.mpg.de/grubmueller/kutzner
http://www.mpibpc.mpg.de/grubmueller/sppexa
More information about the gromacs.org_gmx-users
mailing list