[gmx-users] problem with GROMACS parallelization

Carsten Kutzner ckutzne at gwdg.de
Fri Jan 18 08:28:52 CET 2008


Hi Anna,

apart from what Mark already mentioned, you might want to investigate  
if the network is the bottleneck. What kind of network do you use? If  
it is Gbit-Ethernet, you could directly back-to-back connect two  
nodes and see if the scaling on 4 CPUs thereby gets significantly  
better. For benchmarks on more than two nodes, make shure flow  
control is enabled on your network. For the 3200 atom system the  
scaling you see is probably normal on todays fast CPUs - there is no  
performance to gain by sending away a few hundred atoms to a distant  
processor.  But the 62k atom system should indeed scale better unless  
you are using 100Mbit Ethernet ...

Regards,
   Carsten


Am 17.01.2008 um 16:34 schrieb Anna Marabotti:

> Dear all,
> we installed GROMACS v. 3.3.2 on a cluster formed by 20 biprocessor  
> nodes with Centos 4 x86_64, following instructions on the
> GROMACS web site. We compiled it in single precision, parallel  
> version using the --enable-mpi option (LAM MPI was already present on
> the cluster). After the final installation, I made calculations and  
> found that the parallel scalability of the software is very bad:
> in fact, when I use 3 or more processors, the performance is the  
> same (or even worse) than using a single processor.
> As benchmarks, I used two systems: the spider toxin peptide  
> (GROMACS tutorial) in a cubic box of 0.7 nm filled with 2948 water
> molecules (model: spc216) (Total: ~3200 atoms), and a dimeric  
> protein of 718 aa in a cubic box (box edge 0.7 nm) filled with 18425
> water molecules (model: spc216) and 6 Na+ ions (Total: ~62700  
> atoms). In both cases I set the Gromos 96 FF (G43a1). I used the
> following commands:
>
> grompp_mpi -np "nofprocessors" -f pr.mdp -c systemmini.gro -o  
> systempr.tpr -p system.top
> mpirun -np "nofprocessors" mdrun_mpi -s systempr.tpr -o  
> systempr.trr -c systempr.gro -e systempr.edr -g systempr.log
>
> I paste here the results of the performances of position-restrained  
> MD:
>
> spider toxin peptide:
>
> np	time		Mnbf/s	GFlops	ns/day	hours/ns
> 1	1m 51s 237	10.797	2.425	7.806	3.074
> 2	0m 59s 614	19.561	4.402	14.664	1.694
> 3	2m 11s 622	9.261	2.082	6.698	3.583
> 4	1m 58s 722	10.283	2.315	7.448	3.222
> 5	1m 40s 580	11.813	2.659	8.554	2.806
> 6	1m 50s 830	10.859	2.442	7.855	3.056
> 7	1m 49s 232	10.878	2.442	7.855	3.056
> 8	2m 2s 292	6.190	1.392	4.477	5.361
> 9	2m 5s 778	9.533	2.150	6.912	3.472
> 10	2m 22s 540	8.349	1.879	6.042	3.972
>
>
> dimeric protein:
>
> np	time		Mnbf/s	GFlops	ns/day	hours/ns
> 1	13m 51s 391	14.041	2.179	1.043	23.016
> 2	7m 45s 321	25.011	3.883	1.858	12.917
> 3	15m 26s 621	12.586	1.954	0.935	25.667
> 4	15m 15s 3	12.749	1.978	0.946	25.361
> 5	12m 35s 274	15.481	2.401	1.149	20.889
> 6	13m 23s 750	14.517	2.255	1.079	22.250
> 7	12m 25s 659	15.669	2.435	1.164	20.611
> 8	13m 0s 434	14.977	2.325	1.112	21.583
> 9	12m 12s 601	15.949	2.475	1.184	20.278
> 10	13m 1s 724	14.962	2.322	1.111	21.611
>
> I saw in the GROMACS mailing list that it could be due to a problem  
> of communication between nodes, but it seems to me that nobody
> obtained so bad results before. Has anybody some suggestions -  
> apart waiting for GROMACS 4.0 version ;-) - about some further checks
> to do on the system or different compilation/installation to try?
>
> Many thanks in advance and regards
>
> Anna
> ______________________________________________
> Anna Marabotti, Ph.D.
> Laboratorio di Bioinformatica e Biologia Computazionale
> Istituto di Scienze dell'Alimentazione, CNR
> Via Roma 52 A/C
> 83100 Avellino (Italy)
> Tel: +39 0825 299651
> Fax: +39 0825 781585
> Skype: annam1972
> E-mail: amarabotti at isa.cnr.it
> Web page: http://bioinformatica.isa.cnr.it/anna.htm
> ____________________________________________________
> "If you think you are too small to make a difference, try sleeping  
> with a mosquito"
>
> _______________________________________________
> gmx-users mailing list    gmx-users at gromacs.org
> http://www.gromacs.org/mailman/listinfo/gmx-users
> Please search the archive at http://www.gromacs.org/search before  
> posting!
> Please don't post (un)subscribe requests to the list. Use the
> www interface or send it to gmx-users-request at gromacs.org.
> Can't post? Read http://www.gromacs.org/mailing_lists/users.php




More information about the gromacs.org_gmx-users mailing list