[gmx-users] Best performace with 0 core for PME calcuation

Nicolas nsapay at ucalgary.ca
Sat Jan 10 01:53:42 CET 2009


Hello,

I'm trying to do a benchmark with Gromacs 4 on our cluster, but I don't 
completely understand the results I obtain. The system I used is a 128 
DOPC bilayer hydrated by ~18800 SPC for a total of ~70200 atoms. The 
size of the system is 9.6x9.6x10.1 nm^3. I'm using the following parameters:

        * nstlist = 10
        * rlist = 1
        * Coulombtype = PME
        * rcoulomb = 1
        * fourier spacing = 0.12
        * vdwtype = Cutoff
        * rvdw = 1

The cluster itself has got 2 procs/node connected by Ethernet 100 MB/s. 
I'm using mpiexec to run Gromacs.  When I use -npme 2 -ddorder 
interleave, I get:
ncore    Perf (ns/day)    PME (%)

    1    0,00    0
    2    0,00    0
    3    0,00    0
    4    1,35    28
    5    1,84    31
    6    2,08    27
    8    2,09    21
    10    2,25    17
    12    2,02    15
    14    2,20    13
    16    2,04    11
    18    2,18    10
    20    2,29    9

So, above 6-8 cores, the PP nodes are spending too much time waiting for 
the PME nodes and the perf forms a plateau. When I use -npme 0, I get:

     ncore    Perf (ns/day)    PME (%)
    1    0,43    33
    2    0,92    34
    3    1,34    35
    4    1,69    36
    5    2,17    33
    6    2,56    32
    8    3,24    33
    10    3,84    34
    12    4,34    35
    14    5,05    32
    16    5,47    34
    18    5,54    37
    20    6,13    36

I obtain much better performances when there is no PME nodes, while I 
was expecting the opposite. Does someone have an explanation for that? 
Does that means domain decomposition is useless below a certain real 
space cutoff?  I'm quite confused.

Thanks,
Nicolas


-------------- next part --------------
A non-text attachment was scrubbed...
Name: nsapay.vcf
Type: text/x-vcard
Size: 310 bytes
Desc: not available
URL: <http://maillist.sys.kth.se/pipermail/gromacs.org_gmx-users/attachments/20090109/7bda5453/attachment.vcf>


More information about the gromacs.org_gmx-users mailing list