[gmx-users] nvidia tesla p100

Irem Altan irem.altan at duke.edu
Tue Nov 1 13:38:11 CET 2016


Hi,

I created a dropbox folder containing different tests (https://www.dropbox.com/sh/cj0vbt0gk5xtch5/AAD-Re5djT5EtahFl6M5jnoFa?dl=0). Let me know if you have trouble accessing it. I also included the relevant input files. Is there any other test that I can try?

Best,
Irem

On Oct 31, 2016, at 6:51 PM, Mark Abraham <mark.j.abraham at gmail.com<mailto:mark.j.abraham at gmail.com>> wrote:

Hi,

OK, well we probably can't help further unless you post a series of
suitably named log files to a file-sharing service and share the links
(list can't take attachments).

Mark

On Mon, Oct 31, 2016 at 11:48 PM Irem Altan <irem.altan at duke.edu<mailto:irem.altan at duke.edu>> wrote:

+ echo 2= 2
2= 2

I mean, it does have two CPUs, with 16 cores each, so maybe that’s the
problem?


There's no detection in any of this. You chose a single node and two tasks
per node, so you're getting what you asked for. That's just not a probably
not good thing to ask for.

I’m assigning two tasks because I was following an example submission
script they had. I assumed this was just because the node has 2 GPUs.
Should this number be increased/decreased? If so, the highest it will allow
me is 16, not 32. Which leads me to think that they actually have 16 cores
with hyperthreading. On their website, they report that the nodes have


 *   2 Intel Xeon v4 CPUs (16 cores, 2.1 GHz base frequency)

which I assumed meant 2x16, but maybe not.

So that's probably related to the thing that that error message is actually
reporting, which is the range of hardware cores on which each thread might
run. See background at

https://urldefense.proofpoint.com/v2/url?u=http-3A__manual.gromacs.org_documentation_2016.1_user-2Dguide_mdrun-2Dperformance.html&d=CwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=r1Wl_e-3DAvYeqhtCRi2Mbok8HBpo_RH4ll0E7Hffr4&m=EYuVMjgf1K6BRzgOcLZqJMzvwTCpzoYs5utDSI2FB3o&s=EpGR3DExshuuzLLbUc2lkDN_UAuQtXEggHnq4s_I-48&e=
.
If they're allowed to move all over the place, then the memory cache is
trashed. Since MPI libraries tend to set these in response to job
schedulers and users, by default mdrun respects affinity masks if set. A
quick test is

mpirun -np 32 gmx_mpi mdrun -ntomp 1 -v -deffnm npt -pin on

which directs mdrun to do something we think is good, rather than you
working out how to do things with SLURM+MPI. Should be a dramatic
improvement, but the hint about using fewer ranks to get more threads per
rank is probably better still.


Well, that did not result in any change in the speed.

The best result I was able to get was with these settings:

!/bin/bash
#SBATCH -N 1 --tasks-per-node=16
#SBATCH -t 00:30:00
#SBATCH -p GPU_100-debug --gres=gpu:2

# Setup the module command
set echo
set -x

module load gromacs/5.1.2

cd $SLURM_SUBMIT_DIR
echo "$SLURM_NPROCS=" $SLURM_NPROCS
mpirun -np $SLURM_NPROCS gmx_mpi mdrun -ntomp 2 -v -deffnm npt

Even then, it’s slower than our local machine that has 24 cores and a
single K20c. Is this normal?

(bridges)
              Core t (s)   Wall t (s)        (%)
      Time:    47432.357      741.131     6400.0
                (ns/day)    (hour/ns)
Performance:      116.579        0.206



(local machine)
              Core t (s)   Wall t (s)        (%)
      Time:    12450.387      519.447     2396.9
                (ns/day)    (hour/ns)
Performance:      166.331        0.144


Best,
Irem
--
Gromacs Users mailing list

* Please search the archive at
https://urldefense.proofpoint.com/v2/url?u=http-3A__www.gromacs.org_Support_Mailing-5FLists_GMX-2DUsers-5FList&d=CwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=r1Wl_e-3DAvYeqhtCRi2Mbok8HBpo_RH4ll0E7Hffr4&m=EEDASIDeUgQJ4ckOutdzmC8XJW6wt1FtTKkUqXZY45E&s=S4evrj2AObpr-VNny5__YLAu8LNEgc8RHhqJewV2Qp0&e=  before
posting!

* Can't post? Read https://urldefense.proofpoint.com/v2/url?u=http-3A__www.gromacs.org_Support_Mailing-5FLists&d=CwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=r1Wl_e-3DAvYeqhtCRi2Mbok8HBpo_RH4ll0E7Hffr4&m=EEDASIDeUgQJ4ckOutdzmC8XJW6wt1FtTKkUqXZY45E&s=bco7PM9oFpz_bPVXLAyWP0E-000R8SqjCU1jioj-7No&e=

* For (un)subscribe requests visit
https://urldefense.proofpoint.com/v2/url?u=https-3A__maillist.sys.kth.se_mailman_listinfo_gromacs.org-5Fgmx-2Dusers&d=CwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=r1Wl_e-3DAvYeqhtCRi2Mbok8HBpo_RH4ll0E7Hffr4&m=EEDASIDeUgQJ4ckOutdzmC8XJW6wt1FtTKkUqXZY45E&s=vgGZ0L-Tax3RdvQ3cLWjbduoNV4GPI_jmpAHrhl63H0&e=  or
send a mail to gmx-users-request at gromacs.org.
--
Gromacs Users mailing list

* Please search the archive at https://urldefense.proofpoint.com/v2/url?u=http-3A__www.gromacs.org_Support_Mailing-5FLists_GMX-2DUsers-5FList&d=CwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=r1Wl_e-3DAvYeqhtCRi2Mbok8HBpo_RH4ll0E7Hffr4&m=EEDASIDeUgQJ4ckOutdzmC8XJW6wt1FtTKkUqXZY45E&s=S4evrj2AObpr-VNny5__YLAu8LNEgc8RHhqJewV2Qp0&e=  before posting!

* Can't post? Read https://urldefense.proofpoint.com/v2/url?u=http-3A__www.gromacs.org_Support_Mailing-5FLists&d=CwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=r1Wl_e-3DAvYeqhtCRi2Mbok8HBpo_RH4ll0E7Hffr4&m=EEDASIDeUgQJ4ckOutdzmC8XJW6wt1FtTKkUqXZY45E&s=bco7PM9oFpz_bPVXLAyWP0E-000R8SqjCU1jioj-7No&e=

* For (un)subscribe requests visit
https://urldefense.proofpoint.com/v2/url?u=https-3A__maillist.sys.kth.se_mailman_listinfo_gromacs.org-5Fgmx-2Dusers&d=CwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=r1Wl_e-3DAvYeqhtCRi2Mbok8HBpo_RH4ll0E7Hffr4&m=EEDASIDeUgQJ4ckOutdzmC8XJW6wt1FtTKkUqXZY45E&s=vgGZ0L-Tax3RdvQ3cLWjbduoNV4GPI_jmpAHrhl63H0&e=  or send a mail to gmx-users-request at gromacs.org<mailto:gmx-users-request at gromacs.org>.



More information about the gromacs.org_gmx-users mailing list