[gmx-users] Groms 2019.0, simulations using 2 GPUs: RTX 2080Ti
Pragati Sharma
pragati2325 at gmail.com
Fri Dec 13 06:22:37 CET 2019
Hello all,
I am running a polymer melt with 100000 atoms, 2 fs time step, PME, on a
workstation with specifications:
2X Intel Xeon 6128 3.4 2666 MHz 6-core CPU
2X16B DDR4 RAM
2XRTX 2080Ti 11 GB
I have installed *GPU and thread_mpi *enabled gromacs 2019.0 version using:
cmake .. -DGMX_BUILD_OWN_FFTW=ON -DREGRESSIONTEST_DOWNLOAD=ON
*-DGMX_THREAD_MPI=ON
-DGMX_GPU=ON*
While running a single job with below command, I am getting a performance
of *65 ns/day. *
*gmx_tmpi mdrun -v -s t1.tpr -c t1.pdb -gpu_id 0 -ntmpi 1 -ntomp 24*
*Q. However I want to run two different simulations at a time using CPU
cores and one GPU for each, Can somebody help me with mdrun command (what
combination of ntmpi and ntomp) I should use to run two simulations with
efficient utilization of CPU cores and 1 GPU each.*
*Q.* I have also tried utilising GPU for PME calculations using -pme GPU,
as in the command
gmx_tmpi mdrun -v -s t1.tpr -c t1.pdb -ntmpi 1 -ntomp 24 -gputasks 01* -nb
gpu -pme gpu*
but i get the below error,
*"Feature not implemented:The input simulation did not use PME in a way
that is supported on the GPU."*
why is this error coming? Should I put extra attributes while compiling
gromacs.
Thanks
More information about the gromacs.org_gmx-users
mailing list