[gmx-users] mdp file help
Muhammad Ayaz Anwar
ayazanwar at hotmail.com
Sat May 11 04:49:59 CEST 2013
Hi guys, I'm new to Gromacs and seeking some input on my .mdp file for the production run. I want to perform simulation to check protein stability over time and the after that the stable protein will be used for protein-protein docking/interactions. I'm using cubic box (with 1nm pbc) with water and Cl ions to neutralize on a GPU accelerated system. my mdp file is as: title = Protein in water ; Run parameters integrator = md ; leap-frog integrator nsteps = 2000 ; 2 * 500000 = 1000 ps, 1 ns dt = 0.002 ; 2 fs cutoff-scheme = Verlet ; for GPU acceleration verlet-buffer-drift = -1 ; now use nstlist ; Output control nstxout = 1000 ; save coordinates every 2 ps nstvout = 1000 ; save velocities every 2 ps nstxtcout = 1000 ; xtc compressed trajectory output every 2 ps nstenergy = 1000 ; save energies every 2 ps nstlog = 1000 ; update log file every 2 ps ; Bond parameters continuation = yes ; Restarting after NPT constraint_algorithm = lincs ; holonomic constraints constraints = all-bonds ; all bonds (even heavy atom-H bonds) constrained lincs_iter = 1 ; accuracy of LINCS lincs_order = 4 ; also related to accuracy ; Neighborsearching ns_type = grid ; search neighboring grid cells nstlist = 30 ; 10 fs rlist = 0.6 ; short-range neighborlist cutoff (in nm) rcoulomb = 0.6 ; short-range electrostatic cutoff (in nm) rvdw = 0.6 ; short-range van der Waals cutoff (in nm) ; Electrostatics coulombtype = PME ; Particle Mesh Ewald for long-range electrostatics pme_order = 4 ; cubic interpolation fourierspacing = 0.12 ; grid spacing for FFT ; Temperature coupling is on tcoupl = V-rescale ; modified Berendsen thermostat tc-grps = Protein Non-Protein ; two coupling groups - more accurate tau_t = 0.1 0.1 ; time constant, in ps ref_t = 300 300 ; reference temperature, one for each group, in K ; Pressure coupling is on pcoupl = Parrinello-Rahman ; Pressure coupling on in NPT pcoupltype = isotropic ; uniform scaling of box vectors tau_p = 2.0 ; time constant, in ps ref_p = 1.0 ; reference pressure, in bar compressibility = 4.5e-5 ; isothermal compressibility of water, bar^-1 ; Periodic boundary conditions pbc = xyz ; 3-D PBC ; Dispersion correction DispCorr = EnerPres ; account for cut-off vdW scheme ; Velocity generation gen_vel = no ; Velocity generation is off By performing a test run with ff99SB force field, I got an efficiency as follow: Reading file test11.tpr, VERSION 4.6.1 (single precision) Using 1 MPI thread Using 4 OpenMP threads 1 GPU detected: #0: NVIDIA GeForce GT 630, compute cap.: 3.0, ECC: no, stat: compatible 1 GPU auto-selected for this run: #0 starting mdrun 'Protein in water' 2000 steps, 4.0 ps. step 60: timed with pme grid 104 104 104, coulomb cutoff 0.600: 6760.6 M-cycles step 120: timed with pme grid 96 96 96, coulomb cutoff 0.643: 7826.0 M-cycles step 180: timed with pme grid 104 104 104, coulomb cutoff 0.600: 6716.3 M-cycles step 240: timed with pme grid 100 100 100, coulomb cutoff 0.617: 7248.5 M-cycles optimal pme grid 104 104 104, coulomb cutoff 0.600 step 1900, remaining runtime: 7 s Writing final coordinates. step 2000, remaining runtime: 0 s NOTE: The GPU has >20% more load than the CPU. This imbalance causes performance loss, consider using a shorter cut-off and a finer PME grid. Core t (s) Wall t (s) (%) Time: 427.620 144.021 296.9 (ns/day) (hour/ns) Performance: 2.401 9.996 By changing just the pme_order=6, I got this: Reading file test7.tpr, VERSION 4.6.1 (single precision) Using 1 MPI thread Using 4 OpenMP threads 1 GPU detected: #0: NVIDIA GeForce GT 630, compute cap.: 3.0, ECC: no, stat: compatible 1 GPU auto-selected for this run: #0 starting mdrun 'Protein in water' 2000 steps, 4.0 ps. step 60: timed with pme grid 104 104 104, coulomb cutoff 0.600: 6818.0 M-cycles step 120: timed with pme grid 96 96 96, coulomb cutoff 0.643: 7821.2 M-cycles step 180: timed with pme grid 104 104 104, coulomb cutoff 0.600: 6718.4 M-cycles step 240: timed with pme grid 100 100 100, coulomb cutoff 0.617: 7257.1 M-cycles optimal pme grid 104 104 104, coulomb cutoff 0.600 step 1900, remaining runtime: 7 s Writing final coordinates. step 2000, remaining runtime: 0 s Core t (s) Wall t (s) (%) Time: 550.020 144.580 380.4 (ns/day) (hour/ns) Performance: 2.392 10.035 I have run many test simulations by changing rlist, rcoulomb, rvdw, pme_order and fourierspacing (that I have concluded after reading papers and gromacs user list), but in most of the time, I got performance loss. I am a bit worry about using the very short rlist, rvdw and rcoulomb (most of the papers used 1.0 for each for different proteins). If I increase these cut-offs, there is performance loss, otherwise these cutoff are short (i guess). can any body guide me where I should improve the mdp file or the current one is reasonable. Thanks in advance. mayaz
More information about the gromacs.org_gmx-users
mailing list