[gmx-users] MPI error in gromacs 4.6

Mark Abraham mark.j.abraham at gmail.com
Mon Mar 24 13:01:19 CET 2014


On Mon, Mar 24, 2014 at 12:48 PM, Ankita Naithani
<ankitanaithani at gmail.com>wrote:

> Hi Pavan,
> Thank you for your response. I am trying to generate the tpr file with the
> following parameter;
> ; Neighborsearching
>  ns_type        = grid        ; search neighboring grid cells
>  nstlist        = 5        ; 25 fs
>  rlist        = 1.0        ; short-range neighborlist cutoff (in nm)
>  rcoulomb    = 1.0        ; short-range electrostatic cutoff (in nm)
>  rvdw        = 1.0        ; short-range van der Waals cutoff (in nm)
>  rlistlong    = 1.0        ; long-range neighborlist cutoff (in nm)
> cutoff-scheme = Verlet
>
> But, I get a warning of Unknown left-hand 'cutoff-scheme' in parameter
> file.
>

If you want to prepare a run for GROMACS 4.6, use a 4.6 version of grompp!

Mark


>
>
> On Mon, Mar 24, 2014 at 11:26 AM, Pavan Kumar <kumar.pavan006 at gmail.com
> >wrote:
>
> > Hello Ankita
> > You have to just include the following line in your mdp file
> > cutoff-scheme=Verlet
> > And run your grompp with the modfied mdp file to generate tpr file and
> then
> > mdrun.
> > Hope this doesn't give you the same error
> >
> >
> > On Mon, Mar 24, 2014 at 4:47 PM, Ankita Naithani
> > <ankitanaithani at gmail.com>wrote:
> >
> > > Hi,
> > >
> > > I am trying to run a simulation of my protein (monomer ~500 residues).
> I
> > > had few questions and erors regarding the same.
> > > I have previously run the simulation of the apo form of the same
> protein
> > > using Gromacs 4.5.5 which was available at the cluster facility I was
> > using
> > > and also which is installed in my system. However, when I tried to run
> > the
> > > holo form, I got error :
> > > Fatal error:
> > > 11 particles communicated to PME node 106 are more than 2/3 times the
> > > cut-off out of the domain decomposition cell of their charge group in
> > > dimension y.
> > > This usually means that your system is not well equilibrated.
> > > For more information and tips for troubleshooting, please check the
> > GROMACS
> > > website at http://www.gromacs.org/Documentation/Errors
> > >
> > > This I figured out could be solved using a lower timestep as my
> previous
> > > timestep was 4fs and now I have reduced it to 3fs which should work
> fine
> > > now.
> > > However, after producing the tpr file for production run in my GROMACS
> > > 4.5.5, I realised that the grant for the cluster facility is over and
> the
> > > new clusters which I am trying to set up the same protein for support
> > only
> > > gromacs 4.6. I am trying to run the code in these clusters and I get he
> > > following error:
> > >
> > >
> > > -------------------------------------------------------
> > > Program mdrun_mpi, VERSION 4.6.3
> > > Source code file: /home/gromacs-4.6.3/src/kernel/runner
> > > .c, line: 824
> > >
> > > Fatal error:
> > > OpenMP threads have been requested with cut-off scheme Group, but these
> > are
> > > only
> > >  supported with cut-off scheme Verlet
> > > For more information and tips for troubleshooting, please check the
> > GROMACS
> > > website at http://www.gromacs.org/Documentation/Errors
> > >
> > >
> >
> ---------------------------------------------------------------------------------
> > >
> > > 1. I wanted help with my mdp options to make it compatible.
> > > 2. Since my pevious calculations were based on gromacs 4.5.5, switching
> > to
> > > gromacs 4.6, would that break the continuity of the run or would that
> > bring
> > > about differences in the way the trajectories would be analysed?
> > >
> > >
> > > Below, is my mdp file
> > > title        = production MD
> > > ; Run parameters
> > > integrator    = md        ; leap-frog algorithm
> > > nsteps        = 33333333    ; 0.003 * 33333333 = 100000 ps or 100 n
> > > dt        = 0.003        ; 3 fs
> > > ; Output control
> > > nstxout        = 0        ; save coordinates every 2 ps
> > > nstvout        = 0        ; save velocities every 2 ps
> > > nstxtcout    = 1000        ; xtc compressed trajectory output every 5
> ps
> > > nstenergy    = 1000            ; save energies every 5 ps
> > > nstlog        = 1000            ; update log file every 5 ps
> > > energygrps      = Protein ATP
> > > ; Bond parameters
> > > constraint_algorithm = lincs    ; holonomic constraints
> > > constraints    = all-bonds    ; all bonds (even heavy atom-H bonds)
> > > constrained
> > > lincs_iter    = 1        ; accuracy of LINCS
> > > lincs_order    = 4        ; also related to accuracy
> > > ; Neighborsearching
> > > ns_type        = grid        ; search neighboring grid cells
> > > nstlist        = 5        ; 25 fs
> > > rlist        = 1.0        ; short-range neighborlist cutoff (in nm)
> > > rcoulomb    = 1.0        ; short-range electrostatic cutoff (in nm)
> > > rvdw        = 1.0        ; short-range van der Waals cutoff (in nm)
> > > rlistlong    = 1.0        ; long-range neighborlist cutoff (in nm)
> > > ; Electrostatics
> > > coulombtype    = PME        ; Particle Mesh Ewald for long-range
> > > electrostatics
> > > pme_order    = 4        ; cubic interpolation
> > > fourierspacing    = 0.16        ; grid spacing for FFT
> > > nstcomm = 10                    ; remove com every 10 steps
> > > ; Temperature coupling is on
> > > tcoupl        = V-rescale    ; modified Berendsen thermostat
> > > tc-grps        = Protein Non-Protein    ; two coupling groups - more
> > > accurate
> > > tau_t        = 0.1    0.1    ; time constant, in ps
> > > ref_t        = 318     318    ; reference temperature, one for each
> > group,
> > > in K
> > > ; Pressure coupling is off
> > > pcoupl          = berendsen    ; Berendsen thermostat
> > > pcoupltype    = isotropic    ; uniform scaling of box vectors
> > > tau_p        = 1.0        ; time constant, in ps
> > > ref_p        = 1.0        ; reference pressure, in bar
> > > compressibility = 4.5e-5    ; isothermal compressibility of water,
> bar^-1
> > > ; Periodic boundary conditions
> > > pbc        = xyz        ; 3-D PBC
> > > ; Dispersion correction
> > > DispCorr    = EnerPres    ; account for cut-off vdW scheme
> > > ; Velocity generation
> > > gen_vel        = yes        ; Velocity generation is on
> > > gen_temp    = 318        ; reference temperature, for protein in K
> > >
> > >
> > >
> > >
> > > Kind regards--
> > > Ankita Naithani
> > > --
> > > Gromacs Users mailing list
> > >
> > > * Please search the archive at
> > > http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before
> > > posting!
> > >
> > > * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
> > >
> > > * For (un)subscribe requests visit
> > > https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or
> > > send a mail to gmx-users-request at gromacs.org.
> > >
> >
> >
> >
> > --
> > Cheers
> > Pavan
> > --
> > Gromacs Users mailing list
> >
> > * Please search the archive at
> > http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before
> > posting!
> >
> > * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
> >
> > * For (un)subscribe requests visit
> > https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or
> > send a mail to gmx-users-request at gromacs.org.
> >
>
>
>
> --
> Ankita Naithani
> --
> Gromacs Users mailing list
>
> * Please search the archive at
> http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before
> posting!
>
> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>
> * For (un)subscribe requests visit
> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or
> send a mail to gmx-users-request at gromacs.org.
>


More information about the gromacs.org_gmx-users mailing list