[gmx-users] Problem with the mdrun_openmpi on cluster
James Starlight
jmsstarlight at gmail.com
Mon Mar 14 17:31:43 CET 2016
also below my mdp options which might be relevant- I am using here a
MArtini ff of the system for simulating of GPCR embedded within
lidips.
title = production run for GPCR
dt = 0.02
nsteps = 500000000
nstxout = 0
nstvout = 0
nstlog = 5000
nstxtcout = 5000
xtc-precision = 5000
nstlist = 1 ; Frequency to update the neighbor
list and long range forces
ns_type = grid ; Method to determine neighbor list
(simple, grid)
rlist = 1.2 ; Cut-off for making neighbor list
(short range forces)
coulombtype = PME ; Treatment of long range
electrostatic interactions
rcoulomb = 1.2 ; Short-range electrostatic cut-off
rvdw = 1.2 ; Short-range Van der Waals cut-off
pbc = xyz ; Periodic Boundary
Conditionscoulomb = 1.2
what -rcon flag provided with mdrub should do following the advises of
error mesage7
2016-03-14 17:27 GMT+01:00 James Starlight <jmsstarlight at gmail.com>:
> forgot to add that I tried to decrease number of CPUs to 16 and error
> was the same
>
> 2016-03-14 17:26 GMT+01:00 James Starlight <jmsstarlight at gmail.com>:
>> the error is likely when I try to run not very big system on large
>> number of CPUs in parallel
>>
>>
>> my system is receptor embedded within the membrane consisted of 120 lipids
>> that was produced by grompp
>>
>> Initializing Domain Decomposition on 64 nodes
>> Dynamic load balancing: no
>> Will sort the charge groups at every domain (re)decomposition
>> Initial maximum inter charge-group distances:
>> two-body bonded interactions: 1.174 nm, Bond, atoms 3610 3611
>> multi-body bonded interactions: 1.174 nm, Improper Dih., atoms 3604 3610
>> Minimum cell size due to bonded interactions: 1.200 nm
>> Maximum distance for 5 constraints, at 120 deg. angles, all-trans: 2.626 nm
>> Estimated maximum distance required for P-LINCS: 2.626 nm
>> This distance will limit the DD cell size, you can override this with -rcon
>> Guess for relative PME load: 0.87
>> Using 0 separate PME nodes, as guessed by mdrun
>> Optimizing the DD grid for 64 cells with a minimum initial size of 2.626 nm
>> The maximum allowed number of cells is: X 3 Y 3 Z 3
>>
>>
>> are there sollutions for this kind of system probably altering some cutoffs etc/
>>
>> 2016-03-14 17:22 GMT+01:00 Smith, Micholas D. <smithmd at ornl.gov>:
>>> What is your box size (x, y, z)?
>>>
>>> What happens if you use half that number of nodes?
>>>
>>> ===================
>>> Micholas Dean Smith, PhD.
>>> Post-doctoral Research Associate
>>> University of Tennessee/Oak Ridge National Laboratory
>>> Center for Molecular Biophysics
>>>
>>> ________________________________________
>>> From: gromacs.org_gmx-users-bounces at maillist.sys.kth.se <gromacs.org_gmx-users-bounces at maillist.sys.kth.se> on behalf of James Starlight <jmsstarlight at gmail.com>
>>> Sent: Monday, March 14, 2016 12:19 PM
>>> To: Discussion list for GROMACS users
>>> Subject: [gmx-users] Problem with the mdrun_openmpi on cluster
>>>
>>> Hello,
>>>
>>> I am trying to submit job on 64 nodes on mu local cluster using below
>>> combination of software
>>>
>>>
>>> DO="mpiexec -np 64"
>>> PROG="g_mdrun_openmpi"
>>>
>>>
>>> $DO $PROG -deffnm sim
>>>
>>>
>>> obtaining error
>>>
>>>
>>>
>>> Program g_mdrun_openmpi, VERSION 4.5.7
>>> Source code file:
>>> /builddir/build/BUILD/gromacs-4.5.7/src/mdlib/domdec.c, line: 6436
>>>
>>> Fatal error:
>>> There is no domain decomposition for 64 nodes that is compatible with
>>> the given box and a minimum cell size of 2.6255 nm
>>> Change the number of nodes or mdrun option -rcon or your LINCS settings
>>>
>>> Could someone provide me trivial sollution/
>>>
>>> Thanks!
>>>
>>> J.
>>> --
>>> Gromacs Users mailing list
>>>
>>> * Please search the archive at http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before posting!
>>>
>>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>>
>>> * For (un)subscribe requests visit
>>> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or send a mail to gmx-users-request at gromacs.org.
>>>
>>> --
>>> Gromacs Users mailing list
>>>
>>> * Please search the archive at http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before posting!
>>>
>>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>>
>>> * For (un)subscribe requests visit
>>> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or send a mail to gmx-users-request at gromacs.org.
More information about the gromacs.org_gmx-users
mailing list