Fwd: [gmx-users] Hardware-specific crash with 4.5.1

Justin A. Lemkul jalemkul at vt.edu
Wed Oct 20 12:55:24 CEST 2010


Please keep all Gromacs-related correspondence on the gmx-users list, and do not 
hijack an old, unrelated thread to ask questions.  I am CC'ing the list and I 
would ask that all further discussion take place there.

I have no answer to your question, aside from advising you to read the primary 
literature for the force fields for which implicit solvent parameters were derived.

-Justin

weixin wrote:
> 
> Hi, Justin,
>  
>  
> Where to get those parameters for GB simulation which should be put in 
> the top file:
>  
> 
> [ implicit_genborn_params ]
> 
>  
> 
>  
> 
> Thanks,
> 
> weixin
> 
>  
> 
>  
> 
>  
> ---------- Forwarded message ----------
> From: *Justin A. Lemkul* <jalemkul at vt.edu <mailto:jalemkul at vt.edu>>
> Date: 2010/9/27
> Subject: [gmx-users] Hardware-specific crash with 4.5.1
> To: Gromacs Users' List <gmx-users at gromacs.org 
> <mailto:gmx-users at gromacs.org>>
> 
> 
> 
> Hi All,
> 
> I'm hoping I might get some tips in tracking down the source of an issue 
> that appears to be hardware-specific, leading to crashes in my system. 
>  The failures are occurring on our supercomputer (Mac OSX 10.3, 
> PowerPC).  Running the same .tpr file on my laptop (Mac OSX 10.5.8, 
> Intel Core2Duo) and on another workstation (Ubuntu 10.04, AMD64) produce 
> identical results.  I suspect the problem stems from unsuccessful energy 
> minimization, which then leads to a crash when running full MD.  All 
> jobs were run in parallel on two cores.  The supercomputer does not 
> support threading, so MPI is invoked using MPICH-1.2.5 (native MPI 
> implementation on the cluster).
> 
> 
> Details as follows:
> 
> EM md.log file: successful run (Intel Core2Duo or AMD64)
> 
> Steepest Descents converged to Fmax < 1000 in 7 steps
> Potential Energy  = -4.8878180e+04
> Maximum force     =  8.7791553e+02 on atom 5440
> Norm of force     =  1.1781271e+02
> 
> 
> EM md.log file: unsuccessful run (PowerPC)
> 
> Steepest Descents converged to Fmax < 1000 in 1 steps
> Potential Energy  = -2.4873273e+04
> Maximum force     =  0.0000000e+00 on atom 0
> Norm of force     =            nan
> 
> 
> MD invoked from the minimized structure generated on my laptop or AMD64 
> runs successfully (at least for a few hundred steps in my test), but the 
> MD on the PowerPC cluster fails immediately:
> 
>           Step           Time         Lambda
>              0        0.00000        0.00000
> 
>   Energies (kJ/mol)
>            U-B    Proper Dih.  Improper Dih.      CMAP Dih.GB Polarization
>    7.93559e+03    9.34958e+03    2.24036e+02   -2.47750e+03   -7.83599e+04
>          LJ-14     Coulomb-14        LJ (SR)   Coulomb (SR)      Potential
>    7.70042e+03    9.94520e+04   -1.17168e+04   -5.79783e+04   -2.55780e+04
>    Kinetic En.   Total Energy    Temperature Pressure (bar)   Constr. rmsd
>            nan            nan            nan    0.00000e+00            nan
>  Constr.2 rmsd
>            nan
> 
> DD  step 9 load imb.: force  3.0%
> 
> 
> -------------------------------------------------------
> Program mdrun_4.5.1_mpi, VERSION 4.5.1
> Source code file: nsgrid.c, line: 601
> 
> Range checking error:
> Explanation: During neighborsearching, we assign each particle to a grid
> based on its coordinates. If your system contains collisions or parameter
> errors that give particles very high velocities you might end up with some
> coordinates being +-Infinity or NaN (not-a-number). Obviously, we cannot
> put these on a grid, so this is usually where we detect those errors.
> Make sure your system is properly energy-minimized and that the potential
> energy seems reasonable before trying again.
> Variable ind has value 7131. It should have been within [ 0 .. 7131 ]
> 
> For more information and tips for troubleshooting, please check the GROMACS
> website at http://www.gromacs.org/Documentation/Errors
> -------------------------------------------------------
> 
> It seems as if the crash really shouldn't be happening, if the value 
> range is inclusive.
> 
> Running with all-vs-all kernels works, but the performance is 
> horrendously slow (<300 ps per day for a 7131-atom system) so I am 
> attempting to use long cutoffs (2.0 nm) as others on the list have 
> suggested.
> 
> Details of the installations and .mdp files are appended below.
> 
> -Justin
> 
> === em.mdp ===
> ; Run parameters
> integrator      = steep         ; EM
> emstep      = 0.005
> emtol       = 1000
> nsteps      = 50000
> nstcomm         = 1
> comm_mode   = angular       ; non-periodic system
> ; Bond parameters
> constraint_algorithm    = lincs
> constraints             = all-bonds
> continuation    = no            ; starting up
> ; required cutoffs for implicit
> nstlist         = 1
> ns_type         = grid
> rlist           = 2.0
> rcoulomb        = 2.0
> rvdw            = 2.0
> ; cutoffs required for qq and vdw
> coulombtype     = cut-off
> vdwtype     = cut-off
> ; temperature coupling
> tcoupl          = no
> ; Pressure coupling is off
> Pcoupl          = no
> ; Periodic boundary conditions are off for implicit
> pbc                 = no
> ; Settings for implicit solvent
> implicit_solvent    = GBSA
> gb_algorithm        = OBC
> rgbradii            = 2.0
> 
> 
> === md.mdp ===
> 
> ; Run parameters
> integrator      = sd            ; velocity Langevin dynamics
> dt                  = 0.002
> nsteps          = 2500000               ; 5000 ps (5 ns)
> nstcomm         = 1
> comm_mode   = angular       ; non-periodic system
> ; Output parameters
> nstxout         = 0             ; nst[xvf]out = 0 to suppress useless 
> .trr output
> nstvout         = 0
> nstfout         = 0
> nstlog      = 5000          ; 10 ps
> nstenergy   = 5000          ; 10 ps
> nstxtcout   = 5000          ; 10 ps
> ; Bond parameters
> constraint_algorithm    = lincs
> constraints             = all-bonds
> continuation    = no            ; starting up
> ; required cutoffs for implicit
> nstlist         = 10
> ns_type         = grid
> rlist           = 2.0
> rcoulomb        = 2.0
> rvdw            = 2.0
> ; cutoffs required for qq and vdw
> coulombtype     = cut-off
> vdwtype     = cut-off
> ; temperature coupling
> tc_grps         = System
> tau_t           = 1.0   ; inverse friction coefficient for Langevin (ps^-1)
> ref_t           = 310
> ; Pressure coupling is off
> Pcoupl          = no
> ; Generate velocities is on
> gen_vel         = yes          
> gen_temp        = 310
> gen_seed        = 173529
> ; Periodic boundary conditions are off for implicit
> pbc                 = no
> ; Free energy must be off to use all-vs-all kernels
> ; default, but just for the sake of being pedantic
> free_energy = no
> ; Settings for implicit solvent
> implicit_solvent    = GBSA
> gb_algorithm        = OBC
> rgbradii            = 2.0
> 
> 
> === Installation commands for the cluster ===
> 
> $ ./configure --prefix=/home/rdiv1001/gromacs-4.5 
> CPPFLAGS="-I/home/rdiv1001/fftw-3.0.1-osx/include" 
> LDFLAGS="-L/home/rdiv1001/fftw-3.0.1-osx/lib" --disable-threads 
> --without-x --program-suffix=_4.5.1_s
> 
> $ make
> 
> $ make install
> 
> $ make distclean
> 
> $ ./configure --prefix=/home/rdiv1001/gromacs-4.5 
> CPPFLAGS="-I/home/rdiv1001/fftw-3.0.1-osx/include" 
> LDFLAGS="-L/home/rdiv1001/fftw-3.0.1-osx/lib" --disable-threads 
> --without-x --program-suffix=_4.5.1_mpi --enable-mpi 
> CXXCPP="/nfs/compilers/mpich-1.2.5/bin/mpicxx -E"
> 
> $ make mdrun
> 
> $ make install-mdrun
> 
> 
> -- 
> ========================================
> 
> Justin A. Lemkul
> Ph.D. Candidate
> ICTAS Doctoral Scholar
> MILES-IGERT Trainee
> Department of Biochemistry
> Virginia Tech
> Blacksburg, VA
> jalemkul[at]vt.edu <http://vt.edu/> | (540) 231-9080
> http://www.bevanlab.biochem.vt.edu/Pages/Personal/justin
> 
> ========================================
> -- 
> gmx-users mailing list    gmx-users at gromacs.org 
> <mailto:gmx-users at gromacs.org>
> http://lists.gromacs.org/mailman/listinfo/gmx-users
> Please search the archive at 
> http://www.gromacs.org/Support/Mailing_Lists/Search before posting!
> Please don't post (un)subscribe requests to the list. Use the www 
> interface or send it to gmx-users-request at gromacs.org 
> <mailto:gmx-users-request at gromacs.org>.
> Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
> 

-- 
========================================

Justin A. Lemkul
Ph.D. Candidate
ICTAS Doctoral Scholar
MILES-IGERT Trainee
Department of Biochemistry
Virginia Tech
Blacksburg, VA
jalemkul[at]vt.edu | (540) 231-9080
http://www.bevanlab.biochem.vt.edu/Pages/Personal/justin

========================================



More information about the gromacs.org_gmx-users mailing list