[gmx-users] Signal:11 info.si_errno:0(Success) si_code:1(SEGV_MAPERR)
David van der Spoel
spoel at xray.bmc.uu.se
Mon Jan 22 19:05:24 CET 2007
Mike Hanby wrote:
> Howdy, a user of a cluster that I manage sent me an output file of their
> Gromacs 3.3.1 job. It looks like the job runs for approx 11 hours and
> then abruptly terminates with a SEGV_MAPERR. What could cause this
> (gromacs, or a system config / resource?).
>
> The output of the log file is below, thanks for any suggestions. Mike
>
> ########################################################################
> #############
> STARTED AT: Wed Jan 17 11:40:03 CST 2007
>
> NSLOTS: 16
> :-)
> /share/apps/gromacs/intel/gromacs-3.3.1-s64-openmpi/bin/mdrun_ompi (-:
>
> Option Filename Type Description
> ------------------------------------------------------------
> -s production-Npt-300K_16CPU.tpr Input Generic run input: tpr
> tpb
> tpa xml
> -o production-Npt-300K_16CPU.trr Output Full precision
> trajectory:
> trr trj
> -x production-Npt-300K_16CPU.xtc Output, Opt! Compressed trajectory
> (portable xdr format)
> -c production-Npt-300K_16CPU.gro Output Generic structure: gro
> g96
> pdb xml
> -e ener.edr Output Generic energy: edr ene
>
> Getting Loaded...
> Reading file production-Npt-300K_16CPU.tpr, VERSION 3.3.1 (single
> precision)
> Loaded with Money
>
> Back Off! I just backed up ener.edr to ./#ener.edr.1#
> starting mdrun 'Protein in water'
> 2500000 steps, 2500.0 ps.
>
> ...... steps ......
>
> step 650170, will finish at Fri Jan 19 06:52:24 2007
>
> step 650180, will finish at Fri Jan 19 06:52:21 2007
> Signal:11 info.si_errno:0(Success) si_code:1(SEGV_MAPERR)
> Failing at addr:0x7c3fefc
> [0] func:/share/apps/openmpi/intel/openmpi-1.1.2-64/lib/libopal.so.0
> [0x2a966102a5]
> [1] func:/lib64/tls/libpthread.so.0 [0x3793b0c430]
> [2]
> func:/share/apps/gromacs/intel/gromacs-3.3.1-s64-openmpi/bin/mdrun_ompi(
> spread_q_bsplines+0x1c0) [0x45c80c]
We've seen quite some problems with MPI libraries previously. I don't
know what the current status of openmpi is. As the cluster manager you
may have run some mpi tests, if so did they work flawless?
The gromacs function (2 in your call tree) could also generate an
overflow due to physics like large forces. Hence it is not possible to
determine from this output what causes the problem.
--
David.
________________________________________________________________________
David van der Spoel, PhD, Assoc. Prof., Molecular Biophysics group,
Dept. of Cell and Molecular Biology, Uppsala University.
Husargatan 3, Box 596, 75124 Uppsala, Sweden
phone: 46 18 471 4205 fax: 46 18 511 755
spoel at xray.bmc.uu.se spoel at gromacs.org http://folding.bmc.uu.se
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
More information about the gromacs.org_gmx-users
mailing list