[gmx-developers] segfault goes away when restarting from cpt
Michael Brunsteiner
mbx0009 at yahoo.com
Fri May 5 11:14:51 CEST 2017
hi,
I post this here as it might be developers rather than a user issue ...I ran an NPT sim with simulated annealing of an amorphous solid sample withsome organic molecules,as in:
gmx grompp -f md-1bar-353-253.mdp -p sis3-7-simp.top -c up-nr2-3.gro -o do-nr2-3.tpr
nohup gmx mdrun -v -deffnm do-nr2-3 > er 2>&1 &
after around 30 nano secs the simulation stops without further notice.neither in the log-file nor in stdout or stderr there are any indicators of what happenedbut when i look into the relevant syslog file i find:
May 5 03:24:38 rcpe-sbd-node03 kernel: [82541302.295784] gmx[2218]:
segfault at ffffffff9d3ebea0 ip 00007f5b708be3a1 sp 00007f5b657f9dc0 error 7 in libgromacs.so.2.3.0[7f5b706d9000+1d1e000]
when i restart the sim on the same computer and from the last cpt file, as in:
nohup gmx mdrun -v -deffnm do-nr2-3 -cpi do-nr2-3.cpt -noappend > er 2>&1 &
the sim happily continues beyond the point where it previously seg-faulted without any further issues ...
tpr file is too large to attach (if anybody's interested i can upload it somewhere)below i put the last 30 or so lines of both stderr+stdout and the log-fileI believe the warning at the end of stderr is harmless, but even if it actually is the reasonfor the segfault this still does not explain why nothing is written to stderr when it happensand why the sim works when restarted from the cpt file ... can it be that this is a hardware issue??
regards,Michael
stderr+stdout:[..] Brand: Intel(R) Core(TM) i7-4930K CPU @ 3.40GHz
SIMD instructions most likely to fit this hardware: AVX_256
SIMD instructions selected at GROMACS compile time: AVX_256
Hardware topology: Full, with devices
GPU info:
Number of GPUs detected: 1
#0: NVIDIA GeForce GTX 780, compute cap.: 3.5, ECC: no, stat: compatible
Reading file do-nr2-3.tpr, VERSION 2016.3 (single precision)
Changing nstlist from 20 to 40, rlist from 1.2 to 1.2
Using 1 MPI thread
Using 12 OpenMP threads
1 compatible GPU is present, with ID 0
1 GPU auto-selected for this run.
Mapping of GPU ID to the 1 PP rank in this node: 0
starting mdrun 'system'
110000000 steps, 110000.0 ps.
step 80: timed with pme grid 40 40 24, coulomb cutoff 1.200: 81.2 M-cycles
step 80: the box size limits the PME load balancing to a coulomb cut-off of 1.368
step 160: timed with pme grid 32 36 24, coulomb cutoff 1.368: 72.9 M-cycles
step 240: timed with pme grid 36 36 24, coulomb cutoff 1.264: 75.7 M-cycles
step 320: timed with pme grid 36 40 24, coulomb cutoff 1.216: 78.6 M-cycles
step 400: timed with pme grid 40 40 24, coulomb cutoff 1.200: 81.2 M-cycles
optimal pme grid 32 36 24, coulomb cutoff 1.368
step 31031000, will finish Fri May 5 14:27:08 2017
Step 31031061 Warning: pressure scaling more than 1%, mu: 0.999153 0.982333 0.997814
log-file:[..]
Step Time
31030000 31030.00000
Current ref_t for group System: 327.9
Energies (kJ/mol)
Bond Angle Proper Dih. Improper Dih. LJ-14
8.65826e+03 1.38650e+04 1.08781e+04 4.26101e+02 6.01339e+03
Coulomb-14 LJ (SR) Coulomb (SR) Coul. recip. Potential
-2.98635e+04 -1.12638e+04 1.63267e+04 2.11350e+02 1.52514e+04
Kinetic En. Total Energy Temperature Pressure (bar)
2.22541e+04 3.75055e+04 3.28069e+02 6.14802e+02
Step Time
31031000 31031.00000
Current ref_t for group System: 327.8
Energies (kJ/mol)
Bond Angle Proper Dih. Improper Dih. LJ-14
8.41290e+03 1.38660e+04 1.08950e+04 3.51583e+02 5.79937e+03
Coulomb-14 LJ (SR) Coulomb (SR) Coul. recip. Potential
-2.99386e+04 -1.15255e+04 1.64549e+04 2.30994e+02 1.45468e+04
Kinetic En. Total Energy Temperature Pressure (bar)
2.27307e+04 3.72775e+04 3.35095e+02 -1.54620e+00
--
Gromacs Developers mailing list
* Please search the archive at http://www.gromacs.org/Support/Mailing_Lists/GMX-developers_List before posting!
* Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
* For (un)subscribe requests visit
https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-developers or send a mail to gmx-developers-request at gromacs.org.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://maillist.sys.kth.se/pipermail/gromacs.org_gmx-developers/attachments/20170505/5927754d/attachment.html>
More information about the gromacs.org_gmx-developers
mailing list