[gmx-users] mdrun: In option s, required option was not provided and the default file 'topol' does not exist or not accessible and non-integer charges
Mohammed I Sorour
Mohammed.Sorour at temple.edu
Fri Aug 2 17:19:00 CEST 2019
Yeah, I deeply appreciate your help. But any idea/recommendation why the
same command doesn't work through the script? I can't run any jobs,
especially such a big calculation, out of the queue system.
On Fri, Aug 2, 2019 at 11:13 AM Justin Lemkul <jalemkul at vt.edu> wrote:
>
>
> On 8/2/19 11:09 AM, Mohammed I Sorour wrote:
> > This is the output
> >
> >
> > Begin Batch Job Epilogue Sat Aug 2 09:07:19 EDT 2019
> > Job ID: 341185
> > Username: tuf73544
> > Group: chem
> > Job Name: NVT
> > Session: 45173
> > Limits: walltime=01:00:00,neednodes=1:ppn=28,nodes=1:ppn=28
> > Resources:
> >
> cput=19:59:39,vmem=1986008kb,walltime=00:42:53,mem=198008kb,energy_used=0
> > Queue: normal
> > Account:
> > Deleting /dev/shm/*...
> > ----------------------------------------
> > End Batch Job Epilogue Sat Aug 2 09:08:39 EDT 2019
> > ----------------------------------------
> > Command line:
> > gmx mdrun -deffnm nvt
> >
> >
> > Running on 1 node with total 28 cores, 28 logical cores
> > Hardware detected:
> > CPU info:
> > Vendor: Intel
> > Brand: Intel(R) Xeon(R) CPU E5-2690 v4 @ 2.60GHz
> > SIMD instructions most likely to fit this hardware: AVX2_256
> > SIMD instructions selected at GROMACS compile time: SSE4.1
> >
> > Hardware topology: Full, with devices
> >
> > Compiled SIMD instructions: SSE4.1, GROMACS could use AVX2_256 on this
> > machine, which is better.
> >
> > *Reading file nvt.tpr, VERSION 2016.3 (single precision)*
> > Changing nstlist from 10 to 20, rlist from 1 to 1.029
> >
> > The number of OpenMP threads was set by environment variable
> > OMP_NUM_THREADS to 1
> >
> > Will use 24 particle-particle and 4 PME only ranks
> > This is a guess, check the performance at the end of the log file
> > Using 28 MPI threads
> > Using 1 OpenMP thread per tMPI thread
> >
> > starting mdrun 'DNA in water'
> > 500000 steps, 1000.0 ps.
> >
> > step 40 Turning on dynamic load balancing, because the performance loss
> due
> > to load imbalance is 4.1 %.
> >
> >
> > Writing final coordinates.
> >
> > Average load imbalance: 1.5 %
> > Part of the total run time spent waiting due to load imbalance: 1.2 %
> > Steps where the load balancing was limited by -rdd, -rcon and/or -dds:
> X 0
> > % Y 0 % Z 0 %
> > Average PME mesh/force load: 0.752
> > Part of the total run time spent waiting due to PP/PME imbalance: 3.1 %
> >
> >
> > Core t (s) Wall t (s) (%)
> > Time: 71996.002 2571.286 2800.0
> > 42:51
> > (ns/day) (hour/ns)
> > Performance: 33.602 0.714
>
> This output indicates that the job finished successfully and did not
> produce the original error you posted.
>
> -Justin
>
> --
> ==================================================
>
> Justin A. Lemkul, Ph.D.
> Assistant Professor
> Office: 301 Fralin Hall
> Lab: 303 Engel Hall
>
> Virginia Tech Department of Biochemistry
> 340 West Campus Dr.
> Blacksburg, VA 24061
>
> jalemkul at vt.edu | (540) 231-3129
> http://www.thelemkullab.com
>
> ==================================================
>
> --
> Gromacs Users mailing list
>
> * Please search the archive at
> http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before
> posting!
>
> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>
> * For (un)subscribe requests visit
> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or
> send a mail to gmx-users-request at gromacs.org.
>
More information about the gromacs.org_gmx-users
mailing list