[gmx-users] multiple processor running of gromacs-4.0.4

Mark Abraham Mark.Abraham at anu.edu.au
Tue Oct 20 02:06:48 CEST 2009


Jianhui Tian wrote:
> Hi,
> 
> I have a small system of fullerene ball with waters. I want to 
> simulation the system without pbc and pme, treating the coulomb and vdw 
> without cutoff. The system can be run on single processor. But when I 
> try to run it with multiple processors, it can't proceed. I am including 
> the error message at the end of this mail.

Those messages are generic failure messages from the MPI system. The 
stack trace therein suggests you should read the stderr and/or log file 
to get some diagnostic information from GROMACS. Possibly your system is 
blowing up because you've started from an unsuitable configuration.

> There are some of my guesses for the possible reason of the error:
> 1. Do we have to use pbc and pme to use multiple processors for simulation?

Wouldn't think so. However because you have no cut-offs, there will be 
no advantage to DD because there is no data locality - every processor 
needs the position of every atom. mdrun -pd may work. It strikes me as 
possible that this scenario doesn't work in parallel on GROMACS.

> 2. Can wen use restraints when use multiple processors?

Yes.

> 3. If a molecule is "divided" into two parts in domain decomposition, 
> will this be a problem for simulation?

No, that's routine.

Mark

> Thanks for any suggestion about this error message.
> 
> logfile:--------------------------------------------------------------------------
> [0,1,1]: OpenIB on host compute-1-12.local was unable to find any HCAs.
> Another transport will be used instead, although this may result in
> lower performance.
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> [0,1,0]: OpenIB on host compute-1-12.local was unable to find any HCAs.
> Another transport will be used instead, although this may result in
> lower performance.
> --------------------------------------------------------------------------
> NNODES=2, MYRANK=0, HOSTNAME=compute-1-12.local
> NODEID=0 argc=13
> NNODES=2, MYRANK=1, HOSTNAME=compute-1-12.local
> NODEID=1 argc=13
> .
> .
> .
> .
> Reading file 
> /data/disk04/tianj/trp_remd_091012/equilibrium/test/trp_full_run1.tpr, 
> VERSION 4.0 (single precision)
> [compute-1-12:08033] *** Process received signal ***
> [compute-1-12:08033] Signal: Segmentation fault (11)
> [compute-1-12:08033] Signal code: Address not mapped (1)
> [compute-1-12:08033] Failing at address: 0xc0
> [compute-1-12:08033] [ 0] /lib64/libpthread.so.0 [0x378fe0de80]
> [compute-1-12:08033] [ 1] /lib64/libc.so.6(_IO_vfprintf+0x39) [0x378f642309]
> [compute-1-12:08033] [ 2] /lib64/libc.so.6(_IO_fprintf+0x88) [0x378f64cf68]
> [compute-1-12:08033] [ 3] 
> /usr/local/gromacs-4.0.4/bin/mdrun_sm(mk_mshift+0x315) [0x516593]
> [compute-1-12:08033] [ 4] /usr/local/gromacs-4.0.4/bin/mdrun_sm [0x45fa97]
> [compute-1-12:08033] [ 5] 
> /usr/local/gromacs-4.0.4/bin/mdrun_sm(dd_bonded_cg_distance+0x36c) 
> [0x45ff53]
> [compute-1-12:08033] [ 6] 
> /usr/local/gromacs-4.0.4/bin/mdrun_sm(init_domain_decomposition+0x780) 
> [0x44d10c]
> [compute-1-12:08033] [ 7] 
> /usr/local/gromacs-4.0.4/bin/mdrun_sm(mdrunner+0x89c) [0x429f6e]
> [compute-1-12:08033] [ 8] 
> /usr/local/gromacs-4.0.4/bin/mdrun_sm(main+0x7ba) [0x4306b6]
> [compute-1-12:08033] [ 9] /lib64/libc.so.6(__libc_start_main+0xf4) 
> [0x378f61d8b4]
> [compute-1-12:08033] [10] /usr/local/gromacs-4.0.4/bin/mdrun_sm [0x4199a9]
> [compute-1-12:08033] *** End of error message ***
> mpirun noticed that job rank 0 with PID 8033 on node compute-1-12.local 
> exited on signal 11 (Segmentation fault).
> 1 additional process aborted (not shown)
> 
> 
> ------------------------------------------------------------------------
> 
> _______________________________________________
> gmx-users mailing list    gmx-users at gromacs.org
> http://lists.gromacs.org/mailman/listinfo/gmx-users
> Please search the archive at http://www.gromacs.org/search before posting!
> Please don't post (un)subscribe requests to the list. Use the 
> www interface or send it to gmx-users-request at gromacs.org.
> Can't post? Read http://www.gromacs.org/mailing_lists/users.php



More information about the gromacs.org_gmx-users mailing list