[gmx-users] problems in MPI run for REMD in GROMACS

Pabitra Mohan uniquepabs at gmail.com
Tue May 3 08:50:52 CEST 2016


Dear Gromacs Users,
Can any body suggest the solutions to my problem appearing during REMD run
in GROMACS

While executing the following command these error messages are coming

pabitra at pabitra-Dell-System-XPS-L502X:~/Desktop/ctld_remd/stage2$ mpirun
-np 4 mdrun_mpi -v -multidir sim0 sim1 sim2 sim3 sim4 -replex 500
GROMACS:    mdrun_mpi, VERSION 5.0.6

GROMACS is written by:
Emile Apol         Rossen Apostolov   Herman J.C. Berendsen Par
Bjelkmar
Aldert van Buuren  Rudi van Drunen    Anton Feenstra     Sebastian Fritsch
Gerrit Groenhof    Christoph Junghans Peter Kasson       Carsten Kutzner
Per Larsson        Justin A. Lemkul   Magnus Lundborg    Pieter Meulenhoff
Erik Marklund      Teemu Murtola      Szilard Pall       Sander Pronk
Roland Schulz      Alexey Shvetsov    Michael Shirts     Alfons Sijbers
Peter Tieleman     Christian Wennberg Maarten Wolf
and the project leaders:
Mark Abraham, Berk Hess, Erik Lindahl, and David van der Spoel

Copyright (c) 1991-2000, University of Groningen, The Netherlands.
Copyright (c) 2001-2014, The GROMACS development team at
Uppsala University, Stockholm University and
the Royal Institute of Technology, Sweden.
check out http://www.gromacs.org for more information.

GROMACS is free software; you can redistribute it and/or modify it
under the terms of the GNU Lesser General Public License
as published by the Free Software Foundation; either version 2.1
of the License, or (at your option) any later version.

GROMACS:      mdrun_mpi, VERSION 5.0.6
Executable:   /usr/bin/mdrun_mpi.openmpi
Library dir:  /usr/share/gromacs/top
Command line:
  mdrun_mpi -v -multidir sim0 sim1 sim2 sim3 sim4 -replex 500


-------------------------------------------------------
Program mdrun_mpi, VERSION 5.0.6
Source code file:
/build/gromacs-EDw7D5/gromacs-5.0.6/src/gromacs/gmxlib/main.cpp, line: 383

Fatal error:
The number of ranks (4) is not a multiple of the number of simulations (5)
For more information and tips for troubleshooting, please check the GROMACS
website at http://www.gromacs.org/Documentation/Errors
-------------------------------------------------------

Error on rank 2, will try to stop all ranks
Halting parallel program mdrun_mpi on CPU 2 out of 4

-------------------------------------------------------
Program mdrun_mpi, VERSION 5.0.6
Source code file:
/build/gromacs-EDw7D5/gromacs-5.0.6/src/gromacs/gmxlib/main.cpp, line: 383

Fatal error:
The number of ranks (4) is not a multiple of the number of simulations (5)
For more information and tips for troubleshooting, please check the GROMACS
website at http://www.gromacs.org/Documentation/Errors
-------------------------------------------------------

Error on rank 0, will try to stop all ranks
Halting parallel program mdrun_mpi on CPU 0 out of 4

-------------------------------------------------------
Program mdrun_mpi, VERSION 5.0.6
Source code file:
/build/gromacs-EDw7D5/gromacs-5.0.6/src/gromacs/gmxlib/main.cpp, line: 383

Fatal error:
The number of ranks (4) is not a multiple of the number of simulations (5)
For more information and tips for troubleshooting, please check the GROMACS
website at http://www.gromacs.org/Documentation/Errors
-------------------------------------------------------

Error on rank 1, will try to stop all ranks
Halting parallel program mdrun_mpi on CPU 1 out of 4

-------------------------------------------------------
Program mdrun_mpi, VERSION 5.0.6
Source code file:
/build/gromacs-EDw7D5/gromacs-5.0.6/src/gromacs/gmxlib/main.cpp, line: 383

Fatal error:
The number of ranks (4) is not a multiple of the number of simulations (5)
For more information and tips for troubleshooting, please check the GROMACS
website at http://www.gromacs.org/Documentation/Errors
-------------------------------------------------------

Error on rank 3, will try to stop all ranks
Halting parallel program mdrun_mpi on CPU 3 out of 4

gcq#11: "She's Not Bad, She's Just Genetically Mean" (Captain Beefheart)

--------------------------------------------------------------------------
MPI_ABORT was invoked on rank 2 in communicator MPI_COMM_WORLD
with errorcode -1.

NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
You may or may not see output from other processes, depending on
exactly when Open MPI kills them.
--------------------------------------------------------------------------

gcq#13: "If Life Seems Jolly Rotten, There's Something You've Forgotten !"
(Monty Python)


gcq#268: "It's Not Dark Yet, But It's Getting There" (Bob Dylan)


gcq#49: "I'll Master Your Language, and In the Meantime I'll Create My Own"
(Tricky)

--------------------------------------------------------------------------
mpirun has exited due to process rank 2 with PID 3232 on
node pabitra-Dell-System-XPS-L502X exiting improperly. There are two
reasons this could occur:

1. this process did not call "init" before exiting, but others in
the job did. This can cause a job to hang indefinitely while it waits
for all processes to call "init". By rule, if one process calls "init",
then ALL processes must call "init" prior to termination.

2. this process called "init", but exited without calling "finalize".
By rule, all processes that call "init" MUST call "finalize" prior to
exiting or it will be considered an "abnormal termination"

This may have caused other processes in the application to be
terminated by signals sent by mpirun (as reported here).
--------------------------------------------------------------------------
[pabitra-Dell-System-XPS-L502X:03229] 3 more processes have sent help
message help-mpi-api.txt / mpi-abort
[pabitra-Dell-System-XPS-L502X:03229] Set MCA parameter
"orte_base_help_aggregate" to 0 to see all help / error messages


Thank you very much for helping

-- 
Dr. Pabitra Mohan Behera
National Post-Doctoral Fellow (DST-SERB)
Computational Biology and Bioinformatics Lab
Institute of Life Sciences, Bhubaneswar
An Autonomous Organization of
Department of Biotechnology, Government of India
Nalco Square, Bhubaneswar
Odisha, Pin-751023 (India)
Mobile No: +91-9776503664, +91-9439770247
E-mail: pabitra at ils.res.in, uniquepabs at gmail.com


More information about the gromacs.org_gmx-users mailing list