[gmx-users] About Plumed+Gromacs+MPI
yong zhou
zy_dut at hotmail.com
Sun Dec 2 04:13:37 CET 2018
Dear all,
I have compiled the plumed+gromacs with the following options
CC=/usr/lib64/openmpi3/bin/mpicc FC=/usr/lib64/openmpi3/bin/mpif90 F77=/usr/lib64/openmpi3/bin/mpif90 CXX=/usr/lib64/openmpi3/bin/mpicxx CMAKE_PREFIX_PATH=//usr/lib64/openmpi3/ cmake3 .. -DGMX_BUILD_OWN_FFTW=on -DGMX_MPI=on -DCMAKE_C_COMPILER=/usr/lib64/openmpi3/bin/mpicc -DCMAKE_CXX_COMPILER=/usr/lib64/openmpi3/bin/mpicxx -DGMX_GPU=on -DNVML_INCLUDE_DIR=/usr/local/cuda-9.0/include -DNVML_LIBRARY=/usr/lib64/libnvidia-ml.so -DBUILD_SHARED_LIBS=OFF -DGMX_PREFER_STATIC_LIBS=ON -DCMAKE_INSTALL_PREFIX=/home/yzhou/xtal/gromacs
But with running the tutorial of Plumed, I got the following errors.
:-) GROMACS - gmx mdrun, 2016.5 (-:
GROMACS is written by:
Emile Apol Rossen Apostolov Herman J.C. Berendsen Par Bjelkmar
Aldert van Buuren Rudi van Drunen Anton Feenstra Gerrit Groenhof
Christoph Junghans Anca Hamuraru Vincent Hindriksen Dimitrios Karkoulis
Peter Kasson Jiri Kraus Carsten Kutzner Per Larsson
Justin A. Lemkul Magnus Lundborg Pieter Meulenhoff Erik Marklund
Teemu Murtola Szilard Pall Sander Pronk Roland Schulz
Alexey Shvetsov Michael Shirts Alfons Sijbers Peter Tieleman
Teemu Virolainen Christian Wennberg Maarten Wolf
and the project leaders:
Mark Abraham, Berk Hess, Erik Lindahl, and David van der Spoel
Copyright (c) 1991-2000, University of Groningen, The Netherlands.
Copyright (c) 2001-2017, The GROMACS development team at
Uppsala University, Stockholm University and
the Royal Institute of Technology, Sweden.
check out http://www.gromacs.org for more information.
GROMACS is free software; you can redistribute it and/or modify it
under the terms of the GNU Lesser General Public License
as published by the Free Software Foundation; either version 2.1
of the License, or (at your option) any later version.
GROMACS: gmx mdrun, version 2016.5
Executable: /home/yzhou/xtal/gromacs/bin/gmx_mpi
Data prefix: /home/yzhou/xtal/gromacs
Working dir: /home/yzhou/data/simulation/plumed/belfast-8/second
Command line:
gmx_mpi mdrun -s topol -plumed plumed -multi 2 -replex 20
Back Off! I just backed up md1.log to ./#md1.log.18#
Back Off! I just backed up md0.log to ./#md0.log.18#
Running on 1 node with total 10 cores, 10 logical cores, 1 compatible GPU
Hardware detected on host localhost.localdomain (the node of MPI rank 0):
CPU info:
Vendor: Intel
Brand: Intel(R) Xeon(R) CPU E5-2650 v3 @ 2.30GHz
SIMD instructions most likely to fit this hardware: AVX2_256
SIMD instructions selected at GROMACS compile time: AVX2_256
Hardware topology: Basic
GPU info:
Number of GPUs detected: 1
#0: NVIDIA GeForce GTX 1080 Ti, compute cap.: 6.1, ECC: no, stat: compatible
Reading file topol1.tpr, VERSION 4.6.5 (single precision)
Reading file topol0.tpr, VERSION 4.6.5 (single precision)
Note: file tpx version 83, software tpx version 110
Note: file tpx version 83, software tpx version 110
This is simulation 1 out of 2 running as a composite GROMACS
multi-simulation job. Setup for this simulation:
Using 1 MPI process
Using 5 OpenMP threads
This is simulation 0 out of 2 running as a composite GROMACS
multi-simulation job. Setup for this simulation:
Using 1 MPI process
Using 5 OpenMP threads
1 compatible GPU is present, with ID 0
1 GPU auto-selected for this run.
Mapping of GPU ID to the 2 PP ranks in this node: 0,0
1 compatible GPU is present, with ID 0
1 GPU auto-selected for this run.
Mapping of GPU ID to the 2 PP ranks in this node: 0,0
Non-default thread affinity set probably by the OpenMP library,
disabling internal thread affinity
Non-default thread affinity set probably by the OpenMP library,
disabling internal thread affinity
-------------------------------------------------------
Program: gmx mdrun, version 2016.5
Source file: src/programs/mdrun/runner.cpp (line 677)
Function: double (* gmx::my_integrator(unsigned int))(FILE*, t_commrec*, int, const t_filenm*, const gmx_output_env_t*, gmx_bool, int, gmx_vsite_t*, gmx_constr_t, int, t_inputrec*, gmx_mtop_t*, t_fcdata*, t_state*, t_mdatoms*, t_nrnb*, gmx_wallcycle_t, gmx_edsam_t, t_forcerec*, int, int, int, gmx_membed_t*, real, real, int, long unsigned int, gmx_walltime_accounting_t)
MPI rank: 1 (out of 2)
Feature not implemented:
-------------------------------------------------------
Program: gmx mdrun, version 2016.5
Source file: src/programs/mdrun/runner.cpp (line 677)
Function: double (* gmx::my_integrator(unsigned int))(FILE*, t_commrec*, int, const t_filenm*, const gmx_output_env_t*, gmx_bool, int, gmx_vsite_t*, gmx_constr_t, int, t_inputrec*, gmx_mtop_t*, t_fcdata*, t_state*, t_mdatoms*, t_nrnb*, gmx_wallcycle_t, gmx_edsam_t, t_forcerec*, int, int, int, gmx_membed_t*, real, real, int, long unsigned int, gmx_walltime_accounting_t)
MPI rank: 0 (out of 2)
Feature not implemented:
SD2 integrator has been removed
For more information and tips for troubleshooting, please check the GROMACS
website at http://www.gromacs.org/Documentation/Errors
-------------------------------------------------------
SD2 integrator has been removed
For more information and tips for troubleshooting, please check the GROMACS
website at http://www.gromacs.org/Documentation/Errors
-------------------------------------------------------
--------------------------------------------------------------------------
MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD
with errorcode 1.
NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
You may or may not see output from other processes, depending on
exactly when Open MPI kills them.
--------------------------------------------------------------------------
--------------------------------------------------------------------------
MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD
with errorcode 1.
NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
You may or may not see output from other processes, depending on
exactly when Open MPI kills them.
--------------------------------------------------------------------------
Would you please tell me what is wrong with the process and how to solve the problem?
Thanks a lot.
Best regards
Zhou Yong
More information about the gromacs.org_gmx-users
mailing list