[gmx-developers] Possible bug in gmx

Alexey Shvetsov alexxy at omrb.pnpi.spb.ru
Sun Sep 28 15:19:03 CEST 2014


Hi Berk!

Its not a cut and paste error, also there are no pdb dumps.
Also I see this error before with other systems.

I can provide tpr file for that system


https://biod.pnpi.spb.ru/~alexxy/gmx/psa_pep_ctrl.md_npt.tpr


Berk Hess писал 28-09-2014 16:37:
> Hi,
> 
> I assume that your old and be coordinates being identical is correct
> and not a cut-and-paste error.
> This seems a bit strange or do you freeze part of the system?
> The only things moving here are then the domain boundaries and I don't
> see an issue there, since they only moved a little.
> 
> Do you have any more output besides the error message? PDB dump files 
> maybe?
> 
> Cheers,
> 
> Berk
> 
> On 09/28/2014 02:22 PM, Alexey Shvetsov wrote:
>> Hi,
>> 
>> just wanna add that this error seems to be reproducable even on single 
>> node. Also i get same error for gpu runs.
>> However i dont see it in large systems (800k+ atoms) running on large 
>> number of cpus (512+)
>> 
>> Alexey Shvetsov писал 28-09-2014 13:44:
>>> Hi,
>>> 
>>> DD grid is
>>> 
>>> Domain decomposition grid 4 x 1 x 1, separate PME ranks 0
>>> PME domain decomposition: 4 x 1 x 1
>>> 
>>> for 4 node setup
>>> 
>>> and
>>> 
>>> Domain decomposition grid 4 x 2 x 1, separate PME ranks 0
>>> PME domain decomposition: 4 x 2 x 1
>>> 
>>> for 8 node setup
>>> 
>>> It's reproducable with 5.0 release and latest git master. I try to
>>> check if its reproducable with 1 node. Also i can provide tpr file 
>>> for
>>> this system
>>> 
>>> Mark Abraham писал 28-09-2014 13:28:
>>>> Hi,
>>>> 
>>>> It's hard to say on that information. There were some issues fixed 
>>>> in
>>>> the lead-up to GROMACS 5 with DD not always working with 2 domains 
>>>> in
>>>> a direction, but that's a pure guess. I'd assume you can reproduce
>>>> this with release-5-0 branch. Do you observe it with a single 
>>>> domain?
>>>> If not, then it's surely a bug (and should be submitted to redmine).
>>>> 
>>>> Mark
>>>> 
>>>> On Sun, Sep 28, 2014 at 11:18 AM, Alexey Shvetsov
>>>> <alexxy at omrb.pnpi.spb.ru> wrote:
>>>> 
>>>>> Hi all!
>>>>> 
>>>>> I'm doing some tests with small peptide and constantly getting this
>>>>> error.
>>>>> I get it with few systems.
>>>>> 
>>>>> Systems sizes are around 10k or 20k
>>>>> I run it on 4 or 8 old nodes each with two xeon 54xx series
>>>>> 
>>>>> starting mdrun '2ZCH_3 in water'
>>>>> 50000000 steps, 100000.0 ps (continuing from step 1881000, 3762.0
>>>>> ps).
>>>>> 
>>>>> Step 13514000:
>>>>> The charge group starting at atom 6608 moved more than the distance
>>>>> allowed by the domain decomposition (1.112924) in direction X
>>>>> distance out of cell -1.193103
>>>>> Old coordinates: 5.467 0.298 3.636
>>>>> New coordinates: 5.467 0.298 3.636
>>>>> Old cell boundaries in direction X: 4.037 5.382
>>>>> New cell boundaries in direction X: 4.089 5.452
>>>>> 
>>>> --------------------------------------------------------------------------
>>>>> MPI_ABORT was invoked on rank 3 in communicator MPI_COMM_WORLD
>>>>> with errorcode 1.
>>>>> 
>>>>> NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
>>>>> You may or may not see output from other processes, depending on
>>>>> exactly when Open MPI kills them.
>>>>> 
>>>> --------------------------------------------------------------------------
>>>>> 
>>>>> -------------------------------------------------------
>>>>> Program mdrun_mpi, VERSION 5.1-dev-20140922-20c00a9-dirty-unknown
>>>>> Source code file:
>>>>> 
>>>> /var/tmp/alexxy/portage/sci-chemistry/gromacs-9999/work/gromacs-9999/src/gromacs/mdlib/domdec.cpp,
>>>>> line: 4388
>>>>> 
>>>>> Fatal error:
>>>>> A charge group moved too far between two domain decomposition steps
>>>>> This usually means that your system is not well equilibrated
>>>>> For more information and tips for troubleshooting, please check the
>>>>> GROMACS
>>>>> website at http://www.gromacs.org/Documentation/Errors [1]
>>>>> -------------------------------------------------------
>>>>> 
>>>>> -- Best Regards,
>>>>> Alexey 'Alexxy' Shvetsov, PhD
>>>>> Department of Molecular and Radiation Biophysics
>>>>> FSBI Petersburg Nuclear Physics Institute, NRC Kurchatov Institute,
>>>>> Leningrad region, Gatchina, Russia
>>>>> mailto:alexxyum at gmail.com
>>>>> mailto:alexxy at omrb.pnpi.spb.ru
>>>>> -- Gromacs Developers mailing list
>>>>> 
>>>>> * Please search the archive at
>>>>> http://www.gromacs.org/Support/Mailing_Lists/GMX-developers_List 
>>>>> [2]
>>>>> before posting!
>>>>> 
>>>>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists [3]
>>>>> 
>>>>> * For (un)subscribe requests visit
>>>>> 
>>>> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-developers
>>>>> [4] or send a mail to gmx-developers-request at gromacs.org.
>>>> 
>>>> 
>>>> 
>>>> Links:
>>>> ------
>>>> [1] http://www.gromacs.org/Documentation/Errors
>>>> [2] http://www.gromacs.org/Support/Mailing_Lists/GMX-developers_List
>>>> [3] http://www.gromacs.org/Support/Mailing_Lists
>>>> [4] 
>>>> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-developers
>>> 
>>> -- Best Regards,
>>> Alexey 'Alexxy' Shvetsov, PhD
>>> Department of Molecular and Radiation Biophysics
>>> FSBI Petersburg Nuclear Physics Institute, NRC Kurchatov Institute,
>>> Leningrad region, Gatchina, Russia
>>> mailto:alexxyum at gmail.com
>>> mailto:alexxy at omrb.pnpi.spb.ru
>> 

-- 
Best Regards,
Alexey 'Alexxy' Shvetsov, PhD
Department of Molecular and Radiation Biophysics
FSBI Petersburg Nuclear Physics Institute, NRC Kurchatov Institute,
Leningrad region, Gatchina, Russia
mailto:alexxyum at gmail.com
mailto:alexxy at omrb.pnpi.spb.ru


More information about the gromacs.org_gmx-developers mailing list