[gmx-developers] ICC 14 support
Kevin Chen
fch6699 at gmail.com
Mon Sep 29 16:35:17 CEST 2014
Hi Guys,
Was wondering if GROMACS will support ICC 14 in the near future?
Kevin Chen, Ph.D.
HPC Applications, TACC
-----Original Message-----
From: gromacs.org_gmx-developers-bounces at maillist.sys.kth.se [mailto:gromacs.org_gmx-developers-bounces at maillist.sys.kth.se] On Behalf Of Alexey Shvetsov
Sent: Sunday, September 28, 2014 3:09 PM
To: gmx-developers at gromacs.org
Subject: Re: [gmx-developers] Possible bug in gmx
Mark Abraham писал 28-09-2014 18:17:
> How about a redmine issue - this thread's not about GROMACS
> development, per se ;-)
Sorry about that =D
Redmine issue http://redmine.gromacs.org/issues/1607
With relevant tpr file attached
>
> Mark
>
> On Sun, Sep 28, 2014 at 3:18 PM, Alexey Shvetsov
> <alexxy at omrb.pnpi.spb.ru> wrote:
>
>> Hi Berk!
>>
>> Its not a cut and paste error, also there are no pdb dumps.
>> Also I see this error before with other systems.
>>
>> I can provide tpr file for that system
>>
>> https://biod.pnpi.spb.ru/~alexxy/gmx/psa_pep_ctrl.md_npt.tpr [1]
>>
>> Berk Hess писал 28-09-2014 16:37:
>>
>> Hi,
>>
>> I assume that your old and be coordinates being identical is correct
>> and not a cut-and-paste error.
>> This seems a bit strange or do you freeze part of the system?
>> The only things moving here are then the domain boundaries and I
>> don't see an issue there, since they only moved a little.
>>
>> Do you have any more output besides the error message? PDB dump files
>> maybe?
>>
>> Cheers,
>>
>> Berk
>>
>> On 09/28/2014 02:22 PM, Alexey Shvetsov wrote:
>> Hi,
>>
>> just wanna add that this error seems to be reproducable even on
>> single node. Also i get same error for gpu runs.
>> However i dont see it in large systems (800k+ atoms) running on large
>> number of cpus (512+)
>>
>> Alexey Shvetsov писал 28-09-2014 13:44:
>> Hi,
>>
>> DD grid is
>>
>> Domain decomposition grid 4 x 1 x 1, separate PME ranks 0 PME domain
>> decomposition: 4 x 1 x 1
>>
>> for 4 node setup
>>
>> and
>>
>> Domain decomposition grid 4 x 2 x 1, separate PME ranks 0 PME domain
>> decomposition: 4 x 2 x 1
>>
>> for 8 node setup
>>
>> It's reproducable with 5.0 release and latest git master. I try to
>> check if its reproducable with 1 node. Also i can provide tpr file
>> for this system
>>
>> Mark Abraham писал 28-09-2014 13:28:
>> Hi,
>>
>> It's hard to say on that information. There were some issues fixed in
>> the lead-up to GROMACS 5 with DD not always working with 2 domains in
>> a direction, but that's a pure guess. I'd assume you can reproduce
>> this with release-5-0 branch. Do you observe it with a single domain?
>> If not, then it's surely a bug (and should be submitted to redmine).
>>
>> Mark
>>
>> On Sun, Sep 28, 2014 at 11:18 AM, Alexey Shvetsov
>> <alexxy at omrb.pnpi.spb.ru> wrote:
>>
>> Hi all!
>>
>> I'm doing some tests with small peptide and constantly getting this
>> error.
>> I get it with few systems.
>>
>> Systems sizes are around 10k or 20k
>> I run it on 4 or 8 old nodes each with two xeon 54xx series
>>
>> starting mdrun '2ZCH_3 in water'
>> 50000000 steps, 100000.0 ps (continuing from step 1881000, 3762.0
>> ps).
>>
>> Step 13514000:
>> The charge group starting at atom 6608 moved more than the distance
>> allowed by the domain decomposition (1.112924) in direction X
>> distance out of cell -1.193103 Old coordinates: 5.467 0.298 3.636 New
>> coordinates: 5.467 0.298 3.636 Old cell boundaries in direction X:
>> 4.037 5.382 New cell boundaries in direction X: 4.089 5.452
>>
>>
> ----------------------------------------------------------------------
> ----
>> MPI_ABORT was invoked on rank 3 in communicator MPI_COMM_WORLD with
>> errorcode 1.
>>
>> NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
>> You may or may not see output from other processes, depending on
>> exactly when Open MPI kills them.
>>
>>
> ----------------------------------------------------------------------
> ----
>>
>> -------------------------------------------------------
>> Program mdrun_mpi, VERSION 5.1-dev-20140922-20c00a9-dirty-unknown
>> Source code file:
>>
>>
> /var/tmp/alexxy/portage/sci-chemistry/gromacs-9999/work/gromacs-9999/s
> rc/gromacs/mdlib/domdec.cpp,
>> line: 4388
>>
>> Fatal error:
>> A charge group moved too far between two domain decomposition steps
>> This usually means that your system is not well equilibrated For more
>> information and tips for troubleshooting, please check the GROMACS
>> website at http://www.gromacs.org/Documentation/Errors [2] [1]
>> -------------------------------------------------------
>>
>> -- Best Regards,
>> Alexey 'Alexxy' Shvetsov, PhD
>> Department of Molecular and Radiation Biophysics FSBI Petersburg
>> Nuclear Physics Institute, NRC Kurchatov Institute, Leningrad region,
>> Gatchina, Russia mailto:alexxyum at gmail.com
>> mailto:alexxy at omrb.pnpi.spb.ru
>> -- Gromacs Developers mailing list
>>
>> * Please search the archive at
>> http://www.gromacs.org/Support/Mailing_Lists/GMX-developers_List
>> [3] [2]
>> before posting!
>>
>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists [4]
>> [3]
>>
>> * For (un)subscribe requests visit
>>
>>
> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-developer
> s
>> [5]
>> [4] or send a mail to gmx-developers-request at gromacs.org.
>>
>> Links:
>> ------
>> [1] http://www.gromacs.org/Documentation/Errors [2] [2]
>> http://www.gromacs.org/Support/Mailing_Lists/GMX-developers_List [3]
>> [3] http://www.gromacs.org/Support/Mailing_Lists [4] [4]
>>
> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-developer
> s
>> [5]
>
> -- Best Regards,
> Alexey 'Alexxy' Shvetsov, PhD
> Department of Molecular and Radiation Biophysics FSBI Petersburg
> Nuclear Physics Institute, NRC Kurchatov Institute, Leningrad region,
> Gatchina, Russia mailto:alexxyum at gmail.com
> mailto:alexxy at omrb.pnpi.spb.ru
>
> --
> Best Regards,
> Alexey 'Alexxy' Shvetsov, PhD
> Department of Molecular and Radiation Biophysics FSBI Petersburg
> Nuclear Physics Institute, NRC Kurchatov Institute, Leningrad region,
> Gatchina, Russia mailto:alexxyum at gmail.com
> mailto:alexxy at omrb.pnpi.spb.ru
> --
> Gromacs Developers mailing list
>
> * Please search the archive at
> http://www.gromacs.org/Support/Mailing_Lists/GMX-developers_List [3]
> before posting!
>
> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists [4]
>
> * For (un)subscribe requests visit
>
> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-developer
> s [5] or send a mail to gmx-developers-request at gromacs.org.
>
>
> Links:
> ------
> [1] https://biod.pnpi.spb.ru/~alexxy/gmx/psa_pep_ctrl.md_npt.tpr
> [2] http://www.gromacs.org/Documentation/Errors
> [3] http://www.gromacs.org/Support/Mailing_Lists/GMX-developers_List
> [4] http://www.gromacs.org/Support/Mailing_Lists
> [5]
> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-developer
> s
--
Best Regards,
Alexey 'Alexxy' Shvetsov, PhD
Department of Molecular and Radiation Biophysics FSBI Petersburg Nuclear Physics Institute, NRC Kurchatov Institute, Leningrad region, Gatchina, Russia mailto:alexxyum at gmail.com mailto:alexxy at omrb.pnpi.spb.ru
--
Gromacs Developers mailing list
* Please search the archive at http://www.gromacs.org/Support/Mailing_Lists/GMX-developers_List before posting!
* Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
* For (un)subscribe requests visit
https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-developers or send a mail to gmx-developers-request at gromacs.org.
More information about the gromacs.org_gmx-developers
mailing list