[gmx-users] handling particle decomposition with distance restraints
Chris Neale
chris.neale at utoronto.ca
Fri Jun 26 00:59:05 CEST 2009
We are running SUSE Linux
It's best to keep all of this on the mailing list in case it becomes
useful to somebody else.
jayant james wrote:
> Hi!
> thanks for your mail. I have never used pull code before and so I am a
> bit apprehensive but I do accept your suggestion and I am working on
> that. By the wat what is the OS that you are using? Is it Suse or Fedora?
> Thanks
> Jayant James
>
> On Thu, Jun 25, 2009 at 4:31 PM, Chris Neale <chris.neale at utoronto.ca
> <mailto:chris.neale at utoronto.ca>> wrote:
>
> Let me re-emphasize that the pull code may be a good solution for you.
>
> As per your request, I currently use the following without any
> problems:
>
> fftw 3.1.2
> gromacs 3.3.1 or 4.0.4
> openmpi 1.2.6
>
> Be especially aware that openmpi 1.3.0 and 1.3.1 are broken, as I
> posted here:
>
> http://lists.gromacs.org/pipermail/gmx-users/2009-March/040844.html
>
>
> To be clear, I have never experienced any openmpi-based problems
> with any version of gromacs 4 and openmpi 1.2.6.
>
> I posted the original notice of our problems with openmpi (1.2.1)
> that were solved by using lam here.
> http://www.mail-archive.com/gmx-users@gromacs.org/msg08257.html
>
> Chris
>
> jayant james wrote:
>
> Hi!
> thanks for your mail. Could you please share what OS and
> versions of fftw, openmpi and gmx you are currently using.
> Thanks you
> JJ
>
> On Thu, Jun 25, 2009 at 12:28 PM, <chris.neale at utoronto.ca
> <mailto:chris.neale at utoronto.ca>
> <mailto:chris.neale at utoronto.ca
> <mailto:chris.neale at utoronto.ca>>> wrote:
>
> Why not use the pull code? If you haev to use distance
> restraints,
> then try LAM mpi with your pd run. We had similar error
> messages
> with vanilla .mdp files using openmpi with large and complex
> systems that went away when we switched to LAM MPI. Our
> problems
> disappeared in gmx 4 so we went back to openmpi for all
> systems as
> that mdrun_mpi version is faster in our hands.
>
> I admit, there is no good reason why LAM would work and openMPI
> would not, but I have seen it happen before so it's worth a
> shot.
>
> -- original message--
>
> The energy minimization went on without any problem on 4
> processors but the
> problem occurs when I perform the MD run. Also, I did not
> get any
> error
> message with relevance to LINCS etc...
> JJ
>
> _______________________________________________
> gmx-users mailing list gmx-users at gromacs.org
> <mailto:gmx-users at gromacs.org>
> <mailto:gmx-users at gromacs.org <mailto:gmx-users at gromacs.org>>
>
> http://lists.gromacs.org/mailman/listinfo/gmx-users
> Please search the archive at http://www.gromacs.org/search
> before
> posting!
> Please don't post (un)subscribe requests to the list. Use
> thewww
> interface or send it to gmx-users-request at gromacs.org
> <mailto:gmx-users-request at gromacs.org>
> <mailto:gmx-users-request at gromacs.org
> <mailto:gmx-users-request at gromacs.org>>.
>
> Can't post? Read http://www.gromacs.org/mailing_lists/users.php
>
>
>
>
> --
> Jayasundar Jayant James
>
> www.chick.com/reading/tracts/0096/0096_01.asp
> <http://www.chick.com/reading/tracts/0096/0096_01.asp>
> <http://www.chick.com/reading/tracts/0096/0096_01.asp>)
>
>
>
>
>
> --
> Jayasundar Jayant James
>
> www.chick.com/reading/tracts/0096/0096_01.asp
> <http://www.chick.com/reading/tracts/0096/0096_01.asp>)
>
More information about the gromacs.org_gmx-users
mailing list