[gmx-users] Dell PowerEdge M710 with Intel Xeon 5667 processor

Mark Abraham Mark.Abraham at anu.edu.au
Wed Jan 19 22:30:36 CET 2011


On 20/01/2011 3:33 AM, Maryam Hamzehee wrote:
> Dear Szilárd,
>
> Many thanks for your reply. I've got following reply from my question 
> from Linux-PowerEdge mailing list. I was wondering which one applies 
> to GROMACS parallel computation (I mean CPU bound, disk bound, etc).
>

In serial, GROMACS is very much CPU-bound, and a lot of work has gone 
into making the most of the CPU. In parallel, that CPU-optimization work 
is so effective that smallish packets of information have to be 
transferred regularly without much possibility of effectively 
overlapping communication and computation, and so a low-latency 
communication network is essential in order to continue making effective 
use of all the CPUs. Something like Infiniband or NUMAlink is definitely 
required.

Mark

> >There shouldn't be any linux compatibility issues with any PowerEdge
> >system.  At Duke we have a large compute cluster using a variety of
> >PowerEdge blades (including M710's) all running on linux.
>
> >What interconnect are you using?  And are your jobs memory bound, cpu
> >bound, disk bound, or network bound?
>
> >If your computation is more dependent on the interlink and communication
> >between the nodes, its more important to worry about your interconnect.
>
> >If Inter-node communication is highly important, you may also want to
> >consider something like the M910.  The M910 can be configured with 4
> >8-core CPUs, thus giving you 32 NUMA-connected cores.  Or 64 logical
> >processors if your job is one that can benefit from HT.  Note that when
> >going with more cores-per chip, your max clockrate tends to be lower.
> >As such, its really important to know how your jobs are bound so that
> >you can order a cluster configuration that'll be best for that job.
>
>
> Cheers, Maryam
>
> --- On *Tue, 18/1/11, Szilárd Páll /<szilard.pall at cbr.su.se>/* wrote:
>
>
>     From: Szilárd Páll <szilard.pall at cbr.su.se>
>     Subject: Re: [gmx-users] Dell PowerEdge M710 with Intel Xeon 5667
>     processor
>     To: "Discussion list for GROMACS users" <gmx-users at gromacs.org>
>     Received: Tuesday, 18 January, 2011, 10:31 PM
>
>     Hi,
>
>     Although the question is a bit fuzzy, I might be able to give you a
>     useful answer.
>
>     >From what I see on the whitepaper of the Poweredge m710 baldes, among
>     other (not so interesting :) OS-es, Dell provides the options of Red
>     Had or SUSE Linux as factory installed OS-es. If you have any of
>     these, you can rest assured that Gromacs will run just fine -- on a
>     single node.
>
>     Parallel runs are little bit different story and depends on the
>     interconnect. If you have Infiniband, than you'll have a very good
>     scaling over multiple nodes. This is true especially if it's the I/O
>     cards are the Mellanox QDR-s.
>
>     Cheers,
>     --
>     Szilárd
>
>
>     On Tue, Jan 18, 2011 at 4:48 PM, Maryam Hamzehee
>     <maryam_h_7860 at yahoo.com </mc/compose?to=maryam_h_7860 at yahoo.com>>
>     wrote:
>     >
>     > Dear list,
>     >
>     > I will appreciate it if I can get your expert opinion on doing
>     parallel computation (I will use GROMACS and AMBER molecular
>     mechanics packages and some other programs like CYANA, ARIA and
>     CNS to do structure calculations based on NMR experimental data)
>     using a cluster based on Dell PowerEdge M710 with Intel Xeon 5667
>     processor architecture which
>     > apparently each blade has two quad-core cpus. I was wondering if
>     I can get some information about LINUX compatibility and parallel
>     computation on this system.
>     > Cheers,
>     > Maryam
>     >
>     > --
>     > gmx-users mailing list gmx-users at gromacs.org
>     </mc/compose?to=gmx-users at gromacs.org>
>     > http://lists.gromacs.org/mailman/listinfo/gmx-users
>     > Please search the archive at
>     http://www.gromacs.org/Support/Mailing_Lists/Search before posting!
>     > Please don't post (un)subscribe requests to the list. Use the
>     > www interface or send it to gmx-users-request at gromacs.org
>     </mc/compose?to=gmx-users-request at gromacs.org>.
>     > Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>     --
>     gmx-users mailing list gmx-users at gromacs.org
>     </mc/compose?to=gmx-users at gromacs.org>
>     http://lists.gromacs.org/mailman/listinfo/gmx-users
>     Please search the archive at
>     http://www.gromacs.org/Support/Mailing_Lists/Search before posting!
>     Please don't post (un)subscribe requests to the list. Use the
>     www interface or send it to gmx-users-request at gromacs.org
>     </mc/compose?to=gmx-users-request at gromacs.org>.
>     Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://maillist.sys.kth.se/pipermail/gromacs.org_gmx-users/attachments/20110120/b374eca8/attachment.html>


More information about the gromacs.org_gmx-users mailing list