[gmx-users] Can we set the number of pure PME nodes when using GPU&CPU?

Theodore Si sjyzhxw at gmail.com
Mon Aug 11 11:45:44 CEST 2014


Hi Mark,

This is information of our cluster, could you give us some advice as 
regards to our cluster so that we can make GMX run faster on our system?

Each CPU node has 2 CPUs and each GPU node has 2 CPUs and 2 Nvidia K20M


Device Name 	Device Type 	Specifications 	Number
CPU Node 	IntelH2216JFFKRNodes 	CPU: 2×Intel Xeon E5-2670(8 Cores, 
2.6GHz, 20MB Cache, 8.0GT)
Mem: 64GB(8×8GB) ECC Registered DDR3 1600MHz Samsung Memory 	332
Fat Node 	IntelH2216WPFKRNodes 	CPU: 2×Intel Xeon E5-2670(8 Cores, 
2.6GHz, 20MB Cache, 8.0GT)
Mem: 256G(16×16G) ECC Registered DDR3 1600MHz Samsung Memory 	20
GPU Node 	IntelR2208GZ4GC 	CPU: 2×Intel Xeon E5-2670(8 Cores, 2.6GHz, 
20MB Cache, 8.0GT)
Mem: 64GB(8×8GB) ECC Registered DDR3 1600MHz Samsung Memory 	50
MIC Node 	IntelR2208GZ4GC 	CPU: 2×Intel Xeon E5-2670(8 Cores, 2.6GHz, 
20MB Cache, 8.0GT)
Mem: 64GB(8×8GB) ECC Registered DDR3 1600MHz Samsung Memory 	5
Computing Network Switch 	Mellanox Infiniband FDR Core Switch 	648× FDR 
Core Switch MSX6536-10R, Mellanox Unified Fabric Manager 	1
Mellanox SX1036 40Gb Switch 	36× 40Gb Ethernet Switch SX1036, 36× QSFP 
Interface 	1
Management Network Switch 	Extreme Summit X440-48t-10G 2-layer Switch 
48× 1Giga Switch Summit X440-48t-10G, authorized by ExtremeXOS 	9
Extreme Summit X650-24X 3-layer Switch 	24× 10Giga 3-layer Ethernet 
Switch Summit X650-24X, authorized by ExtremeXOS 	1
Parallel Storage 	DDN Parallel Storage System 	DDN SFA12K Storage System 	1
GPU 	GPU Accelerator 	NVIDIA Tesla Kepler K20M 	70
MIC 	MIC 	Intel Xeon Phi 5110P Knights Corner 	10
40Gb Ethernet Card 	MCX314A-BCBT 	Mellanox ConnextX-3 Chip 40Gb Ethernet 
Card
2× 40Gb Ethernet ports, enough QSFP cables 	16
SSD 	Intel SSD910 	Intel SSD910 Disk, 400GB, PCIE 	80






On 8/10/2014 5:50 AM, Mark Abraham wrote:
> That's not what I said.... "You can set..."
>
> -npme behaves the same whether or not GPUs are in use. Using separate ranks
> for PME caters to trying to minimize the cost of the all-to-all
> communication of the 3DFFT. That's still relevant when using GPUs, but if
> separate PME ranks are used, any GPUs on nodes that only have PME ranks are
> left idle. The most effective approach depends critically on the hardware
> and simulation setup, and whether you pay money for your hardware.
>
> Mark
>
>
> On Sat, Aug 9, 2014 at 2:56 AM, Theodore Si <sjyzhxw at gmail.com> wrote:
>
>> Hi,
>>
>> You mean no matter we use GPU acceleration or not, -npme is just a
>> reference?
>> Why we can't set that to a exact value?
>>
>>
>> On 8/9/2014 5:14 AM, Mark Abraham wrote:
>>
>>> You can set the number of PME-only ranks with -npme. Whether it's useful
>>> is
>>> another matter :-) The CPU-based PME offload and the GPU-based PP offload
>>> do not combine very well.
>>>
>>> Mark
>>>
>>>
>>> On Fri, Aug 8, 2014 at 7:24 AM, Theodore Si <sjyzhxw at gmail.com> wrote:
>>>
>>>   Hi,
>>>> Can we set the number manually with -npme when using GPU acceleration?
>>>>
>>>>
>>>> --
>>>> Gromacs Users mailing list
>>>>
>>>> * Please search the archive at http://www.gromacs.org/
>>>> Support/Mailing_Lists/GMX-Users_List before posting!
>>>>
>>>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>>>
>>>> * For (un)subscribe requests visit
>>>> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or
>>>> send a mail to gmx-users-request at gromacs.org.
>>>>
>>>>
>> --
>> Gromacs Users mailing list
>>
>> * Please search the archive at http://www.gromacs.org/
>> Support/Mailing_Lists/GMX-Users_List before posting!
>>
>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>
>> * For (un)subscribe requests visit
>> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or
>> send a mail to gmx-users-request at gromacs.org.
>>



More information about the gromacs.org_gmx-users mailing list