[gmx-users] Can we set the number of pure PME nodes when using GPU&CPU?
Theodore Si
sjyzhxw at gmail.com
Mon Aug 11 11:45:44 CEST 2014
Hi Mark,
This is information of our cluster, could you give us some advice as
regards to our cluster so that we can make GMX run faster on our system?
Each CPU node has 2 CPUs and each GPU node has 2 CPUs and 2 Nvidia K20M
Device Name Device Type Specifications Number
CPU Node IntelH2216JFFKRNodes CPU: 2×Intel Xeon E5-2670(8 Cores,
2.6GHz, 20MB Cache, 8.0GT)
Mem: 64GB(8×8GB) ECC Registered DDR3 1600MHz Samsung Memory 332
Fat Node IntelH2216WPFKRNodes CPU: 2×Intel Xeon E5-2670(8 Cores,
2.6GHz, 20MB Cache, 8.0GT)
Mem: 256G(16×16G) ECC Registered DDR3 1600MHz Samsung Memory 20
GPU Node IntelR2208GZ4GC CPU: 2×Intel Xeon E5-2670(8 Cores, 2.6GHz,
20MB Cache, 8.0GT)
Mem: 64GB(8×8GB) ECC Registered DDR3 1600MHz Samsung Memory 50
MIC Node IntelR2208GZ4GC CPU: 2×Intel Xeon E5-2670(8 Cores, 2.6GHz,
20MB Cache, 8.0GT)
Mem: 64GB(8×8GB) ECC Registered DDR3 1600MHz Samsung Memory 5
Computing Network Switch Mellanox Infiniband FDR Core Switch 648× FDR
Core Switch MSX6536-10R, Mellanox Unified Fabric Manager 1
Mellanox SX1036 40Gb Switch 36× 40Gb Ethernet Switch SX1036, 36× QSFP
Interface 1
Management Network Switch Extreme Summit X440-48t-10G 2-layer Switch
48× 1Giga Switch Summit X440-48t-10G, authorized by ExtremeXOS 9
Extreme Summit X650-24X 3-layer Switch 24× 10Giga 3-layer Ethernet
Switch Summit X650-24X, authorized by ExtremeXOS 1
Parallel Storage DDN Parallel Storage System DDN SFA12K Storage System 1
GPU GPU Accelerator NVIDIA Tesla Kepler K20M 70
MIC MIC Intel Xeon Phi 5110P Knights Corner 10
40Gb Ethernet Card MCX314A-BCBT Mellanox ConnextX-3 Chip 40Gb Ethernet
Card
2× 40Gb Ethernet ports, enough QSFP cables 16
SSD Intel SSD910 Intel SSD910 Disk, 400GB, PCIE 80
On 8/10/2014 5:50 AM, Mark Abraham wrote:
> That's not what I said.... "You can set..."
>
> -npme behaves the same whether or not GPUs are in use. Using separate ranks
> for PME caters to trying to minimize the cost of the all-to-all
> communication of the 3DFFT. That's still relevant when using GPUs, but if
> separate PME ranks are used, any GPUs on nodes that only have PME ranks are
> left idle. The most effective approach depends critically on the hardware
> and simulation setup, and whether you pay money for your hardware.
>
> Mark
>
>
> On Sat, Aug 9, 2014 at 2:56 AM, Theodore Si <sjyzhxw at gmail.com> wrote:
>
>> Hi,
>>
>> You mean no matter we use GPU acceleration or not, -npme is just a
>> reference?
>> Why we can't set that to a exact value?
>>
>>
>> On 8/9/2014 5:14 AM, Mark Abraham wrote:
>>
>>> You can set the number of PME-only ranks with -npme. Whether it's useful
>>> is
>>> another matter :-) The CPU-based PME offload and the GPU-based PP offload
>>> do not combine very well.
>>>
>>> Mark
>>>
>>>
>>> On Fri, Aug 8, 2014 at 7:24 AM, Theodore Si <sjyzhxw at gmail.com> wrote:
>>>
>>> Hi,
>>>> Can we set the number manually with -npme when using GPU acceleration?
>>>>
>>>>
>>>> --
>>>> Gromacs Users mailing list
>>>>
>>>> * Please search the archive at http://www.gromacs.org/
>>>> Support/Mailing_Lists/GMX-Users_List before posting!
>>>>
>>>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>>>
>>>> * For (un)subscribe requests visit
>>>> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or
>>>> send a mail to gmx-users-request at gromacs.org.
>>>>
>>>>
>> --
>> Gromacs Users mailing list
>>
>> * Please search the archive at http://www.gromacs.org/
>> Support/Mailing_Lists/GMX-Users_List before posting!
>>
>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>
>> * For (un)subscribe requests visit
>> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or
>> send a mail to gmx-users-request at gromacs.org.
>>
More information about the gromacs.org_gmx-users
mailing list