[gmx-users] GMX 2018 regression tests: cufftPlanMany R2C plan failure (error code 5)

Alex nedomacho at gmail.com
Wed Feb 7 17:18:43 CET 2018


Hi Mark,

Nothing has been installed yet, so the commands were issued from 
/build/bin and so I am not sure about the output of that mdrun-test (let 
me know what exact command could make it more informative).

Thank you,

Alex

***

 > ./gmx -version

GROMACS version:    2018
Precision:          single
Memory model:       64 bit
MPI library:        thread_mpi
OpenMP support:     enabled (GMX_OPENMP_MAX_THREADS = 64)
GPU support:        CUDA
SIMD instructions:  AVX2_256
FFT library:        fftw-3.3.5-fma-sse2-avx-avx2-avx2_128-avx512
RDTSCP usage:       enabled
TNG support:        enabled
Hwloc support:      hwloc-1.11.0
Tracing support:    disabled
Built on:           2018-02-06 19:30:36
Built by:           smolyan at 647trc-md1 [CMAKE]
Build OS/arch:      Linux 4.4.0-112-generic x86_64
Build CPU vendor:   Intel
Build CPU brand:    Intel(R) Xeon(R) CPU E5-2699 v4 @ 2.20GHz
Build CPU family:   6   Model: 79   Stepping: 1
Build CPU features: aes apic avx avx2 clfsh cmov cx8 cx16 f16c fma hle 
htt intel lahf mmx msr nonstop_tsc pcid pclmuldq pdcm pdpe1gb popcnt pse 
rdrnd rdtscp rtm sse2 sse3 sse4.1 sse4.2 ssse3 tdt x2apic
C compiler:         /usr/bin/cc GNU 5.4.0
C compiler flags:    -march=core-avx2     -O3 -DNDEBUG 
-funroll-all-loops -fexcess-precision=fast
C++ compiler:       /usr/bin/c++ GNU 5.4.0
C++ compiler flags:  -march=core-avx2    -std=c++11   -O3 -DNDEBUG 
-funroll-all-loops -fexcess-precision=fast
CUDA compiler:      /usr/local/cuda/bin/nvcc nvcc: NVIDIA (R) Cuda 
compiler driver;Copyright (c) 2005-2017 NVIDIA Corporation;Built on 
Fri_Nov__3_21:07:56_CDT_2017;Cuda compilation tools, release 9.1, V9.1.85
CUDA compiler 
flags:-gencode;arch=compute_30,code=sm_30;-gencode;arch=compute_35,code=sm_35;-gencode;arch=compute_37,code=sm_37;-gencode;arch=compute_50,code=sm_50;-gencode;arch=compute_52,code=sm_52;-gencode;arch=compute_60,code=sm_60;-gencode;arch=compute_61,code=sm_61;-gencode;arch=compute_70,code=sm_70;-gencode;arch=compute_70,code=compute_70;-use_fast_math;-D_FORCE_INLINES;; 
;-march=core-avx2;-std=c++11;-O3;-DNDEBUG;-funroll-all-loops;-fexcess-precision=fast;
CUDA driver:        9.10
CUDA runtime:       9.10

 > ldd -r ./mdrun-test
         linux-vdso.so.1 =>  (0x00007ffcfcc3e000)
         libgromacs.so.3 => 
/home/smolyan/scratch/gmx2018_install_temp/gromacs-2018/build/bin/./../lib/libgromacs.so.3 
(0x00007faa58f8f000)
         libpthread.so.0 => /lib/x86_64-linux-gnu/libpthread.so.0 
(0x00007faa58d72000)
         libstdc++.so.6 => /usr/lib/x86_64-linux-gnu/libstdc++.so.6 
(0x00007faa589f0000)
         libm.so.6 => /lib/x86_64-linux-gnu/libm.so.6 (0x00007faa586e7000)
         libgcc_s.so.1 => /lib/x86_64-linux-gnu/libgcc_s.so.1 
(0x00007faa584d1000)
         libc.so.6 => /lib/x86_64-linux-gnu/libc.so.6 (0x00007faa58107000)
         libdl.so.2 => /lib/x86_64-linux-gnu/libdl.so.2 (0x00007faa57f03000)
         librt.so.1 => /lib/x86_64-linux-gnu/librt.so.1 (0x00007faa57cfb000)
         libcufft.so.9.1 => /usr/local/cuda/lib64/libcufft.so.9.1 
(0x00007faa5080e000)
         libhwloc.so.5 => /usr/lib/x86_64-linux-gnu/libhwloc.so.5 
(0x00007faa505d4000)
         libgomp.so.1 => /usr/lib/x86_64-linux-gnu/libgomp.so.1 
(0x00007faa503b2000)
         /lib64/ld-linux-x86-64.so.2 (0x00007faa5c1ad000)
         libnuma.so.1 => /usr/lib/x86_64-linux-gnu/libnuma.so.1 
(0x00007faa501a7000)
         libltdl.so.7 => /usr/lib/x86_64-linux-gnu/libltdl.so.7 
(0x00007faa4ff9d000)


On 2/7/2018 5:13 AM, Mark Abraham wrote:
> Hi,
>
> I checked back with the CUDA-facing GROMACS developers. They've run the
> code with 9.1 and believe there's no intrinsic problem within GROMACS.
>
>> So I don't have much to suggest other then rebuilding everything cleanly,
> as this is an internal non-descript cuFFT/driver error that is not supposed
> to happen,
> especially in mdrun-test with its single input system, and it will prevent
> him from using -pme gpu.
>> The only thing PME could do better is to show more meaningful error
> messages (which would have to be hardcoded anyway as cuFFT doesn't even
> have human readable strings for error codes).
>
> If you could share the output of
> * gmx -version
> * ldd -r mdrun-test
> then perhaps we can find an issue (or at least report to nvidia usefully).
> Ensuring you are using the CUDA driver that came with the CUDA runtime is
> most likely to work smoothly.
>
> Mark
>
> On Tue, Feb 6, 2018 at 9:24 PM Alex <nedomacho at gmail.com> wrote:
>
>> And this is with:
>>> gcc --version
>>> gcc (Ubuntu 5.4.0-6ubuntu1~16.04.6) 5.4.0 20160609 <020-16%2006%2009>
>>
>>
>> On Tue, Feb 6, 2018 at 1:18 PM, Alex <nedomacho at gmail.com> wrote:
>>
>>> Hi all,
>>>
>>> I've just built the latest version and regression tests are running. Here
>>> is one error:
>>>
>>> "Program:     mdrun-test, version 2018
>>> Source file: src/gromacs/ewald/pme-3dfft.cu (line 56)
>>>
>>> Fatal error:
>>> cufftPlanMany R2C plan failure (error code 5)"
>>>
>>> This is with CUDA 9.1.
>>>
>>> Anything to worry about?
>>>
>>> Thank you,
>>>
>>> Alex
>>>
>> --
>> Gromacs Users mailing list
>>
>> * Please search the archive at
>> http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before
>> posting!
>>
>> * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
>>
>> * For (un)subscribe requests visit
>> https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or
>> send a mail to gmx-users-request at gromacs.org.
>>



More information about the gromacs.org_gmx-users mailing list