[petsc-dev] Error running on Titan with GPUs & GNU

Karl Rupp rupp at iue.tuwien.ac.at
Wed Oct 31 04:05:09 CDT 2018


Hi Mark,

please comment or remove lines 83 and 84 in
  config/BuildSystem/config/packages/cuda.py

Is there a compiler newer than GCC 4.3 available?

Best regards,
Karli



On 10/31/18 8:15 AM, Mark Adams via petsc-dev wrote:
> After loading a cuda module ...
> 
> On Wed, Oct 31, 2018 at 2:58 AM Mark Adams <mfadams at lbl.gov 
> <mailto:mfadams at lbl.gov>> wrote:
> 
>     I get an error with --with-cuda=1
> 
>     On Tue, Oct 30, 2018 at 4:44 PM Smith, Barry F. <bsmith at mcs.anl.gov
>     <mailto:bsmith at mcs.anl.gov>> wrote:
> 
>         --with-cudac=1 should be --with-cuda=1
> 
> 
> 
>          > On Oct 30, 2018, at 12:35 PM, Smith, Barry F. via petsc-dev
>         <petsc-dev at mcs.anl.gov <mailto:petsc-dev at mcs.anl.gov>> wrote:
>          >
>          >
>          >
>          >> On Oct 29, 2018, at 8:09 PM, Mark Adams <mfadams at lbl.gov
>         <mailto:mfadams at lbl.gov>> wrote:
>          >>
>          >> And a debug build seems to work:
>          >
>          >    Well ok.
>          >
>          >    Are there newer versions of the Gnu compiler for this
>         system? Are there any other compilers on the system that would
>         likely be less buggy? IBM compilers? If this simple code
>         generates a gross error with optimization who's to say how many
>         more subtle bugs may be induced in the library by the buggy
>         optimizer (there may be none but IMHO probability says there
>         will be others).
>          >
>          >    Is there any chance that valgrind runs on this machine;
>         you could run the optimized version through it and see what it says.
>          >
>          >   Barry
>          >
>          >>
>          >> 21:04 1 master= /lustre/atlas/proj-shared/geo127/petsc$ make
>         PETSC_DIR=/lustre/atlas/proj-shared/geo127/petsc_titan_dbg64idx_gnu_cuda
>         PETSC_ARCH="" test
>          >> Running test examples to verify correct installation
>          >> Using
>         PETSC_DIR=/lustre/atlas/proj-shared/geo127/petsc_titan_dbg64idx_gnu_cuda
>         and PETSC_ARCH=
>          >> *******************Error detected during compile or
>         link!*******************
>          >> See http://www.mcs.anl.gov/petsc/documentation/faq.html
>          >>
>         /lustre/atlas/proj-shared/geo127/petsc/src/snes/examples/tutorials
>         ex19
>          >>
>         *********************************************************************************
>          >> cc -o ex19.o -c -g 
>           -I/lustre/atlas/proj-shared/geo127/petsc_titan_dbg64idx_gnu_cuda/include    `pwd`/ex19.c
>          >> cc -g  -o ex19 ex19.o 
>         -L/lustre/atlas/proj-shared/geo127/petsc_titan_dbg64idx_gnu_cuda/lib
>         -Wl,-rpath,/lustre/atlas/proj-shared/geo127/petsc_titan_dbg64idx_gnu_cuda/lib
>         -L/lustre/atlas/proj-shared/geo127/petsc_titan_dbg64idx_gnu_cuda/lib
>         -lpetsc -lHYPRE -lflapack -lfblas -lparmetis -lmetis -ldl
>          >>
>         /lustre/atlas/proj-shared/geo127/petsc_titan_dbg64idx_gnu_cuda/lib/libpetsc.a(dlimpl.o):
>         In function `PetscDLOpen':
>          >>
>         /lustre/atlas1/geo127/proj-shared/petsc/src/sys/dll/dlimpl.c:108: warning:
>         Using 'dlopen' in statically linked applications requires at
>         runtime the shared libraries from the glibc version used for linking
>          >>
>         /lustre/atlas/proj-shared/geo127/petsc_titan_dbg64idx_gnu_cuda/lib/libpetsc.a(send.o):
>         In function `PetscOpenSocket':
>          >>
>         /lustre/atlas1/geo127/proj-shared/petsc/src/sys/classes/viewer/impls/socket/send.c:108:
>         warning: Using 'gethostbyname' in statically linked applications
>         requires at runtime the shared libraries from the glibc version
>         used for linking
>          >> true ex19
>          >> rm ex19.o
>          >> Possible error running C/C++
>         src/snes/examples/tutorials/ex19 with 1 MPI process
>          >> See http://www.mcs.anl.gov/petsc/documentation/faq.html
>          >> lid velocity = 0.0016, prandtl # = 1., grashof # = 1.
>          >> Number of SNES iterations = 2
>          >> Application 19081049 resources: utime ~1s, stime ~1s, Rss
>         ~17112, inblocks ~36504, outblocks ~111043
>          >> Possible error running C/C++
>         src/snes/examples/tutorials/ex19 with 2 MPI processes
>          >> See http://www.mcs.anl.gov/petsc/documentation/faq.html
>          >> lid velocity = 0.0016, prandtl # = 1., grashof # = 1.
>          >> Number of SNES iterations = 2
>          >> Application 19081050 resources: utime ~1s, stime ~1s, Rss
>         ~19816, inblocks ~36527, outblocks ~111043
>          >> 5a6
>          >>> Application 19081051 resources: utime ~1s, stime ~0s, Rss
>         ~13864, inblocks ~36527, outblocks ~111043
>          >>
>         /lustre/atlas/proj-shared/geo127/petsc/src/snes/examples/tutorials
>          >> Possible problem with ex19_hypre, diffs above
>          >> =========================================
>          >> *******************Error detected during compile or
>         link!*******************
>          >> See http://www.mcs.anl.gov/petsc/documentation/faq.html
>          >>
>         /lustre/atlas/proj-shared/geo127/petsc/src/snes/examples/tutorials
>         ex5f
>          >> *********************************************************
>          >> ftn -c -g   
>         -I/lustre/atlas/proj-shared/geo127/petsc_titan_dbg64idx_gnu_cuda/include 
>            -o ex5f.o ex5f.F90
>          >> ftn -g   -o ex5f ex5f.o 
>         -L/lustre/atlas/proj-shared/geo127/petsc_titan_dbg64idx_gnu_cuda/lib
>         -Wl,-rpath,/lustre/atlas/proj-shared/geo127/petsc_titan_dbg64idx_gnu_cuda/lib
>         -L/lustre/atlas/proj-shared/geo127/petsc_titan_dbg64idx_gnu_cuda/lib
>         -lpetsc -lHYPRE -lflapack -lfblas -lparmetis -lmetis -ldl
>          >>
>         /lustre/atlas/proj-shared/geo127/petsc_titan_dbg64idx_gnu_cuda/lib/libpetsc.a(dlimpl.o):
>         In function `PetscDLOpen':
>          >>
>         /lustre/atlas1/geo127/proj-shared/petsc/src/sys/dll/dlimpl.c:108: warning:
>         Using 'dlopen' in statically linked applications requires at
>         runtime the shared libraries from the glibc version used for linking
>          >>
>         /lustre/atlas/proj-shared/geo127/petsc_titan_dbg64idx_gnu_cuda/lib/libpetsc.a(send.o):
>         In function `PetscOpenSocket':
>          >>
>         /lustre/atlas1/geo127/proj-shared/petsc/src/sys/classes/viewer/impls/socket/send.c:108:
>         warning: Using 'gethostbyname' in statically linked applications
>         requires at runtime the shared libraries from the glibc version
>         used for linking
>          >> rm ex5f.o
>          >> Possible error running Fortran example
>         src/snes/examples/tutorials/ex5f with 1 MPI process
>          >> See http://www.mcs.anl.gov/petsc/documentation/faq.html
>          >> Number of SNES iterations =     4
>          >> Application 19081055 resources: utime ~1s, stime ~0s, Rss
>         ~12760, inblocks ~36800, outblocks ~111983
>          >> Completed test examples
>          >> 21:06 master= /lustre/atlas/proj-shared/geo127/petsc$
>          >
> 


More information about the petsc-dev mailing list