[petsc-users] Compiling PETSc in Polaris with gnu
Satish Balay
balay at mcs.anl.gov
Thu May 2 16:51:47 CDT 2024
Perhaps you need to:
module load craype-accel-nvidia80
And then rebuild PETSc, your application
And have the same list of modules loaded at runtime.
Satish
On Thu, 2 May 2024, Vanella, Marcos (Fed) via petsc-users wrote:
> Thank you Satish and Junchao! I was able to compile PETSc with your configure options + suitesparse and hypre, and then compile my fortran code linking to PETSc.
> But when I try to run my test run I'm picking up an error at the very beginning:
>
> MPICH ERROR [Rank 0] [job id 01eb3c4a-28a7-4178-aced-512b4fb704c6] [Thu May 2 20:44:26 2024] [x3006c0s19b1n0] - Abort(-1) (rank 0 in comm 0): MPIDI_CRAY_init: GPU_SUPPORT_ENABLED is requested, but GTL library is not linked
> (Other MPI error)
>
> aborting job:
> MPIDI_CRAY_init: GPU_SUPPORT_ENABLED is requested, but GTL library is not linked
>
> It says in the Polaris user guide that:
>
> The environment variable MPICH_GPU_SUPPORT_ENABLED=1 needs to be set if your application requires MPI-GPU support whereby the MPI library sends and receives data directly from GPU buffers. In this case, it will be important to have the craype-accel-nvidia80 module loaded both when compiling your application and during runtime to correctly link against a GPU Transport Layer (GTL) MPI library. Otherwise, you'll likely see GPU_SUPPORT_ENABLED is requested, but GTL library is not linked errors during runtime.
>
> I tried adding loading this module (also needed to add nvhpc-mixed) in my submission script but I get the same result.
> I'll get in touch with alcf help on this.
>
>
>
> ________________________________
> From: Satish Balay <balay at mcs.anl.gov>
> Sent: Thursday, May 2, 2024 11:58 AM
> To: Junchao Zhang <junchao.zhang at gmail.com>
> Cc: petsc-users <petsc-users at mcs.anl.gov>; Vanella, Marcos (Fed) <marcos.vanella at nist.gov>; Mueller, Eric V. (Fed) <eric.mueller at nist.gov>
> Subject: Re: [petsc-users] Compiling PETSc in Polaris with gnu
>
> I just tried a build (used default versions) - and the following builds for me [on the login node].
>
>
> module use /soft/modulefiles
> module load PrgEnv-gnu
> module load cudatoolkit-standalone
> module load cray-libsci
> ./configure --with-cc=cc --with-fc=ftn --with-cxx=CC --with-make-np=4 --with-cuda=1 --with-cudac=nvcc --with-cuda-arch=80 \
> --with-debugging=0 COPTFLAGS=-O2 CXXOPTFLAGS=-O2 FOPTFLAGS=-O2 CUDAOPTFLAGS=-O2 --download-kokkos --download-kokkos-kernels
> make
>
> Satish
>
> ---
>
> balay at polaris-login-01:~> module list
>
> Currently Loaded Modules:
> 1) libfabric/1.15.2.0 4) darshan/3.4.4 7) cray-dsmml/0.2.2 10) cray-pals/1.3.4 13) PrgEnv-gnu/8.5.0
> 2) craype-network-ofi 5) gcc-native/12.3 8) cray-mpich/8.1.28 11) cray-libpals/1.3.4 14) cudatoolkit-standalone/12.2.2
> 3) perftools-base/23.12.0 6) craype/2.7.30 9) cray-pmi/6.1.13 12) craype-x86-milan 15) cray-libsci/23.12.5
>
>
> On Thu, 2 May 2024, Junchao Zhang wrote:
>
> > I used cudatoolkit-standalone/12.4.1 and gcc-12.3.
> >
> > Be sure to use the latest petsc/main or petsc/release, which contains fixes
> > for Polaris.
> >
> > --Junchao Zhang
> >
> >
> > On Thu, May 2, 2024 at 10:23 AM Satish Balay via petsc-users <
> > petsc-users at mcs.anl.gov> wrote:
> >
> > > Try:
> > >
> > > module use /soft/modulefiles
> > >
> > > Satish
> > >
> > > On Thu, 2 May 2024, Vanella, Marcos (Fed) via petsc-users wrote:
> > >
> > > > Hi all, it seems the modules in Polaris have changed (can't find
> > > cudatoolkit-standalone anymore).
> > > > Does anyone have recent experience compiling the library with gnu and
> > > cuda in the machine?
> > > > Thank you!
> > > > Marcos
> > > >
> > >
> > >
> >
>
More information about the petsc-users
mailing list