[petsc-dev] sm_70
Mark Adams
mfadams at lbl.gov
Sat Sep 26 11:44:47 CDT 2020
0:30 adams/feature-mat-cuda *= ~/petsc$ nvidia-smi
Sat Sep 26 12:42:37 2020
+-----------------------------------------------------------------------------+
| NVIDIA-SMI 418.116.00 Driver Version: 418.116.00 CUDA Version: 10.1
|
|-------------------------------+----------------------+----------------------+
| GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr.
ECC |
| Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute
M. |
|===============================+======================+======================|
| 0 Tesla V100-SXM2... On | 00000004:05:00.0 Off |
0 |
| N/A 38C P0 36W / 300W | 0MiB / 16130MiB | 0% E.
Process |
+-------------------------------+----------------------+----------------------+
| 1 Tesla V100-SXM2... On | 00000035:04:00.0 Off |
0 |
| N/A 40C P0 38W / 300W | 0MiB / 16130MiB | 0% E.
Process |
+-------------------------------+----------------------+----------------------+
+-----------------------------------------------------------------------------+
| Processes: GPU
Memory |
| GPU PID Type Process name Usage
|
|=============================================================================|
| No running processes found
|
+-----------------------------------------------------------------------------+
On Fri, Sep 25, 2020 at 9:36 PM Jacob Faibussowitsch <jacob.fai at gmail.com>
wrote:
> Configure by default should find out the available GPU and build for that
> sm_* it should not require the user to set this (how the heck is the user
> going to know what to set?) If I remember correctly there is a utility
> available that gives this information.
>
> For CUDA I believe the tool is nvidia-smi. Should make sure this automatic
> detection works when configuring —with-batch though since login nodes might
> have different arch than compute.
>
> Best regards,
>
> Jacob Faibussowitsch
> (Jacob Fai - booss - oh - vitch)
> Cell: (312) 694-3391
>
> On Sep 25, 2020, at 21:09, Barry Smith <bsmith at petsc.dev> wrote:
>
>
> Configure by default should find out the available GPU and build for
> that sm_* it should not require the user to set this (how the heck is the
> user going to know what to set?) If I remember correctly there is a
> utility available that gives this information.
>
> For generic builds like in package distributions I don't know how it
> should work, ideally all the possibilities would be available in the
> library and at run time the correct one will be utilized.
>
> Barry
>
>
> On Sep 25, 2020, at 5:49 PM, Mark Adams <mfadams at lbl.gov> wrote:
>
> '--CUDAFLAGS=-arch=sm_70',
>
> seems to fix this.
>
> On Fri, Sep 25, 2020 at 6:31 PM Mark Adams <mfadams at lbl.gov> wrote:
>
>> I see kokkos and hyper have a sm_70 flag, but I don't see one for PETSc.
>>
>> It looks like you have to specify this to get modern atomics to work in
>> Cuda. I get:
>>
>> /ccs/home/adams/petsc/include/petscaijdevice.h(99): error: no instance of
>> overloaded function "atomicAdd" matches the argument list
>> argument types are: (double *, double)
>>
>> I tried using a Kokkos configuration, thinking I could get these sm_70
>> flags, but that did not work.
>>
>> Any ideas?
>>
>> Mark
>>
>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-dev/attachments/20200926/69518327/attachment-0001.html>
More information about the petsc-dev
mailing list