[petsc-dev] sm_70
Mark Adams
mfadams at lbl.gov
Sat Sep 26 11:46:16 CDT 2020
That was a login node. Here is a compute node on SUMMIT:
12:45 130 adams/feature-mat-cuda *= ~/petsc/src/mat/tutorials$ nvidia-smi
Sat Sep 26 12:45:28 2020
+-----------------------------------------------------------------------------+
| NVIDIA-SMI 418.116.00 Driver Version: 418.116.00 CUDA Version: 10.1
|
|-------------------------------+----------------------+----------------------+
| GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr.
ECC |
| Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute
M. |
|===============================+======================+======================|
| 0 Tesla V100-SXM2... On | 00000004:04:00.0 Off |
0 |
| N/A 36C P0 36W / 300W | 0MiB / 16130MiB | 0% E.
Process |
+-------------------------------+----------------------+----------------------+
| 1 Tesla V100-SXM2... On | 00000004:05:00.0 Off |
0 |
| N/A 41C P0 36W / 300W | 0MiB / 16130MiB | 0% E.
Process |
+-------------------------------+----------------------+----------------------+
| 2 Tesla V100-SXM2... On | 00000035:03:00.0 Off |
0 |
| N/A 34C P0 36W / 300W | 0MiB / 16130MiB | 0% E.
Process |
+-------------------------------+----------------------+----------------------+
| 3 Tesla V100-SXM2... On | 00000035:04:00.0 Off |
0 |
| N/A 42C P0 37W / 300W | 0MiB / 16130MiB | 0% E.
Process |
+-------------------------------+----------------------+----------------------+
+-----------------------------------------------------------------------------+
| Processes: GPU
Memory |
| GPU PID Type Process name Usage
|
|=============================================================================|
| No running processes found
|
+-----------------------------------------------------------------------------+
On Sat, Sep 26, 2020 at 12:44 PM Mark Adams <mfadams at lbl.gov> wrote:
> 0:30 adams/feature-mat-cuda *= ~/petsc$ nvidia-smi
> Sat Sep 26 12:42:37 2020
>
> +-----------------------------------------------------------------------------+
> | NVIDIA-SMI 418.116.00 Driver Version: 418.116.00 CUDA Version: 10.1
> |
>
> |-------------------------------+----------------------+----------------------+
> | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr.
> ECC |
> | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute
> M. |
>
> |===============================+======================+======================|
> | 0 Tesla V100-SXM2... On | 00000004:05:00.0 Off |
> 0 |
> | N/A 38C P0 36W / 300W | 0MiB / 16130MiB | 0% E.
> Process |
>
> +-------------------------------+----------------------+----------------------+
> | 1 Tesla V100-SXM2... On | 00000035:04:00.0 Off |
> 0 |
> | N/A 40C P0 38W / 300W | 0MiB / 16130MiB | 0% E.
> Process |
>
> +-------------------------------+----------------------+----------------------+
>
>
> +-----------------------------------------------------------------------------+
> | Processes: GPU
> Memory |
> | GPU PID Type Process name Usage
> |
>
> |=============================================================================|
> | No running processes found
> |
>
> +-----------------------------------------------------------------------------+
>
> On Fri, Sep 25, 2020 at 9:36 PM Jacob Faibussowitsch <jacob.fai at gmail.com>
> wrote:
>
>> Configure by default should find out the available GPU and build for
>> that sm_* it should not require the user to set this (how the heck is the
>> user going to know what to set?) If I remember correctly there is a
>> utility available that gives this information.
>>
>> For CUDA I believe the tool is nvidia-smi. Should make sure this
>> automatic detection works when configuring —with-batch though since login
>> nodes might have different arch than compute.
>>
>> Best regards,
>>
>> Jacob Faibussowitsch
>> (Jacob Fai - booss - oh - vitch)
>> Cell: (312) 694-3391
>>
>> On Sep 25, 2020, at 21:09, Barry Smith <bsmith at petsc.dev> wrote:
>>
>>
>> Configure by default should find out the available GPU and build for
>> that sm_* it should not require the user to set this (how the heck is the
>> user going to know what to set?) If I remember correctly there is a
>> utility available that gives this information.
>>
>> For generic builds like in package distributions I don't know how it
>> should work, ideally all the possibilities would be available in the
>> library and at run time the correct one will be utilized.
>>
>> Barry
>>
>>
>> On Sep 25, 2020, at 5:49 PM, Mark Adams <mfadams at lbl.gov> wrote:
>>
>> '--CUDAFLAGS=-arch=sm_70',
>>
>> seems to fix this.
>>
>> On Fri, Sep 25, 2020 at 6:31 PM Mark Adams <mfadams at lbl.gov> wrote:
>>
>>> I see kokkos and hyper have a sm_70 flag, but I don't see one for PETSc.
>>>
>>> It looks like you have to specify this to get modern atomics to work in
>>> Cuda. I get:
>>>
>>> /ccs/home/adams/petsc/include/petscaijdevice.h(99): error: no instance
>>> of overloaded function "atomicAdd" matches the argument list
>>> argument types are: (double *, double)
>>>
>>> I tried using a Kokkos configuration, thinking I could get these sm_70
>>> flags, but that did not work.
>>>
>>> Any ideas?
>>>
>>> Mark
>>>
>>
>>
>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-dev/attachments/20200926/5f7a13e1/attachment.html>
More information about the petsc-dev
mailing list