[petsc-dev] Petsc "make test" have more failures for --with-openmp=1

Eric Chamberland Eric.Chamberland at giref.ulaval.ca
Sat Mar 13 07:17:53 CST 2021


Hi Pierre!

On 2021-03-13 3:17 a.m., Pierre Jolivet wrote:
> Hello Eric,
> I’ve made an “interesting” discovery, so I’ll put back the list in c/c.
> It appears the following snippet of code which uses Allreduce() + 
> lambda function + MPI_IN_PLACE is:
> - Valgrind-clean with MPICH;
> - Valgrind-clean with OpenMPI 4.0.5;
> - not Valgrind-clean with OpenMPI 4.1.0.
> I’m not sure who is to blame here, I’ll need to look at the MPI 
> specification for what is required by the implementors and users in 
> that case.
>
> In the meantime, I’ll do the following:
> - update config/BuildSystem/config/packages/OpenMPI.py to use OpenMPI 
> 4.1.0, see if any other error appears;

ok,  I think it is a good idea since 4.1 is the "stable" version...


> - provide a hotfix to bypass the segfaults;

If OpenMPI guys fix the issue, maybe it will be included in the 4.1.1 
release (4.1.1rc1 is still open for modifications I think)...

An idea: I our code, we "block" compilations/usages of buggy MPIs 
versions...


> - look at the hypre issue and whether they should be deferred to the 
> hypre team.
Oh thanks for this! :)
>
> Thank you for the Docker files, they were really useful.
> If you want to avoid oversubscription failures, you can edit the file 
> /opt/openmpi-4.1.0/etc/openmpi-default-hostfile and append the line:
> localhost slots=12
> If you want to increase the timeout limit of PETSc test suite for each 
> test, you can add the extra flag in your command line TIMEOUT=180 
> (default is 60, units are seconds).
Noted, I will add this to my scripts and the dockerfiles too...
>
> Thanks, I’ll ping you on GitLab when I’ve got something ready for you 
> to try,

Thank you for all your work!

Eric

> Pierre
>
-- 
Eric Chamberland, ing., M. Ing
Professionnel de recherche
GIREF/Université Laval
(418) 656-2131 poste 41 22 42



More information about the petsc-dev mailing list