[petsc-dev] Petsc "make test" have more failures for --with-openmp=1

Pierre Jolivet pierre at joliv.et
Wed Mar 3 08:22:06 CST 2021


Here is a fix for PETSc disregarding what kind of MPI implementation is being used with cluster PARDISO: https://gitlab.com/petsc/petsc/-/merge_requests/3678 <https://gitlab.com/petsc/petsc/-/merge_requests/3678>
This doesn’t fix src/mat/tests/ex242, whose backtrace looks suspicious to me…
(gdb) #0  0x00002ac152c9c6ed in mkl_trans_avx2_mkl_domatcopy2_t ()
   from /ccc/products/mkl-19.0.5.281/intel--19.0.5.281__openmpi--4.0.1/default/19.0.5.281/mkl/lib/intel64/libmkl_avx2.so
#1  0x00002ba345400dc7 in mkl_trans_avx2_mkl_domatcopy ()
   from /ccc/products/mkl-19.0.5.281/intel--19.0.5.281__openmpi--4.0.1/default/19.0.5.281/mkl/lib/intel64/libmkl_avx2.so
#2  0x00002ba31564fac2 in dmmdatdl_ ()
   from /ccc/products/mkl-19.0.5.281/intel--19.0.5.281__openmpi--4.0.1/default/19.0.5.281/mkl/lib/intel64/libmkl_scalapack_lp64.so
#3  0x00002ba3156d70de in PB_Cptran_DC ()
   from /ccc/products/mkl-19.0.5.281/intel--19.0.5.281__openmpi--4.0.1/default/19.0.5.281/mkl/lib/intel64/libmkl_scalapack_lp64.so
#4  0x00002ba315740062 in pdtran_ ()
   from /ccc/products/mkl-19.0.5.281/intel--19.0.5.281__openmpi--4.0.1/default/19.0.5.281/mkl/lib/intel64/libmkl_scalapack_lp64.so
#5  0x00002ba31370211b in MatTranspose_ScaLAPACK ()
   from /ccc/work/cont003/rndm/rndm/petsc/arch-linux2-c-opt-ompi/lib/libpetsc.so.3.014

I’m now realizing that this is also segfaulting with BLACS IntelMPI, so it’s probably unrelated, and simply some Intel mess.
(gdb) #0  0x00002ab818c4f3de in mkl_trans_def_mkl_domatcopy2_t ()
   from /ccc/products/mkl-20.0.4/system/default/20.0.4/mkl/lib/intel64/libmkl_def.so
#1  0x00002ab818c51247 in mkl_trans_def_mkl_domatcopy ()
   from /ccc/products/mkl-20.0.4/system/default/20.0.4/mkl/lib/intel64/libmkl_def.so
#2  0x00002ab76e0ba062 in dmmdatdl_ ()
   from /ccc/products/mkl-20.0.4/system/default/20.0.4/mkl/lib/intel64/libmkl_scalapack_lp64.so
#3  0x00002ab76e14157e in PB_Cptran_DC ()
   from /ccc/products/mkl-20.0.4/system/default/20.0.4/mkl/lib/intel64/libmkl_scalapack_lp64.so
#4  0x00002ab76e1aa502 in pdtran_ ()
   from /ccc/products/mkl-20.0.4/system/default/20.0.4/mkl/lib/intel64/libmkl_scalapack_lp64.so
#5  0x00002b7f87c833a1 in MatTranspose_ScaLAPACK ()
   from /ccc/work/cont003/rndm/rndm/petsc/arch-linux2-c-opt-impi/lib/libpetsc.so.3.014

(Notice the switch from arch-linux2-c-opt-ompi to arch-linux2-c-opt-impi)

Thanks,
Pierre

> On 3 Mar 2021, at 8:42 AM, Pierre Jolivet <pierre at joliv.et> wrote:
> 
> 
>> If it ends that there is a problem combining MKL + openMP that relies on linking configuration for example, should it be a good thing to have this (--with-openmp=1) tested into the pipelines (with external packages of course)?
>> 
> As Barry said, there is not much (if any) OpenMP in PETSc.
> There is however some workers with the MKL (+ Intel compilers) turned on, but I don’t think we test MKL + GNU compilers (which I feel like is a very niche combination, hence not really worth testing, IMHO).
> 
>> Does the guys who maintain all these libs are reading petsc-dev? ;)
>> 
> I don’t think they are, but don’t worry, we do forward the appropriate messages to them :)
> 
> About yesterday’s failures…
> 1) I cannot reproduce any of the PCHYPRE/PCBDDC/PCHPDDM errors (sorry I didn’t bother putting the SuperLU_DIST tarball on my cluster)
> 2) I can reproduce the src/mat/tests/ex242.c error (which explicitly uses ScaLAPACK, none of the above PC uses it explicitly, except PCBDDC/PCHPDDM when using MUMPS on “big” problems where root nodes are factorized using ScaLAPACK, see -mat_mumps_icntl_13)
> 3) I’m seeing that both on your machine and mine, PETSc BuildSystem insist on linking libmkl_blacs_intelmpi_lp64.so even though we supply explicitly libmkl_blacs_openmpi_lp64.so
> This for example yields a wrong Makefile.inc for MUMPS:
> $ cat arch-linux2-c-opt-ompi/externalpackages/MUMPS_5.3.5/Makefile.inc|grep blacs
> SCALAP  = […] -lmkl_blacs_openmpi_lp64
> LIBBLAS = […] -lmkl_blacs_intelmpi_lp64 -lgomp -ldl -lpthread -lm […]
> 
> Despite what Barry says, I think PETSc is partially to blame as well (why use libmkl_blacs_intelmpi_lp64.so even though BuildSystem is capable of detecting we are using OpenMPI).
> I’ll try to fix this to see if it solves 2).
> 
> Thanks,
> Pierre
> 
> http://joliv.et/irene-rome-configure.log <http://joliv.et/irene-rome-configure.log>
> $ /usr/bin/gmake -f gmakefile test test-fail=1
> Using MAKEFLAGS: test-fail=1
>         TEST arch-linux2-c-opt-ompi/tests/counts/snes_tutorials-ex12_quad_hpddm_reuse_baij.counts
>  ok snes_tutorials-ex12_quad_hpddm_reuse_baij
>  ok diff-snes_tutorials-ex12_quad_hpddm_reuse_baij
>         TEST arch-linux2-c-opt-ompi/tests/counts/ksp_ksp_tutorials-ex50_tut_2.counts
>  ok ksp_ksp_tutorials-ex50_tut_2 # SKIP PETSC_HAVE_SUPERLU_DIST requirement not met
>         TEST arch-linux2-c-opt-ompi/tests/counts/snes_tutorials-ex56_hypre.counts
>  ok snes_tutorials-ex56_hypre
>  ok diff-snes_tutorials-ex56_hypre
>         TEST arch-linux2-c-opt-ompi/tests/counts/snes_tutorials-ex17_3d_q3_trig_elas.counts
>  ok snes_tutorials-ex17_3d_q3_trig_elas
>  ok diff-snes_tutorials-ex17_3d_q3_trig_elas
>         TEST arch-linux2-c-opt-ompi/tests/counts/snes_tutorials-ex12_quad_hpddm_reuse_threshold_baij.counts
>  ok snes_tutorials-ex12_quad_hpddm_reuse_threshold_baij
>  ok diff-snes_tutorials-ex12_quad_hpddm_reuse_threshold_baij
>         TEST arch-linux2-c-opt-ompi/tests/counts/snes_tutorials-ex12_tri_parmetis_hpddm_baij.counts
>  ok snes_tutorials-ex12_tri_parmetis_hpddm_baij
>  ok diff-snes_tutorials-ex12_tri_parmetis_hpddm_baij
>         TEST arch-linux2-c-opt-ompi/tests/counts/snes_tutorials-ex19_tut_3.counts
>  ok snes_tutorials-ex19_tut_3
>  ok diff-snes_tutorials-ex19_tut_3
>         TEST arch-linux2-c-opt-ompi/tests/counts/mat_tests-ex242_3.counts
> not ok mat_tests-ex242_3 # Error code: 137
> #	[1]PETSC ERROR: ------------------------------------------------------------------------
> #	[1]PETSC ERROR: Caught signal number 11 SEGV: Segmentation Violation, probably memory access out of range
> #	[1]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger
> #	[1]PETSC ERROR: or see https://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind <https://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind>
> #	[1]PETSC ERROR: or try http://valgrind.org <http://valgrind.org/> on GNU/linux and Apple Mac OS X to find memory corruption errors
> #	[1]PETSC ERROR: configure using --with-debugging=yes, recompile, link, and run
> #	[1]PETSC ERROR: to get more information on the crash.
> #	[1]PETSC ERROR: --------------------- Error Message --------------------------------------------------------------
> #	[1]PETSC ERROR: Signal received
> #	[1]PETSC ERROR: See https://www.mcs.anl.gov/petsc/documentation/faq.html <https://www.mcs.anl.gov/petsc/documentation/faq.html> for trouble shooting.
> #	[1]PETSC ERROR: Petsc Development GIT revision: v3.14.4-733-g7ab9467ef9  GIT Date: 2021-03-02 16:15:11 +0000
> #	[2]PETSC ERROR: ------------------------------------------------------------------------
> #	[2]PETSC ERROR: Caught signal number 11 SEGV: Segmentation Violation, probably memory access out of range
> #	[2]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger
> #	[2]PETSC ERROR: or see https://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind <https://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind>
> #	[2]PETSC ERROR: or try http://valgrind.org <http://valgrind.org/> on GNU/linux and Apple Mac OS X to find memory corruption errors
> #	[2]PETSC ERROR: configure using --with-debugging=yes, recompile, link, and run
> #	[2]PETSC ERROR: to get more information on the crash.
> #	[2]PETSC ERROR: --------------------- Error Message --------------------------------------------------------------
> #	[2]PETSC ERROR: Signal received
> #	[2]PETSC ERROR: See https://www.mcs.anl.gov/petsc/documentation/faq.html <https://www.mcs.anl.gov/petsc/documentation/faq.html> for trouble shooting.
> #	[2]PETSC ERROR: Petsc Development GIT revision: v3.14.4-733-g7ab9467ef9  GIT Date: 2021-03-02 16:15:11 +0000
> #	[2]PETSC ERROR: /ccc/work/cont003/rndm/rndm/petsc/arch-linux2-c-opt-ompi/tests/mat/tests/runex242_3/../ex242 on a arch-linux2-c-opt-ompi named irene4047 by jolivetp Wed Mar  3 08:21:20 2021
> #	[2]PETSC ERROR: Configure options --download-hpddm --download-hpddm-commit=origin/main --download-hypre --download-metis --download-mumps --download-parmetis --download-ptscotch --download-slepc --download-slepc-commit=origin/main --download-tetgen --known-mpi-c-double-complex --known-mpi-int64_t --known-mpi-long-double --with-avx512-kernels=1 --with-blaslapack-dir=/ccc/products/mkl-19.0.5.281/intel--19.0.5.281__openmpi--4.0.1/default/19.0.5.281/mkl/lib/intel64 --with-cc=mpicc --with-cxx=mpicxx --with-debugging=0 --with-fc=mpifort --with-fortran-bindings=0 --with-make-np=40 --with-mkl_cpardiso-dir=/ccc/products/mkl-19.0.5.281/intel--19.0.5.281__openmpi--4.0.1/default/19.0.5.281 --with-mkl_cpardiso=1 --with-mkl_pardiso-dir=/ccc/products/mkl-19.0.5.281/intel--19.0.5.281__openmpi--4.0.1/default/19.0.5.281/mkl --with-mkl_pardiso=1 --with-mpiexec=ccc_mprun --with-openmp=1 --with-packages-download-dir=/ccc/cont003/home/enseeiht/jolivetp/Dude/externalpackages/ --with-scalapack-include=/ccc/products/mkl-19.0.5.281/intel--19.0.5.281__openmpi--4.0.1/default/19.0.5.281/mkl/include --with-scalapack-lib="[/ccc/products/mkl-19.0.5.281/intel--19.0.5.281__openmpi--4.0.1/default/19.0.5.281/mkl/lib/intel64/libmkl_scalapack_lp64.so,/ccc/products/mkl-19.0.5.281/intel--19.0.5.281__openmpi--4.0.1/default/19.0.5.281/mkl/lib/intel64/libmkl_blacs_openmpi_lp64.so]" --with-scalar-type=real --with-x=0 COPTFLAGS="-O3 -fp-model fast -mavx2" CXXOPTFLAGS="-O3 -fp-model fast -mavx2" FOPTFLAGS="-O3 -fp-model fast -mavx2" PETSC_ARCH=arch-linux2-c-opt-ompi
> #	[2]PETSC ERROR: #1 User provided function() line 0 in  unknown file
> #	[2]PETSC ERROR: Run with -malloc_debug to check if memory corruption is causing the crash.
> #	[1]PETSC ERROR: /ccc/work/cont003/rndm/rndm/petsc/arch-linux2-c-opt-ompi/tests/mat/tests/runex242_3/../ex242 on a arch-linux2-c-opt-ompi named irene4047 by jolivetp Wed Mar  3 08:21:20 2021
> #	[1]PETSC ERROR: Configure options --download-hpddm --download-hpddm-commit=origin/main --download-hypre --download-metis --download-mumps --download-parmetis --download-ptscotch --download-slepc --download-slepc-commit=origin/main --download-tetgen --known-mpi-c-double-complex --known-mpi-int64_t --known-mpi-long-double --with-avx512-kernels=1 --with-blaslapack-dir=/ccc/products/mkl-19.0.5.281/intel--19.0.5.281__openmpi--4.0.1/default/19.0.5.281/mkl/lib/intel64 --with-cc=mpicc --with-cxx=mpicxx --with-debugging=0 --with-fc=mpifort --with-fortran-bindings=0 --with-make-np=40 --with-mkl_cpardiso-dir=/ccc/products/mkl-19.0.5.281/intel--19.0.5.281__openmpi--4.0.1/default/19.0.5.281 --with-mkl_cpardiso=1 --with-mkl_pardiso-dir=/ccc/products/mkl-19.0.5.281/intel--19.0.5.281__openmpi--4.0.1/default/19.0.5.281/mkl --with-mkl_pardiso=1 --with-mpiexec=ccc_mprun --with-openmp=1 --with-packages-download-dir=/ccc/cont003/home/enseeiht/jolivetp/Dude/externalpackages/ --with-scalapack-include=/ccc/products/mkl-19.0.5.281/intel--19.0.5.281__openmpi--4.0.1/default/19.0.5.281/mkl/include --with-scalapack-lib="[/ccc/products/mkl-19.0.5.281/intel--19.0.5.281__openmpi--4.0.1/default/19.0.5.281/mkl/lib/intel64/libmkl_scalapack_lp64.so,/ccc/products/mkl-19.0.5.281/intel--19.0.5.281__openmpi--4.0.1/default/19.0.5.281/mkl/lib/intel64/libmkl_blacs_openmpi_lp64.so]" --with-scalar-type=real --with-x=0 COPTFLAGS="-O3 -fp-model fast -mavx2" CXXOPTFLAGS="-O3 -fp-model fast -mavx2" FOPTFLAGS="-O3 -fp-model fast -mavx2" PETSC_ARCH=arch-linux2-c-opt-ompi
> #	[1]PETSC ERROR: #1 User provided function() line 0 in  unknown file
> #	[1]PETSC ERROR: Run with -malloc_debug to check if memory corruption is causing the crash.
> #	--------------------------------------------------------------------------
> #	MPI_ABORT was invoked on rank 2 in communicator MPI_COMM_WORLD
> #	with errorcode 50176059.
> #
> #	NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
> #	You may or may not see output from other processes, depending on
> #	exactly when Open MPI kills them.
> #	--------------------------------------------------------------------------
> #	--------------------------------------------------------------------------
> #	MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD
> #	with errorcode 50176059.
> #
> #	NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
> #	You may or may not see output from other processes, depending on
> #	exactly when Open MPI kills them.
> #	--------------------------------------------------------------------------
> #	srun: Job step aborted: Waiting up to 302 seconds for job step to finish.
> #	slurmstepd-irene4047: error: *** STEP 1374176.36 ON irene4047 CANCELLED AT 2021-03-03T08:21:20 ***
> #	srun: error: irene4047: task 0: Killed
> #	srun: error: irene4047: tasks 1-2: Exited with exit code 16
>  ok mat_tests-ex242_3 # SKIP Command failed so no diff
>         TEST arch-linux2-c-opt-ompi/tests/counts/snes_tutorials-ex17_3d_q3_trig_vlap.counts
>  ok snes_tutorials-ex17_3d_q3_trig_vlap
>  ok diff-snes_tutorials-ex17_3d_q3_trig_vlap
>         TEST arch-linux2-c-opt-ompi/tests/counts/snes_tutorials-ex56_attach_mat_nearnullspace-1_bddc_approx_hypre.counts
>  ok snes_tutorials-ex56_attach_mat_nearnullspace-1_bddc_approx_hypre
>  ok diff-snes_tutorials-ex56_attach_mat_nearnullspace-1_bddc_approx_hypre
>         TEST arch-linux2-c-opt-ompi/tests/counts/ksp_ksp_tutorials-ex49_hypre_nullspace.counts
>  ok ksp_ksp_tutorials-ex49_hypre_nullspace
>  ok diff-ksp_ksp_tutorials-ex49_hypre_nullspace
>         TEST arch-linux2-c-opt-ompi/tests/counts/ts_tutorials-ex18_p1p1_xper_ref.counts
>  ok ts_tutorials-ex18_p1p1_xper_ref
>  ok diff-ts_tutorials-ex18_p1p1_xper_ref
>         TEST arch-linux2-c-opt-ompi/tests/counts/ts_tutorials-ex18_p1p1_xyper_ref.counts
>  ok ts_tutorials-ex18_p1p1_xyper_ref
>  ok diff-ts_tutorials-ex18_p1p1_xyper_ref
>         TEST arch-linux2-c-opt-ompi/tests/counts/snes_tutorials-ex56_attach_mat_nearnullspace-0_bddc_approx_hypre.counts
>  ok snes_tutorials-ex56_attach_mat_nearnullspace-0_bddc_approx_hypre
>  ok diff-snes_tutorials-ex56_attach_mat_nearnullspace-0_bddc_approx_hypre
>         TEST arch-linux2-c-opt-ompi/tests/counts/ksp_ksp_tutorials-ex64_1.counts
>  ok ksp_ksp_tutorials-ex64_1 # SKIP PETSC_HAVE_SUPERLU_DIST requirement not met
> 
>> On 3 Mar 2021, at 6:21 AM, Eric Chamberland <Eric.Chamberland at giref.ulaval.ca <mailto:Eric.Chamberland at giref.ulaval.ca>> wrote:
>> 
>> Just started a discussion on the side:
>> 
>> https://community.intel.com/t5/Intel-oneAPI-Math-Kernel-Library/Intel-MKL-Link-Line-Advisor-as-external-tool/m-p/1260895#M30974 <https://community.intel.com/t5/Intel-oneAPI-Math-Kernel-Library/Intel-MKL-Link-Line-Advisor-as-external-tool/m-p/1260895#M30974>
>> Eric
>> 
>> On 2021-03-02 3:50 p.m., Pierre Jolivet wrote:
>>> Hello Eric,
>>> src/mat/tests/ex237.c is a recent test with some code paths that should be disabled for “old” MKL versions. It’s tricky to check directly in the source (we do check in BuildSystem) because there is no such thing as PETSC_PKG_MKL_VERSION_LT, but I guess we can change if defined(PETSC_HAVE_MKL) to if defined(PETSC_HAVE_MKL) && defined(PETSC_HAVE_MKL_SPARSE_OPTIMIZE), I’ll make a MR, thanks for reporting this.
>>> 
>>> For the other issues, I’m sensing this is a problem with gomp + intel_gnu_thread, but this is pure speculation… sorry.
>>> I’ll try to reproduce some of these problems if you are not given a more meaningful answer.
>>> 
>>> Thanks,
>>> Pierre
>>> 
>>>> On 2 Mar 2021, at 9:14 PM, Eric Chamberland <Eric.Chamberland at giref.ulaval.ca <mailto:Eric.Chamberland at giref.ulaval.ca>> wrote:
>>>> 
>>>> Hi,
>>>> 
>>>> It all started when I wanted to test PETSC/CUDA compatibility for our code.
>>>> 
>>>> I had to activate --with-openmp to configure with --with-cuda=1 successfully.
>>>> 
>>>> I then saw that PETSC_HAVE_OPENMP  is used at least in MUMPS (and some other places).
>>>> 
>>>> So, I configured and tested petsc with openmp activated, without CUDA.
>>>> 
>>>> The first thing I see is that our code CI pipelines now fails for many tests.
>>>> 
>>>> After looking deeper, it seems that PETSc itself fails many tests when I activate openmp!
>>>> 
>>>> Here are all the configurations I have results for, after/before activating OpenMP for PETSc:
>>>> ==============================================================================
>>>> 
>>>> ==============================================================================
>>>> 
>>>> For petsc/master + OpenMPI 4.0.4 + MKL 2019.4.243:
>>>> 
>>>> With OpenMP=1
>>>> 
>>>> https://giref.ulaval.ca/~cmpgiref/petsc-master-debug/2021.03.02.02h00m02s_make_test.log <https://giref.ulaval.ca/~cmpgiref/petsc-master-debug/2021.03.02.02h00m02s_make_test.log>
>>>> https://giref.ulaval.ca/~cmpgiref/petsc-master-debug/2021.03.02.02h00m02s_configure.log <https://giref.ulaval.ca/~cmpgiref/petsc-master-debug/2021.03.02.02h00m02s_configure.log>
>>>> # -------------
>>>> #   Summary    
>>>> # -------------
>>>> # FAILED snes_tutorials-ex12_quad_hpddm_reuse_baij diff-ksp_ksp_tests-ex33_superlu_dist_2 diff-ksp_ksp_tests-ex49_superlu_dist+nsize-1herm-0_conv-0 diff-ksp_ksp_tests-ex49_superlu_dist+nsize-1herm-0_conv-1 diff-ksp_ksp_tests-ex49_superlu_dist+nsize-1herm-1_conv-0 diff-ksp_ksp_tests-ex49_superlu_dist+nsize-1herm-1_conv-1 diff-ksp_ksp_tests-ex49_superlu_dist+nsize-4herm-0_conv-0 diff-ksp_ksp_tests-ex49_superlu_dist+nsize-4herm-0_conv-1 diff-ksp_ksp_tests-ex49_superlu_dist+nsize-4herm-1_conv-0 diff-ksp_ksp_tests-ex49_superlu_dist+nsize-4herm-1_conv-1 ksp_ksp_tutorials-ex50_tut_2 diff-ksp_ksp_tests-ex33_superlu_dist diff-snes_tutorials-ex56_hypre snes_tutorials-ex17_3d_q3_trig_elas snes_tutorials-ex12_quad_hpddm_reuse_threshold_baij ksp_ksp_tutorials-ex5_superlu_dist_3 ksp_ksp_tutorials-ex5f_superlu_dist snes_tutorials-ex12_tri_parmetis_hpddm_baij diff-snes_tutorials-ex19_tut_3 mat_tests-ex242_3 snes_tutorials-ex17_3d_q3_trig_vlap ksp_ksp_tutorials-ex5f_superlu_dist_3 snes_tutorials-ex19_superlu_dist diff-snes_tutorials-ex56_attach_mat_nearnullspace-1_bddc_approx_hypre diff-ksp_ksp_tutorials-ex49_hypre_nullspace ts_tutorials-ex18_p1p1_xper_ref ts_tutorials-ex18_p1p1_xyper_ref snes_tutorials-ex19_superlu_dist_2 ksp_ksp_tutorials-ex5_superlu_dist_2 diff-snes_tutorials-ex56_attach_mat_nearnullspace-0_bddc_approx_hypre ksp_ksp_tutorials-ex64_1 ksp_ksp_tutorials-ex5_superlu_dist ksp_ksp_tutorials-ex5f_superlu_dist_2
>>>> # success 8275/10003 tests (82.7%)
>>>> # failed 33/10003 tests (0.3%)
>>>> With OpenMP=0
>>>> 
>>>> https://giref.ulaval.ca/~cmpgiref/petsc-master-debug/2021.02.26.02h00m16s_make_test.log <https://giref.ulaval.ca/~cmpgiref/petsc-master-debug/2021.02.26.02h00m16s_make_test.log>
>>>> https://giref.ulaval.ca/~cmpgiref/petsc-master-debug/2021.02.26.02h00m16s_configure.log <https://giref.ulaval.ca/~cmpgiref/petsc-master-debug/2021.02.26.02h00m16s_configure.log>
>>>> # -------------
>>>> #   Summary    
>>>> # -------------
>>>> # FAILED tao_constrained_tutorials-tomographyADMM_6 snes_tutorials-ex17_3d_q3_trig_elas mat_tests-ex242_3 snes_tutorials-ex17_3d_q3_trig_vlap tao_leastsquares_tutorials-tomography_1 tao_constrained_tutorials-tomographyADMM_5
>>>> # success 8262/9983 tests (82.8%)
>>>> # failed 6/9983 tests (0.1%)
>>>> ==============================================================================
>>>> 
>>>> ==============================================================================
>>>> 
>>>> For OpenMPI 3.1.x/master:
>>>> 
>>>> With OpenMP=1:
>>>> 
>>>> https://giref.ulaval.ca/~cmpgiref/ompi_3.x/2021.03.01.22h00m01s_make_test.log <https://giref.ulaval.ca/~cmpgiref/ompi_3.x/2021.03.01.22h00m01s_make_test.log>
>>>> https://giref.ulaval.ca/~cmpgiref/ompi_3.x/2021.03.01.22h00m01s_configure.log <https://giref.ulaval.ca/~cmpgiref/ompi_3.x/2021.03.01.22h00m01s_configure.log>
>>>> # -------------
>>>> #   Summary    
>>>> # -------------
>>>> # FAILED mat_tests-ex242_3 mat_tests-ex242_2 diff-mat_tests-ex219f_1 diff-dm_tutorials-ex11f90_1 ksp_ksp_tutorials-ex5_superlu_dist_3 diff-ksp_ksp_tutorials-ex49_hypre_nullspace ksp_ksp_tutorials-ex5f_superlu_dist_3 snes_tutorials-ex17_3d_q3_trig_vlap diff-snes_tutorials-ex56_attach_mat_nearnullspace-1_bddc_approx_hypre diff-snes_tutorials-ex19_tut_3 diff-snes_tutorials-ex56_hypre diff-snes_tutorials-ex56_attach_mat_nearnullspace-0_bddc_approx_hypre tao_leastsquares_tutorials-tomography_1 tao_constrained_tutorials-tomographyADMM_4 tao_constrained_tutorials-tomographyADMM_6 diff-tao_constrained_tutorials-toyf_1
>>>> # success 8142/9765 tests (83.4%)
>>>> # failed 16/9765 tests (0.2%)
>>>> With OpenMP=0:
>>>> 
>>>> https://giref.ulaval.ca/~cmpgiref/ompi_3.x/2021.02.28.22h00m02s_make_test.log <https://giref.ulaval.ca/~cmpgiref/ompi_3.x/2021.02.28.22h00m02s_make_test.log>
>>>> https://giref.ulaval.ca/~cmpgiref/ompi_3.x/2021.02.28.22h00m02s_configure.log <https://giref.ulaval.ca/~cmpgiref/ompi_3.x/2021.02.28.22h00m02s_configure.log>
>>>> # -------------
>>>> #   Summary    
>>>> # -------------
>>>> # FAILED mat_tests-ex242_3 mat_tests-ex242_2 diff-mat_tests-ex219f_1 diff-dm_tutorials-ex11f90_1 ksp_ksp_tutorials-ex56_2 snes_tutorials-ex17_3d_q3_trig_vlap tao_leastsquares_tutorials-tomography_1 tao_constrained_tutorials-tomographyADMM_4 diff-tao_constrained_tutorials-toyf_1
>>>> # success 8151/9767 tests (83.5%)
>>>> # failed 9/9767 tests (0.1%)
>>>> ==============================================================================
>>>> 
>>>> ==============================================================================
>>>> 
>>>> For OpenMPI 4.0.x/master:
>>>> 
>>>> With OpenMP=1:
>>>> 
>>>> https://giref.ulaval.ca/~cmpgiref/ompi_4.x/2021.03.01.20h00m01s_make_test.log <https://giref.ulaval.ca/~cmpgiref/ompi_4.x/2021.03.01.20h00m01s_make_test.log>
>>>> https://giref.ulaval.ca/~cmpgiref/ompi_4.x/2021.03.01.20h00m01s_configure.log <https://giref.ulaval.ca/~cmpgiref/ompi_4.x/2021.03.01.20h00m01s_configure.log>
>>>> # FAILED snes_tutorials-ex17_3d_q3_trig_elas snes_tutorials-ex19_hypre ksp_ksp_tutorials-ex56_2 tao_leastsquares_tutorials-tomography_1 tao_constrained_tutorials-tomographyADMM_5 mat_tests-ex242_3 ksp_ksp_tutorials-ex55_hypre ksp_ksp_tutorials-ex5_superlu_dist_2 tao_constrained_tutorials-tomographyADMM_6 snes_tutorials-ex56_hypre snes_tutorials-ex56_attach_mat_nearnullspace-0_bddc_approx_hypre ksp_ksp_tutorials-ex5f_superlu_dist_3 ksp_ksp_tutorials-ex34_hyprestruct diff-ksp_ksp_tutorials-ex49_hypre_nullspace snes_tutorials-ex56_attach_mat_nearnullspace-1_bddc_approx_hypre ksp_ksp_tutorials-ex5f_superlu_dist ksp_ksp_tutorials-ex5f_superlu_dist_2 ksp_ksp_tutorials-ex5_superlu_dist snes_tutorials-ex19_tut_3 snes_tutorials-ex19_superlu_dist ksp_ksp_tutorials-ex50_tut_2 snes_tutorials-ex17_3d_q3_trig_vlap ksp_ksp_tutorials-ex5_superlu_dist_3 snes_tutorials-ex19_superlu_dist_2 tao_constrained_tutorials-tomographyADMM_4 ts_tutorials-ex26_2
>>>> # success 8125/9753 tests (83.3%)
>>>> # failed 26/9753 tests (0.3%)
>>>> With OpenMP=0
>>>> 
>>>> https://giref.ulaval.ca/~cmpgiref/ompi_4.x/2021.02.28.20h00m04s_make_test.log <https://giref.ulaval.ca/~cmpgiref/ompi_4.x/2021.02.28.20h00m04s_make_test.log>
>>>> https://giref.ulaval.ca/~cmpgiref/ompi_4.x/2021.02.28.20h00m04s_configure.log <https://giref.ulaval.ca/~cmpgiref/ompi_4.x/2021.02.28.20h00m04s_configure.log>
>>>> # FAILED mat_tests-ex242_3
>>>> # success 8174/9777 tests (83.6%)
>>>> # failed 1/9777 tests (0.0%)
>>>> 
>>>> ==============================================================================
>>>> 
>>>> ==============================================================================
>>>> 
>>>> Is that known and normal?
>>>> 
>>>> In all cases, I am using MKL and I suspect it  may come from there... :/
>>>> 
>>>> I also saw a second problem, "make test" fails to compile petsc examples on older versions of MKL (but that's less important for me, I just upgraded to OneAPI to avoid this, but you may want to know):
>>>> 
>>>> https://giref.ulaval.ca/~cmpgiref/dernier_ompi/2021.03.02.02h16m01s_make_test.log <https://giref.ulaval.ca/~cmpgiref/dernier_ompi/2021.03.02.02h16m01s_make_test.log>
>>>> https://giref.ulaval.ca/~cmpgiref/dernier_ompi/2021.03.02.02h16m01s_configure.log <https://giref.ulaval.ca/~cmpgiref/dernier_ompi/2021.03.02.02h16m01s_configure.log>
>>>> Thanks,
>>>> 
>>>> Eric
>>>> 
>>>> -- 
>>>> Eric Chamberland, ing., M. Ing
>>>> Professionnel de recherche
>>>> GIREF/Université Laval
>>>> (418) 656-2131 poste 41 22 42
>>> 
>> -- 
>> Eric Chamberland, ing., M. Ing
>> Professionnel de recherche
>> GIREF/Université Laval
>> (418) 656-2131 poste 41 22 42
> 

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-dev/attachments/20210303/d1d844b6/attachment-0001.html>


More information about the petsc-dev mailing list