<html><head><meta http-equiv="Content-Type" content="text/html; charset=utf-8"></head><body style="word-wrap: break-word; -webkit-nbsp-mode: space; line-break: after-white-space;" class=""><div class=""><br class=""></div> Ok, so it could be a bug in PETSc, but if it appears with particular MPI implementations shouldn't we turn off the support in those cases we know it will fail?<div class=""><br class=""></div><div class=""> Barry</div><div class=""><br class=""><div><br class=""><blockquote type="cite" class=""><div class="">On Sep 14, 2021, at 11:10 AM, Junchao Zhang <<a href="mailto:junchao.zhang@gmail.com" class="">junchao.zhang@gmail.com</a>> wrote:</div><br class="Apple-interchange-newline"><div class=""><div dir="ltr" class="">MPI one-sided is tricky and needs careful synchronization (like OpenMP). An incorrect code could work in one interface but fail in another.<div class=""><div class=""><br clear="all" class=""><div class=""><div dir="ltr" class="gmail_signature" data-smartmail="gmail_signature"><div dir="ltr" class="">--Junchao Zhang</div></div></div><br class=""></div></div></div><br class=""><div class="gmail_quote"><div dir="ltr" class="gmail_attr">On Tue, Sep 14, 2021 at 10:01 AM Barry Smith <<a href="mailto:bsmith@petsc.dev" class="">bsmith@petsc.dev</a>> wrote:<br class=""></div><blockquote class="gmail_quote" style="margin:0px 0px 0px 0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex"><div style="overflow-wrap: break-word;" class=""><div class=""><br class=""></div><div class=""> It sounds reproducible and related to using a particular versions of OpenMPI and even particular interfaces.</div><div class=""><br class=""></div><div class=""> Barry</div><div class=""><br class=""></div> On Tue, Sep 14, 2021 at 2:35 AM Stefano Zampini <<a href="mailto:stefano.zampini@gmail.com" target="_blank" class="">stefano.zampini@gmail.com</a>> wrote:<blockquote class="gmail_quote" style="margin:0px 0px 0px 0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex"><div dir="ltr" class="">I can reproduce it even with OpenMPI 4.1.1 on a different machine (Ubuntu 18 + AMD milan + clang from AOCC) and it is definitely an OpenMPI bug in the vader BTL If I use tcp, everything runs smooth</div><div dir="ltr" class=""><br class=""></div><div dir="ltr" class=""><br class=""></div><div dir="ltr" class=""><br class=""></div></blockquote><div class=""><br class=""><blockquote type="cite" class=""><div class="">On Sep 14, 2021, at 10:54 AM, Junchao Zhang <<a href="mailto:junchao.zhang@gmail.com" target="_blank" class="">junchao.zhang@gmail.com</a>> wrote:</div><br class=""><div class=""><div dir="ltr" class="">Without a standalone & valid mpi example to reproduce the error, we are not assured to say it is an OpenMPI bug. <div class=""><br clear="all" class=""><div class=""><div dir="ltr" class=""><div dir="ltr" class="">--Junchao Zhang</div></div></div><br class=""></div></div><br class=""><div class="gmail_quote"><div dir="ltr" class="gmail_attr">On Tue, Sep 14, 2021 at 6:17 AM Matthew Knepley <<a href="mailto:knepley@gmail.com" target="_blank" class="">knepley@gmail.com</a>> wrote:<br class=""></div><blockquote class="gmail_quote" style="margin:0px 0px 0px 0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex"><div dir="ltr" class="">Okay, we have to send this to OpenMPI. Volunteers?<div class=""><br class=""></div><div class="">Maybe we should note this in the FAQ, or installation, so we remember how to fix it if someone else asks?</div><div class=""><br class=""></div><div class=""> Thanks,</div><div class=""><br class=""></div><div class=""> Matt</div></div><br class=""><div class="gmail_quote"><div dir="ltr" class="gmail_attr">On Tue, Sep 14, 2021 at 2:35 AM Stefano Zampini <<a href="mailto:stefano.zampini@gmail.com" target="_blank" class="">stefano.zampini@gmail.com</a>> wrote:<br class=""></div><blockquote class="gmail_quote" style="margin:0px 0px 0px 0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex"><div dir="ltr" class=""><div class="">I can reproduce it even with OpenMPI 4.1.1 on a different machine (Ubuntu 18 + AMD milan + clang from AOCC) and it is definitely an OpenMPI bug in the vader BTL If I use tcp, everything runs smooth</div><div class=""><br class=""></div><div class="">zampins@kanary:~/Devel/petsc$ cat /home/zampins/local/etc/openmpi-mca-params.conf | grep btl<br class="">btl = tcp,self<br class="">zampins@kanary:~/Devel/petsc$ make -f gmakefile.test vec_is_sf_tutorials-ex1_4 <br class="">Using MAKEFLAGS:<br class=""> TEST arch-debug/tests/counts/vec_is_sf_tutorials-ex1_4.counts<br class=""> ok vec_is_sf_tutorials-ex1_4+sf_window_sync-fence_sf_window_flavor-create<br class=""> ok diff-vec_is_sf_tutorials-ex1_4+sf_window_sync-fence_sf_window_flavor-create<br class=""> ok vec_is_sf_tutorials-ex1_4+sf_window_sync-fence_sf_window_flavor-dynamic<br class=""> ok diff-vec_is_sf_tutorials-ex1_4+sf_window_sync-fence_sf_window_flavor-dynamic<br class=""> ok vec_is_sf_tutorials-ex1_4+sf_window_sync-fence_sf_window_flavor-allocate<br class=""> ok diff-vec_is_sf_tutorials-ex1_4+sf_window_sync-fence_sf_window_flavor-allocate<br class=""> ok vec_is_sf_tutorials-ex1_4+sf_window_sync-active_sf_window_flavor-create<br class=""> ok diff-vec_is_sf_tutorials-ex1_4+sf_window_sync-active_sf_window_flavor-create<br class=""> ok vec_is_sf_tutorials-ex1_4+sf_window_sync-active_sf_window_flavor-dynamic<br class=""> ok diff-vec_is_sf_tutorials-ex1_4+sf_window_sync-active_sf_window_flavor-dynamic<br class=""> ok vec_is_sf_tutorials-ex1_4+sf_window_sync-active_sf_window_flavor-allocate<br class=""> ok diff-vec_is_sf_tutorials-ex1_4+sf_window_sync-active_sf_window_flavor-allocate<br class=""> ok vec_is_sf_tutorials-ex1_4+sf_window_sync-lock_sf_window_flavor-create<br class=""> ok diff-vec_is_sf_tutorials-ex1_4+sf_window_sync-lock_sf_window_flavor-create<br class=""> ok vec_is_sf_tutorials-ex1_4+sf_window_sync-lock_sf_window_flavor-dynamic<br class=""> ok diff-vec_is_sf_tutorials-ex1_4+sf_window_sync-lock_sf_window_flavor-dynamic<br class=""> ok vec_is_sf_tutorials-ex1_4+sf_window_sync-lock_sf_window_flavor-allocate<br class=""> ok diff-vec_is_sf_tutorials-ex1_4+sf_window_sync-lock_sf_window_flavor-allocate</div><div class=""><br class=""></div><div class=""><br class=""></div><div class="">zampins@kanary:~/Devel/petsc$ cat /home/zampins/local/etc/openmpi-mca-params.conf | grep btl<br class="">btl = vader,tcp,self<br class="">zampins@kanary:~/Devel/petsc$ make -f gmakefile.test vec_is_sf_tutorials-ex1_4 <br class="">Using MAKEFLAGS:<br class=""> TEST arch-debug/tests/counts/vec_is_sf_tutorials-ex1_4.counts<br class=""> ok vec_is_sf_tutorials-ex1_4+sf_window_sync-fence_sf_window_flavor-create<br class="">not ok diff-vec_is_sf_tutorials-ex1_4+sf_window_sync-fence_sf_window_flavor-create # Error code: 1<br class=""># 43,46c43,46<br class=""># < [0] 0: 4001 2000 2002 3002 4002<br class=""># < [1] 0: 1001 3000<br class=""># < [2] 0: 2001 4000<br class=""># < [3] 0: 3001 1000<br class=""># ---<br class=""># > [0] 0: 2002 2146435072 2 2146435072 38736240<br class=""># > [1] 0: 3000 2146435072<br class=""># > [2] 0: 2001 2146435072<br class=""># > [3] 0: 3001 2146435072<br class=""> ok vec_is_sf_tutorials-ex1_4+sf_window_sync-fence_sf_window_flavor-dynamic<br class="">not ok diff-vec_is_sf_tutorials-ex1_4+sf_window_sync-fence_sf_window_flavor-dynamic # Error code: 1<br class=""># 43,46c43,46<br class=""># < [0] 0: 4001 2000 2002 3002 4002<br class=""># < [1] 0: 1001 3000<br class=""># < [2] 0: 2001 4000<br class=""># < [3] 0: 3001 1000<br class=""># ---<br class=""># > [0] 0: 2002 2146435072 2 2146435072 0<br class=""># > [1] 0: 3000 2146435072<br class=""># > [2] 0: 2001 2146435072<br class=""># > [3] 0: 3001 2146435072<br class=""> ok vec_is_sf_tutorials-ex1_4+sf_window_sync-fence_sf_window_flavor-allocate<br class=""> ok diff-vec_is_sf_tutorials-ex1_4+sf_window_sync-fence_sf_window_flavor-allocate<br class=""># retrying vec_is_sf_tutorials-ex1_4+sf_window_sync-active_sf_window_flavor-create<br class="">not ok vec_is_sf_tutorials-ex1_4+sf_window_sync-active_sf_window_flavor-create # Error code: 98<br class=""># [1]PETSC ERROR: --------------------- Error Message --------------------------------------------------------------<br class=""># [1]PETSC ERROR: General MPI error <br class=""># [1]PETSC ERROR: MPI error 6 MPI_ERR_RANK: invalid rank<br class=""># [1]PETSC ERROR: See <a href="https://petsc.org/release/faq/" target="_blank" class="">https://petsc.org/release/faq/</a> for trouble shooting.<br class=""># [1]PETSC ERROR: Petsc Development GIT revision: v3.15.4-783-g168bb9f76b GIT Date: 2021-09-13 14:01:22 +0000<br class=""># [1]PETSC ERROR: ../ex1 on a arch-debug named <a href="http://kanary.kaust.edu.sa/" target="_blank" class="">kanary.kaust.edu.sa</a> by zampins Tue Sep 14 09:31:42 2021<br class=""># [1]PETSC ERROR: [2]PETSC ERROR: --------------------- Error Message --------------------------------------------------------------<br class=""># [2]PETSC ERROR: General MPI error <br class=""># [2]PETSC ERROR: MPI error 6 MPI_ERR_RANK: invalid rank<br class=""># [2]PETSC ERROR: See <a href="https://petsc.org/release/faq/" target="_blank" class="">https://petsc.org/release/faq/</a> for trouble shooting.<br class=""># [2]PETSC ERROR: Petsc Development GIT revision: v3.15.4-783-g168bb9f76b GIT Date: 2021-09-13 14:01:22 +0000<br class=""># [2]PETSC ERROR: ../ex1 on a arch-debug named <a href="http://kanary.kaust.edu.sa/" target="_blank" class="">kanary.kaust.edu.sa</a> by zampins Tue Sep 14 09:31:42 2021<br class=""># [2]PETSC ERROR: Configure options --with-cc=/home/zampins/local/bin/mpicc --with-cxx-dialect=c++14 --with-cxx=/home/zampins/local/bin/mpicxx --with-debugging=1 --with-fc=/home/zampins/local/bin/mpifort --with-fortran-bindings=0 --with-hip-dir=/opt/rocm --with-hip=1 --with-hypre-dir=/home/zampins/local-petsc --with-kokkos-dir=/home/zampins/local-petsc --with-kokkos-kernels-dir=/home/zampins/local-petsc --with-blaslapack-include=/home/zampins/local-aocl/aocc/3.0-6/include --with-blaslapack-lib="[/home/zampins/local-aocl/aocc/3.0-6/lib/liblapacke.a,/home/zampins/local-aocl/aocc/3.0-6/lib/libflame.a,/home/zampins/local-aocl/aocc/3.0-6/lib/libblis-mt.a]" HIPFLAGS=--amdgpu-target=gfx908 HIPPPFLAGS=-I/home/zampins/local-petsc/include PETSC_ARCH=arch-debug<br class=""># [2]PETSC ERROR: #1 PetscSFGetGroups() at /home/zampins/Devel/petsc/src/vec/is/sf/interface/sf.c:1166<br class=""># [2]PETSC ERROR: #2 PetscSFSetUp_Window() at /home/zampins/Devel/petsc/src/vec/is/sf/impls/window/sfwindow.c:708<br class=""># [2]PETSC ERROR: #3 PetscSFSetUp() at /home/zampins/Devel/petsc/src/vec/is/sf/interface/sf.c:318<br class=""># [2]PETSC ERROR: #4 main() at /home/zampins/Devel/petsc/src/vec/is/sf/tutorials/ex1.c:172<br class=""># [2]PETSC ERROR: PETSc Option Table entries:<br class=""># [2]PETSC ERROR: -sf_type window<br class=""># [2]PETSC ERROR: -sf_window_flavor create<br class=""># [2]PETSC ERROR: -sf_window_sync active<br class=""># [2]PETSC ERROR: -test_gather<br class=""># [2]PETSC ERROR: ----------------End of Error Message -------send entire error message to <a href="mailto:petsc-maint@mcs.anl.gov" target="_blank" class="">petsc-maint@mcs.anl.gov</a>----------<br class=""># [3]PETSC ERROR: --------------------- Error Message --------------------------------------------------------------<br class=""># [3]PETSC ERROR: General MPI error <br class=""># [3]PETSC ERROR: MPI error 6 MPI_ERR_RANK: invalid rank<br class=""># [3]PETSC ERROR: See <a href="https://petsc.org/release/faq/" target="_blank" class="">https://petsc.org/release/faq/</a> for trouble shooting.<br class=""># [3]PETSC ERROR: Petsc Development GIT revision: v3.15.4-783-g168bb9f76b GIT Date: 2021-09-13 14:01:22 +0000<br class=""># [3]PETSC ERROR: ../ex1 on a arch-debug named <a href="http://kanary.kaust.edu.sa/" target="_blank" class="">kanary.kaust.edu.sa</a> by zampins Tue Sep 14 09:31:42 2021<br class=""># [3]PETSC ERROR: Configure options --with-cc=/home/zampins/local/bin/mpicc --with-cxx-dialect=c++14 --with-cxx=/home/zampins/local/bin/mpicxx --with-debugging=1 --with-fc=/home/zampins/local/bin/mpifort --with-fortran-bindings=0 --with-hip-dir=/opt/rocm --with-hip=1 --with-hypre-dir=/home/zampins/local-petsc --with-kokkos-dir=/home/zampins/local-petsc --with-kokkos-kernels-dir=/home/zampins/local-petsc --with-blaslapack-include=/home/zampins/local-aocl/aocc/3.0-6/include --with-blaslapack-lib="[/home/zampins/local-aocl/aocc/3.0-6/lib/liblapacke.a,/home/zampins/local-aocl/aocc/3.0-6/lib/libflame.a,/home/zampins/local-aocl/aocc/3.0-6/lib/libblis-mt.a]" HIPFLAGS=--amdgpu-target=gfx908 HIPPPFLAGS=-I/home/zampins/local-petsc/include PETSC_ARCH=arch-debug<br class=""># [3]PETSC ERROR: #1 PetscSFGetGroups() at /home/zampins/Devel/petsc/src/vec/is/sf/interface/sf.c:1166<br class=""># [3]PETSC ERROR: #2 PetscSFSetUp_Window() at /home/zampins/Devel/petsc/src/vec/is/sf/impls/window/sfwindow.c:708<br class=""># [3]PETSC ERROR: #3 PetscSFSetUp() at /home/zampins/Devel/petsc/src/vec/is/sf/interface/sf.c:318<br class=""># [3]PETSC ERROR: #4 main() at /home/zampins/Devel/petsc/src/vec/is/sf/tutorials/ex1.c:172<br class=""># [3]PETSC ERROR: PETSc Option Table entries:<br class=""># [3]PETSC ERROR: -sf_type window<br class=""># [3]PETSC ERROR: -sf_window_flavor create<br class=""># [3]PETSC ERROR: Configure options --with-cc=/home/zampins/local/bin/mpicc --with-cxx-dialect=c++14 --with-cxx=/home/zampins/local/bin/mpicxx --with-debugging=1 --with-fc=/home/zampins/local/bin/mpifort --with-fortran-bindings=0 --with-hip-dir=/opt/rocm --with-hip=1 --with-hypre-dir=/home/zampins/local-petsc --with-kokkos-dir=/home/zampins/local-petsc --with-kokkos-kernels-dir=/home/zampins/local-petsc --with-blaslapack-include=/home/zampins/local-aocl/aocc/3.0-6/include --with-blaslapack-lib="[/home/zampins/local-aocl/aocc/3.0-6/lib/liblapacke.a,/home/zampins/local-aocl/aocc/3.0-6/lib/libflame.a,/home/zampins/local-aocl/aocc/3.0-6/lib/libblis-mt.a]" HIPFLAGS=--amdgpu-target=gfx908 HIPPPFLAGS=-I/home/zampins/local-petsc/include PETSC_ARCH=arch-debug<br class=""># [1]PETSC ERROR: #1 PetscSFGetGroups() at /home/zampins/Devel/petsc/src/vec/is/sf/interface/sf.c:1166<br class=""># [1]PETSC ERROR: #2 PetscSFSetUp_Window() at /home/zampins/Devel/petsc/src/vec/is/sf/impls/window/sfwindow.c:708<br class=""># [1]PETSC ERROR: #3 PetscSFSetUp() at /home/zampins/Devel/petsc/src/vec/is/sf/interface/sf.c:318<br class=""># [1]PETSC ERROR: #4 main() at /home/zampins/Devel/petsc/src/vec/is/sf/tutorials/ex1.c:172<br class=""># [1]PETSC ERROR: PETSc Option Table entries:<br class=""># [1]PETSC ERROR: -sf_type window<br class=""># [1]PETSC ERROR: -sf_window_flavor create<br class=""># [1]PETSC ERROR: -sf_window_sync active<br class=""># [1]PETSC ERROR: -test_gather<br class=""># [1]PETSC ERROR: ----------------End of Error Message -------send entire error message to <a href="mailto:petsc-maint@mcs.anl.gov" target="_blank" class="">petsc-maint@mcs.anl.gov</a>----------<br class=""># -sf_window_sync active<br class=""># [3]PETSC ERROR: -test_gather<br class=""># [3]PETSC ERROR: ----------------End of Error Message -------send entire error message to <a href="mailto:petsc-maint@mcs.anl.gov" target="_blank" class="">petsc-maint@mcs.anl.gov</a>----------<br class=""># --------------------------------------------------------------------------<br class=""># MPI_ABORT was invoked on rank 2 in communicator MPI_COMM_WORLD<br class=""># with errorcode 98.<br class=""># <br class=""># NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.<br class=""># You may or may not see output from other processes, depending on<br class=""># exactly when Open MPI kills them.<br class=""># --------------------------------------------------------------------------<br class=""># [<a href="http://kanary.kaust.edu.sa:115527" target="_blank" class="">kanary.kaust.edu.sa:115527</a>] 2 more processes have sent help message help-mpi-api.txt / mpi-abort<br class=""># [<a href="http://kanary.kaust.edu.sa:115527" target="_blank" class="">kanary.kaust.edu.sa:115527</a>] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages<br class=""> ok vec_is_sf_tutorials-ex1_4 # SKIP Command failed so no diff<br class=""># retrying vec_is_sf_tutorials-ex1_4+sf_window_sync-active_sf_window_flavor-dynamic<br class="">not ok vec_is_sf_tutorials-ex1_4+sf_window_sync-active_sf_window_flavor-dynamic # Error code: 98<br class=""># [1]PETSC ERROR: --------------------- Error Message --------------------------------------------------------------<br class=""># [1]PETSC ERROR: General MPI error <br class=""># [1]PETSC ERROR: MPI error 6 MPI_ERR_RANK: invalid rank<br class=""># [1]PETSC ERROR: See <a href="https://petsc.org/release/faq/" target="_blank" class="">https://petsc.org/release/faq/</a> for trouble shooting.<br class=""># [1]PETSC ERROR: Petsc Development GIT revision: v3.15.4-783-g168bb9f76b GIT Date: 2021-09-13 14:01:22 +0000<br class=""># [1]PETSC ERROR: ../ex1 on a arch-debug named <a href="http://kanary.kaust.edu.sa/" target="_blank" class="">kanary.kaust.edu.sa</a> by zampins Tue Sep 14 09:31:45 2021<br class=""># [1]PETSC ERROR: Configure options --with-cc=/home/zampins/local/bin/mpicc --with-cxx-dialect=c++14 --with-cxx=/home/zampins/local/bin/mpicxx --with-debugging=1 --with-fc=/home/zampins/local/bin/mpifort --with-fortran-bindings=0 --with-hip-dir=/opt/rocm --with-hip=1 --with-hypre-dir=/home/zampins/local-petsc --with-kokkos-dir=/home/zampins/local-petsc --with-kokkos-kernels-dir=/home/zampins/local-petsc --with-blaslapack-include=/home/zampins/local-aocl/aocc/3.0-6/include --with-blaslapack-lib="[/home/zampins/local-aocl/aocc/3.0-6/lib/liblapacke.a,/home/zampins/local-aocl/aocc/3.0-6/lib/libflame.a,/home/zampins/local-aocl/aocc/3.0-6/lib/libblis-mt.a]" HIPFLAGS=--amdgpu-target=gfx908 HIPPPFLAGS=-I/home/zampins/local-petsc/include PETSC_ARCH=arch-debug<br class=""># [1]PETSC ERROR: #1 PetscSFGetGroups() at /home/zampins/Devel/petsc/src/vec/is/sf/interface/sf.c:1166<br class=""># [1]PETSC ERROR: #2 PetscSFSetUp_Window() at /home/zampins/Devel/petsc/src/vec/is/sf/impls/window/sfwindow.c:708<br class=""># [1]PETSC ERROR: #3 PetscSFSetUp() at /home/zampins/Devel/petsc/src/vec/is/sf/interface/sf.c:318<br class=""># [1]PETSC ERROR: #4 main() at /home/zampins/Devel/petsc/src/vec/is/sf/tutorials/ex1.c:172<br class=""># [1]PETSC ERROR: PETSc Option Table entries:<br class=""># [1]PETSC ERROR: -sf_type window<br class=""># [1]PETSC ERROR: [2]PETSC ERROR: --------------------- Error Message --------------------------------------------------------------<br class=""># [2]PETSC ERROR: General MPI error <br class=""># [2]PETSC ERROR: MPI error 6 MPI_ERR_RANK: invalid rank<br class=""># [2]PETSC ERROR: See <a href="https://petsc.org/release/faq/" target="_blank" class="">https://petsc.org/release/faq/</a> for trouble shooting.<br class=""># [2]PETSC ERROR: Petsc Development GIT revision: v3.15.4-783-g168bb9f76b GIT Date: 2021-09-13 14:01:22 +0000<br class=""># [2]PETSC ERROR: ../ex1 on a arch-debug named <a href="http://kanary.kaust.edu.sa/" target="_blank" class="">kanary.kaust.edu.sa</a> by zampins Tue Sep 14 09:31:45 2021<br class=""># [2]PETSC ERROR: Configure options --with-cc=/home/zampins/local/bin/mpicc --with-cxx-dialect=c++14 --with-cxx=/home/zampins/local/bin/mpicxx --with-debugging=1 --with-fc=/home/zampins/local/bin/mpifort --with-fortran-bindings=0 --with-hip-dir=/opt/rocm --with-hip=1 --with-hypre-dir=/home/zampins/local-petsc --with-kokkos-dir=/home/zampins/local-petsc --with-kokkos-kernels-dir=/home/zampins/local-petsc --with-blaslapack-include=/home/zampins/local-aocl/aocc/3.0-6/include --with-blaslapack-lib="[/home/zampins/local-aocl/aocc/3.0-6/lib/liblapacke.a,/home/zampins/local-aocl/aocc/3.0-6/lib/libflame.a,/home/zampins/local-aocl/aocc/3.0-6/lib/libblis-mt.a]" HIPFLAGS=--amdgpu-target=gfx908 HIPPPFLAGS=-I/home/zampins/local-petsc/include PETSC_ARCH=arch-debug<br class=""># [2]PETSC ERROR: #1 PetscSFGetGroups() at /home/zampins/Devel/petsc/src/vec/is/sf/interface/sf.c:1166<br class=""># [2]PETSC ERROR: #2 PetscSFSetUp_Window() at /home/zampins/Devel/petsc/src/vec/is/sf/impls/window/sfwindow.c:708<br class=""># [2]PETSC ERROR: #3 PetscSFSetUp() at /home/zampins/Devel/petsc/src/vec/is/sf/interface/sf.c:318<br class=""># [2]PETSC ERROR: #4 main() at /home/zampins/Devel/petsc/src/vec/is/sf/tutorials/ex1.c:172<br class=""># [2]PETSC ERROR: PETSc Option Table entries:<br class=""># [2]PETSC ERROR: -sf_type window<br class=""># [2]PETSC ERROR: -sf_window_flavor dynamic<br class=""># [2]PETSC ERROR: -sf_window_sync active<br class=""># [2]PETSC ERROR: -test_gather<br class=""># [2]PETSC ERROR: ----------------End of Error Message -------send entire error message to <a href="mailto:petsc-maint@mcs.anl.gov" target="_blank" class="">petsc-maint@mcs.anl.gov</a>----------<br class=""># [3]PETSC ERROR: --------------------- Error Message --------------------------------------------------------------<br class=""># [3]PETSC ERROR: General MPI error <br class=""># [3]PETSC ERROR: MPI error 6 MPI_ERR_RANK: invalid rank<br class=""># [3]PETSC ERROR: See <a href="https://petsc.org/release/faq/" target="_blank" class="">https://petsc.org/release/faq/</a> for trouble shooting.<br class=""># [3]PETSC ERROR: Petsc Development GIT revision: v3.15.4-783-g168bb9f76b GIT Date: 2021-09-13 14:01:22 +0000<br class=""># [3]PETSC ERROR: ../ex1 on a arch-debug named <a href="http://kanary.kaust.edu.sa/" target="_blank" class="">kanary.kaust.edu.sa</a> by zampins Tue Sep 14 09:31:45 2021<br class=""># [3]PETSC ERROR: Configure options --with-cc=/home/zampins/local/bin/mpicc --with-cxx-dialect=c++14 --with-cxx=/home/zampins/local/bin/mpicxx --with-debugging=1 --with-fc=/home/zampins/local/bin/mpifort --with-fortran-bindings=0 --with-hip-dir=/opt/rocm --with-hip=1 --with-hypre-dir=/home/zampins/local-petsc --with-kokkos-dir=/home/zampins/local-petsc --with-kokkos-kernels-dir=/home/zampins/local-petsc --with-blaslapack-include=/home/zampins/local-aocl/aocc/3.0-6/include --with-blaslapack-lib="[/home/zampins/local-aocl/aocc/3.0-6/lib/liblapacke.a,/home/zampins/local-aocl/aocc/3.0-6/lib/libflame.a,/home/zampins/local-aocl/aocc/3.0-6/lib/libblis-mt.a]" HIPFLAGS=--amdgpu-target=gfx908 HIPPPFLAGS=-I/home/zampins/local-petsc/include PETSC_ARCH=arch-debug<br class=""># [3]PETSC ERROR: #1 PetscSFGetGroups() at /home/zampins/Devel/petsc/src/vec/is/sf/interface/sf.c:1166<br class=""># [3]PETSC ERROR: #2 PetscSFSetUp_Window() at /home/zampins/Devel/petsc/src/vec/is/sf/impls/window/sfwindow.c:708<br class=""># [3]PETSC ERROR: #3 PetscSFSetUp() at /home/zampins/Devel/petsc/src/vec/is/sf/interface/sf.c:318<br class=""># [3]PETSC ERROR: #4 main() at /home/zampins/Devel/petsc/src/vec/is/sf/tutorials/ex1.c:172<br class=""># [3]PETSC ERROR: PETSc Option Table entries:<br class=""># [3]PETSC ERROR: -sf_type window<br class=""># [3]PETSC ERROR: -sf_window_flavor dynamic<br class=""># [3]PETSC ERROR: -sf_window_sync active<br class=""># [3]PETSC ERROR: -test_gather<br class=""># [3]PETSC ERROR: ----------------End of Error Message -------send entire error message to <a href="mailto:petsc-maint@mcs.anl.gov" target="_blank" class="">petsc-maint@mcs.anl.gov</a>----------<br class=""># -sf_window_flavor dynamic<br class=""># [1]PETSC ERROR: -sf_window_sync active<br class=""># [1]PETSC ERROR: -test_gather<br class=""># [1]PETSC ERROR: ----------------End of Error Message -------send entire error message to <a href="mailto:petsc-maint@mcs.anl.gov" target="_blank" class="">petsc-maint@mcs.anl.gov</a>----------<br class=""># --------------------------------------------------------------------------<br class=""># MPI_ABORT was invoked on rank 3 in communicator MPI_COMM_WORLD<br class=""># with errorcode 98.<br class=""># <br class=""># NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.<br class=""># You may or may not see output from other processes, depending on<br class=""># exactly when Open MPI kills them.<br class=""># --------------------------------------------------------------------------<br class=""># [<a href="http://kanary.kaust.edu.sa:115572" target="_blank" class="">kanary.kaust.edu.sa:115572</a>] 2 more processes have sent help message help-mpi-api.txt / mpi-abort<br class=""># [<a href="http://kanary.kaust.edu.sa:115572" target="_blank" class="">kanary.kaust.edu.sa:115572</a>] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages<br class=""> ok vec_is_sf_tutorials-ex1_4 # SKIP Command failed so no diff<br class=""> ok vec_is_sf_tutorials-ex1_4+sf_window_sync-active_sf_window_flavor-allocate<br class=""> ok diff-vec_is_sf_tutorials-ex1_4+sf_window_sync-active_sf_window_flavor-allocate<br class=""> ok vec_is_sf_tutorials-ex1_4+sf_window_sync-lock_sf_window_flavor-create<br class="">not ok diff-vec_is_sf_tutorials-ex1_4+sf_window_sync-lock_sf_window_flavor-create # Error code: 1<br class=""># 43,46c43,46<br class=""># < [0] 0: 4001 2000 2002 3002 4002<br class=""># < [1] 0: 1001 3000<br class=""># < [2] 0: 2001 4000<br class=""># < [3] 0: 3001 1000<br class=""># ---<br class=""># > [0] 0: 4002 2146435072 2 2146435072 34619728<br class=""># > [1] 0: 3000 2146435072<br class=""># > [2] 0: 4000 2146435072<br class=""># > [3] 0: 3001 2146435072<br class=""> ok vec_is_sf_tutorials-ex1_4+sf_window_sync-lock_sf_window_flavor-dynamic<br class="">not ok diff-vec_is_sf_tutorials-ex1_4+sf_window_sync-lock_sf_window_flavor-dynamic # Error code: 1<br class=""># 43,46c43,46<br class=""># < [0] 0: 4001 2000 2002 3002 4002<br class=""># < [1] 0: 1001 3000<br class=""># < [2] 0: 2001 4000<br class=""># < [3] 0: 3001 1000<br class=""># ---<br class=""># > [0] 0: 4002 2146435072 2 2146435072 0<br class=""># > [1] 0: 3000 2146435072<br class=""># > [2] 0: 4000 2146435072<br class=""># > [3] 0: 3001 2146435072<br class=""> ok vec_is_sf_tutorials-ex1_4+sf_window_sync-lock_sf_window_flavor-allocate<br class=""> ok diff-vec_is_sf_tutorials-ex1_4+sf_window_sync-lock_sf_window_flavor-allocate<br class=""></div><div class=""><br class=""></div><div class=""><br class=""></div></div><br class=""><div class="gmail_quote"><div dir="ltr" class="gmail_attr">Il giorno mar 14 set 2021 alle ore 07:44 Stefano Zampini <<a href="mailto:stefano.zampini@gmail.com" target="_blank" class="">stefano.zampini@gmail.com</a>> ha scritto:<br class=""></div><blockquote class="gmail_quote" style="margin:0px 0px 0px 0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex"><div dir="auto" class="">I'll see if I can reproduce </div><br class=""><div class="gmail_quote"><div dir="ltr" class="gmail_attr">Il Mar 14 Set 2021, 06:58 Junchao Zhang <<a href="mailto:junchao.zhang@gmail.com" target="_blank" class="">junchao.zhang@gmail.com</a>> ha scritto:<br class=""></div><blockquote class="gmail_quote" style="margin:0px 0px 0px 0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex"><div dir="ltr" class="">Hi, Stefano,<div class=""> Ping you again to see if you want to resolve this problem before petsc-3.16 </div><div class=""><br class=""></div><div class=""><div class=""><div dir="ltr" class=""><div dir="ltr" class="">--Junchao Zhang</div></div></div><br class=""></div></div><br class=""><div class="gmail_quote"><div dir="ltr" class="gmail_attr">On Sun, Sep 12, 2021 at 3:06 PM Antonio T. sagitter <<a href="mailto:sagitter@fedoraproject.org" rel="noreferrer" target="_blank" class="">sagitter@fedoraproject.org</a>> wrote:<br class=""></div><blockquote class="gmail_quote" style="margin:0px 0px 0px 0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex">Unfortunately, it's not possible. I must use the OpenMPI provided by <br class="">
Fedora build-system (these rpm builds of PETSc are for Fedora's <br class="">
repositories), downloading external software is not permitted.<br class="">
<br class="">
On 9/12/21 21:10, Pierre Jolivet wrote:<br class="">
> <br class="">
>> On 12 Sep 2021, at 8:56 PM, Matthew Knepley <<a href="mailto:knepley@gmail.com" rel="noreferrer" target="_blank" class="">knepley@gmail.com</a> <br class="">
>> <mailto:<a href="mailto:knepley@gmail.com" rel="noreferrer" target="_blank" class="">knepley@gmail.com</a>>> wrote:<br class="">
>><br class="">
>> On Sun, Sep 12, 2021 at 2:49 PM Antonio T. sagitter <br class="">
>> <<a href="mailto:sagitter@fedoraproject.org" rel="noreferrer" target="_blank" class="">sagitter@fedoraproject.org</a> <mailto:<a href="mailto:sagitter@fedoraproject.org" rel="noreferrer" target="_blank" class="">sagitter@fedoraproject.org</a>>> wrote:<br class="">
>><br class="">
>> Those attached are configure.log/make.log from a MPI build in<br class="">
>> Fedora 34<br class="">
>> x86_64 where the error below occurred.<br class="">
>><br class="">
>><br class="">
>> This is OpenMPI 4.1.0. Is that the only MPI you build? My first <br class="">
>> inclination is that this is an MPI implementation bug.<br class="">
>><br class="">
>> Junchao, do we have an OpenMPI build in the CI?<br class="">
> <br class="">
> config/examples/arch-ci-linux-cuda-double-64idx.py: <br class="">
> '--download-openmpi=1',<br class="">
> config/examples/arch-ci-linux-pkgs-dbg-ftn-interfaces.py: <br class="">
> '--download-openmpi=1',<br class="">
> config/examples/arch-ci-linux-pkgs-opt.py: '--download-openmpi=1',<br class="">
> <br class="">
> config/BuildSystem/config/packages/OpenMPI.py uses version 4.1.0 as well.<br class="">
> I’m not sure PETSc is to blame here Antonio. You may want to try to <br class="">
> ditch the OpenMPI shipped by your packet manager and try <br class="">
> --download-openmpi as well, just for a quick sanity check.<br class="">
> <br class="">
> Thanks,<br class="">
> Pierre<br class="">
> <br class="">
<br class="">
-- <br class="">
---<br class="">
Antonio Trande<br class="">
Fedora Project<br class="">
mailto: <a href="mailto:sagitter@fedoraproject.org" rel="noreferrer" target="_blank" class="">sagitter@fedoraproject.org</a><br class="">
GPG key: 0x29FBC85D7A51CC2F<br class="">
GPG key server: <a href="https://keyserver1.pgp.com/" rel="noreferrer noreferrer" target="_blank" class="">https://keyserver1.pgp.com/</a><br class="">
</blockquote></div>
</blockquote></div>
</blockquote></div><br clear="all" class=""><br class="">-- <br class=""><div dir="ltr" class="">Stefano</div>
</blockquote></div><br clear="all" class=""><div class=""><br class=""></div>-- <br class=""><div dir="ltr" class=""><div dir="ltr" class=""><div class=""><div dir="ltr" class=""><div class=""><div dir="ltr" class=""><div class="">What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead.<br class="">-- Norbert Wiener</div><div class=""><br class=""></div><div class=""><a href="http://www.cse.buffalo.edu/~knepley/" target="_blank" class="">https://www.cse.buffalo.edu/~knepley/</a><br class=""></div></div></div></div></div></div></div>
</blockquote></div>
</div></blockquote></div><br class=""></div></blockquote></div>
</div></blockquote></div><br class=""></div></body></html>