[petsc-dev] Possible error running C/C++ src/snes/examples/tutorials/ex19 with 1 MPI process
Satish Balay
balay at mcs.anl.gov
Sun Oct 9 23:15:42 CDT 2016
On Sun, 9 Oct 2016, Antonio Trande wrote:
> There is still a bad exit status of OpenMPI libs on *32bit architectures
> only*:
>
> make: Entering directory
> '/builddir/build/BUILD/petsc-3.7.4/buildopenmpi_dir'
> Running test examples to verify correct installation
> Using PETSC_DIR=/builddir/build/BUILD/petsc-3.7.4/buildopenmpi_dir and
> PETSC_ARCH=i386
> Possible error running C/C++ src/snes/examples/tutorials/ex19 with 1 MPI
> process
> See http://www.mcs.anl.gov/petsc/documentation/faq.html
> [0]PETSC ERROR: #1 PetscOptionsInsert() line 717 in
> /builddir/build/BUILD/petsc-3.7.4/buildopenmpi_dir/src/sys/objects/options.c
> [0]PETSC ERROR: #2 PetscInitialize() line 871 in
> /builddir/build/BUILD/petsc-3.7.4/buildopenmpi_dir/src/sys/objects/pinit.c
> -------------------------------------------------------
> Primary job terminated normally, but 1 process returned
> a non-zero exit code.. Per user-direction, the job has been aborted.
> -------------------------------------------------------
> --------------------------------------------------------------------------
> mpiexec detected that one or more processes exited with non-zero status,
> thus causing
> the job to be terminated. The first process to do so was:
> Process name: [[16865,1],0]
> Exit code: 1
> --------------------------------------------------------------------------
> Possible error running C/C++ src/snes/examples/tutorials/ex19 with 2 MPI
> processes
> See http://www.mcs.anl.gov/petsc/documentation/faq.html
> [0]PETSC ERROR: #1 PetscOptionsInsert() line 717 in
> /builddir/build/BUILD/petsc-3.7.4/buildopenmpi_dir/src/sys/objects/options.c
> [0]PETSC ERROR: #2 PetscInitialize() line 871 in
> /builddir/build/BUILD/petsc-3.7.4/buildopenmpi_dir/src/sys/objects/pinit.c
> [1]PETSC ERROR: #1 PetscOptionsInsert() line 717 in
> /builddir/build/BUILD/petsc-3.7.4/buildopenmpi_dir/src/sys/objects/options.c
> [1]PETSC ERROR: #2 PetscInitialize() line 871 in
> /builddir/build/BUILD/petsc-3.7.4/buildopenmpi_dir/src/sys/objects/pinit.c
> -------------------------------------------------------
> Primary job terminated normally, but 1 process returned
> a non-zero exit code.. Per user-direction, the job has been aborted.
> -------------------------------------------------------
> --------------------------------------------------------------------------
> mpiexec detected that one or more processes exited with non-zero status,
> thus causing
> the job to be terminated. The first process to do so was:
> Process name: [[16891,1],0]
> Exit code: 1
> --------------------------------------------------------------------------
> Fortran example src/snes/examples/tutorials/ex5f run successfully with 1
> MPI process
> Completed test examples
> =========================================
>
> Full build log (32bit):
> https://copr-be.cloud.fedoraproject.org/results/sagitter/petsc/fedora-rawhide-i386/00462843-petsc/build.log.gz
>
> Full build log (64bit):
> https://copr-be.cloud.fedoraproject.org/results/sagitter/petsc/fedora-rawhide-x86_64/00462843-petsc/build.log.gz
>
>
>>>>>>>>
sundials:
Includes: -I/usr/include/openmpi-i386
Library: -L/usr/lib/openmpi/lib -lsundials_nvecparallel -L/usr/lib -lsundials_cvode
<<<<<<<
Don't know if this is the reason - but I would avoid -L/usr/lib in
here [and let the compiler locate it on its own].
Also -L/usr/lib/openmpi/lib and -I/usr/include/openmpi-i386 are in
default compiler search path for
/usr/lib/openmpi/bin/mpicc/mpicxx/mpif90 - so I would also not specify
them..
>>>>>>>
BLAS/LAPACK: -lopenblas -lopenblas
<<<<<<<
BTW: you could replace '--with-blas-lib=libopenblas.so --with-lapack-lib=libopenblas.so' with a single '--with-blas-lapack-lib=libopenblas.so'
to avoid this duplicate listing.
Also - you have '--with-64-bit-indices=0 ' configure option listed twice. But this should not cause any harm..
And I see the sequential build also crashing. Perhaps you can run this in valgrind again - to see what
the issue is.. [that line corresonds to a call to MPI_Comm_rank() - and I don't see any reason why this should not work]
[0]PETSC ERROR: #1 PetscOptionsInsert() line 717 in /builddir/build/BUILD/petsc-3.7.4/petsc-3.7.4/src/sys/objects/options.c
[0]PETSC ERROR: #2 PetscInitialize() line 871 in /builddir/build/BUILD/petsc-3.7.4/petsc-3.7.4/src/sys/objects/pinit.c
Satish
More information about the petsc-dev
mailing list