[petsc-users] [EXTERNAL] PETSc Installation Assistance
Matthew Knepley
knepley at gmail.com
Mon Jul 17 14:36:47 CDT 2023
Hi Jesus,
I think you are on main. Did everything you have get rebuilt? Toby just
rewrote large sections of logging
and this is right where it fails for you. It should be easy to see what is
wrong by running in the debugger.
Thanks,
Matt
On Mon, Jul 17, 2023 at 3:11 PM Pierre Jolivet <pierre.jolivet at lip6.fr>
wrote:
>
> On 17 Jul 2023, at 9:00 PM, Ferrand, Jesus A. <FERRANJ2 at my.erau.edu>
> wrote:
>
> Pierre:
> Setting the environment variable allows make check to complete without
> errors.
> It only seems to run three checks.
> I recall that in the past, make check would run 100+ tests.
>
>
> You are probably thinking of make test, which is running thousands of
> tests indeed.
>
> This is what I get from make check:
> =========================================================================
> Running check examples to verify correct installation
> Using PETSC_DIR=/home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc and
> PETSC_ARCH=arch-linux-c-debug
> C/C++ example src/snes/tutorials/ex19 run successfully with 1 MPI process
> C/C++ example src/snes/tutorials/ex19 run successfully with 2 MPI processes
>
> C/C++ example src/vec/vec/tests/ex47 run successfully with hdf5
> Fortran example src/snes/tutorials/ex5f run successfully with 1 MPI process
>
> Completed test examples
> =========================================================================
>
> Anyhow, some of the APIs that worked before now break.
> For example "DMPlexCreateGmshFromFile()" breaks when I read simple test
> mesh.
> I tried configuring and installing both the newest release of PETSc and
> that tarball of version 3.18.3.
> Both seem to be breaking APIs that used to work in the older OS.
>
>
> There has been some internal changes to the behavior of DMPlex, they
> should be documented in the change log.
> But you are saying that with the same version, just by switching OS, you
> get a different behavior?
> Do as Satish suggested then, because it’s unlikely that the root cause is
> in the library itself.
>
> Thanks,
> Pierre
>
> =========================================================================
> [0]PETSC ERROR:
> ------------------------------------------------------------------------
> [0]PETSC ERROR: Caught signal number 11 SEGV: Segmentation Violation,
> probably memory access out of range
> [0]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger
>
> [0]PETSC ERROR: or see https://petsc.org/release/faq/#valgrind and
> https://petsc.org/release/faq/
> [0]PETSC ERROR: --------------------- Stack Frames
> ------------------------------------
> [0]PETSC ERROR: The line numbers in the error traceback are not always
> exact.
> [0]PETSC ERROR: #1 DMPlexSetMigrationSF() at
> /home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc/src/dm/impls/plex/ple
> xnatural.c:21
> [0]PETSC ERROR: #2 JAF_DMPlexCreateFromMesh() at
> /home/jesus/Desktop/JAF_NML/ApplicationCode/PETSc/PETScCGH5.c:5865
> application called MPI_Abort(MPI_COMM_WORLD, 59) - process 0
> [unset]: PMIU_write error; fd=-1 buf=:cmd=abort exitcode=59
> message=application called MPI_Abort(MPI_COMM_WORLD, 59)
> - process 0
> :
> system msg for write_line failure : Bad file descriptor
> [0]PETSC ERROR:
> ------------------------------------------------------------------------
> [0]PETSC ERROR: Caught signal number 11 SEGV: Segmentation Violation,
> probably memory access out of range
> [0]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger
>
> [0]PETSC ERROR: or see https://petsc.org/release/faq/#valgrind and
> https://petsc.org/release/faq/
> [0]PETSC ERROR: --------------------- Stack Frames
> ------------------------------------
> [0]PETSC ERROR: The line numbers in the error traceback are not always
> exact.
> [0]PETSC ERROR: #1 DMPlexSetMigrationSF() at
> /home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc/src/dm/impls/plex/ple
> xnatural.c:21
> [0]PETSC ERROR: #2 JAF_DMPlexCreateFromMesh() at
> /home/jesus/Desktop/JAF_NML/ApplicationCode/PETSc/PETScCGH5.c:5865
> application called MPI_Abort(MPI_COMM_WORLD, 59) - process 0
> [unset]: PMIU_write error; fd=-1 buf=:cmd=abort exitcode=59
> message=application called MPI_Abort(MPI_COMM_WORLD, 59)
> - process 0
> :
> system msg for write_line failure : Bad file descriptor
> --------------------------------------------------------------------------
>
> Primary job terminated normally, but 1 process returned
> a non-zero exit code. Per user-direction, the job has been aborted.
> --------------------------------------------------------------------------
>
> [0]PETSC ERROR:
> ------------------------------------------------------------------------
> [0]PETSC ERROR: Caught signal number 15 Terminate: Some process (or the
> batch system) has told this process to end
> [0]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger
>
> [0]PETSC ERROR: or see https://petsc.org/release/faq/#valgrind and
> https://petsc.org/release/faq/
> [0]PETSC ERROR: --------------------- Stack Frames
> ------------------------------------
> [0]PETSC ERROR: The line numbers in the error traceback are not always
> exact.
> [0]PETSC ERROR: #1 PetscEventRegLogGetEvent() at
> /home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc/src/sys/logging/u
> tils/eventlog.c:622
> [0]PETSC ERROR: #2 PetscLogEventRegister() at
> /home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc/src/sys/logging/plog
> .c:802
> [0]PETSC ERROR: #3 VecInitializePackage() at
> /home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc/src/vec/vec/interface
> /dlregisvec.c:187
> [0]PETSC ERROR: #4 VecCreate() at
> /home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc/src/vec/vec/interface/veccreate.
> c:32
> [0]PETSC ERROR: #5 DMCreateLocalVector_Section_Private() at
> /home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc/src/dm
> /interface/dmi.c:80
> [0]PETSC ERROR: #6 DMCreateLocalVector_Plex() at
> /home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc/src/dm/impls/plex
> /plexcreate.c:4432
> [0]PETSC ERROR: #7 DMCreateLocalVector() at
> /home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc/src/dm/interface/dm.c:
> 1056
> [0]PETSC ERROR: #8 DMPlexCreateGmsh() at
> /home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc/src/dm/impls/plex/plexgms
> h.c:1933
> [0]PETSC ERROR: #9 DMPlexCreateGmshFromFile() at
> /home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc/src/dm/impls/plex
> /plexgmsh.c:1433
> [0]PETSC ERROR: #10 JAF_DMPlexCreateFromMesh() at
> /home/jesus/Desktop/JAF_NML/ApplicationCode/PETSc/PETScCGH5.c:5845
> application called MPI_Abort(MPI_COMM_WORLD, 59) - process 0
> [unset]: PMIU_write error; fd=-1 buf=:cmd=abort exitcode=59
> message=application called MPI_Abort(MPI_COMM_WORLD, 59)
> - process 0
> :
> system msg for write_line failure : Bad file descriptor
> [0]PETSC ERROR:
> ------------------------------------------------------------------------
> [0]PETSC ERROR: Caught signal number 15 Terminate: Some process (or the
> batch system) has told this process to end
> [0]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger
>
> [0]PETSC ERROR: or see https://petsc.org/release/faq/#valgrind and
> https://petsc.org/release/faq/
> [0]PETSC ERROR: --------------------- Stack Frames
> ------------------------------------
> [0]PETSC ERROR: The line numbers in the error traceback are not always
> exact.
> [0]PETSC ERROR: #1 PetscStrcasecmp() at
> /home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc/src/sys/utils/str.c:285
> [0]PETSC ERROR: #2 PetscEventRegLogGetEvent() at
> /home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc/src/sys/logging/u
> tils/eventlog.c:622
> [0]PETSC ERROR: #3 PetscLogEventRegister() at
> /home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc/src/sys/logging/plog
> .c:802
> [0]PETSC ERROR: #4 VecInitializePackage() at
> /home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc/src/vec/vec/interface
> /dlregisvec.c:188
> [0]PETSC ERROR: #5 VecCreate() at
> /home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc/src/vec/vec/interface/veccreate.
> c:32
> [0]PETSC ERROR: #6 DMCreateLocalVector_Section_Private() at
> /home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc/src/dm
> /interface/dmi.c:80
> [0]PETSC ERROR: #7 DMCreateLocalVector_Plex() at
> /home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc/src/dm/impls/plex
> /plexcreate.c:4432
> [0]PETSC ERROR: #8 DMCreateLocalVector() at
> /home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc/src/dm/interface/dm.c:
> 1056
> [0]PETSC ERROR: #9 DMPlexCreateGmsh() at
> /home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc/src/dm/impls/plex/plexgms
> h.c:1933
> [0]PETSC ERROR: #10 DMPlexCreateGmshFromFile() at
> /home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc/src/dm/impls/ple
> x/plexgmsh.c:1433
> [0]PETSC ERROR: #11 JAF_DMPlexCreateFromMesh() at
> /home/jesus/Desktop/JAF_NML/ApplicationCode/PETSc/PETScCGH5.c:5845
> application called MPI_Abort(MPI_COMM_WORLD, 59) - process 0
> [unset]: PMIU_write error; fd=-1 buf=:cmd=abort exitcode=59
> message=application called MPI_Abort(MPI_COMM_WORLD, 59)
> - process 0
> :
> system msg for write_line failure : Bad file descriptor
> --------------------------------------------------------------------------
>
> mpiexec detected that one or more processes exited with non-zero status,
> thus causing
> the job to be terminated. The first process to do so was:
>
> Process name: [[33478,1],2]
> Exit code: 15
> ===========================================================================
>
> Here are the options I give ./configure:
>
> ./configure --download-mpich=yes --download-viennacl=yes
> --download-hdf5=yes --download-chaco=yes --download-metis=yes
> --download-parmetis=yes --download-cgns=yes
> ------------------------------
> *From:* Pierre Jolivet <pierre.jolivet at lip6.fr>
> *Sent:* Monday, July 17, 2023 1:58 PM
> *To:* Ferrand, Jesus A. <FERRANJ2 at my.erau.edu>
> *Cc:* petsc-users at mcs.anl.gov <petsc-users at mcs.anl.gov>
> *Subject:* [EXTERNAL] Re: [petsc-users] PETSc Installation Assistance
>
> *CAUTION:* This email originated outside of Embry-Riddle Aeronautical
> University. Do not click links or open attachments unless you recognize the
> sender and know the content is safe.
>
> https://petsc.org/release/faq/#what-does-the-message-hwloc-linux-ignoring-pci-device-with-non-16bit-domain-mean
>
> Thanks,
> Pierre
>
> On 17 Jul 2023, at 7:51 PM, Ferrand, Jesus A. <FERRANJ2 at my.erau.edu>
> wrote:
>
> Greetings.
>
> I recently changed operating systems (Ubuntu 20.04 -> Debian 12
> "Bookworm") and tried to reinstall PETSc.
> I tried doing the usual as described in (
> https://petsc.org/release/install/download/#recommended-obtain-release-version-with-git
> ):
>
> - git clone/pull
> - ./configure --<option_1> ... --<option_N>
> - make install
> - make check
>
> Everything proceeds smoothly until the "make check" step, where I get the
> following error:
>
> ==============================================================================
> Running check examples to verify correct installation
> Using PETSC_DIR=/home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc and
> PETSC_ARCH=arch-linux-c-debug
> Possible error running C/C++ src/snes/tutorials/ex19 with 1 MPI process
> See https://petsc.org/release/faq/
> hwloc/linux: Ignoring PCI device with non-16bit domain.
> Pass --enable-32bits-pci-domain to configure to support such devices
> (warning: it would break the library ABI, don't enable unless really
> needed).
> lid velocity = 0.0016, prandtl # = 1., grashof # = 1.
> Number of SNES iterations = 2
> Possible error running C/C++ src/snes/tutorials/ex19 with 2 MPI processes
> See https://petsc.org/release/faq/
> hwloc/linux: Ignoring PCI device with non-16bit domain.
> Pass --enable-32bits-pci-domain to configure to support such devices
> (warning: it would break the library ABI, don't enable unless really
> needed).
> lid velocity = 0.0016, prandtl # = 1., grashof # = 1.
> Number of SNES iterations = 2
> 0a1,3
> > hwloc/linux: Ignoring PCI device with non-16bit domain.
> > Pass --enable-32bits-pci-domain to configure to support such devices
> > (warning: it would break the library ABI, don't enable unless really
> needed).
> /home/jesus/Desktop/JAF_NML/3rd_Party/PETSc/petsc/src/vec/vec/tests
> Possible problem with ex47 running with hdf5, diffs above
> =========================================
> Possible error running Fortran example src/snes/tutorials/ex5f with 1 MPI
> process
> See https://petsc.org/release/faq/
> hwloc/linux: Ignoring PCI device with non-16bit domain.
> Pass --enable-32bits-pci-domain to configure to support such devices
> (warning: it would break the library ABI, don't enable unless really
> needed).
> Number of SNES iterations = 3
> Completed test examples
> Error while running make check
> gmake[1]: *** [makefile:123: check] Error 1
> make: *** [GNUmakefile:17: check] Error 2
>
> ==============================================================================
>
>
> I tried reinstalling the same version I was able to use prior to changing
> OS's (PETSc 3.18.3, via tarball) and get a similar error.
>
> ==============================================================================
> make PETSC_DIR=/home/jesus/Desktop/JAF_NML/3rd_P
> arty/newPETSC/petsc-3.18.3 PETSC_ARCH=arch-linux-c-debug check
> Running check examples to verify correct installation
> Using
> PETSC_DIR=/home/jesus/Desktop/JAF_NML/3rd_Party/newPETSC/petsc-3.18.3 and
> PETSC_ARCH=arch-linux-c-debug
> Possible error running C/C++ src/snes/tutorials/ex19 with 1 MPI process
> See https://petsc.org/release/faq/
> hwloc/linux: Ignoring PCI device with non-16bit domain.
> Pass --enable-32bits-pci-domain to configure to support such devices
> (warning: it would break the library ABI, don't enable unless really
> needed).
> lid velocity = 0.0016, prandtl # = 1., grashof # = 1.
> Number of SNES iterations = 2
> Possible error running C/C++ src/snes/tutorials/ex19 with 2 MPI processes
> See https://petsc.org/release/faq/
> hwloc/linux: Ignoring PCI device with non-16bit domain.
> Pass --enable-32bits-pci-domain to configure to support such devices
> (warning: it would break the library ABI, don't enable unless really
> needed).
> hwloc/linux: Ignoring PCI device with non-16bit domain.
> Pass --enable-32bits-pci-domain to configure to support such devices
> (warning: it would break the library ABI, don't enable unless really
> needed).
> lid velocity = 0.0016, prandtl # = 1., grashof # = 1.
> Number of SNES iterations = 2
> 0a1,3
> > hwloc/linux: Ignoring PCI device with non-16bit domain.
> > Pass --enable-32bits-pci-domain to configure to support such devices
> > (warning: it would break the library ABI, don't enable unless really
> needed).
>
> /home/jesus/Desktop/JAF_NML/3rd_Party/newPETSC/petsc-3.18.3/src/vec/vec/tests
>
> Possible problem with ex47 running with hdf5, diffs above
> =========================================
> Possible error running Fortran example src/snes/tutorials/ex5f with 1 MPI
> process
> See https://petsc.org/release/faq/
> hwloc/linux: Ignoring PCI device with non-16bit domain.
> Pass --enable-32bits-pci-domain to configure to support such devices
> (warning: it would break the library ABI, don't enable unless really
> needed).
> Number of SNES iterations = 3
> Completed test examples
> Error while running make check
> gmake[1]: *** [makefile:149: check] Error 1
> make: *** [GNUmakefile:17: check] Error 2
>
> ==============================================================================
>
>
> I checked the FAQs (https://petsc.org/release/faq/#table-of-contents) and
> tried the suggestions there (specifically the points under
> https://petsc.org/release/faq/#what-does-it-mean-when-make-check-hangs-or-errors-on-petscoptionsinsertfile)
> to no avail.
>
> - I don't use a VPN.
> - I tried running make check with Wifi disabled.
> - Ran ./configure with --download-mpich-device=ch3:nemesis
> --download-mpich
>
>
> Surprisingly, I can compile my PETSc programs using a makefile, and run
> them.
> However, the programs are now broken (I get SIGSEGV originating from
> within DM/DMPlex APIs).
>
>
> Sincerely:
> *J.A. Ferrand*
> Embry-Riddle Aeronautical University - Daytona Beach - FL
> Ph.D. Candidate, Aerospace Engineering
> M.Sc. Aerospace Engineering
> B.Sc. Aerospace Engineering
> B.Sc. Computational Mathematics
>
> *Phone:* (386)-843-1829
> *Email(s):* ferranj2 at my.erau.edu
> jesus.ferrand at gmail.com
>
>
>
--
What most experimenters take for granted before they begin their
experiments is infinitely more interesting than any results to which their
experiments lead.
-- Norbert Wiener
https://www.cse.buffalo.edu/~knepley/ <http://www.cse.buffalo.edu/~knepley/>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20230717/2d7f61ad/attachment-0001.html>
More information about the petsc-users
mailing list