[petsc-users] Error using Metis with PETSc installed with MUMPS
Victoria Rolandi
victoria.rolandi93 at gmail.com
Thu Nov 2 11:29:01 CDT 2023
Pierre,
Yes, sorry, I'll keep the list in copy.
Launching with those options (-mat_mumps_icntl_28 2 -mat_mumps_icntl_29 2)
I get an error during the analysis step. I also launched increasing the
memory and I still have the error.
*The calculations stops at :*
Entering CMUMPS 5.4.1 from C interface with JOB, N = 1 699150
executing #MPI = 2, without OMP
=================================================
MUMPS compiled with option -Dmetis
MUMPS compiled with option -Dparmetis
=================================================
L U Solver for unsymmetric matrices
Type of parallelism: Working host
****** ANALYSIS STEP ********
** Maximum transversal (ICNTL(6)) not allowed because matrix is distributed
Using ParMETIS for parallel ordering
Structural symmetry is: 90%
*The error:*
[0]PETSC ERROR: Caught signal number 11 SEGV: Segmentation Violation,
probably memory access out of range
[0]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger
[0]PETSC ERROR: or see https://petsc.org/release/faq/#valgrind
[0]PETSC ERROR: or try http://valgrind.org on GNU/linux and Apple MacOS to
find memory corruption errors
[0]PETSC ERROR: configure using --with-debugging=yes, recompile, link, and
run
[0]PETSC ERROR: to get more information on the crash.
[0]PETSC ERROR: --------------------- Error Message
--------------------------------------------------------------
[0]PETSC ERROR: Signal received
[0]PETSC ERROR: See https://petsc.org/release/faq/ for trouble shooting.
[0]PETSC ERROR: Petsc Release Version 3.17.0, unknown
[0]PETSC ERROR: ./charlin.exe on a named n1056 by vrolandi Wed Nov 1
11:38:28 2023
[0]PETSC ERROR: Configure options
--prefix=/u/home/v/vrolandi/CODES/LIBRARY/packages/petsc/installationDir
--with-cc=mpiicc --with-cxx=mpiicpc --with-fc=mpiifort CXXOPTFLAGS=-O3
--with-scalar-type=complex --with-debugging=0 --with-precision=single
--download-mumps --download-scalapack --download-parmetis --download-metis
[0]PETSC ERROR: #1 User provided function() at unknown file:0
[0]PETSC ERROR: Run with -malloc_debug to check if memory corruption is
causing the crash.
Abort(59) on node 0 (rank 0 in comm 0): application called
MPI_Abort(MPI_COMM_WORLD, 59) - process 0
Thanks,
Victoria
Il giorno mer 1 nov 2023 alle ore 10:33 Pierre Jolivet <pierre at joliv.et> ha
scritto:
> Victoria, please keep the list in copy.
>
> I am not understanding how can I switch to ParMetis if it does not appear
> in the options of -mat_mumps_icntl_7.In the options I only have Metis and
> not ParMetis.
>
>
> You need to use -mat_mumps_icntl_28 2 -mat_mumps_icntl_29 2
>
> Barry, I don’t think we can programmatically shut off this warning, it’s
> guarded by a bunch of KEEP() values, see src/dana_driver.F:4707, which are
> only settable/gettable by people with access to consortium releases.
> I’ll ask the MUMPS people for confirmation.
> Note that this warning is only printed to screen with the option
> -mat_mumps_icntl_4 2 (or higher), so this won’t show up for standard runs.
>
> Thanks,
> Pierre
>
> On 1 Nov 2023, at 5:52 PM, Barry Smith <bsmith at petsc.dev> wrote:
>
>
> Pierre,
>
> Could the PETSc MUMPS interface "turn-off" ICNTL(6) in this situation
> so as to not trigger the confusing warning message from MUMPS?
>
> Barry
>
> On Nov 1, 2023, at 12:17 PM, Pierre Jolivet <pierre at joliv.et> wrote:
>
>
>
> On 1 Nov 2023, at 3:33 PM, Zhang, Hong via petsc-users <
> petsc-users at mcs.anl.gov> wrote:
>
> Victoria,
> "** Maximum transversal (ICNTL(6)) not allowed because matrix is
> distributed
> Ordering based on METIS"
>
>
> This warning is benign and appears for every run using a sequential
> partitioner in MUMPS with a MATMPIAIJ.
> (I’m not saying switching to ParMETIS will not make the issue go away)
>
> Thanks,
> Pierre
>
> $ ../../../../arch-darwin-c-debug-real/bin/mpirun -n 2 ./ex2 -pc_type lu
> -mat_mumps_icntl_4 2
> Entering DMUMPS 5.6.2 from C interface with JOB, N = 1 56
> executing #MPI = 2, without OMP
>
> =================================================
> MUMPS compiled with option -Dmetis
> MUMPS compiled with option -Dparmetis
> MUMPS compiled with option -Dpord
> MUMPS compiled with option -Dptscotch
> MUMPS compiled with option -Dscotch
> =================================================
> L U Solver for unsymmetric matrices
> Type of parallelism: Working host
>
> ****** ANALYSIS STEP ********
>
> ** Maximum transversal (ICNTL(6)) not allowed because matrix is
> distributed
> Processing a graph of size: 56 with 194 edges
> Ordering based on AMF
> WARNING: Largest root node of size 26 not selected for parallel
> execution
>
> Leaving analysis phase with ...
> INFOG(1) = 0
> INFOG(2) = 0
> […]
>
> Try parmetis.
> Hong
> ------------------------------
> *From:* petsc-users <petsc-users-bounces at mcs.anl.gov> on behalf of
> Victoria Rolandi <victoria.rolandi93 at gmail.com>
> *Sent:* Tuesday, October 31, 2023 10:30 PM
> *To:* petsc-users at mcs.anl.gov <petsc-users at mcs.anl.gov>
> *Subject:* [petsc-users] Error using Metis with PETSc installed with MUMPS
>
> Hi,
>
> I'm solving a large sparse linear system in parallel and I am using PETSc
> with MUMPS. I am trying to test different options, like the ordering of the
> matrix. Everything works if I use the *-mat_mumps_icntl_7 2 *or *-mat_mumps_icntl_7
> 0 *options (with the first one, AMF, performing better than AMD), however
> when I test METIS *-mat_mumps_icntl_7** 5 *I get an error (reported at
> the end of the email).
>
> I have configured PETSc with the following options:
>
> --with-cc=mpiicc --with-cxx=mpiicpc --with-fc=mpiifort
> --with-scalar-type=complex --with-debugging=0 --with-precision=single
> --download-mumps --download-scalapack --download-parmetis --download-metis
>
> and the installation didn't give any problems.
>
> Could you help me understand why metis is not working?
>
> Thank you in advance,
> Victoria
>
> Error:
>
> ****** ANALYSIS STEP ********
> ** Maximum transversal (ICNTL(6)) not allowed because matrix is
> distributed
> Processing a graph of size: 699150 with 69238690 edges
> Ordering based on METIS
> 510522 37081376 [100] [10486 699150]
> Error! Unknown CType: -1
>
>
>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20231102/54fd7657/attachment-0001.html>
More information about the petsc-users
mailing list