[petsc-dev] [petsc-users] gamg failure with petsc-dev

Stephan Kramer s.kramer at imperial.ac.uk
Thu Sep 18 09:05:46 CDT 2014


On 18/09/14 13:45, Mark Adams wrote:
> Can you remind me how this problem occurred?  I should make it a switch
> in the code because it is not super cheap to do and you are the only
> ones to have had this problem.

1) The problem is easily reproducible; just run:

./ex49 -elas_pc_type gamg -mx 100 -my 100 -mat_no_inode

in src/ksp/ksp/examples/tutorials. See output below. (The -mat_no_inode is not necessary but simplifies things).

2) The problem is caused by the fact that the aggregates at the finest level (that form the nodes of the one-but-finest level) cannot be guaranteed to be large enough to *independently* support all the modes of the near nullspace.

For example in a 2D Stokes velocity problem, if one of the aggregates is just an isolated node (and afaics this is unavoidable) even if you use blocking there are only 2 DOFs at the finest level associated with that aggregate whereas it is represented by 3 DOFs at the coarser level (1 for each of the near nullspace modes). Therefore the prolongation operator from the coarser to the finest level is necessarily rank deficient. In fact the orthogonalisation will simply produce a zero column in this case.

The zero column in the prolongator results in a zero diagonal entry on the coarse-level operator (P^T A P), which causes the smoothing at that level to fail (only with SOR, as Jacobi magically fixes this)

3) The likelihood of hitting this failure increases if you do not set the block size (MatSetBlockSize). As result there was a whole, only tangently related, discussion on the fact that we got very poor performance when setting the block size (order of magnitude more iterations with exactly the same matrix). Interesting as it may be, that whole discussion in itself had nothing to do with the issue above, which may occur whether or not you set the block size.

IMHO the method is simply not robust when using the default smoothing settings. However if you prefer the fix to be optional, we would be very happy with that solution as well
Thanks for looking into this

Cheers
Stephan

skramer at gyre:~/git/petsc/src/ksp/ksp/examples/tutorials$ ./ex49 -elas_pc_type gamg -mx 100 -my 100 -mat_no_inode
[0]PETSC ERROR: --------------------- Error Message --------------------------------------------------------------
[0]PETSC ERROR: Arguments are incompatible
[0]PETSC ERROR: Zero diagonal on row 1
[0]PETSC ERROR: See http://www.mcs.anl.gov/petsc/documentation/faq.html for trouble shooting.
[0]PETSC ERROR: Petsc Development GIT revision: v3.5.2-300-gdaea54c  GIT Date: 2014-09-17 20:06:25 -0500
[0]PETSC ERROR: ./ex49 on a arch-linux2-c-debug named gyre by skramer Thu Sep 18 14:51:52 2014
[0]PETSC ERROR: Configure options --download-fblaslapack=1 --download-blacs=1 --download-scalapack=1 --download-ptscotch=1 --download-mumps=1 --download-hypre=1 --download-suitesparse=1 --download-ml=1 --with-debugging=1 --prefix=/home/skramer/petsc/master-debug-18092014
[0]PETSC ERROR: #1 MatInvertDiagonal_SeqAIJ() line 1728 in /home/skramer/git/petsc/src/mat/impls/aij/seq/aij.c
[0]PETSC ERROR: #2 MatSOR_SeqAIJ() line 1760 in /home/skramer/git/petsc/src/mat/impls/aij/seq/aij.c
[0]PETSC ERROR: #3 MatSOR() line 3644 in /home/skramer/git/petsc/src/mat/interface/matrix.c
[0]PETSC ERROR: #4 PCApply_SOR() line 35 in /home/skramer/git/petsc/src/ksp/pc/impls/sor/sor.c
[0]PETSC ERROR: #5 PCApply() line 440 in /home/skramer/git/petsc/src/ksp/pc/interface/precon.c
[0]PETSC ERROR: #6 KSP_PCApply() line 230 in /home/skramer/git/petsc/include/petsc-private/kspimpl.h
[0]PETSC ERROR: #7 KSPSolve_Chebyshev() line 414 in /home/skramer/git/petsc/src/ksp/ksp/impls/cheby/cheby.c
[0]PETSC ERROR: #8 KSPSolve() line 460 in /home/skramer/git/petsc/src/ksp/ksp/interface/itfunc.c
[0]PETSC ERROR: #9 PCMGMCycle_Private() line 19 in /home/skramer/git/petsc/src/ksp/pc/impls/mg/mg.c
[0]PETSC ERROR: #10 PCMGMCycle_Private() line 48 in /home/skramer/git/petsc/src/ksp/pc/impls/mg/mg.c
[0]PETSC ERROR: #11 PCApply_MG() line 337 in /home/skramer/git/petsc/src/ksp/pc/impls/mg/mg.c
[0]PETSC ERROR: #12 PCApply() line 440 in /home/skramer/git/petsc/src/ksp/pc/interface/precon.c
[0]PETSC ERROR: #13 KSP_PCApply() line 230 in /home/skramer/git/petsc/include/petsc-private/kspimpl.h
[0]PETSC ERROR: #14 KSPInitialResidual() line 63 in /home/skramer/git/petsc/src/ksp/ksp/interface/itres.c
[0]PETSC ERROR: #15 KSPSolve_GMRES() line 234 in /home/skramer/git/petsc/src/ksp/ksp/impls/gmres/gmres.c
[0]PETSC ERROR: #16 KSPSolve() line 460 in /home/skramer/git/petsc/src/ksp/ksp/interface/itfunc.c
[0]PETSC ERROR: #17 solve_elasticity_2d() line 1053 in /home/skramer/git/petsc/src/ksp/ksp/examples/tutorials/ex49.c
[0]PETSC ERROR: #18 main() line 1104 in /home/skramer/git/petsc/src/ksp/ksp/examples/tutorials/ex49.c
[0]PETSC ERROR: ----------------End of Error Message -------send entire error message to petsc-maint at mcs.anl.gov----------





More information about the petsc-dev mailing list