[petsc-users] Poor multigrid convergence in parallel
Jed Brown
jed at jedbrown.org
Mon Jul 21 12:29:41 CDT 2014
Lawrence Mitchell <lawrence.mitchell at imperial.ac.uk> writes:
> Below I show output from a run on 1 process and then two (along with ksp_view) for the following options:
>
> -pc_type mg -ksp_rtol 1e-8 -ksp_max_it 6 -pc_mg_levels 2 -mg_levels_pc_type sor -ksp_monitor
>
> On 1 process:
> 0 KSP Residual norm 5.865090856053e+02
> 1 KSP Residual norm 1.293159126247e+01
> 2 KSP Residual norm 5.181199296299e-01
> 3 KSP Residual norm 1.268870802643e-02
> 4 KSP Residual norm 5.116058930806e-04
> 5 KSP Residual norm 3.735036960550e-05
> 6 KSP Residual norm 1.755288530515e-06
> KSP Object: 1 MPI processes
> type: gmres
> GMRES: restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement
> GMRES: happy breakdown tolerance 1e-30
> maximum iterations=6, initial guess is zero
> tolerances: relative=1e-08, absolute=1e-50, divergence=10000
> left preconditioning
> using PRECONDITIONED norm type for convergence test
> PC Object: 1 MPI processes
> type: mg
> MG: type is MULTIPLICATIVE, levels=2 cycles=v
> Cycles per PCApply=1
> Not using Galerkin computed coarse grid matrices
How are you sure the rediscretized matrices are correct in parallel?
I would stick with the redundant coarse solve and use
-mg_levels_ksp_type chebyshev -mg_levels_pc_type jacobi -ksp_monitor_true_residual
Use of Jacobi here is to make the smoother the same in parallel as
serial. (Usually SOR is a bit stronger, though I think the Cheby/SOR
combination is somewhat peculiar and usually overkill.)
Compare convergence with and without -pc_mg_galerkin.
-------------- next part --------------
A non-text attachment was scrubbed...
Name: not available
Type: application/pgp-signature
Size: 818 bytes
Desc: not available
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20140721/2dadf970/attachment.pgp>
More information about the petsc-users
mailing list