[petsc-users] subprocess (block) tolerance.
HeeHo Park
heeho.park at gmail.com
Thu Oct 4 18:09:51 CDT 2018
Barry and Jed,
Thank you for your answers. I think I need to learn more about domain
decomposition as I am a bit confused.
Is it true that we are using BiCGstab here to solve the system of
equations, using Additive Schwartz as a domain decomposition
preconditioner, and that precondition matrix for each block is
preconditioned by ilu[0]?
Thanks,
On Thu, Oct 4, 2018 at 5:53 PM Smith, Barry F. <bsmith at mcs.anl.gov> wrote:
>
> Since
>
> KSP Object: (flow_sub_) 1 MPI processes
> type: preonly
>
> this means only a single iteration of the inner solver is used so the
> numbers in red are not used.
>
> You could do something like -flow_ksp_type fgmres -flow_sub_ksp_type gmres
> -flow_sub_ksp_rtol 1.e-2 but it wouldn't help maters. Likely the current
> values are the best.
>
> Barry
>
>
> > On Oct 4, 2018, at 5:44 PM, HeeHo Park <heeho.park at gmail.com> wrote:
> >
> > Hi, I'm running PFLOTRAN and in PFLOTRAN, we have flow_ and flow_sub_
> processes. I was wondering what the red underlined values meant (each block
> tolerance?) and how to change them (would it affect convergence?). Blue
> marked bold values are changed from the default values for linear solvers.
> >
> > FLOW Linear Solver
> > solver: bcgs
> > preconditioner: asm
> > atol: 1.000000E-10
> > rtol: 1.000000E-10
> > dtol: 1.000000E+04
> > maximum iteration: 10000
> > KSP Object: (flow_) 8 MPI processes
> > type: bcgs
> > maximum iterations=10000, initial guess is zero
> > tolerances: relative=1e-10, absolute=1e-10, divergence=10000.
> > left preconditioning
> > using PRECONDITIONED norm type for convergence test
> > PC Object: (flow_) 8 MPI processes
> > type: asm
> > Additive Schwarz: total subdomain blocks = 8, amount of overlap = 1
> > Additive Schwarz: restriction/interpolation type - RESTRICT
> > [0] number of local blocks = 1
> > [1] number of local blocks = 1
> > [2] number of local blocks = 1
> > [3] number of local blocks = 1
> > [4] number of local blocks = 1
> > [5] number of local blocks = 1
> > [6] number of local blocks = 1
> > [7] number of local blocks = 1
> > Local solve info for each block is in the following KSP and PC
> objects:
> > - - - - - - - - - - - - - - - - - -
> > [0] local block number 0, size = 1389
> > KSP Object: (flow_sub_) 1 MPI processes
> > type: preonly
> > maximum iterations=10000, initial guess is zero
> > >>> tolerances: relative=1e-05, absolute=1e-50, divergence=10000.
> > left preconditioning
> > using DEFAULT norm type for convergence test
> > PC Object: (flow_sub_) 1 MPI processes
> > type: ilu
> > PC has not been set up so information may be incomplete
> > out-of-place factorization
> > 0 levels of fill
> > tolerance for zero pivot 2.22045e-14
> > using diagonal shift on blocks to prevent zero pivot [INBLOCKS]
> > matrix ordering: natural
> > linear system matrix = precond matrix:
> > Mat Object: (flow_) 1 MPI processes
> > type: seqbaij
> > rows=1389, cols=1389, bs=3
> > total: nonzeros=20025, allocated nonzeros=20025
> > total number of mallocs used during MatSetValues calls =0
> > block size is 3
> > - - - - - - - - - - - - - - - - - -
> >
> > --
> > HeeHo Daniel Park
>
>
--
HeeHo Daniel Park
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20181004/1b0884dd/attachment-0001.html>
More information about the petsc-users
mailing list