<html>
<head>
<meta content="text/html; charset=utf-8" http-equiv="Content-Type">
</head>
<body bgcolor="#FFFFFF" text="#000000">
<br>
<div class="moz-cite-prefix">On 06/12/2017 07:50 AM, Matthew Knepley
wrote:<br>
</div>
<blockquote
cite="mid:CAMYG4GknfUFxjy66mZ5OQ3Rx6iSGfKDFd3nkaVyJPSUD8RiWgw@mail.gmail.com"
type="cite">
<div dir="ltr">
<div class="gmail_extra">
<div class="gmail_quote">On Sun, Jun 11, 2017 at 11:06 PM,
David Nolte <span dir="ltr"><<a moz-do-not-send="true"
href="mailto:dnolte@dim.uchile.cl" target="_blank">dnolte@dim.uchile.cl</a>></span>
wrote:<br>
<blockquote class="gmail_quote" style="margin:0px 0px 0px
0.8ex;border-left:1px solid
rgb(204,204,204);padding-left:1ex">
<div bgcolor="#FFFFFF"> Thanks Matt, makes sense to me!<br>
<br>
I skipped direct solvers at first because for these
'real' configurations LU (mumps/superlu_dist) usally
goes out of memory (got 32GB RAM). It would be
reasonable to take one more step back and play with
synthetic examples.<br>
I managed to run one case though with 936k dofs using:
("user" =pressure mass matrix)<br>
<br>
<tt><...><br>
-pc_fieldsplit_schur_fact_type upper</tt><tt><br>
</tt><tt>-pc_fieldsplit_schur_<wbr>precondition user</tt><tt><br>
</tt><tt>-fieldsplit_0_ksp_type preonly </tt><tt><br>
</tt><tt>-fieldsplit_0_pc_type lu</tt><tt><br>
</tt><tt>-fieldsplit_0_pc_factor_mat_<wbr>solver_package
mumps</tt><tt><br>
</tt><tt><br>
</tt><tt> -fieldsplit_1_ksp_type gmres<br>
-fieldsplit_1_ksp_monitor_<wbr>true_residuals<br>
-fieldsplit_1_ksp_rtol 1e-10<br>
</tt><tt>-fieldsplit_1_pc_type lu</tt><tt><br>
</tt><tt> -fieldsplit_1_pc_factor_mat_<wbr>solver_package
mumps</tt><tt><br>
</tt><br>
It takes 2 outer iterations, as expected. However the
fieldsplit_1 solve takes very long.<br>
</div>
</blockquote>
<div><br>
</div>
<div>1) It should take 1 outer iterate, not two. The problem
is that your Schur tolerance is way too high. Use</div>
<div><br>
</div>
<div> -fieldsplit_1_ksp_rtol 1e-10</div>
<div><br>
</div>
<div>or something like that. Then it will take 1 iterate.</div>
</div>
</div>
</div>
</blockquote>
<br>
Shouldn't it take 2 with a triangular Schur factorization and exact
preconditioners, and 1 with a full factorization? (cf. Benzi et al
2005, p.66, <a class="moz-txt-link-freetext" href="http://www.mathcs.emory.edu/~benzi/Web_papers/bgl05.pdf">http://www.mathcs.emory.edu/~benzi/Web_papers/bgl05.pdf</a>)<br>
<br>
That's exactly what I set: <tt> -fieldsplit_1_ksp_rtol 1e-10 </tt>and
the Schur solver does drop below "rtol < 1e-10"<br>
<br>
<blockquote
cite="mid:CAMYG4GknfUFxjy66mZ5OQ3Rx6iSGfKDFd3nkaVyJPSUD8RiWgw@mail.gmail.com"
type="cite">
<div dir="ltr">
<div class="gmail_extra">
<div class="gmail_quote">
<div><br>
</div>
<div>2) There is a problem with the Schur solve. Now from
the iterates</div>
<div><br>
</div>
<div><span style="font-family:monospace">423 KSP
preconditioned resid norm 2.638419658982e-02 true resid
norm 7.229653211635e-11 ||r(i)||/||b||
7.229653211635e-11</span><br>
</div>
<div><br>
</div>
<div>it is clear that the preconditioner is really screwing
stuff up. For testing, you can use</div>
<div><br>
</div>
<div> -pc_fieldsplit_schur_precondition full</div>
<div><br>
</div>
<div>and your same setup here. It should take one iterate. I
think there is something wrong with your</div>
<div>mass matrix.</div>
</div>
</div>
</div>
</blockquote>
<br>
I agree. I forgot to mention that I am considering an "enclosed
flow" problem, with u=0 on all the boundary and a Dirichlet
condition for the pressure in one point for fixing the constant
pressure. Maybe the preconditioner is not consistent with this
setup, need to check this..<br>
<br>
Thanks a lot<br>
<br>
<br>
<blockquote
cite="mid:CAMYG4GknfUFxjy66mZ5OQ3Rx6iSGfKDFd3nkaVyJPSUD8RiWgw@mail.gmail.com"
type="cite">
<div dir="ltr">
<div class="gmail_extra">
<div class="gmail_quote">
<div><br>
</div>
<div> Thanks,</div>
<div><br>
</div>
<div> Matt</div>
<div><br>
</div>
<blockquote class="gmail_quote" style="margin:0px 0px 0px
0.8ex;border-left:1px solid
rgb(204,204,204);padding-left:1ex">
<div bgcolor="#FFFFFF"> <br>
<tt> 0 KSP unpreconditioned resid norm
4.038466809302e-03 true resid norm 4.038466809302e-03
||r(i)||/||b|| 1.000000000000e+00</tt><tt><br>
</tt><tt> Residual norms for fieldsplit_1_ solve.</tt><tt><br>
</tt><tt> 0 KSP preconditioned resid norm
0.000000000000e+00 true resid norm 0.000000000000e+00
||r(i)||/||b|| -nan</tt><tt><br>
</tt><tt> Linear fieldsplit_1_ solve converged due to
CONVERGED_ATOL iterations 0</tt><tt><br>
</tt><tt> 1 KSP unpreconditioned resid norm
4.860095964831e-06 true resid norm 4.860095964831e-06
||r(i)||/||b|| 1.203450763452e-03</tt><tt><br>
</tt><tt> Residual norms for fieldsplit_1_ solve.</tt><tt><br>
</tt><tt> 0 KSP preconditioned resid norm
2.965546249872e+08 true resid norm 1.000000000000e+00
||r(i)||/||b|| 1.000000000000e+00</tt><tt><br>
</tt><tt> 1 KSP preconditioned resid norm
1.347596594634e+08 true resid norm 3.599678801575e-01
||r(i)||/||b|| 3.599678801575e-01</tt><tt><br>
</tt><tt> 2 KSP preconditioned resid norm
5.913230136403e+07 true resid norm 2.364916760834e-01
||r(i)||/||b|| 2.364916760834e-01</tt><tt><br>
</tt><tt> 3 KSP preconditioned resid norm
4.629700028930e+07 true resid norm 1.984444715595e-01
||r(i)||/||b|| 1.984444715595e-01</tt><tt><br>
</tt><tt> 4 KSP preconditioned resid norm
3.804431276819e+07 true resid norm 1.747224559120e-01
||r(i)||/||b|| 1.747224559120e-01</tt><tt><br>
</tt><tt> 5 KSP preconditioned resid norm
3.178769422140e+07 true resid norm 1.402254864444e-01
||r(i)||/||b|| 1.402254864444e-01</tt><tt><br>
</tt><tt> 6 KSP preconditioned resid norm
2.648669043919e+07 true resid norm 1.191164310866e-01
||r(i)||/||b|| 1.191164310866e-01</tt><tt><br>
</tt><tt> 7 KSP preconditioned resid norm
2.203522108614e+07 true resid norm 9.690500018007e-02
||r(i)||/||b|| 9.690500018007e-02</tt><tt><br>
<...><br>
422 KSP preconditioned resid norm
2.984888715147e-02 true resid norm 8.598401046494e-11
||r(i)||/||b|| 8.598401046494e-11<br>
423 KSP preconditioned resid norm
2.638419658982e-02 true resid norm 7.229653211635e-11
||r(i)||/||b|| 7.229653211635e-11<br>
Linear fieldsplit_1_ solve converged due to
CONVERGED_RTOL iterations 423<br>
2 KSP unpreconditioned resid norm 3.539889585599e-16
true resid norm 3.542279617063e-16 ||r(i)||/||b||
8.771347603759e-14<br>
Linear solve converged due to CONVERGED_RTOL
iterations 2<br>
</tt><tt><br>
</tt><br>
Does the slow convergence of the Schur block mean that
my preconditioning matrix Sp is a poor choice?<br>
<br>
Thanks,<br>
David<br>
<br>
<br>
<div class="gmail-m_5328507656823621836moz-cite-prefix">On
06/11/2017 08:53 AM, Matthew Knepley wrote:<br>
</div>
<blockquote type="cite">
<div dir="ltr">
<div class="gmail_extra">
<div class="gmail_quote">On Sat, Jun 10, 2017 at
8:25 PM, David Nolte <span dir="ltr"><<a
moz-do-not-send="true"
href="mailto:dnolte@dim.uchile.cl"
target="_blank">dnolte@dim.uchile.cl</a>></span>
wrote:<br>
<blockquote class="gmail_quote"
style="margin:0px 0px 0px
0.8ex;border-left:1px solid
rgb(204,204,204);padding-left:1ex">Dear all,<br>
<br>
I am solving a Stokes problem in 3D aorta
geometries, using a P2/P1<br>
finite elements discretization on tetrahedral
meshes resulting in<br>
~1-1.5M DOFs. Viscosity is uniform (can be
adjusted arbitrarily), and<br>
the right hand side is a function of noisy
measurement data.<br>
<br>
In other settings of "standard" Stokes flow
problems I have obtained<br>
good convergence with an "upper" Schur
complement preconditioner, using<br>
AMG (ML or Hypre) on the velocity block and
approximating the Schur<br>
complement matrix by the diagonal of the
pressure mass matrix:<br>
<br>
-ksp_converged_reason<br>
-ksp_monitor_true_residual<br>
-ksp_initial_guess_nonzero<br>
-ksp_diagonal_scale<br>
-ksp_diagonal_scale_fix<br>
-ksp_type fgmres<br>
-ksp_rtol 1.0e-8<br>
<br>
-pc_type fieldsplit<br>
-pc_fieldsplit_type schur<br>
-pc_fieldsplit_detect_saddle_p<wbr>oint<br>
-pc_fieldsplit_schur_fact_type upper<br>
-pc_fieldsplit_schur_precondit<wbr>ion
user # <-- pressure mass matrix<br>
<br>
-fieldsplit_0_ksp_type preonly<br>
-fieldsplit_0_pc_type ml<br>
<br>
-fieldsplit_1_ksp_type preonly<br>
-fieldsplit_1_pc_type jacobi<br>
</blockquote>
<div><br>
</div>
<div>1) I always recommend starting from an
exact solver and backing off in small steps
for optimization. Thus</div>
<div> I would start with LU on the upper
block and GMRES/LU with toelrance 1e-10 on the
Schur block.</div>
<div> This should converge in 1 iterate.</div>
<div><br>
</div>
<div>2) I don't think you want preonly on the
Schur system. You might want GMRES/Jacobi to
invert the mass matrix.</div>
<div><br>
</div>
<div>3) You probably want to tighten the
tolerance on the Schur solve, at least to
start, and then slowly let it out. The</div>
<div> tight tolerance will show you how
effective the preconditioner is using that
Schur operator. Then you can start</div>
<div> to evaluate how effective the Schur
linear sovler is.</div>
<div><br>
</div>
<div>Does this make sense?</div>
<div><br>
</div>
<div> Thanks,</div>
<div><br>
</div>
<div> Matt</div>
<div> </div>
<blockquote class="gmail_quote"
style="margin:0px 0px 0px
0.8ex;border-left:1px solid
rgb(204,204,204);padding-left:1ex"> In my
present case this setup gives rather slow
convergence (varies for<br>
different geometries between 200-500 or
several thousands!). I obtain<br>
better convergence with
"-pc_fieldsplit_schur_precondi<wbr>tion
selfp"and<br>
using multigrid on S, with
"-fieldsplit_1_pc_type ml" (I don't think<br>
this is optimal, though).<br>
<br>
I don't understand why the pressure mass
matrix approach performs so<br>
poorly and wonder what I could try to improve
the convergence. Until now<br>
I have been using ML and Hypre BoomerAMG
mostly with default parameters.<br>
Surely they can be improved by tuning some
parameters. Which could be a<br>
good starting point? Are there other options I
should consider?<br>
<br>
With the above setup (jacobi) for a case that
works better than others,<br>
the KSP terminates with<br>
467 KSP unpreconditioned resid norm
2.072014323515e-09 true resid norm<br>
2.072014322600e-09 ||r(i)||/||b||
9.939098100674e-09<br>
<br>
You can find the output of -ksp_view below.
Let me know if you need more<br>
details.<br>
<br>
Thanks in advance for your advice!<br>
Best wishes<br>
David<br>
<br>
<br>
KSP Object: 1 MPI processes<br>
type: fgmres<br>
GMRES: restart=30, using Classical
(unmodified) Gram-Schmidt<br>
Orthogonalization with no iterative refinement<br>
GMRES: happy breakdown tolerance 1e-30<br>
maximum iterations=10000<br>
tolerances: relative=1e-08, absolute=1e-50,
divergence=10000.<br>
right preconditioning<br>
diagonally scaled system<br>
using nonzero initial guess<br>
using UNPRECONDITIONED norm type for
convergence test<br>
PC Object: 1 MPI processes<br>
type: fieldsplit<br>
FieldSplit with Schur preconditioner,
factorization UPPER<br>
Preconditioner for the Schur complement
formed from user provided matrix<br>
Split info:<br>
Split number 0 Defined by IS<br>
Split number 1 Defined by IS<br>
KSP solver for A00 block<br>
KSP Object: (fieldsplit_0_) 1
MPI processes<br>
type: preonly<br>
maximum iterations=10000, initial
guess is zero<br>
tolerances: relative=1e-05,
absolute=1e-50, divergence=10000.<br>
left preconditioning<br>
using NONE norm type for convergence
test<br>
PC Object: (fieldsplit_0_) 1
MPI processes<br>
type: ml<br>
MG: type is MULTIPLICATIVE, levels=5
cycles=v<br>
Cycles per PCApply=1<br>
Using Galerkin computed coarse
grid matrices<br>
Coarse grid solver -- level
------------------------------<wbr>-<br>
KSP Object:
(fieldsplit_0_mg_coarse_) 1 MPI<br>
processes<br>
type: preonly<br>
maximum iterations=10000, initial
guess is zero<br>
tolerances: relative=1e-05,
absolute=1e-50, divergence=10000.<br>
left preconditioning<br>
using NONE norm type for
convergence test<br>
PC Object:
(fieldsplit_0_mg_coarse_) 1 MPI<br>
processes<br>
type: lu<br>
LU: out-of-place factorization<br>
tolerance for zero pivot
2.22045e-14<br>
using diagonal shift on blocks
to prevent zero pivot<br>
[INBLOCKS]<br>
matrix ordering: nd<br>
factor fill ratio given 5.,
needed 1.<br>
Factored matrix follows:<br>
Mat Object:
1 MPI processes<br>
type: seqaij<br>
rows=3, cols=3<br>
package used to perform
factorization: petsc<br>
total: nonzeros=3,
allocated nonzeros=3<br>
total number of mallocs
used during MatSetValues<br>
calls =0<br>
not using I-node
routines<br>
linear system matrix = precond
matrix:<br>
Mat Object: 1 MPI
processes<br>
type: seqaij<br>
rows=3, cols=3<br>
total: nonzeros=3, allocated
nonzeros=3<br>
total number of mallocs used
during MatSetValues calls =0<br>
not using I-node routines<br>
Down solver (pre-smoother) on level 1<br>
------------------------------<wbr>-<br>
KSP Object:
(fieldsplit_0_mg_levels_1_) 1<br>
MPI processes<br>
type: richardson<br>
Richardson: damping factor=1.<br>
maximum iterations=2<br>
tolerances: relative=1e-05,
absolute=1e-50, divergence=10000.<br>
left preconditioning<br>
using nonzero initial guess<br>
using NONE norm type for
convergence test<br>
PC Object:
(fieldsplit_0_mg_levels_1_) 1<br>
MPI processes<br>
type: sor<br>
SOR: type = local_symmetric,
iterations = 1, local<br>
iterations = 1, omega = 1.<br>
linear system matrix = precond
matrix:<br>
Mat Object: 1 MPI
processes<br>
type: seqaij<br>
rows=15, cols=15<br>
total: nonzeros=69, allocated
nonzeros=69<br>
total number of mallocs used
during MatSetValues calls =0<br>
not using I-node routines<br>
Up solver (post-smoother) same as down
solver (pre-smoother)<br>
Down solver (pre-smoother) on level 2<br>
------------------------------<wbr>-<br>
KSP Object:
(fieldsplit_0_mg_levels_2_) 1<br>
MPI processes<br>
type: richardson<br>
Richardson: damping factor=1.<br>
maximum iterations=2<br>
tolerances: relative=1e-05,
absolute=1e-50, divergence=10000.<br>
left preconditioning<br>
using nonzero initial guess<br>
using NONE norm type for
convergence test<br>
PC Object:
(fieldsplit_0_mg_levels_2_) 1<br>
MPI processes<br>
type: sor<br>
SOR: type = local_symmetric,
iterations = 1, local<br>
iterations = 1, omega = 1.<br>
linear system matrix = precond
matrix:<br>
Mat Object: 1 MPI
processes<br>
type: seqaij<br>
rows=304, cols=304<br>
total: nonzeros=7354, allocated
nonzeros=7354<br>
total number of mallocs used
during MatSetValues calls =0<br>
not using I-node routines<br>
Up solver (post-smoother) same as down
solver (pre-smoother)<br>
Down solver (pre-smoother) on level 3<br>
------------------------------<wbr>-<br>
KSP Object:
(fieldsplit_0_mg_levels_3_) 1<br>
MPI processes<br>
type: richardson<br>
Richardson: damping factor=1.<br>
maximum iterations=2<br>
tolerances: relative=1e-05,
absolute=1e-50, divergence=10000.<br>
left preconditioning<br>
using nonzero initial guess<br>
using NONE norm type for
convergence test<br>
PC Object:
(fieldsplit_0_mg_levels_3_) 1<br>
MPI processes<br>
type: sor<br>
SOR: type = local_symmetric,
iterations = 1, local<br>
iterations = 1, omega = 1.<br>
linear system matrix = precond
matrix:<br>
Mat Object: 1 MPI
processes<br>
type: seqaij<br>
rows=30236, cols=30236<br>
total: nonzeros=2730644,
allocated nonzeros=2730644<br>
total number of mallocs used
during MatSetValues calls =0<br>
not using I-node routines<br>
Up solver (post-smoother) same as down
solver (pre-smoother)<br>
Down solver (pre-smoother) on level 4<br>
------------------------------<wbr>-<br>
KSP Object:
(fieldsplit_0_mg_levels_4_) 1<br>
MPI processes<br>
type: richardson<br>
Richardson: damping factor=1.<br>
maximum iterations=2<br>
tolerances: relative=1e-05,
absolute=1e-50, divergence=10000.<br>
left preconditioning<br>
using nonzero initial guess<br>
using NONE norm type for
convergence test<br>
PC Object:
(fieldsplit_0_mg_levels_4_) 1<br>
MPI processes<br>
type: sor<br>
SOR: type = local_symmetric,
iterations = 1, local<br>
iterations = 1, omega = 1.<br>
linear system matrix = precond
matrix:<br>
Mat Object:
(fieldsplit_0_) 1 MPI<br>
processes<br>
type: seqaij<br>
rows=894132, cols=894132<br>
total: nonzeros=70684164,
allocated nonzeros=70684164<br>
total number of mallocs used
during MatSetValues calls =0<br>
not using I-node routines<br>
Up solver (post-smoother) same as down
solver (pre-smoother)<br>
linear system matrix = precond matrix:<br>
Mat Object: (fieldsplit_0_)
1 MPI processes<br>
type: seqaij<br>
rows=894132, cols=894132<br>
total: nonzeros=70684164, allocated
nonzeros=70684164<br>
total number of mallocs used during
MatSetValues calls =0<br>
not using I-node routines<br>
KSP solver for S = A11 - A10 inv(A00) A01<br>
KSP Object: (fieldsplit_1_) 1
MPI processes<br>
type: preonly<br>
maximum iterations=10000, initial
guess is zero<br>
tolerances: relative=1e-05,
absolute=1e-50, divergence=10000.<br>
left preconditioning<br>
using NONE norm type for convergence
test<br>
PC Object: (fieldsplit_1_) 1
MPI processes<br>
type: jacobi<br>
linear system matrix followed by
preconditioner matrix:<br>
Mat Object: (fieldsplit_1_)
1 MPI processes<br>
type: schurcomplement<br>
rows=42025, cols=42025<br>
Schur complement A11 - A10
inv(A00) A01<br>
A11<br>
Mat Object:
(fieldsplit_1_) 1<br>
MPI processes<br>
type: seqaij<br>
rows=42025, cols=42025<br>
total: nonzeros=554063,
allocated nonzeros=554063<br>
total number of mallocs used
during MatSetValues calls =0<br>
not using I-node routines<br>
A10<br>
Mat Object: 1 MPI
processes<br>
type: seqaij<br>
rows=42025, cols=894132<br>
total: nonzeros=6850107,
allocated nonzeros=6850107<br>
total number of mallocs used
during MatSetValues calls =0<br>
not using I-node routines<br>
KSP of A00<br>
KSP Object:
(fieldsplit_0_) 1<br>
MPI processes<br>
type: preonly<br>
maximum iterations=10000,
initial guess is zero<br>
tolerances: relative=1e-05,
absolute=1e-50,<br>
divergence=10000.<br>
left preconditioning<br>
using NONE norm type for
convergence test<br>
PC Object:
(fieldsplit_0_) 1<br>
MPI processes<br>
type: ml<br>
MG: type is MULTIPLICATIVE,
levels=5 cycles=v<br>
Cycles per PCApply=1<br>
Using Galerkin computed
coarse grid matrices<br>
Coarse grid solver -- level
------------------------------<wbr>-<br>
KSP Object:<br>
(fieldsplit_0_mg_coarse_) 1
MPI processes<br>
type: preonly<br>
maximum iterations=10000,
initial guess is zero<br>
tolerances:
relative=1e-05, absolute=1e-50,<br>
divergence=10000.<br>
left preconditioning<br>
using NONE norm type for
convergence test<br>
PC Object:<br>
(fieldsplit_0_mg_coarse_) 1
MPI processes<br>
type: lu<br>
LU: out-of-place
factorization<br>
tolerance for zero pivot
2.22045e-14<br>
using diagonal shift on
blocks to prevent zero<br>
pivot [INBLOCKS]<br>
matrix ordering: nd<br>
factor fill ratio given
5., needed 1.<br>
Factored matrix
follows:<br>
Mat Object:
1 MPI<br>
processes<br>
type: seqaij<br>
rows=3, cols=3<br>
package used to
perform factorization: petsc<br>
total: nonzeros=3,
allocated nonzeros=3<br>
total number of
mallocs used during<br>
MatSetValues calls =0<br>
not using I-node
routines<br>
linear system matrix =
precond matrix:<br>
Mat Object:
1 MPI processes<br>
type: seqaij<br>
rows=3, cols=3<br>
total: nonzeros=3,
allocated nonzeros=3<br>
total number of mallocs
used during MatSetValues<br>
calls =0<br>
not using I-node
routines<br>
Down solver (pre-smoother) on
level 1<br>
------------------------------<wbr>-<br>
KSP Object:<br>
(fieldsplit_0_mg_levels_1_)
1 MPI processes<br>
type: richardson<br>
Richardson: damping
factor=1.<br>
maximum iterations=2<br>
tolerances:
relative=1e-05, absolute=1e-50,<br>
divergence=10000.<br>
left preconditioning<br>
using nonzero initial
guess<br>
using NONE norm type for
convergence test<br>
PC Object:<br>
(fieldsplit_0_mg_levels_1_)
1 MPI processes<br>
type: sor<br>
SOR: type =
local_symmetric, iterations = 1, local<br>
iterations = 1, omega = 1.<br>
linear system matrix =
precond matrix:<br>
Mat Object:
1 MPI processes<br>
type: seqaij<br>
rows=15, cols=15<br>
total: nonzeros=69,
allocated nonzeros=69<br>
total number of mallocs
used during MatSetValues<br>
calls =0<br>
not using I-node
routines<br>
Up solver (post-smoother) same
as down solver (pre-smoother)<br>
Down solver (pre-smoother) on
level 2<br>
------------------------------<wbr>-<br>
KSP Object:<br>
(fieldsplit_0_mg_levels_2_)
1 MPI processes<br>
type: richardson<br>
Richardson: damping
factor=1.<br>
maximum iterations=2<br>
tolerances:
relative=1e-05, absolute=1e-50,<br>
divergence=10000.<br>
left preconditioning<br>
using nonzero initial
guess<br>
using NONE norm type for
convergence test<br>
PC Object:<br>
(fieldsplit_0_mg_levels_2_)
1 MPI processes<br>
type: sor<br>
SOR: type =
local_symmetric, iterations = 1, local<br>
iterations = 1, omega = 1.<br>
linear system matrix =
precond matrix:<br>
Mat Object:
1 MPI processes<br>
type: seqaij<br>
rows=304, cols=304<br>
total: nonzeros=7354,
allocated nonzeros=7354<br>
total number of mallocs
used during MatSetValues<br>
calls =0<br>
not using I-node
routines<br>
Up solver (post-smoother) same
as down solver (pre-smoother)<br>
Down solver (pre-smoother) on
level 3<br>
------------------------------<wbr>-<br>
KSP Object:<br>
(fieldsplit_0_mg_levels_3_)
1 MPI processes<br>
type: richardson<br>
Richardson: damping
factor=1.<br>
maximum iterations=2<br>
tolerances:
relative=1e-05, absolute=1e-50,<br>
divergence=10000.<br>
left preconditioning<br>
using nonzero initial
guess<br>
using NONE norm type for
convergence test<br>
PC Object:<br>
(fieldsplit_0_mg_levels_3_)
1 MPI processes<br>
type: sor<br>
SOR: type =
local_symmetric, iterations = 1, local<br>
iterations = 1, omega = 1.<br>
linear system matrix =
precond matrix:<br>
Mat Object:
1 MPI processes<br>
type: seqaij<br>
rows=30236, cols=30236<br>
total: nonzeros=2730644,
allocated nonzeros=2730644<br>
total number of mallocs
used during MatSetValues<br>
calls =0<br>
not using I-node
routines<br>
Up solver (post-smoother) same
as down solver (pre-smoother)<br>
Down solver (pre-smoother) on
level 4<br>
------------------------------<wbr>-<br>
KSP Object:<br>
(fieldsplit_0_mg_levels_4_)
1 MPI processes<br>
type: richardson<br>
Richardson: damping
factor=1.<br>
maximum iterations=2<br>
tolerances:
relative=1e-05, absolute=1e-50,<br>
divergence=10000.<br>
left preconditioning<br>
using nonzero initial
guess<br>
using NONE norm type for
convergence test<br>
PC Object:<br>
(fieldsplit_0_mg_levels_4_)
1 MPI processes<br>
type: sor<br>
SOR: type =
local_symmetric, iterations = 1, local<br>
iterations = 1, omega = 1.<br>
linear system matrix =
precond matrix:<br>
Mat Object:<br>
(fieldsplit_0_) 1 MPI
processes<br>
type: seqaij<br>
rows=894132, cols=894132<br>
total:
nonzeros=70684164, allocated nonzeros=70684164<br>
total number of mallocs
used during MatSetValues<br>
calls =0<br>
not using I-node
routines<br>
Up solver (post-smoother) same
as down solver (pre-smoother)<br>
linear system matrix = precond
matrix:<br>
Mat Object:<br>
(fieldsplit_0_) 1 MPI
processes<br>
type: seqaij<br>
rows=894132, cols=894132<br>
total: nonzeros=70684164,
allocated nonzeros=70684164<br>
total number of mallocs used
during MatSetValues calls =0<br>
not using I-node routines<br>
A01<br>
Mat Object: 1 MPI
processes<br>
type: seqaij<br>
rows=894132, cols=42025<br>
total: nonzeros=6850107,
allocated nonzeros=6850107<br>
total number of mallocs used
during MatSetValues calls =0<br>
not using I-node routines<br>
Mat Object: 1 MPI processes<br>
type: seqaij<br>
rows=42025, cols=42025<br>
total: nonzeros=554063, allocated
nonzeros=554063<br>
total number of mallocs used during
MatSetValues calls =0<br>
not using I-node routines<br>
linear system matrix = precond matrix:<br>
Mat Object: 1 MPI processes<br>
type: seqaij<br>
rows=936157, cols=936157<br>
total: nonzeros=84938441, allocated
nonzeros=84938441<br>
total number of mallocs used during
MatSetValues calls =0<br>
not using I-node routines<br>
<br>
<br>
</blockquote>
</div>
<br>
<br clear="all">
<span class="gmail-HOEnZb"><font color="#888888">
<div><br>
</div>
-- <br>
<div
class="gmail-m_5328507656823621836gmail_signature">
<div dir="ltr">
<div>What most experimenters take for
granted before they begin their
experiments is infinitely more
interesting than any results to which
their experiments lead.<br>
-- Norbert Wiener</div>
<div><br>
</div>
<div><a moz-do-not-send="true"
href="http://www.caam.rice.edu/%7Emk51/"
target="_blank">http://www.caam.rice.edu/~<wbr>mk51/</a><br>
</div>
</div>
</div>
</font></span></div>
</div>
</blockquote>
<br>
</div>
</blockquote>
</div>
<br>
<br clear="all">
<div><br>
</div>
-- <br>
<div class="gmail_signature">
<div dir="ltr">
<div>What most experimenters take for granted before they
begin their experiments is infinitely more interesting
than any results to which their experiments lead.<br>
-- Norbert Wiener</div>
<div><br>
</div>
<div><a moz-do-not-send="true"
href="http://www.caam.rice.edu/%7Emk51/"
target="_blank">http://www.caam.rice.edu/~mk51/</a><br>
</div>
</div>
</div>
</div>
</div>
</blockquote>
<br>
</body>
</html>