On Mon, Dec 12, 2011 at 5:35 AM, Alexander Grayver <span dir="ltr"><<a href="mailto:agrayver@gfz-potsdam.de">agrayver@gfz-potsdam.de</a>></span> wrote:<br><div class="gmail_quote"><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">
<u></u>
<div bgcolor="#ffffff" text="#000000">
Hello,<br>
<br>
I use PETSs with MUMPS and looking carefully at the -ksp_view
-ksp_monitor results I see:<br>
<br>
KSP Object:(fwd_) 64 MPI processes<br>
type: preonly<br>
maximum iterations=10000, initial guess is zero<br>
tolerances: relative=1e-05, absolute=1e-50, divergence=10000<br>
left preconditioning<br>
using NONE norm type for convergence test<br>
PC Object:(fwd_) 64 MPI processes<br>
type: cholesky<br>
Cholesky: out-of-place factorization<br>
tolerance for zero pivot 2.22045e-14<br>
matrix ordering: natural<br>
factor fill ratio given 0, needed 0<br>
Factored matrix follows:<br>
Matrix Object: 64 MPI processes<br>
type: mpiaij<br>
rows=1048944, cols=1048944<br>
package used to perform factorization: mumps<br>
total: nonzeros=1266866685, allocated nonzeros=1266866685<br>
total number of mallocs used during MatSetValues calls =0<br>
MUMPS run parameters:<br>
SYM (matrix type): 1 <br>
PAR (host participation): 1 <br>
ICNTL(1) (output for error): 6 <br>
ICNTL(2) (output of diagnostic msg): 0 <br>
ICNTL(3) (output for global info): 0 <br>
ICNTL(4) (level of printing): 0 <br>
ICNTL(5) (input mat struct): 0 <br>
ICNTL(6) (matrix prescaling): 0 <br>
ICNTL(7) (sequentia matrix ordering):5 <br>
ICNTL(8) (scalling strategy): 77 <br>
ICNTL(10) (max num of refinements): 0 <br>
ICNTL(11) (error analysis): 0 <br>
ICNTL(12) (efficiency
control): 1 <br>
ICNTL(13) (efficiency
control): 0 <br>
ICNTL(14) (percentage of estimated workspace
increase): 30 <br>
ICNTL(18) (input mat
struct): 3 <br>
ICNTL(19) (Shur complement
info): 0 <br>
ICNTL(20) (rhs sparse
pattern): 0 <br>
ICNTL(21) (solution
struct): 1 <br>
ICNTL(22) (in-core/out-of-core
facility): 0 <br>
ICNTL(23) (max size of memory can be allocated
locally):0 <br>
ICNTL(24) (detection of null pivot
rows): 0 <br>
ICNTL(25) (computation of a null space
basis): 0 <br>
ICNTL(26) (Schur options for rhs or
solution): 0 <br>
ICNTL(27) (experimental
parameter): -8 <br>
ICNTL(28) (use parallel or sequential
ordering): 2 <br>
ICNTL(29) (parallel
ordering): 0 <br>
ICNTL(30) (user-specified set of entries in
inv(A)): 0 <br>
ICNTL(31) (factors is discarded in the solve
phase): 0 <br>
ICNTL(33) (compute
determinant): 0 <br>
...<br>
linear system matrix = precond matrix:<br>
Matrix Object: 64 MPI processes<br>
type: mpiaij<br>
rows=1048944, cols=1048944<br>
<b> total: nonzeros=7251312, allocated nonzeros=11554449<br>
total number of mallocs used during MatSetValues calls =1071</b><br>
not using I-node (on process 0) routines<br>
<br>
The particularly interesting part are last 3 lines.<br>
Where do these mallocs come from? Is it possible to reduce this
number?<br></div></blockquote><div><br></div><div>Yes, it looks like you are not preallocating correctly.</div><div><br></div><div> Matt</div><div> </div><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">
<div bgcolor="#ffffff" text="#000000">
Regards,<br>
Alexander<br>
<br>
</div>
</blockquote></div><br><br clear="all"><div><br></div>-- <br>What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead.<br>
-- Norbert Wiener<br>