[petsc-dev] Seeming performance regression with GAMG
Lawrence Mitchell
lawrence.mitchell at imperial.ac.uk
Mon Apr 27 10:06:30 CDT 2015
-----BEGIN PGP SIGNED MESSAGE-----
Hash: SHA1
Dear all,
we recently noticed a slowdown when using GAMG that I'm trying to
track down in a little more detail. I'm solving an Hdiv-L2
"helmholtz" pressure correction using a schur complement. I
precondition the schur complement with 'selfp', which morally looks
like a normal Helmholtz operator (except in the DG space). The domain
is very anisotropic (a thin atmospheric shell), so getting round to
trying Toby's column-based coarsening plugin is on the horizon but I
haven't done it yet.
I don't have a good feel for exactly when things go worse, but here
are two data points:
A recentish master (e4b003c), and master from 26th Feb (30ab49e4). I
notice in the former that MatPtAP takes significantly longer (full
logs below), different coarsening maybe? As a point of comparison,
the PCSetup for Hypre takes ballpark half a second on the same operator.
I test with KSP ex6 (with a constant RHS):
Any ideas?
Cheers,
Lawrence
$ ./ex6-e4b003c -f helmholtz-sphere.dat -ksp_type cg
- -ksp_convergence_test skip -ksp_max_it 2 -ksp_monitor -table
- -pc_type gamg -log_summary -ksp_view
0 KSP Residual norm 3.676132751311e-11
1 KSP Residual norm 1.764616084171e-14
2 KSP Residual norm 9.253867842133e-14
KSP Object: 1 MPI processes
type: cg
maximum iterations=2, initial guess is zero
tolerances: relative=1e-05, absolute=1e-50, divergence=10000
left preconditioning
using PRECONDITIONED norm type for convergence test
PC Object: 1 MPI processes
type: gamg
MG: type is MULTIPLICATIVE, levels=5 cycles=v
Cycles per PCApply=1
Using Galerkin computed coarse grid matrices
GAMG specific options
Threshold for dropping small values from graph 0
AGG specific options
Symmetric graph false
Coarse grid solver -- level -------------------------------
KSP Object: (mg_coarse_) 1 MPI processes
type: gmres
GMRES: restart=30, using Classical (unmodified) Gram-Schmidt
Orthogonalization with no iterative refinement
GMRES: happy breakdown tolerance 1e-30
maximum iterations=1, initial guess is zero
tolerances: relative=1e-05, absolute=1e-50, divergence=10000
left preconditioning
using NONE norm type for convergence test
PC Object: (mg_coarse_) 1 MPI processes
type: bjacobi
block Jacobi: number of blocks = 1
Local solve is same for all blocks, in the following KSP and
PC objects:
KSP Object: (mg_coarse_sub_) 1 MPI processes
type: preonly
maximum iterations=1, initial guess is zero
tolerances: relative=1e-05, absolute=1e-50, divergence=10000
left preconditioning
using NONE norm type for convergence test
PC Object: (mg_coarse_sub_) 1 MPI processes
type: lu
LU: out-of-place factorization
tolerance for zero pivot 2.22045e-14
using diagonal shift on blocks to prevent zero pivot
[INBLOCKS]
matrix ordering: nd
factor fill ratio given 5, needed 1
Factored matrix follows:
Mat Object: 1 MPI processes
type: seqaij
rows=3, cols=3
package used to perform factorization: petsc
total: nonzeros=9, allocated nonzeros=9
total number of mallocs used during MatSetValues
calls =0
using I-node routines: found 1 nodes, limit used is 5
linear system matrix = precond matrix:
Mat Object: 1 MPI processes
type: seqaij
rows=3, cols=3
total: nonzeros=9, allocated nonzeros=9
total number of mallocs used during MatSetValues calls =0
using I-node routines: found 1 nodes, limit used is 5
linear system matrix = precond matrix:
Mat Object: 1 MPI processes
type: seqaij
rows=3, cols=3
total: nonzeros=9, allocated nonzeros=9
total number of mallocs used during MatSetValues calls =0
using I-node routines: found 1 nodes, limit used is 5
Down solver (pre-smoother) on level 1 -------------------------------
KSP Object: (mg_levels_1_) 1 MPI processes
type: chebyshev
Chebyshev: eigenvalue estimates: min = 0.0999929, max = 1.09992
Chebyshev: eigenvalues estimated using gmres with translations
[0 0.1; 0 1.1]
KSP Object: (mg_levels_1_esteig_) 1 MPI processes
type: gmres
GMRES: restart=30, using Classical (unmodified)
Gram-Schmidt Orthogonalization with no iterative refinement
GMRES: happy breakdown tolerance 1e-30
maximum iterations=10
tolerances: relative=1e-05, absolute=1e-50, divergence=10000
left preconditioning
using nonzero initial guess
using NONE norm type for convergence test
maximum iterations=2
tolerances: relative=1e-05, absolute=1e-50, divergence=10000
left preconditioning
using nonzero initial guess
using NONE norm type for convergence test
PC Object: (mg_levels_1_) 1 MPI processes
type: sor
SOR: type = local_symmetric, iterations = 1, local iterations
= 1, omega = 1
linear system matrix = precond matrix:
Mat Object: 1 MPI processes
type: seqaij
rows=93, cols=93
total: nonzeros=8649, allocated nonzeros=8649
total number of mallocs used during MatSetValues calls =0
using I-node routines: found 19 nodes, limit used is 5
Up solver (post-smoother) same as down solver (pre-smoother)
Down solver (pre-smoother) on level 2 -------------------------------
KSP Object: (mg_levels_2_) 1 MPI processes
type: chebyshev
Chebyshev: eigenvalue estimates: min = 0.0998389, max = 1.09823
Chebyshev: eigenvalues estimated using gmres with translations
[0 0.1; 0 1.1]
KSP Object: (mg_levels_2_esteig_) 1 MPI processes
type: gmres
GMRES: restart=30, using Classical (unmodified)
Gram-Schmidt Orthogonalization with no iterative refinement
GMRES: happy breakdown tolerance 1e-30
maximum iterations=10
tolerances: relative=1e-05, absolute=1e-50, divergence=10000
left preconditioning
using nonzero initial guess
using NONE norm type for convergence test
maximum iterations=2
tolerances: relative=1e-05, absolute=1e-50, divergence=10000
left preconditioning
using nonzero initial guess
using NONE norm type for convergence test
PC Object: (mg_levels_2_) 1 MPI processes
type: sor
SOR: type = local_symmetric, iterations = 1, local iterations
= 1, omega = 1
linear system matrix = precond matrix:
Mat Object: 1 MPI processes
type: seqaij
rows=2991, cols=2991
total: nonzeros=8.94608e+06, allocated nonzeros=8.94608e+06
total number of mallocs used during MatSetValues calls =0
using I-node routines: found 599 nodes, limit used is 5
Up solver (post-smoother) same as down solver (pre-smoother)
Down solver (pre-smoother) on level 3 -------------------------------
KSP Object: (mg_levels_3_) 1 MPI processes
type: chebyshev
Chebyshev: eigenvalue estimates: min = 0.0998975, max = 1.09887
Chebyshev: eigenvalues estimated using gmres with translations
[0 0.1; 0 1.1]
KSP Object: (mg_levels_3_esteig_) 1 MPI processes
type: gmres
GMRES: restart=30, using Classical (unmodified)
Gram-Schmidt Orthogonalization with no iterative refinement
GMRES: happy breakdown tolerance 1e-30
maximum iterations=10
tolerances: relative=1e-05, absolute=1e-50, divergence=10000
left preconditioning
using nonzero initial guess
using NONE norm type for convergence test
maximum iterations=2
tolerances: relative=1e-05, absolute=1e-50, divergence=10000
left preconditioning
using nonzero initial guess
using NONE norm type for convergence test
PC Object: (mg_levels_3_) 1 MPI processes
type: sor
SOR: type = local_symmetric, iterations = 1, local iterations
= 1, omega = 1
linear system matrix = precond matrix:
Mat Object: 1 MPI processes
type: seqaij
rows=35419, cols=35419
total: nonzeros=1.55936e+07, allocated nonzeros=1.55936e+07
total number of mallocs used during MatSetValues calls =1
not using I-node routines
Up solver (post-smoother) same as down solver (pre-smoother)
Down solver (pre-smoother) on level 4 -------------------------------
KSP Object: (mg_levels_4_) 1 MPI processes
type: chebyshev
Chebyshev: eigenvalue estimates: min = 0.1, max = 1.1
Chebyshev: eigenvalues estimated using gmres with translations
[0 0.1; 0 1.1]
KSP Object: (mg_levels_4_esteig_) 1 MPI processes
type: gmres
GMRES: restart=30, using Classical (unmodified)
Gram-Schmidt Orthogonalization with no iterative refinement
GMRES: happy breakdown tolerance 1e-30
maximum iterations=10
tolerances: relative=1e-05, absolute=1e-50, divergence=10000
left preconditioning
using nonzero initial guess
using NONE norm type for convergence test
maximum iterations=2
tolerances: relative=1e-05, absolute=1e-50, divergence=10000
left preconditioning
using nonzero initial guess
using NONE norm type for convergence test
PC Object: (mg_levels_4_) 1 MPI processes
type: sor
SOR: type = local_symmetric, iterations = 1, local iterations
= 1, omega = 1
linear system matrix = precond matrix:
Mat Object: 1 MPI processes
type: seqaij
rows=327680, cols=327680
total: nonzeros=3.25828e+06, allocated nonzeros=3.25828e+06
total number of mallocs used during MatSetValues calls =0
not using I-node routines
Up solver (post-smoother) same as down solver (pre-smoother)
linear system matrix = precond matrix:
Mat Object: 1 MPI processes
type: seqaij
rows=327680, cols=327680
total: nonzeros=3.25828e+06, allocated nonzeros=3.25828e+06
total number of mallocs used during MatSetValues calls =0
not using I-node routines
helmholt 2 9e+03 gamg
************************************************************************************************************************
*** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r
- -fCourier9' to print this document ***
************************************************************************************************************************
- ---------------------------------------------- PETSc Performance
Summary: ----------------------------------------------
./ex6-master on a arch-linux2-c-opt named yam.doc.ic.ac.uk with 1
processor, by lmitche1 Mon Apr 27 16:03:36 2015
Using Petsc Development GIT revision: v3.5.3-2602-ga9b180a GIT Date:
2015-04-07 20:34:49 -0500
Max Max/Min Avg Total
Time (sec): 1.072e+02 1.00000 1.072e+02
Objects: 2.620e+02 1.00000 2.620e+02
Flops: 4.582e+10 1.00000 4.582e+10 4.582e+10
Flops/sec: 4.275e+08 1.00000 4.275e+08 4.275e+08
MPI Messages: 0.000e+00 0.00000 0.000e+00 0.000e+00
MPI Message Lengths: 0.000e+00 0.00000 0.000e+00 0.000e+00
MPI Reductions: 0.000e+00 0.00000
Flop counting convention: 1 flop = 1 real number operation of type
(multiply/divide/add/subtract)
e.g., VecAXPY() for real vectors of length
N --> 2N flops
and VecAXPY() for complex vectors of
length N --> 8N flops
Summary of Stages: ----- Time ------ ----- Flops ----- ---
Messages --- -- Message Lengths -- -- Reductions --
Avg %Total Avg %Total counts
%Total Avg %Total counts %Total
0: Main Stage: 1.0898e-01 0.1% 7.4996e+06 0.0% 0.000e+00
0.0% 0.000e+00 0.0% 0.000e+00 0.0%
1: mystage 1: 1.0466e+02 97.6% 4.2348e+10 92.4% 0.000e+00
0.0% 0.000e+00 0.0% 0.000e+00 0.0%
2: mystage 2: 2.4395e+00 2.3% 3.4689e+09 7.6% 0.000e+00
0.0% 0.000e+00 0.0% 0.000e+00 0.0%
-
------------------------------------------------------------------------------------------------------------------------
See the 'Profiling' chapter of the users' manual for details on
interpreting output.
Phase summary info:
Count: number of times phase was executed
Time and Flops: Max - maximum over all processors
Ratio - ratio of maximum to minimum over all processors
Mess: number of messages sent
Avg. len: average message length (bytes)
Reduct: number of global reductions
Global: entire computation
Stage: stages of a computation. Set stages with PetscLogStagePush()
and PetscLogStagePop().
%T - percent time in this phase %F - percent flops in
this phase
%M - percent messages in this phase %L - percent message
lengths in this phase
%R - percent reductions in this phase
Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time
over all processors)
-
------------------------------------------------------------------------------------------------------------------------
Event Count Time (sec) Flops
--- Global --- --- Stage --- Total
Max Ratio Max Ratio Max Ratio Mess Avg
len Reduct %T %F %M %L %R %T %F %M %L %R Mflop/s
-
------------------------------------------------------------------------------------------------------------------------
- --- Event Stage 0: Main Stage
ThreadCommRunKer 2 1.0 1.9073e-06 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatMult 1 1.0 4.4990e-03 1.0 6.19e+06 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 4 83 0 0 0 1376
MatAssemblyBegin 1 1.0 1.1921e-06 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatAssemblyEnd 1 1.0 1.0468e-02 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 10 0 0 0 0 0
MatLoad 1 1.0 9.3672e-02 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 86 0 0 0 0 0
VecNorm 1 1.0 9.1791e-05 1.0 6.55e+05 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 9 0 0 0 7140
VecSet 5 1.0 3.1860e-03 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 3 0 0 0 0 0
VecAXPY 1 1.0 3.9697e-04 1.0 6.55e+05 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 9 0 0 0 1651
- --- Event Stage 1: mystage 1
MatMult 40 1.0 3.5990e-01 1.0 5.52e+08 1.0 0.0e+00
0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 1535
MatConvert 4 1.0 1.2038e-01 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatScale 12 1.0 8.2839e-02 1.0 7.21e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 870
MatAssemblyBegin 31 1.0 1.4067e-05 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatAssemblyEnd 31 1.0 1.0995e-01 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatGetRow 1464732 1.0 8.5680e-02 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatCoarsen 4 1.0 2.4768e-02 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatAXPY 4 1.0 8.1362e-02 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatMatMult 4 1.0 6.0399e-01 1.0 6.39e+07 1.0 0.0e+00
0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 106
MatMatMultSym 4 1.0 4.4536e-01 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatMatMultNum 4 1.0 1.5859e-01 1.0 6.39e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 403
MatPtAP 4 1.0 1.0255e+02 1.0 4.15e+10 1.0 0.0e+00
0.0e+00 0.0e+00 96 91 0 0 0 98 98 0 0 0 405
MatPtAPSymbolic 4 1.0 6.1707e+01 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 58 0 0 0 0 59 0 0 0 0 0
MatPtAPNumeric 4 1.0 4.0840e+01 1.0 4.15e+10 1.0 0.0e+00
0.0e+00 0.0e+00 38 91 0 0 0 39 98 0 0 0 1017
MatTrnMatMult 1 1.0 1.9648e-01 1.0 2.34e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 119
MatTrnMatMultSym 1 1.0 1.3640e-01 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatTrnMatMultNum 1 1.0 6.0079e-02 1.0 2.34e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 389
MatGetSymTrans 5 1.0 8.4374e-02 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecMDot 40 1.0 1.5124e-02 1.0 4.03e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2663
VecNorm 44 1.0 1.1821e-03 1.0 8.06e+06 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 6815
VecScale 44 1.0 1.4737e-03 1.0 4.03e+06 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2733
VecCopy 4 1.0 3.2711e-04 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecSet 143 1.0 2.7236e-02 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecAXPY 4 1.0 3.9601e-04 1.0 7.32e+05 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1849
VecMAXPY 44 1.0 1.9676e-02 1.0 4.76e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2419
VecAssemblyBegin 4 1.0 1.9073e-06 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecAssemblyEnd 4 1.0 0.0000e+00 0.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecPointwiseMult 44 1.0 7.9026e-03 1.0 4.03e+06 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 510
VecSetRandom 4 1.0 3.6092e-03 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecNormalize 44 1.0 2.6934e-03 1.0 1.21e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 4486
KSPGMRESOrthog 40 1.0 3.1765e-02 1.0 8.06e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2536
KSPSetUp 10 1.0 1.1930e-02 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
PCGAMGGraph_AGG 4 1.0 5.7353e-01 1.0 5.56e+07 1.0 0.0e+00
0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 97
PCGAMGCoarse_AGG 4 1.0 2.5846e-01 1.0 2.34e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 90
PCGAMGProl_AGG 4 1.0 4.9806e-02 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
PCGAMGPOpt_AGG 4 1.0 1.2128e+00 1.0 7.38e+08 1.0 0.0e+00
0.0e+00 0.0e+00 1 2 0 0 0 1 2 0 0 0 608
GAMG: createProl 4 1.0 2.0973e+00 1.0 8.17e+08 1.0 0.0e+00
0.0e+00 0.0e+00 2 2 0 0 0 2 2 0 0 0 389
Graph 8 1.0 5.7211e-01 1.0 5.56e+07 1.0 0.0e+00
0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 97
MIS/Agg 4 1.0 2.4861e-02 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
SA: col data 4 1.0 7.6509e-04 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
SA: frmProl0 4 1.0 4.6539e-02 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
SA: smooth 4 1.0 1.2128e+00 1.0 7.38e+08 1.0 0.0e+00
0.0e+00 0.0e+00 1 2 0 0 0 1 2 0 0 0 608
GAMG: partLevel 4 1.0 1.0255e+02 1.0 4.15e+10 1.0 0.0e+00
0.0e+00 0.0e+00 96 91 0 0 0 98 98 0 0 0 405
PCSetUp 1 1.0 1.0465e+02 1.0 4.23e+10 1.0 0.0e+00
0.0e+00 0.0e+00 98 92 0 0 0 100100 0 0 0 405
- --- Event Stage 2: mystage 2
MatMult 121 1.0 1.0144e+00 1.0 1.61e+09 1.0 0.0e+00
0.0e+00 0.0e+00 1 4 0 0 0 42 47 0 0 0 1592
MatMultAdd 12 1.0 3.1757e-02 1.0 4.95e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 1 1 0 0 0 1558
MatMultTranspose 12 1.0 3.7137e-02 1.0 4.95e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 2 1 0 0 0 1332
MatSolve 6 1.0 9.2983e-06 1.0 9.00e+01 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 10
MatSOR 116 1.0 1.2805e+00 1.0 1.61e+09 1.0 0.0e+00
0.0e+00 0.0e+00 1 4 0 0 0 52 47 0 0 0 1260
MatLUFactorSym 1 1.0 1.5020e-05 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatLUFactorNum 1 1.0 5.9605e-06 1.0 1.60e+01 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 3
MatResidual 12 1.0 1.0432e-01 1.0 1.67e+08 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 4 5 0 0 0 1599
MatGetRowIJ 1 1.0 2.1458e-06 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatGetOrdering 1 1.0 4.2915e-05 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatView 8 1.0 6.7115e-04 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecMDot 43 1.0 1.3515e-02 1.0 4.03e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 1 1 0 0 0 2980
VecTDot 4 1.0 1.1048e-03 1.0 2.62e+06 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2373
VecNorm 53 1.0 1.4720e-03 1.0 1.00e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 6809
VecScale 50 1.0 1.4396e-03 1.0 4.03e+06 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2798
VecCopy 21 1.0 3.6387e-03 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecSet 108 1.0 1.0684e-02 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecAXPY 15 1.0 2.1303e-03 1.0 4.09e+06 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1918
VecAYPX 97 1.0 1.1820e-02 1.0 1.16e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 985
VecAXPBYCZ 48 1.0 8.2519e-03 1.0 2.20e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 1 0 0 0 2663
VecMAXPY 50 1.0 1.6957e-02 1.0 4.76e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 1 1 0 0 0 2807
VecNormalize 50 1.0 2.6865e-03 1.0 1.21e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 4498
KSPGMRESOrthog 43 1.0 2.7892e-02 1.0 8.06e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 1 2 0 0 0 2888
KSPSetUp 5 1.0 2.9690e-03 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
KSPSolve 1 1.0 2.4374e+00 1.0 3.47e+09 1.0 0.0e+00
0.0e+00 0.0e+00 2 8 0 0 0 100100 0 0 0 1423
PCSetUp 1 1.0 9.3937e-05 1.0 1.60e+01 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
PCSetUpOnBlocks 3 1.0 9.6798e-05 1.0 1.60e+01 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
PCApply 3 1.0 2.4240e+00 1.0 3.45e+09 1.0 0.0e+00
0.0e+00 0.0e+00 2 8 0 0 0 99 99 0 0 0 1423
-
------------------------------------------------------------------------------------------------------------------------
Memory usage is given in bytes:
Object Type Creations Destructions Memory Descendants'
Mem.
Reports information only for process 0.
- --- Event Stage 0: Main Stage
Viewer 2 2 1520 0
Matrix 1 10 54702412 0
Vector 3 97 72154440 0
Krylov Solver 0 11 146840 0
Preconditioner 0 7 7332 0
Index Set 0 3 2400 0
- --- Event Stage 1: mystage 1
Viewer 1 0 0 0
Matrix 22 14 691989068 0
Matrix Coarsen 4 4 2576 0
Vector 125 91 67210200 0
Krylov Solver 15 4 120864 0
Preconditioner 15 8 7520 0
Index Set 4 4 3168 0
PetscRandom 4 4 2560 0
- --- Event Stage 2: mystage 2
Matrix 1 0 0 0
Vector 60 0 0 0
Index Set 5 2 1592 0
========================================================================================================================
Average time to get PetscTime(): 0
#PETSc Option Table entries:
- -f helmholtz-sphere.dat
- -ksp_convergence_test skip
- -ksp_max_it 2
- -ksp_monitor
- -ksp_type cg
- -ksp_view
- -log_summary
- -matload_block_size 1
- -pc_type gamg
- -table
#End of PETSc Option Table entries
Compiled without FORTRAN kernels
Compiled with full precision matrices (default)
sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8
sizeof(PetscScalar) 8 sizeof(PetscInt) 4
Configure options: --download-chaco=1 --download-ctetgen=1
- --download-exodusii=1 --download-hdf5=1 --download-hypre=1
- --download-metis=1 --download-ml=1 --download-mumps=1
- --download-netcdf=1 --download-parmetis=1 --download-ptscotch=1
- --download-scalapack=1 --download-superlu=1 --download-superlu_dist=1
- --download-triangle=1 --with-c2html=0 --with-debugging=0
- --with-make-np=32 --with-openmp=0 --with-pthreadclasses=0
- --with-shared-libraries=1 --with-threadcomm=0 PETSC_ARCH=arch-linux2-c-opt
- -----------------------------------------
Libraries compiled on Wed Apr 8 10:00:43 2015 on yam.doc.ic.ac.uk
Machine characteristics:
Linux-3.13.0-45-generic-x86_64-with-Ubuntu-14.04-trusty
Using PETSc directory: /data/lmitche1/src/deps/petsc
Using PETSc arch: arch-linux2-c-opt
- -----------------------------------------
Using C compiler: mpicc -fPIC -Wall -Wwrite-strings
- -Wno-strict-aliasing -Wno-unknown-pragmas -O ${COPTFLAGS} ${CFLAGS}
Using Fortran compiler: mpif90 -fPIC -Wall -Wno-unused-variable
- -ffree-line-length-0 -Wno-unused-dummy-argument -O ${FOPTFLAGS}
${FFLAGS}
- -----------------------------------------
Using include paths:
- -I/data/lmitche1/src/deps/petsc/arch-linux2-c-opt/include
- -I/data/lmitche1/src/deps/petsc/include
- -I/data/lmitche1/src/deps/petsc/include
- -I/data/lmitche1/src/deps/petsc/arch-linux2-c-opt/include
- -I/usr/lib/openmpi/include -I/usr/lib/openmpi/include/openmpi
- -----------------------------------------
Using C linker: mpicc
Using Fortran linker: mpif90
Using libraries:
- -Wl,-rpath,/data/lmitche1/src/deps/petsc/arch-linux2-c-opt/lib
- -L/data/lmitche1/src/deps/petsc/arch-linux2-c-opt/lib -lpetsc
- -Wl,-rpath,/data/lmitche1/src/deps/petsc/arch-linux2-c-opt/lib
- -L/data/lmitche1/src/deps/petsc/arch-linux2-c-opt/lib -lcmumps
- -ldmumps -lsmumps -lzmumps -lmumps_common -lpord -lsuperlu_4.3
- -lsuperlu_dist_4.0 -lHYPRE -Wl,-rpath,/usr/lib/openmpi/lib
- -L/usr/lib/openmpi/lib -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.8
- -L/usr/lib/gcc/x86_64-linux-gnu/4.8
- -Wl,-rpath,/usr/lib/x86_64-linux-gnu -L/usr/lib/x86_64-linux-gnu
- -Wl,-rpath,/lib/x86_64-linux-gnu -L/lib/x86_64-linux-gnu -lmpi_cxx
- -lstdc++ -lscalapack -lml -lmpi_cxx -lstdc++ -lexoIIv2for -lexodus
- -llapack -lblas -lparmetis -ltriangle -lnetcdf -lmetis -lchaco
- -lctetgen -lX11 -lptesmumps -lptscotch -lptscotcherr -lscotch
- -lscotcherr -lhdf5hl_fortran -lhdf5_fortran -lhdf5_hl -lhdf5 -lssl
- -lcrypto -lm -lmpi_f90 -lmpi_f77 -lgfortran -lm -lgfortran -lm
- -lquadmath -lm -lmpi_cxx -lstdc++ -lrt -lm -lz
- -Wl,-rpath,/usr/lib/openmpi/lib -L/usr/lib/openmpi/lib
- -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.8
- -L/usr/lib/gcc/x86_64-linux-gnu/4.8
- -Wl,-rpath,/usr/lib/x86_64-linux-gnu -L/usr/lib/x86_64-linux-gnu
- -Wl,-rpath,/lib/x86_64-linux-gnu -L/lib/x86_64-linux-gnu
- -Wl,-rpath,/usr/lib/x86_64-linux-gnu -L/usr/lib/x86_64-linux-gnu -ldl
- -lmpi -lhwloc -lgcc_s -lpthread -ldl
- -----------------------------------------
$ ./ex6-30ab49e4
0 KSP Residual norm 3.679528502747e-11
1 KSP Residual norm 1.410011347346e-14
2 KSP Residual norm 2.871653636831e-14
KSP Object: 1 MPI processes
type: cg
maximum iterations=2, initial guess is zero
tolerances: relative=1e-05, absolute=1e-50, divergence=10000
left preconditioning
using PRECONDITIONED norm type for convergence test
PC Object: 1 MPI processes
type: gamg
MG: type is MULTIPLICATIVE, levels=3 cycles=v
Cycles per PCApply=1
Using Galerkin computed coarse grid matrices
Coarse grid solver -- level -------------------------------
KSP Object: (mg_coarse_) 1 MPI processes
type: preonly
maximum iterations=1, initial guess is zero
tolerances: relative=1e-05, absolute=1e-50, divergence=10000
left preconditioning
using NONE norm type for convergence test
PC Object: (mg_coarse_) 1 MPI processes
type: bjacobi
block Jacobi: number of blocks = 1
Local solve is same for all blocks, in the following KSP and
PC objects:
KSP Object: (mg_coarse_sub_) 1 MPI processes
type: preonly
maximum iterations=1, initial guess is zero
tolerances: relative=1e-05, absolute=1e-50, divergence=10000
left preconditioning
using NONE norm type for convergence test
PC Object: (mg_coarse_sub_) 1 MPI processes
type: lu
LU: out-of-place factorization
tolerance for zero pivot 2.22045e-14
using diagonal shift on blocks to prevent zero pivot
[INBLOCKS]
matrix ordering: nd
factor fill ratio given 5, needed 1
Factored matrix follows:
Mat Object: 1 MPI processes
type: seqaij
rows=592, cols=592
package used to perform factorization: petsc
total: nonzeros=350464, allocated nonzeros=350464
total number of mallocs used during MatSetValues
calls =0
using I-node routines: found 119 nodes, limit used
is 5
linear system matrix = precond matrix:
Mat Object: 1 MPI processes
type: seqaij
rows=592, cols=592
total: nonzeros=350464, allocated nonzeros=350464
total number of mallocs used during MatSetValues calls =0
using I-node routines: found 119 nodes, limit used is 5
linear system matrix = precond matrix:
Mat Object: 1 MPI processes
type: seqaij
rows=592, cols=592
total: nonzeros=350464, allocated nonzeros=350464
total number of mallocs used during MatSetValues calls =0
using I-node routines: found 119 nodes, limit used is 5
Down solver (pre-smoother) on level 1 -------------------------------
KSP Object: (mg_levels_1_) 1 MPI processes
type: chebyshev
Chebyshev: eigenvalue estimates: min = 0.0871826, max = 1.83084
maximum iterations=2
tolerances: relative=1e-05, absolute=1e-50, divergence=10000
left preconditioning
using nonzero initial guess
using NONE norm type for convergence test
PC Object: (mg_levels_1_) 1 MPI processes
type: sor
SOR: type = local_symmetric, iterations = 1, local iterations
= 1, omega = 1
linear system matrix = precond matrix:
Mat Object: 1 MPI processes
type: seqaij
rows=35419, cols=35419
total: nonzeros=1.55936e+07, allocated nonzeros=1.55936e+07
total number of mallocs used during MatSetValues calls =1
not using I-node routines
Up solver (post-smoother) same as down solver (pre-smoother)
Down solver (pre-smoother) on level 2 -------------------------------
KSP Object: (mg_levels_2_) 1 MPI processes
type: chebyshev
Chebyshev: eigenvalue estimates: min = 0.099472, max = 2.08891
maximum iterations=2
tolerances: relative=1e-05, absolute=1e-50, divergence=10000
left preconditioning
using nonzero initial guess
using NONE norm type for convergence test
PC Object: (mg_levels_2_) 1 MPI processes
type: sor
SOR: type = local_symmetric, iterations = 1, local iterations
= 1, omega = 1
linear system matrix = precond matrix:
Mat Object: 1 MPI processes
type: seqaij
rows=327680, cols=327680
total: nonzeros=3.25828e+06, allocated nonzeros=3.25828e+06
total number of mallocs used during MatSetValues calls =0
not using I-node routines
Up solver (post-smoother) same as down solver (pre-smoother)
linear system matrix = precond matrix:
Mat Object: 1 MPI processes
type: seqaij
rows=327680, cols=327680
total: nonzeros=3.25828e+06, allocated nonzeros=3.25828e+06
total number of mallocs used during MatSetValues calls =0
not using I-node routines
Number of iterations = 2
Residual norm = 8368.22
************************************************************************************************************************
*** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r
- -fCourier9' to print this document ***
************************************************************************************************************************
- ---------------------------------------------- PETSc Performance
Summary: ----------------------------------------------
./ex6-1ddf9fe on a test named yam.doc.ic.ac.uk with 1 processor, by
lmitche1 Mon Apr 27 16:02:36 2015
Using Petsc Release Version 3.5.2, unknown
Max Max/Min Avg Total
Time (sec): 2.828e+01 1.00000 2.828e+01
Objects: 1.150e+02 1.00000 1.150e+02
Flops: 1.006e+10 1.00000 1.006e+10 1.006e+10
Flops/sec: 3.559e+08 1.00000 3.559e+08 3.559e+08
MPI Messages: 0.000e+00 0.00000 0.000e+00 0.000e+00
MPI Message Lengths: 0.000e+00 0.00000 0.000e+00 0.000e+00
MPI Reductions: 0.000e+00 0.00000
Flop counting convention: 1 flop = 1 real number operation of type
(multiply/divide/add/subtract)
e.g., VecAXPY() for real vectors of length
N --> 2N flops
and VecAXPY() for complex vectors of
length N --> 8N flops
Summary of Stages: ----- Time ------ ----- Flops ----- ---
Messages --- -- Message Lengths -- -- Reductions --
Avg %Total Avg %Total counts
%Total Avg %Total counts %Total
0: Main Stage: 9.9010e-02 0.4% 7.4996e+06 0.1% 0.000e+00
0.0% 0.000e+00 0.0% 0.000e+00 0.0%
1: mystage 1: 2.6509e+01 93.7% 8.4700e+09 84.2% 0.000e+00
0.0% 0.000e+00 0.0% 0.000e+00 0.0%
2: mystage 2: 1.6704e+00 5.9% 1.5861e+09 15.8% 0.000e+00
0.0% 0.000e+00 0.0% 0.000e+00 0.0%
-
------------------------------------------------------------------------------------------------------------------------
See the 'Profiling' chapter of the users' manual for details on
interpreting output.
Phase summary info:
Count: number of times phase was executed
Time and Flops: Max - maximum over all processors
Ratio - ratio of maximum to minimum over all processors
Mess: number of messages sent
Avg. len: average message length (bytes)
Reduct: number of global reductions
Global: entire computation
Stage: stages of a computation. Set stages with PetscLogStagePush()
and PetscLogStagePop().
%T - percent time in this phase %F - percent flops in
this phase
%M - percent messages in this phase %L - percent message
lengths in this phase
%R - percent reductions in this phase
Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time
over all processors)
-
------------------------------------------------------------------------------------------------------------------------
Event Count Time (sec) Flops
--- Global --- --- Stage --- Total
Max Ratio Max Ratio Max Ratio Mess Avg
len Reduct %T %F %M %L %R %T %F %M %L %R Mflop/s
-
------------------------------------------------------------------------------------------------------------------------
- --- Event Stage 0: Main Stage
ThreadCommRunKer 2 1.0 3.0994e-06 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatMult 1 1.0 7.2370e-03 1.0 6.19e+06 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 7 83 0 0 0 855
MatAssemblyBegin 1 1.0 1.9073e-06 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatAssemblyEnd 1 1.0 1.0748e-02 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 11 0 0 0 0 0
MatLoad 1 1.0 8.5824e-02 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 87 0 0 0 0 0
VecNorm 1 1.0 9.2983e-05 1.0 6.55e+05 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 9 0 0 0 7048
VecSet 5 1.0 2.9252e-03 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 3 0 0 0 0 0
VecAXPY 1 1.0 4.6611e-04 1.0 6.55e+05 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 9 0 0 0 1406
- --- Event Stage 1: mystage 1
MatMult 20 1.0 2.8699e-01 1.0 3.73e+08 1.0 0.0e+00
0.0e+00 0.0e+00 1 4 0 0 0 1 4 0 0 0 1301
MatConvert 2 1.0 8.3888e-02 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatScale 6 1.0 7.1905e-02 1.0 4.85e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 1 0 0 0 675
MatAssemblyBegin 20 1.0 1.9789e-05 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatAssemblyEnd 20 1.0 1.1295e-01 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatGetRow 1452396 1.0 9.5270e-02 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatCoarsen 2 1.0 3.0676e-02 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatAXPY 2 1.0 8.2162e-02 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatMatMult 2 1.0 4.2625e-01 1.0 4.31e+07 1.0 0.0e+00
0.0e+00 0.0e+00 2 0 0 0 0 2 1 0 0 0 101
MatMatMultSym 2 1.0 3.0257e-01 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 0
MatMatMultNum 2 1.0 1.2364e-01 1.0 4.31e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 1 0 0 0 349
MatPtAP 2 1.0 2.3871e+01 1.0 7.82e+09 1.0 0.0e+00
0.0e+00 0.0e+00 84 78 0 0 0 90 92 0 0 0 328
MatPtAPSymbolic 2 1.0 1.4329e+01 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 51 0 0 0 0 54 0 0 0 0 0
MatPtAPNumeric 2 1.0 9.5422e+00 1.0 7.82e+09 1.0 0.0e+00
0.0e+00 0.0e+00 34 78 0 0 0 36 92 0 0 0 819
MatTrnMatMult 2 1.0 9.7712e-01 1.0 8.24e+07 1.0 0.0e+00
0.0e+00 0.0e+00 3 1 0 0 0 4 1 0 0 0 84
MatTrnMatMultSym 2 1.0 5.0258e-01 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 2 0 0 0 0 2 0 0 0 0 0
MatTrnMatMultNum 2 1.0 4.7454e-01 1.0 8.24e+07 1.0 0.0e+00
0.0e+00 0.0e+00 2 1 0 0 0 2 1 0 0 0 174
MatGetSymTrans 4 1.0 6.2370e-02 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecMDot 20 1.0 1.7304e-02 1.0 3.99e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2308
VecNorm 22 1.0 1.3692e-03 1.0 7.99e+06 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 5834
VecScale 22 1.0 1.8549e-03 1.0 3.99e+06 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2153
VecCopy 2 1.0 5.0211e-04 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecSet 77 1.0 3.6245e-02 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecAXPY 2 1.0 3.7718e-04 1.0 7.26e+05 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1925
VecMAXPY 22 1.0 2.2252e-02 1.0 4.72e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 1 0 0 0 2121
VecAssemblyBegin 2 1.0 9.5367e-07 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecAssemblyEnd 2 1.0 0.0000e+00 0.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecPointwiseMult 22 1.0 9.2957e-03 1.0 3.99e+06 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 430
VecSetRandom 2 1.0 8.8599e-03 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecNormalize 22 1.0 3.2570e-03 1.0 1.20e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 3679
KSPGMRESOrthog 20 1.0 3.6396e-02 1.0 7.99e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 2195
KSPSetUp 6 1.0 1.4364e-02 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
PCGAMGgraph_AGG 2 1.0 4.8670e-01 1.0 3.77e+07 1.0 0.0e+00
0.0e+00 0.0e+00 2 0 0 0 0 2 0 0 0 0 77
PCGAMGcoarse_AGG 2 1.0 1.0664e+00 1.0 8.24e+07 1.0 0.0e+00
0.0e+00 0.0e+00 4 1 0 0 0 4 1 0 0 0 77
PCGAMGProl_AGG 2 1.0 6.4827e-02 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
PCGAMGPOpt_AGG 2 1.0 9.9913e-01 1.0 5.31e+08 1.0 0.0e+00
0.0e+00 0.0e+00 4 5 0 0 0 4 6 0 0 0 532
PCSetUp 1 1.0 2.6505e+01 1.0 8.47e+09 1.0 0.0e+00
0.0e+00 0.0e+00 94 84 0 0 0 100100 0 0 0 320
- --- Event Stage 2: mystage 2
MatMult 38 1.0 5.6846e-01 1.0 6.85e+08 1.0 0.0e+00
0.0e+00 0.0e+00 2 7 0 0 0 34 43 0 0 0 1204
MatMultAdd 6 1.0 2.7303e-02 1.0 3.25e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 2 2 0 0 0 1189
MatMultTranspose 6 1.0 3.2745e-02 1.0 3.25e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 2 2 0 0 0 991
MatSolve 3 1.0 1.6339e-03 1.0 2.10e+06 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1286
MatSOR 36 1.0 9.4071e-01 1.0 6.79e+08 1.0 0.0e+00
0.0e+00 0.0e+00 3 7 0 0 0 56 43 0 0 0 722
MatLUFactorSym 1 1.0 5.9440e-03 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatLUFactorNum 1 1.0 4.0792e-02 1.0 1.15e+08 1.0 0.0e+00
0.0e+00 0.0e+00 0 1 0 0 0 2 7 0 0 0 2820
MatResidual 6 1.0 9.2385e-02 1.0 1.13e+08 1.0 0.0e+00
0.0e+00 0.0e+00 0 1 0 0 0 6 7 0 0 0 1224
MatGetRowIJ 1 1.0 1.4091e-04 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatGetOrdering 1 1.0 2.3508e-04 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatView 6 1.0 5.8508e-04 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecTDot 4 1.0 1.4160e-03 1.0 2.62e+06 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1851
VecNorm 3 1.0 3.5286e-04 1.0 1.97e+06 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 5572
VecCopy 8 1.0 1.1313e-02 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 1 0 0 0 0 0
VecSet 22 1.0 4.3163e-03 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecAXPY 4 1.0 1.6727e-03 1.0 2.62e+06 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1567
VecAYPX 49 1.0 1.7958e-02 1.0 1.15e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 1 1 0 0 0 643
VecAXPBYCZ 24 1.0 1.3117e-02 1.0 2.18e+07 1.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 1 1 0 0 0 1661
KSPSetUp 2 1.0 9.2983e-06 1.0 0.00e+00 0.0 0.0e+00
0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
KSPSolve 1 1.0 1.6690e+00 1.0 1.59e+09 1.0 0.0e+00
0.0e+00 0.0e+00 6 16 0 0 0 100100 0 0 0 950
PCSetUp 1 1.0 4.7009e-02 1.0 1.15e+08 1.0 0.0e+00
0.0e+00 0.0e+00 0 1 0 0 0 3 7 0 0 0 2447
PCSetUpOnBlocks 3 1.0 4.7014e-02 1.0 1.15e+08 1.0 0.0e+00
0.0e+00 0.0e+00 0 1 0 0 0 3 7 0 0 0 2447
PCApply 3 1.0 1.6409e+00 1.0 1.57e+09 1.0 0.0e+00
0.0e+00 0.0e+00 6 16 0 0 0 98 99 0 0 0 954
-
------------------------------------------------------------------------------------------------------------------------
Memory usage is given in bytes:
Object Type Creations Destructions Memory Descendants'
Mem.
Reports information only for process 0.
- --- Event Stage 0: Main Stage
Viewer 1 1 760 0
Matrix 1 6 58874852 0
Vector 3 20 27954224 0
Krylov Solver 0 5 23360 0
Preconditioner 0 5 5332 0
Index Set 0 3 7112 0
- --- Event Stage 1: mystage 1
Viewer 1 0 0 0
Matrix 14 10 477163156 0
Matrix Coarsen 2 2 1288 0
Vector 69 52 66638640 0
Krylov Solver 7 2 60432 0
Preconditioner 7 2 2096 0
Index Set 2 2 1584 0
PetscRandom 2 2 1280 0
- --- Event Stage 2: mystage 2
Matrix 1 0 0 0
Index Set 5 2 2536 0
========================================================================================================================
Average time to get PetscTime(): 0
#PETSc Option Table entries:
- -f helmholtz-sphere.dat
- -ksp_convergence_test skip
- -ksp_max_it 2
- -ksp_monitor
- -ksp_type cg
- -ksp_view
- -log_summary
- -matload_block_size 1
- -pc_type gamg
#End of PETSc Option Table entries
Compiled without FORTRAN kernels
Compiled with full precision matrices (default)
sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8
sizeof(PetscScalar) 8 sizeof(PetscInt) 4
Configure options: PETSC_ARCH=test --with-debugging=0
- -----------------------------------------
Libraries compiled on Mon Apr 27 10:49:24 2015 on yam.doc.ic.ac.uk
Machine characteristics:
Linux-3.13.0-45-generic-x86_64-with-Ubuntu-14.04-trusty
Using PETSc directory: /data/lmitche1/src/deps/petsc
Using PETSc arch: test
- -----------------------------------------
Using C compiler: mpicc -fPIC -Wall -Wwrite-strings
- -Wno-strict-aliasing -Wno-unknown-pragmas -O ${COPTFLAGS} ${CFLAGS}
Using Fortran compiler: mpif90 -fPIC -Wall -Wno-unused-variable
- -ffree-line-length-0 -Wno-unused-dummy-argument -O ${FOPTFLAGS}
${FFLAGS}
- -----------------------------------------
Using include paths: -I/data/lmitche1/src/deps/petsc/test/include
- -I/data/lmitche1/src/deps/petsc/include
- -I/data/lmitche1/src/deps/petsc/include
- -I/data/lmitche1/src/deps/petsc/test/include
- -I/usr/lib/openmpi/include -I/usr/lib/openmpi/include/openmpi
- -----------------------------------------
Using C linker: mpicc
Using Fortran linker: mpif90
Using libraries: -Wl,-rpath,/data/lmitche1/src/deps/petsc/test/lib
- -L/data/lmitche1/src/deps/petsc/test/lib -lpetsc -llapack -lblas -lX11
- -lssl -lcrypto -lpthread -lm -Wl,-rpath,/usr/lib/openmpi/lib
- -L/usr/lib/openmpi/lib -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.8
- -L/usr/lib/gcc/x86_64-linux-gnu/4.8
- -Wl,-rpath,/usr/lib/x86_64-linux-gnu -L/usr/lib/x86_64-linux-gnu
- -Wl,-rpath,/lib/x86_64-linux-gnu -L/lib/x86_64-linux-gnu -lmpi_f90
- -lmpi_f77 -lgfortran -lm -lgfortran -lm -lquadmath -lm -lmpi_cxx
- -lstdc++ -Wl,-rpath,/usr/lib/openmpi/lib -L/usr/lib/openmpi/lib
- -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.8
- -L/usr/lib/gcc/x86_64-linux-gnu/4.8
- -Wl,-rpath,/usr/lib/x86_64-linux-gnu -L/usr/lib/x86_64-linux-gnu
- -Wl,-rpath,/lib/x86_64-linux-gnu -L/lib/x86_64-linux-gnu
- -Wl,-rpath,/usr/lib/x86_64-linux-gnu -L/usr/lib/x86_64-linux-gnu -ldl
- -lmpi -lhwloc -lgcc_s -lpthread -ldl
- -----------------------------------------
-----BEGIN PGP SIGNATURE-----
Version: GnuPG v1
iQEcBAEBAgAGBQJVPlBxAAoJECOc1kQ8PEYv4p4IAJKGFsnbkyUfA45gQgMSoF/p
Ox+0tD3iIJ8bLXFBh30b6/Y8UuJD1jown7iOkc6xE7DZFPZCdVXKqKI7G+0yHkPI
0f5f1cMEuAD/N+zsN7ls9/CsB3YIiSEF4xAOenvLCQ3cai5KFZtv/V1PS/67+npS
3w7iWTE6ylJxi/rrtGzOVFhhR9e+Kf4w7dez11+tra6pfX6sQZvH/2sZ46yzP8td
SvbSmdNtVoXLIMIUWGbVYHOs8qg6IqvyorYYjwll6WOe+9oPgz2s9XFNPVCV7N5G
sKJgezt1ClLEKwLD3DS0Z81lpjFJUGYQC3VMtrLLOujyLxX7gp6RX6Xh/rMfo7g=
=Clvj
-----END PGP SIGNATURE-----
More information about the petsc-dev
mailing list