A 3D example of KSPSolve?

Satish Balay balay at mcs.anl.gov
Fri Feb 9 23:02:14 CST 2007


Just looking at 8 proc run [diffusion stage] we have:

MatMult        :  79 sec
MatMultAdd     :   2 sec
VecScatterBegin:  17 sec
VecScatterEnd  :  51 sec

So basically the communication in MatMult/Add is represented by
VecScatters. Here out of 81 sec total - 68 seconds are used for
communication [with a load imbalance of 11 for vecscaterend]

So - I think MPI performance is reducing scalability here..

Things to try:

* -vecstatter_rr etc options I sugested earlier

* install mpich with '--with-device=ch3:ssm' and see
if it makes a difference

Satish

--- Event Stage 4: Diffusion

[x]rhsLtP            297 1.0 1.1017e+02 1.5 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  7  0  0  0  0  39  0  0  0  0     0
[x]rhsGravity         99 1.0 4.2582e+0083.5 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   1  0  0  0  0     0
VecDot              4657 1.0 2.5748e+01 3.2 7.60e+07 3.2 0.0e+00 0.0e+00 4.7e+03  1  1  0  0  6   5  3  0  0 65   191
VecNorm             2477 1.0 2.2109e+01 2.2 3.22e+07 2.2 0.0e+00 0.0e+00 2.5e+03  1  0  0  0  3   5  2  0  0 35   118
VecScale             594 1.0 2.9330e-02 1.5 2.61e+08 1.5 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0  1361
VecCopy              594 1.0 2.7552e-01 1.3 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
VecSet              3665 1.0 6.0793e-01 1.4 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
VecAXPY             5251 1.0 2.5892e+00 1.2 3.31e+08 1.2 0.0e+00 0.0e+00 0.0e+00  0  1  0  0  0   1  4  0  0  0  2137
VecAYPX             1883 1.0 8.6419e-01 1.3 3.62e+08 1.3 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   0  1  0  0  0  2296
VecScatterBegin     2873 1.0 1.7569e+01 3.0 0.00e+00 0.0 3.8e+04 1.6e+05 0.0e+00  1  0 10 20  0   5  0100100  0     0
VecScatterEnd       2774 1.0 5.1519e+0110.9 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  1  0  0  0  0   7  0  0  0  0     0
MatMult             2477 1.0 7.9186e+01 2.4 2.34e+08 2.4 3.5e+04 1.7e+05 0.0e+00  3 11  9 20  0  20 48 91 98  0   850
MatMultAdd           297 1.0 2.8161e+00 5.4 4.46e+07 2.2 3.6e+03 3.4e+04 0.0e+00  0  0  1  0  0   0  0  9  2  0   125
MatSolve            2477 1.0 6.2245e+01 1.2 1.41e+08 1.2 0.0e+00 0.0e+00 0.0e+00  4 10  0  0  0  22 41  0  0  0   926
MatLUFactorNum         3 1.0 2.7686e-01 1.1 2.79e+08 1.1 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0  2016
MatGetRow        19560420 1.0 5.5195e+01 1.6 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  3  0  0  0  0  20  0  0  0  0     0
KSPSetup               6 1.0 3.0756e-05 2.8 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
KSPSolve             297 1.0 1.3142e+02 1.0 1.31e+08 1.1 3.1e+04 1.7e+05 7.1e+03  8 22  8 18  9  50 93 80 86100  1001
PCSetUp                6 1.0 2.7700e-01 1.1 2.78e+08 1.1 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0  2015
PCSetUpOnBlocks      297 1.0 2.7794e-01 1.1 2.78e+08 1.1 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0  2008
PCApply             2477 1.0 6.2772e+01 1.2 1.39e+08 1.2 0.0e+00 0.0e+00 0.0e+00  4 10  0  0  0  23 41  0  0  0   918




More information about the petsc-users mailing list