[petsc-users] Read in sequential, solve in parallel

Jed Brown jed at 59A2.org
Wed Sep 29 08:16:01 CDT 2010


On Wed, Sep 29, 2010 at 14:51, Moinier, Pierre (UK)
<Pierre.Moinier at baesystems.com> wrote:
> Jed,
>
> The matrix is 1000000x1000000 and I have 4996000 non zeros

These statistics look like perhaps this matrix comes from a 5-point
discretization of an elliptic operator.  Is that true?

> Here is the output for a single proc:

You'll want to compare the time in each event when run in parallel:

> MatMult             1633 1.0 1.6247e+01 1.0 1.47e+10 1.0 0.0e+00 0.0e+00
> 0.0e+00 36 43  0  0  0  36 43  0  0  0   904

16 seconds in this event, 904 Mflop/s is good for a problem like this
on a single core.

> VecDot              3266 1.0 4.4834e+00 1.0 6.53e+09 1.0 0.0e+00 0.0e+00
> 0.0e+00 10 19  0  0  0  10 19  0  0  0  1457
> VecNorm             1634 1.0 1.2968e+01 1.0 3.27e+09 1.0 0.0e+00 0.0e+00
> 0.0e+00 28 10  0  0  0  28 10  0  0  0   252

This is confusing, the norms should take about a tenth of this (they
should have almost double the Mflop/s of VecDot).  Is there something
else running on this machine?  Anyone have other ideas?

> VecAXPY             3266 1.0 5.5580e+00 1.0 6.53e+09 1.0 0.0e+00 0.0e+00
> 0.0e+00 12 19  0  0  0  12 19  0  0  0  1175
> VecAYPX             1632 1.0 2.5961e+00 1.0 3.26e+09 1.0 0.0e+00 0.0e+00
> 0.0e+00  6 10  0  0  0   6 10  0  0  0  1257

These look normal.

> KSPSolve               1 1.0 4.4828e+01 1.0 3.43e+10 1.0 0.0e+00 0.0e+00
> 0.0e+00 98100  0  0  0  98100  0  0  0   765

And here's your solve total, the aggregate numbers look fine.

Jed


More information about the petsc-users mailing list