[petsc-users] Read in sequential, solve in parallel
    Jed Brown 
    jed at 59A2.org
       
    Wed Sep 29 08:16:01 CDT 2010
    
    
  
On Wed, Sep 29, 2010 at 14:51, Moinier, Pierre (UK)
<Pierre.Moinier at baesystems.com> wrote:
> Jed,
>
> The matrix is 1000000x1000000 and I have 4996000 non zeros
These statistics look like perhaps this matrix comes from a 5-point
discretization of an elliptic operator.  Is that true?
> Here is the output for a single proc:
You'll want to compare the time in each event when run in parallel:
> MatMult             1633 1.0 1.6247e+01 1.0 1.47e+10 1.0 0.0e+00 0.0e+00
> 0.0e+00 36 43  0  0  0  36 43  0  0  0   904
16 seconds in this event, 904 Mflop/s is good for a problem like this
on a single core.
> VecDot              3266 1.0 4.4834e+00 1.0 6.53e+09 1.0 0.0e+00 0.0e+00
> 0.0e+00 10 19  0  0  0  10 19  0  0  0  1457
> VecNorm             1634 1.0 1.2968e+01 1.0 3.27e+09 1.0 0.0e+00 0.0e+00
> 0.0e+00 28 10  0  0  0  28 10  0  0  0   252
This is confusing, the norms should take about a tenth of this (they
should have almost double the Mflop/s of VecDot).  Is there something
else running on this machine?  Anyone have other ideas?
> VecAXPY             3266 1.0 5.5580e+00 1.0 6.53e+09 1.0 0.0e+00 0.0e+00
> 0.0e+00 12 19  0  0  0  12 19  0  0  0  1175
> VecAYPX             1632 1.0 2.5961e+00 1.0 3.26e+09 1.0 0.0e+00 0.0e+00
> 0.0e+00  6 10  0  0  0   6 10  0  0  0  1257
These look normal.
> KSPSolve               1 1.0 4.4828e+01 1.0 3.43e+10 1.0 0.0e+00 0.0e+00
> 0.0e+00 98100  0  0  0  98100  0  0  0   765
And here's your solve total, the aggregate numbers look fine.
Jed
    
    
More information about the petsc-users
mailing list