[petsc-users] petsc 3.7.2 memory usage is much higher when compared to 3.6.1

Matthew Knepley knepley at gmail.com
Tue Jul 5 16:13:35 CDT 2016


On Tue, Jul 5, 2016 at 3:42 PM, Hassan Raiesi <
Hassan.Raiesi at aero.bombardier.com> wrote:

> Hi,
>
>
>
> PETSc 3.7.2 seems to have a much higher memory usage when compared with
> PETSc- 3.1.1 c, to a point that it crashes our code for large problems that
> we ran with version 3.6.1 in the past.
>
> I have re-compiled the code with same options, and ran the same code
> linked with the two versions, here are the log-summarie:
>

According to the log_summary (which you NEED to send in full if we are to
understand anything), the memory usage is largely the same.
There are more matrices, which leads me to believe that GAMG is not
coarsening as quickly. You might consider a non-zero threshold for
it.

The best way to understand what is happening is to run Massif (from
valgrind) on both.

  Thanks,

     Matt


> -flow_ksp_max_it 20
>
> -flow_ksp_monitor_true_residual
>
> -flow_ksp_rtol 0.1
>
> -flow_ksp_type fgmres
>
> -flow_mg_coarse_pc_factor_mat_solver_package mumps
>
> -flow_mg_coarse_pc_type lu
>
> -flow_mg_levels_ksp_type richardson
>
> -flow_mg_levels_pc_type sor
>
> -flow_pc_gamg_agg_nsmooths 0
>
> -flow_pc_gamg_coarse_eq_limit 2000
>
> -flow_pc_gamg_process_eq_limit 2500
>
> -flow_pc_gamg_repartition true
>
> -flow_pc_gamg_reuse_interpolation true
>
> -flow_pc_gamg_square_graph 3
>
> -flow_pc_gamg_sym_graph true
>
> -flow_pc_gamg_type agg
>
> -flow_pc_mg_cycle v
>
> -flow_pc_mg_levels 20
>
> -flow_pc_mg_type kaskade
>
> -flow_pc_type gamg
>
> -log_summary
>
>
>
> Note: it is not specific to PCGAMG, even a bjacobi+fgmres would need more
> memory (4.5GB/core in version 3.6.1 compared to 6.8GB/core for 3.7.2).
>
>
>
>
>
>
>
> Using Petsc Development GIT revision: v3.7.2-812-gc68d048  GIT Date:
> 2016-07-05 12:04:34 -0400
>
>
>
>                          Max       Max/Min        Avg      Total
>
> Time (sec):           6.760e+02      1.00006   6.760e+02
>
> Objects:              1.284e+03      1.00469   1.279e+03
>
> Flops:                3.563e+10      1.10884   3.370e+10  1.348e+13
>
> Flops/sec:            5.271e+07      1.10884   4.985e+07  1.994e+10
>
> MPI Messages:         4.279e+04      7.21359   1.635e+04  6.542e+06
>
> MPI Message Lengths:  3.833e+09     17.25274   7.681e+04  5.024e+11
>
> MPI Reductions:       4.023e+03      1.00149
>
>
>
> Flop counting convention: 1 flop = 1 real number operation of type
> (multiply/divide/add/subtract)
>
>                             e.g., VecAXPY() for real vectors of length N
> --> 2N flops
>
>                             and VecAXPY() for complex vectors of length N
> --> 8N flops
>
>
>
> Summary of Stages:   ----- Time ------  ----- Flops -----  --- Messages
> ---  -- Message Lengths --  -- Reductions --
>
>                         Avg     %Total     Avg     %Total   counts
> %Total     Avg         %Total   counts   %Total
>
> 0:      Main Stage: 6.7600e+02 100.0%  1.3478e+13 100.0%  6.533e+06
> 99.9%  7.674e+04       99.9%  4.010e+03  99.7%
>
>
>
>
> ------------------------------------------------------------------------------------------------------------------------
>
> See the 'Profiling' chapter of the users' manual for details on
> interpreting output.
>
> Phase summary info:
>
>    Count: number of times phase was executed
>
>    Time and Flops: Max - maximum over all processors
>
>                    Ratio - ratio of maximum to minimum over all processors
>
>    Mess: number of messages sent
>
>    Avg. len: average message length (bytes)
>
>    Reduct: number of global reductions
>
>    Global: entire computation
>
>    Stage: stages of a computation. Set stages with PetscLogStagePush() and
> PetscLogStagePop().
>
>       %T - percent time in this phase         %F - percent flops in this
> phase
>
>       %M - percent messages in this phase     %L - percent message lengths
> in this phase
>
>       %R - percent reductions in this phase
>
>    Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time
> over all processors)
>
>
> ------------------------------------------------------------------------------------------------------------------------
>
> Event                Count      Time (sec)     Flops
>        --- Global ---  --- Stage ---   Total
>
>                    Max Ratio  Max     Ratio   Max  Ratio  Mess   Avg len
> Reduct  %T %F %M %L %R  %T %F %M %L %R Mflop/s
>
>
> ------------------------------------------------------------------------------------------------------------------------
>
>
>
> --- Event Stage 0: Main Stage
>
>
>
> MatMult              500 1.0 1.0582e+01 1.2 6.68e+09 1.1 1.9e+06 1.0e+04
> 0.0e+00  1 19 28  4  0   1 19 29  4  0 237625
>
> MatMultTranspose     120 1.0 7.6262e-01 1.3 3.58e+08 1.1 2.4e+05 1.5e+04
> 0.0e+00  0  1  4  1  0   0  1  4  1  0 180994
>
> MatSolve             380 1.0 4.1580e+00 1.1 1.17e+09 1.1 8.6e+03 8.8e+01
> 6.0e+01  1  3  0  0  1   1  3  0  0  1 105950
>
> MatSOR               120 1.0 1.4316e+01 1.2 6.75e+09 1.1 9.5e+05 7.4e+03
> 0.0e+00  2 19 15  1  0   2 19 15  1  0 177298
>
> MatLUFactorSym         2 1.0 2.3449e-01 1.1 0.00e+00 0.0 0.0e+00 0.0e+00
> 1.0e+01  0  0  0  0  0   0  0  0  0  0     0
>
> MatLUFactorNum        60 1.0 8.8820e+00 1.0 1.95e+08 1.2 0.0e+00 0.0e+00
> 0.0e+00  1  1  0  0  0   1  1  0  0  0  7877
>
> MatILUFactorSym        1 1.0 1.9795e-02 1.4 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> MatConvert             6 1.0 2.9893e-02 1.1 0.00e+00 0.0 0.0e+00 0.0e+00
> 1.8e+01  0  0  0  0  0   0  0  0  0  0     0
>
> MatScale               6 1.0 1.8810e-02 1.4 4.52e+06 1.1 2.4e+04 1.5e+03
> 0.0e+00  0  0  0  0  0   0  0  0  0  0 90171
>
> MatAssemblyBegin     782 1.0 1.8294e+01 2.9 0.00e+00 0.0 9.2e+05 4.1e+05
> 4.2e+02  2  0 14 75 10   2  0 14 75 10     0
>
> MatAssemblyEnd       782 1.0 1.4283e+01 3.0 0.00e+00 0.0 4.1e+05 8.7e+02
> 4.7e+02  1  0  6  0 12   1  0  6  0 12     0
>
> MatGetRow        6774900 1.1 9.4289e-01 1.2 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> MatGetRowIJ            3 3.0 6.6261e-036948.0 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> MatGetSubMatrix       12 1.0 2.6783e+01 1.0 0.00e+00 0.0 1.1e+05 1.3e+05
> 2.0e+02  4  0  2  3  5   4  0  2  3  5     0
>
> MatGetOrdering         3 3.0 7.7400e-03 7.2 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> MatPartitioning        6 1.0 1.8949e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00
> 1.4e+01  0  0  0  0  0   0  0  0  0  0     0
>
> MatCoarsen             6 1.0 9.5692e-02 1.2 0.00e+00 0.0 2.6e+05 1.1e+03
> 4.1e+01  0  0  4  0  1   0  0  4  0  1     0
>
> MatZeroEntries       142 1.0 9.7085e-01 1.1 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> MatTranspose           6 1.0 2.1740e-01 1.0 0.00e+00 0.0 1.9e+05 8.5e+02
> 7.8e+01  0  0  3  0  2   0  0  3  0  2     0
>
> MatPtAP              120 1.0 6.0157e+01 1.0 1.82e+10 1.1 1.5e+06 2.7e+05
> 4.2e+02  9 51 22 80 10   9 51 22 80 10 114269
>
> MatPtAPSymbolic       12 1.0 8.1081e+00 1.0 0.00e+00 0.0 2.2e+05 3.8e+04
> 8.4e+01  1  0  3  2  2   1  0  3  2  2     0
>
> MatPtAPNumeric       120 1.0 5.2205e+01 1.0 1.82e+10 1.1 1.2e+06 3.1e+05
> 3.4e+02  8 51 19 78  8   8 51 19 78  8 131676
>
> MatTrnMatMult          3 1.0 1.8608e+00 1.0 3.23e+07 1.2 8.3e+04 7.9e+03
> 5.7e+01  0  0  1  0  1   0  0  1  0  1  6275
>
> MatTrnMatMultSym       3 1.0 1.3447e+00 1.0 0.00e+00 0.0 6.9e+04 3.8e+03
> 5.1e+01  0  0  1  0  1   0  0  1  0  1     0
>
> MatTrnMatMultNum       3 1.0 5.1695e-01 1.0 3.23e+07 1.2 1.3e+04 3.0e+04
> 6.0e+00  0  0  0  0  0   0  0  0  0  0 22588
>
> MatGetLocalMat       126 1.0 1.0355e+00 1.8 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> MatGetBrAoCol        120 1.0 9.5921e+0019.2 0.00e+00 0.0 5.7e+05 3.3e+04
> 0.0e+00  1  0  9  4  0   1  0  9  4  0     0
>
> VecDot               320 1.0 1.1400e+00 1.6 2.04e+08 1.1 0.0e+00 0.0e+00
> 3.2e+02  0  1  0  0  8   0  1  0  0  8 68967
>
> VecMDot              260 1.0 1.9577e+00 2.8 3.70e+08 1.1 0.0e+00 0.0e+00
> 2.6e+02  0  1  0  0  6   0  1  0  0  6 72792
>
> VecNorm              440 1.0 2.6273e+00 1.9 5.88e+08 1.1 0.0e+00 0.0e+00
> 4.4e+02  0  2  0  0 11   0  2  0  0 11 86035
>
> VecScale             320 1.0 2.1386e-01 1.2 7.91e+07 1.1 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0 141968
>
> VecCopy              220 1.0 7.0370e-01 1.2 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> VecSet               862 1.0 7.1000e-01 1.6 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> VecAXPY              440 1.0 8.6790e-01 1.1 3.83e+08 1.1 0.0e+00 0.0e+00
> 0.0e+00  0  1  0  0  0   0  1  0  0  0 169857
>
> VecAYPX              280 1.0 5.7766e-01 1.5 1.92e+08 1.1 0.0e+00 0.0e+00
> 0.0e+00  0  1  0  0  0   0  1  0  0  0 127599
>
> VecMAXPY             300 1.0 9.7396e-01 1.2 4.98e+08 1.1 0.0e+00 0.0e+00
> 0.0e+00  0  1  0  0  0   0  1  0  0  0 196768
>
> VecAssemblyBegin     234 1.0 4.6313e+00 5.6 0.00e+00 0.0 0.0e+00 0.0e+00
> 6.8e+02  0  0  0  0 17   0  0  0  0 17     0
>
> VecAssemblyEnd       234 1.0 5.1503e-0319.5 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> VecScatterBegin     1083 1.0 2.9274e-01 4.5 0.00e+00 0.0 3.8e+06 8.5e+03
> 2.0e+01  0  0 59  6  0   0  0 59  6  0     0
>
> VecScatterEnd       1063 1.0 3.9653e+00 5.6 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> KSPGMRESOrthog        20 1.0 1.7405e+00 3.7 1.28e+08 1.1 0.0e+00 0.0e+00
> 2.0e+01  0  0  0  0  0   0  0  0  0  0 28232
>
> KSPSetUp             222 1.0 6.8469e-02 1.1 0.00e+00 0.0 0.0e+00 0.0e+00
> 1.2e+01  0  0  0  0  0   0  0  0  0  0     0
>
> KSPSolve              60 1.0 1.4767e+02 1.0 3.55e+10 1.1 6.3e+06 7.2e+04
> 3.2e+03 22100 96 90 79  22100 96 90 79 91007
>
> PCGAMGGraph_AGG        6 1.0 6.0792e+00 1.0 4.52e+06 1.1 3.8e+05 9.0e+02
> 2.5e+02  1  0  6  0  6   1  0  6  0  6   279
>
> PCGAMGCoarse_AGG       6 1.0 2.0660e+00 1.0 3.23e+07 1.2 4.2e+05 3.1e+03
> 1.5e+02  0  0  6  0  4   0  0  6  0  4  5652
>
> PCGAMGProl_AGG         6 1.0 1.8842e+00 1.0 0.00e+00 0.0 7.3e+05 3.3e+03
> 8.6e+02  0  0 11  0 21   0  0 11  0 22     0
>
> PCGAMGPOpt_AGG         6 1.0 6.4373e-06 0.0 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> GAMG: createProl       6 1.0 1.0036e+01 1.0 3.68e+07 1.2 1.5e+06 2.7e+03
> 1.3e+03  1  0 23  1 31   1  0 23  1 31  1332
>
>   Graph               12 1.0 6.0783e+00 1.0 4.52e+06 1.1 3.8e+05 9.0e+02
> 2.5e+02  1  0  6  0  6   1  0  6  0  6   279
>
>   MIS/Agg              6 1.0 9.5831e-02 1.2 0.00e+00 0.0 2.6e+05 1.1e+03
> 4.1e+01  0  0  4  0  1   0  0  4  0  1     0
>
>   SA: col data         6 1.0 7.7358e-01 1.0 0.00e+00 0.0 6.7e+05 2.9e+03
> 7.8e+02  0  0 10  0 19   0  0 10  0 19     0
>
>   SA: frmProl0         6 1.0 1.0759e+00 1.0 0.00e+00 0.0 6.2e+04 7.6e+03
> 6.0e+01  0  0  1  0  1   0  0  1  0  1     0
>
> GAMG: partLevel        6 1.0 3.8136e+01 1.0 9.09e+08 1.1 3.8e+05 5.0e+04
> 5.4e+02  6  3  6  4 13   6  3  6  4 14  9013
>
>   repartition          6 1.0 2.7910e+00 1.0 0.00e+00 0.0 4.6e+04 1.3e+02
> 1.6e+02  0  0  1  0  4   0  0  1  0  4     0
>
>   Invert-Sort          6 1.0 2.5045e+00 2.0 0.00e+00 0.0 0.0e+00 0.0e+00
> 2.4e+01  0  0  0  0  1   0  0  0  0  1     0
>
>   Move A               6 1.0 1.4832e+01 1.0 0.00e+00 0.0 8.5e+04 1.7e+05
> 1.1e+02  2  0  1  3  3   2  0  1  3  3     0
>
>   Move P               6 1.0 1.2023e+01 1.0 0.00e+00 0.0 2.4e+04 3.8e+03
> 1.1e+02  2  0  0  0  3   2  0  0  0  3     0
>
> PCSetUp              100 1.0 1.1212e+02 1.0 1.84e+10 1.1 3.2e+06 1.3e+05
> 2.2e+03 17 52 49 84 54  17 52 49 84 54 62052
>
> PCSetUpOnBlocks       40 1.0 1.0386e+00 1.2 1.95e+08 1.2 0.0e+00 0.0e+00
> 0.0e+00  0  1  0  0  0   0  1  0  0  0 67368
>
> PCApply              380 1.0 2.0034e+01 1.1 8.60e+09 1.1 1.5e+06 9.9e+03
> 6.0e+01  3 24 22  3  1   3 24 22  3  1 161973
>
> SFSetGraph            12 1.0 4.9813e-0310.2 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> SFBcastBegin          47 1.0 3.3110e-02 2.6 0.00e+00 0.0 2.6e+05 1.1e+03
> 6.0e+00  0  0  4  0  0   0  0  4  0  0     0
>
> SFBcastEnd            47 1.0 1.3497e-02 3.6 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> SFReduceBegin          6 1.0 1.8593e-02 4.2 0.00e+00 0.0 7.2e+04 4.9e+02
> 6.0e+00  0  0  1  0  0   0  0  1  0  0     0
>
> SFReduceEnd            6 1.0 7.1628e-0318.5 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> BuildTwoSided         12 1.0 3.5771e-02 2.5 0.00e+00 0.0 5.0e+04 4.0e+00
> 1.2e+01  0  0  1  0  0   0  0  1  0  0     0
>
>
> ------------------------------------------------------------------------------------------------------------------------
>
>
>
> Memory usage is given in bytes:
>
>
>
> Object Type          Creations   Destructions     Memory  Descendants' Mem.
>
> Reports information only for process 0.
>
>
>
> --- Event Stage 0: Main Stage
>
>
>
>               Matrix   302            299   1992700700     0.
>
> Matrix Partitioning     6              6         3888     0.
>
>       Matrix Coarsen     6              6         3768     0.
>
>               Vector   600            600   1582204168     0.
>
>       Vector Scatter    87             87      5614432     0.
>
>        Krylov Solver    11             11        59472     0.
>
>       Preconditioner    11             11        11120     0.
>
>          PetscRandom     1              1          638     0.
>
>               Viewer     1              0            0     0.
>
>            Index Set   247            247      9008420     0.
>
> Star Forest Bipartite Graph    12             12        10176     0.
>
>
> ========================================================================================================================
>
>
>
> And for  petsc 3.6.1:
>
>
>
> Using Petsc Development GIT revision: v3.6.1-307-g26c82d3  GIT Date:
> 2015-08-06 11:50:34 -0500
>
>
>
>                          Max       Max/Min        Avg      Total
>
> Time (sec):           5.515e+02      1.00001   5.515e+02
>
> Objects:              1.231e+03      1.00490   1.226e+03
>
> Flops:                3.431e+10      1.12609   3.253e+10  1.301e+13
>
> Flops/sec:            6.222e+07      1.12609   5.899e+07  2.359e+10
>
> MPI Messages:         4.432e+04      7.84165   1.504e+04  6.016e+06
>
> MPI Message Lengths:  2.236e+09     12.61261   5.027e+04  3.024e+11
>
> MPI Reductions:       4.012e+03      1.00150
>
>
>
> Flop counting convention: 1 flop = 1 real number operation of type
> (multiply/divide/add/subtract)
>
>                             e.g., VecAXPY() for real vectors of length N
> --> 2N flops
>
>                             and VecAXPY() for complex vectors of length N
> --> 8N flops
>
>
>
> Summary of Stages:   ----- Time ------  ----- Flops -----  --- Messages
> ---  -- Message Lengths --  -- Reductions --
>
>                         Avg     %Total     Avg     %Total   counts
> %Total     Avg         %Total   counts   %Total
>
> 0:      Main Stage: 5.5145e+02 100.0%  1.3011e+13 100.0%  6.007e+06
> 99.9%  5.020e+04       99.9%  3.999e+03  99.7%
>
>
>
>
> ------------------------------------------------------------------------------------------------------------------------
>
> See the 'Profiling' chapter of the users' manual for details on
> interpreting output.
>
> Phase summary info:
>
>    Count: number of times phase was executed
>
>    Time and Flops: Max - maximum over all processors
>
>                    Ratio - ratio of maximum to minimum over all processors
>
>    Mess: number of messages sent
>
>    Avg. len: average message length (bytes)
>
>    Reduct: number of global reductions
>
>    Global: entire computation
>
>    Stage: stages of a computation. Set stages with PetscLogStagePush() and
> PetscLogStagePop().
>
>       %T - percent time in this phase         %F - percent flops in this
> phase
>
>       %M - percent messages in this phase     %L - percent message lengths
> in this phase
>
>       %R - percent reductions in this phase
>
>    Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time
> over all processors)
>
>
> ------------------------------------------------------------------------------------------------------------------------
>
> Event                Count      Time (sec)
> Flops                             --- Global ---  --- Stage ---   Total
>
>                    Max Ratio  Max     Ratio   Max  Ratio  Mess   Avg len
> Reduct  %T %F %M %L %R  %T %F %M %L %R Mflop/s
>
>
> ------------------------------------------------------------------------------------------------------------------------
>
>
>
> --- Event Stage 0: Main Stage
>
>
>
> MatMult              500 1.0 1.0172e+01 1.2 6.68e+09 1.1 1.9e+06 9.9e+03
> 0.0e+00  2 19 31  6  0   2 19 31  6  0 247182
>
> MatMultTranspose     120 1.0 6.9889e-01 1.2 3.56e+08 1.1 2.5e+05 1.4e+04
> 0.0e+00  0  1  4  1  0   0  1  4  1  0 197492
>
> MatSolve             380 1.0 3.9310e+00 1.1 1.17e+09 1.1 1.3e+04 5.7e+01
> 6.0e+01  1  3  0  0  1   1  3  0  0  2 112069
>
> MatSOR               120 1.0 1.3915e+01 1.1 6.73e+09 1.1 9.5e+05 7.4e+03
> 0.0e+00  2 20 16  2  0   2 20 16  2  0 182405
>
> MatLUFactorSym         2 1.0 2.1180e-01 1.8 0.00e+00 0.0 0.0e+00 0.0e+00
> 1.0e+01  0  0  0  0  0   0  0  0  0  0     0
>
> MatLUFactorNum        60 1.0 7.9378e+00 1.0 1.95e+08 1.2 0.0e+00 0.0e+00
> 0.0e+00  1  1  0  0  0   1  1  0  0  0  8814
>
> MatILUFactorSym        1 1.0 2.3076e-02 1.5 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> MatConvert             6 1.0 3.2693e-02 1.5 0.00e+00 0.0 0.0e+00 0.0e+00
> 1.8e+01  0  0  0  0  0   0  0  0  0  0     0
>
> MatScale               6 1.0 2.1923e-02 1.7 4.50e+06 1.1 2.4e+04 1.5e+03
> 0.0e+00  0  0  0  0  0   0  0  0  0  0 77365
>
> MatAssemblyBegin     266 1.0 1.0337e+01 4.4 0.00e+00 0.0 1.8e+05 3.8e+03
> 4.2e+02  1  0  3  0 10   1  0  3  0 10     0
>
> MatAssemblyEnd       266 1.0 3.0336e+00 1.0 0.00e+00 0.0 4.1e+05 8.6e+02
> 4.7e+02  1  0  7  0 12   1  0  7  0 12     0
>
> MatGetRow        6730366 1.1 8.6473e-01 1.2 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> MatGetRowIJ            3 3.0 5.2931e-035550.2 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> MatGetSubMatrix       12 1.0 2.2689e+01 1.0 0.00e+00 0.0 1.1e+05 1.3e+05
> 1.9e+02  4  0  2  5  5   4  0  2  5  5     0
>
> MatGetOrdering         3 3.0 6.5000e-03 5.4 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> MatPartitioning        6 1.0 2.9801e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00
> 1.4e+01  1  0  0  0  0   1  0  0  0  0     0
>
> MatCoarsen             6 1.0 9.5374e-02 1.1 0.00e+00 0.0 2.5e+05 1.1e+03
> 3.8e+01  0  0  4  0  1   0  0  4  0  1     0
>
> MatZeroEntries        22 1.0 6.1185e-01 1.2 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> MatTranspose           6 1.0 1.9780e-01 1.1 0.00e+00 0.0 1.9e+05 8.6e+02
> 7.8e+01  0  0  3  0  2   0  0  3  0  2     0
>
> MatPtAP              120 1.0 5.2996e+01 1.0 1.70e+10 1.1 9.7e+05 2.1e+05
> 4.2e+02 10 49 16 67 10  10 49 16 67 11 120900
>
> MatPtAPSymbolic       12 1.0 5.8209e+00 1.0 0.00e+00 0.0 2.2e+05 3.7e+04
> 8.4e+01  1  0  4  3  2   1  0  4  3  2     0
>
> MatPtAPNumeric       120 1.0 4.7185e+01 1.0 1.70e+10 1.1 7.6e+05 2.6e+05
> 3.4e+02  9 49 13 64  8   9 49 13 64  8 135789
>
> MatTrnMatMult          3 1.0 1.1679e+00 1.0 3.22e+07 1.2 8.2e+04 8.0e+03
> 5.7e+01  0  0  1  0  1   0  0  1  0  1  9997
>
> MatTrnMatMultSym       3 1.0 6.8366e-01 1.0 0.00e+00 0.0 6.9e+04 3.9e+03
> 5.1e+01  0  0  1  0  1   0  0  1  0  1     0
>
> MatTrnMatMultNum       3 1.0 4.8513e-01 1.0 3.22e+07 1.2 1.3e+04 3.0e+04
> 6.0e+00  0  0  0  0  0   0  0  0  0  0 24069
>
> MatGetLocalMat       126 1.0 1.1939e+00 1.3 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> MatGetBrAoCol        120 1.0 5.9887e-01 2.7 0.00e+00 0.0 5.7e+05 3.3e+04
> 0.0e+00  0  0  9  6  0   0  0  9  6  0     0
>
> MatGetSymTrans        24 1.0 1.4878e-01 1.4 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> VecDot               320 1.0 1.5860e+00 1.5 2.04e+08 1.1 0.0e+00 0.0e+00
> 3.2e+02  0  1  0  0  8   0  1  0  0  8 49574
>
> VecMDot              260 1.0 1.8154e+00 2.5 3.70e+08 1.1 0.0e+00 0.0e+00
> 2.6e+02  0  1  0  0  6   0  1  0  0  7 78497
>
> VecNorm              440 1.0 2.8876e+00 1.8 5.88e+08 1.1 0.0e+00 0.0e+00
> 4.4e+02  0  2  0  0 11   0  2  0  0 11 78281
>
> VecScale             320 1.0 2.2738e-01 1.2 7.88e+07 1.1 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0 133517
>
> VecCopy              220 1.0 7.1162e-01 1.2 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> VecSet               862 1.0 7.0683e-01 1.5 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> VecAXPY              440 1.0 9.0657e-01 1.2 3.83e+08 1.1 0.0e+00 0.0e+00
> 0.0e+00  0  1  0  0  0   0  1  0  0  0 162612
>
> VecAYPX              280 1.0 5.8935e-01 1.5 1.92e+08 1.1 0.0e+00 0.0e+00
> 0.0e+00  0  1  0  0  0   0  1  0  0  0 125070
>
> VecMAXPY             300 1.0 9.7644e-01 1.2 4.98e+08 1.1 0.0e+00 0.0e+00
> 0.0e+00  0  1  0  0  0   0  1  0  0  0 196269
>
> VecAssemblyBegin     234 1.0 5.0308e+00 5.5 0.00e+00 0.0 0.0e+00 0.0e+00
> 6.8e+02  1  0  0  0 17   1  0  0  0 17     0
>
> VecAssemblyEnd       234 1.0 1.8253e-03 8.8 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> VecScatterBegin     1083 1.0 2.8195e-01 4.7 0.00e+00 0.0 3.8e+06 8.4e+03
> 2.0e+01  0  0 64 11  0   0  0 64 11  1     0
>
> VecScatterEnd       1063 1.0 3.4924e+00 6.9 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> KSPGMRESOrthog        20 1.0 1.5598e+00 3.2 1.28e+08 1.1 0.0e+00 0.0e+00
> 2.0e+01  0  0  0  0  0   0  0  0  0  1 31503
>
> KSPSetUp             222 1.0 9.7521e-02 1.6 0.00e+00 0.0 0.0e+00 0.0e+00
> 1.2e+01  0  0  0  0  0   0  0  0  0  0     0
>
> KSPSolve              60 1.0 1.3742e+02 1.0 3.42e+10 1.1 5.7e+06 4.4e+04
> 3.2e+03 25100 95 83 79  25100 95 83 79 94396
>
> PCGAMGGraph_AGG        6 1.0 5.7683e+00 1.0 4.50e+06 1.1 3.8e+05 9.1e+02
> 2.5e+02  1  0  6  0  6   1  0  6  0  6   294
>
> PCGAMGCoarse_AGG       6 1.0 1.4101e+00 1.0 3.22e+07 1.2 4.0e+05 3.2e+03
> 1.4e+02  0  0  7  0  4   0  0  7  0  4  8280
>
> PCGAMGProl_AGG         6 1.0 1.8976e+00 1.0 0.00e+00 0.0 7.2e+05 3.4e+03
> 8.6e+02  0  0 12  1 22   0  0 12  1 22     0
>
> PCGAMGPOpt_AGG         6 1.0 5.7220e-06 0.0 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> GAMG: createProl       6 1.0 9.0840e+00 1.0 3.67e+07 1.2 1.5e+06 2.7e+03
> 1.3e+03  2  0 25  1 31   2  0 25  1 31  1472
>
>   Graph               12 1.0 5.7669e+00 1.0 4.50e+06 1.1 3.8e+05 9.1e+02
> 2.5e+02  1  0  6  0  6   1  0  6  0  6   294
>
>   MIS/Agg              6 1.0 9.5481e-02 1.1 0.00e+00 0.0 2.5e+05 1.1e+03
> 3.8e+01  0  0  4  0  1   0  0  4  0  1     0
>
>   SA: col data         6 1.0 8.5414e-01 1.0 0.00e+00 0.0 6.6e+05 3.0e+03
> 7.8e+02  0  0 11  1 19   0  0 11  1 20     0
>
>   SA: frmProl0         6 1.0 1.0123e+00 1.0 0.00e+00 0.0 6.2e+04 7.6e+03
> 6.0e+01  0  0  1  0  1   0  0  1  0  2     0
>
> GAMG: partLevel        6 1.0 3.6150e+01 1.0 8.41e+08 1.1 3.5e+05 5.0e+04
> 5.3e+02  7  2  6  6 13   7  2  6  6 13  8804
>
>   repartition          6 1.0 3.8351e+00 1.0 0.00e+00 0.0 4.7e+04 1.3e+02
> 1.6e+02  1  0  1  0  4   1  0  1  0  4     0
>
>   Invert-Sort          6 1.0 4.4953e+00 1.8 0.00e+00 0.0 0.0e+00 0.0e+00
> 2.4e+01  1  0  0  0  1   1  0  0  0  1     0
>
>   Move A               6 1.0 1.0806e+01 1.0 0.00e+00 0.0 8.5e+04 1.6e+05
> 1.0e+02  2  0  1  5  3   2  0  1  5  3     0
>
>   Move P               6 1.0 1.1953e+01 1.0 0.00e+00 0.0 2.5e+04 3.6e+03
> 1.0e+02  2  0  0  0  3   2  0  0  0  3     0
>
> PCSetUp              100 1.0 1.0166e+02 1.0 1.72e+10 1.1 2.7e+06 8.3e+04
> 2.2e+03 18 50 44 73 54  18 50 44 73 54 63848
>
> PCSetUpOnBlocks       40 1.0 1.0812e+00 1.2 1.95e+08 1.2 0.0e+00 0.0e+00
> 0.0e+00  0  1  0  0  0   0  1  0  0  0 64711
>
> PCApply              380 1.0 1.9359e+01 1.1 8.58e+09 1.1 1.4e+06 9.6e+03
> 6.0e+01  3 25 24  5  1   3 25 24  5  2 167605
>
> SFSetGraph            12 1.0 3.5203e-03 6.8 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> SFBcastBegin          44 1.0 2.4242e-02 3.0 0.00e+00 0.0 2.5e+05 1.1e+03
> 6.0e+00  0  0  4  0  0   0  0  4  0  0     0
>
> SFBcastEnd            44 1.0 3.0994e-02 8.6 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
> SFReduceBegin          6 1.0 1.6784e-02 3.8 0.00e+00 0.0 7.1e+04 5.0e+02
> 6.0e+00  0  0  1  0  0   0  0  1  0  0     0
>
> SFReduceEnd            6 1.0 8.6989e-0332.0 0.00e+00 0.0 0.0e+00 0.0e+00
> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>
>
> ------------------------------------------------------------------------------------------------------------------------
>
>
>
> Memory usage is given in bytes:
>
>
>
> Object Type          Creations   Destructions     Memory  Descendants' Mem.
>
> Reports information only for process 0.
>
>
>
> --- Event Stage 0: Main Stage
>
>
>
>               Matrix   246            243   1730595756     0
>
> Matrix Partitioning     6              6         3816     0
>
>       Matrix Coarsen     6              6         3720     0
>
>               Vector   602            602   1603749672     0
>
>       Vector Scatter    87             87      4291136     0
>
>        Krylov Solver    12             12        60416     0
>
>       Preconditioner    12             12        12040     0
>
>               Viewer     1              0            0     0
>
>            Index Set   247            247      9018060     0
>
> Star Forest Bipartite Graph    12             12        10080     0
>
>
> ========================================================================================================================
>
>
>
> Any idea why there are more matrix created with version 3.7.2? I only have
> 2 MatCreate calls and 4 VecCreate calls in my code!, so I assume the others
> are internally created.
>
>
>
>
>
> Thank you,
>
>
>
>
>
> *Hassan Raiesi, PhD*
>
>
>
> Advanced Aerodynamics Department
>
> Bombardier Aerospace
>
>
>
> hassan.raiesi at aero.bombardier.com
>
>
>
> *2351 boul. Alfred-Nobel (BAN1)*
>
> *Ville Saint-Laurent, Québec, H4S 2A9*
>
>
>
>
>
>
>
> Tél.
>
>   514-855-5001    # 62204
>
>
>
>
>
>
>
>
>
>
>
> *CONFIDENTIALITY NOTICE* - This communication may contain privileged or
> confidential information.
> If you are not the intended recipient or received this communication by
> error, please notify the sender
> and delete the message without copying, forwarding and/or disclosing it.
>
>
>
>
>
>
>



-- 
What most experimenters take for granted before they begin their
experiments is infinitely more interesting than any results to which their
experiments lead.
-- Norbert Wiener
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20160705/dfbe941d/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: image001.png
Type: image/png
Size: 6402 bytes
Desc: not available
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20160705/dfbe941d/attachment-0001.png>


More information about the petsc-users mailing list